Installation of VFVS
Last updated
Last updated
After setting up access to the supercomputing cluster, open up the command line interface (terminal in Linux or macOS, Command Prompt in Windows). At the command line, follow the instructions of the specific supercomputing cluster being used to log in to connect to the supercomputing cluster. Typically ssh
is used to connect to the login node of the Slurm cluster.
Once logged into and interfacing with the supercomputing cluster, enter the following command:
This shows the current folder, also known as the directory, where you are located. The first step is to navigate to the appropriate directory in the supercomputing cluster workspace where VirtualFlow will be installed.
VirtualFlow should be installed on a fast shared scratch file system rather than the home directory. The reasoning behind this and further information regarding the installation of VirtualFlow can be found within the "VirtualFlow" subsection located in the "Installation" section of the .
After choosing the directory where you want VFVS and the workflow to be located, go to this directory. This can be done by entering the following command:
Replace <installation directory>
with the name of the folder (directory) that you chose.
VFVS requires Python 3.x (it has been tested with Python 3.9.4). Additionally, it requires packages to be installed that are not included at many sites. As a result, we recommend that you create a virtualenv for VFVS that can be used to run jobs.
Make sure that virtualenv is installed:
Create a virtualenv (this example creates it under $HOME/vfvs_env):
Enter the virtualenv and install needed packages
Alternatively, all the above steps can be performed using conda
instead of pip.
Further information on setting up a conda
style environment can be found here.
To exit a virtual environment:
When running VFVS commands, you should enter the virtualenv that you have set up using:
Alternatively, all the above steps can be performed using conda
instead of pip.
Further information on setting up a conda
style environment can be found here.
For the remainder of the tutorial and for future use of VFVS, please ensure the virtual environment is activated.
After Python is set up, we can install VFVS 2.0. Normally this is done by cloning the GitHub repo, but for this tutorial, we provide a pre-configured folder for download which is available here:
On the AWS Batch login node, you can download it by running the following command:
This preconfigured folder was prepared by:
Cloning the GitHub repo (branch vfvs-2)
Adding the docking input files to the input-files
folder
Configuring the all.ctrl
file. The following parameters are changed compared to the template file in the VFVS GitHub repo:
job_name
batchsystem
job_storage_mode
docking_scenario_names
docking_scenario_programs
docking_scenario_replicas
docking_scenario_batchsizes
For VFVS to be able to access the ligand libraries stored in AWS Open Data, we need to install the AWS-CLI.
You will need to have an AWS Account. If you don't have one yet, you can create a free one here: https://aws.amazon.com/
You will also need to create access keys for your username, e.g. the root user: https://docs.aws.amazon.com/IAM/latest/UserGuide/id_root-user_manage_add-key.html
First, let's make sure that no prior version of AWS-CLI is installed:
Next, download the archive:
When finished, simply unzip the archive:
and install:
If you don't have sudo rights, you will need to ask the cluster admins to install it for you.
To check that your installation is successful, you can run:
The above command should generate an output similar to:
Next, you must configure AWS-CLI:
You have to define the following (the keys should match the ones you use in your local version of aws-cli):
Congratulations, you now have a fully configured version of VF2.0! We will now proceed to the next step, setting up a docking run.