Skip to content

Analysis and figure code from Alley et al. 2019.

Notifications You must be signed in to change notification settings

shechty2/UniRep-analysis

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

UniRep-analysis

Analysis and figure code from Alley et al. 2019.

Start by cloning the repo:

git clone https://github.com/churchlab/UniRep-analysis.git

Requirements

python: 3.5.2

For reference on how to install, see https://askubuntu.com/questions/682869/how-do-i-install-a-different-python-version-using-apt-get

venv with necessary requirements, can be installed with:

cd UniRep-analysis # root directory of the repository
python3 -m venv venv/
source venv/bin/activate
pip install -r venv_requirements/requirements-py3.txt
deactivate

conda, can be installed with:

wget https://repo.continuum.io/miniconda/Miniconda3-latest-Linux-x86_64.sh
bash Miniconda3-latest-Linux-x86_64.sh
# restart the currrent shell

two conda environments necessary to run different parts of the code:

cd UniRep-analysis # root directory of the repository
conda env create -f ./yml/grig_alldatasets_run.yml
conda env create -f ./yml/ethan_analysis.yml

Getting the data

mkdir data
cd data

wget https://s3.us-east-2.amazonaws.com/unirep-data-storage/unirep_analysis_data_part2.tar.gz
tar -zxvf unirep_analysis_data_part2.tar.gz # this may take some time
mv data/* ./
rm unirep_analysis_data_part2.tar.gz

wget https://s3.amazonaws.com/unirep-public/unirep_analysis_data.zip
unzip unirep_analysis_data.zip # may need to install unzip with sudo apt install unzip
mv unirep_analysis_data/* ./
rm unirep_analysis_data.zip

cd ..

Project Structure

.
├── analysis
│   ├── analysis_unsupervised_clustering_oxbench_homstrad.ipynb # ethan_analysis
│   ├── FINAL_compute_std_by_val_resampling.py # grig_alldatasets_run
│   ├── FINAL_run_l1_regr_quant_function_stability_and_supp_analyses.py # grig_alldatasets_run
│   ├── FINAL_run_RF_homology_detection.py # grig_alldatasets_run
│   └── FINAL_run_transfer_analysis_function_prediction__stability.py # grig_alldatasets_run
├── figures
│   ├── figure2
│   │   ├── fig_1a.ipynb # ethan_analysis
│   │   ├── fig2b_supp_fig2_upper.ipynb # ethan_analysis
│   │   ├── fig2c.ipynb # ethan_analysis
│   │   ├── fig2e_supp_fig4-5.ipynb # ethan_analysis
│   │   ├── FINAL_Fig2d_AND_SupTableS2_Homology_detection.ipynb # grig_alldatasets_run
│   │   ├── FINAL_Fig2g_alpha-beta_neuron.ipynb # grig_alldatasets_run
│   │   ├── supp_fig2.ipynb # ethan_analysis
│   │   └── supp_fig3.ipynb # ethan_analysis
│   ├── figure3
│   │   ├── fig3b.ipynb # ethan_analysis
│   │   ├── fig3c.ipynb # ethan_analysis
│   │   ├── FINAL_Fig_3a_Rosetta_comparison.ipynb # grig_alldatasets_run
│   │   ├── FINAL_Fig3e_Quant_function_prediction_Fig3b_stability_ssm2_significance_SuppTableS4-5.ipynb # grig_alldatasets_run
│   │   └── supp_fig8.ipynb # ethan_analysis
│   ├── figure4 
│   │   ├── A007h_budget_constrained_functional_sequence_recovery_analysis.ipynb # venv
│   │   ├── A007j_pred_v_actual_fpbase_plots.ipynb # venv
│   │   ├── A008c_visualize_ss_feature_predictors_on_protein.ipynb # venv
│   │   ├── common.py
│   │   ├── supp_fig10a_partial_and_e.ipynb # ethan_analysis
│   │   ├── supp_fig10a_partial_and_f.ipynb # ethan_analysis
│   │   ├── supp_fig10b-d_left.ipynb # ethan_analysis
│   │   ├── supp_fig10b-d_right.ipynb # ethan_analysis
│   │   └── supp_fig_10g_10h.ipynb # venv
│   └── other
│       ├── FINAL_supp_data_3_2_1__supp_fig_S9__Supp_fig_s12.ipynb # grig_alldatasets_run
│       ├── FINAL_SuppFigS1_Seq_db_growth.ipynb # grig_alldatasets_run
│       └── supp_fig13.ipynb # ethan_analysis
├── common
├── common_v2
├── README.md
├── venv_requirements
└── yml

Usage

Reproducing figures

To re-generate all the figures in the main text, one should execute jupyter/ipython notebooks in the /figures directory using the right environment.

To run a notebook, do the following:

Activate the right environment (as noted in the Project Structure section for each notebook):

For grig_alldatasets_run:

source activate grig_alldatasets_run

For ethan_analysis:

source activate ethan_analysis

For venv:

source venv/bin/activate

Then execute:

jupyter notebook

This will automatically open a browser window where one can interactively rerun the code generating the figures. Aesthetic components (colors, font sizes) may differ slightly from the final version of the figures in the paper.

Re-training top models and re-generating performance metrics

In order to re-train the models and regenerate the metrics from which the figures are constructed, one can run the python scripts in the /analysis folder. By default these will evaluate all representations and baselines on all available datasets and subsets and will computer metrics (such as MSE and Pearson r) on the test subset.

The easiest way is to do this is to start an AWS instance with sufficient resources (we recommend m5.12xlarge or m5.24xlarge for shorter runtime - the code takes advantage of all the available CPU cores) with Ubuntu Server 18.04 LTS AMI (for example, ami-0f65671a86f061fcd). After performing the initial setup above, create the necessary directories:

cd analysis
mkdir results # folder for various model performance metrics
mkdir predictions # folder for model predictions for various datasets
mkdir models # folder for trained models
mkdir params # folder for recording best parameters after hyperparameter search 

Activate the right environment:

source activate grig_alldatasets_run

To run SCOP 1.67 Superfamily Remote Homology Detection and SCOP 1.67 Fold-level Similarity Detection with Random Forest, execute:

python FINAL_run_RF_homology_detection.py

To run quantitative function prediction, de novo designed mini proteins stability prediction, DMS stability prediction for 17 de novo designed and natural protein datasets from Figure 3, as well as supplementary benchmarks, such as small-scale function prediction (Supp. Table S4):

python FINAL_run_l1_regr_quant_function_stability_and_supp_analyses.py
python FINAL_compute_std_by_val_resampling.py # computes estimates of standard deviations through validation/test set resampling for significance testing (generates std_results_val_resamp.csv)

To run the analyses from Supp. Fig. S10 (generalized stability prediction, generalized quantitative function prediction and a special central to remote generalized stability prediction task):

python FINAL_run_transfer_analysis_function_prediction__stability.py

About

Analysis and figure code from Alley et al. 2019.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 99.7%
  • Python 0.3%