Need help with UniRep?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

churchlab
243 Stars 81 Forks 33 Commits 1 Opened issues

Description

UniRep model, usage, and examples.

Services available

!
?

Need anything else?

Contributors list

UniRep, a mLSTM "babbler" deep representation learner for protein engineering informatics.

We present an interface for training, inferencing representations, generative modelling aka "babbling", and data management. All three architectures (64, 256, and 1900 units) are provided along with the trained architectures, the random initializations used to begin evotuning (to ensure reproducibility) and the evotuned parameters.

For training/finetuning: note that backpropagation of an mLSTM of this size is very memory intensive, and the primary determinant of memory use is the max length of the input sequence rather than the batch size. We have finetuned on GFP-like fluorescent proteins (~120-280aa) on a p3.2xlarge instance (aws) with 16GB GPU memory successfully. Higher memory hardware should accommodate larger sequences, as will using one of the smaller pre-trained models (64 or 256). If you are having difficulty with your use case, please reach out. We are happy to assist you.

Quick-start

First clone or fork this repository and navigate to the repository's top directory (

cd UniRep
). We recommend using our docker environments. Install docker to get started.

CPU-only support

  1. Build docker:
    docker build -f docker/Dockerfile.cpu -t unirep-cpu .
    This step pulls the Tensorflow 1.3 CPU image and installs a few required python packages. Note that Tensorflow pulls from Ubuntu 16.04.
  2. Run docker:
    docker/run_cpu_docker.sh
    . This will launch Jupyter. Copy and paste the provided URL into your browser. Note that if you are running this code on a remote machine you will need to set up port forwarding between your local machine and your remote machine. See this example (note that in our case jupyter is serving port 8888, not 8889 as the example describes).
  3. Open up the
    unirep_tutorial.ipynb
    notebook and get started. The 64-unit model should be OK to run on any machine. The full-sized model will require a machine with more than 16GB of RAM.

GPU support

  1. System requirements: NVIDIA CUDA 8.0 (V8.0.61), NVIDIA cuDNN 6.0.21, NVIDIA GPU Driver 410.79 (though == 361.93 or >= 375.51 should work. Untested), nvidia-docker. We use the AWS Deep Learning Base AMI for Ubuntu (tested on version 17.0 ami-0ff00f007c727c376), which has these requirements pre-configured.
  2. Build docker:
    docker build -f docker/Dockerfile.gpu -t unirep-gpu .
    This step pulls the Tensorflow 1.3 GPU image and installs a few required python packages. Note that Tensorflow pulls from Ubuntu 16.04.
  3. Run docker:
    docker/run_gpu_docker.sh
    . This will launch Jupyter. Copy and paste the provided URL into your browser. Note that if you are running this code on a remote machine you will need to set up port forwarding between your local machine and your remote machine. See this example (note that in our case jupyter is serving port 8888, not 8889 as the example describes).
  4. Open up the
    unirep_tutorial.ipynb
    notebook and get started. The 64-unit model should be OK to run on any machine. The full-sized model will require a machine with more than 16GB of GPU RAM.

Obtaining weight files

The

unirep_tutorial.ipynb
notebook downloads the needed weight files for the 64-unit and 1900-unit UniRep models. However, if you want to download these or other weight files directly, you first need
awscli
(the AWS Command Line Interface tool). If using the docker environments built above, it is included. If working outside docker, first do
pip install awscli
. To grab a set of weights, do:
aws s3 sync --no-sign-request --quiet s3://unirep-public/ 

where

 is one of:

  • 1900_weights/
    : trained weights for the 1900-unit (full) UniRep model
  • 256_weights/
    : trained weights for the 256-unit UniRep model
  • 64_weights/
    : trained weights for the 64-unit UniRep model
  • 1900_weights_random/
    : random weights that were used to initialize the 1900-unit (full) UniRep model for Random Evotuned.
  • 256_weights_random/
    : random weights that could be used to initialize the 256-unit UniRep model (e.g. for evotuning).
  • 64_weights_random/
    : random weights that could be used to initialize the 64-unit UniRep model (e.g. for evotuning).
  • evotuned/unirep/
    : the weights, as a tensorflow checkpoint file, after 13k unsupervised weight updates on fluorescent protein homologs obtained with JackHMMer of the globally pre-trained UniRep (1900-unit model).
  • evotuned/random_init/
    : the weights, as a tensorflow checkpoint file, after 13k unsupervised weight updates on fluorescent protein homologs obtained with JackHMMer of a randomly initialized UniRep (initialized with 1900weightsrandom) that was not pre-trained at all (1900-unit model).

Description of files in this repository

  • unirep_tutorial.ipynb - Start here for examples on loading the model, preparing data, training, and running inference.
  • unirep_tutorial.py - A pure python script version of the tutorial notebook.
  • unirep.py - Interface for most use cases.
  • custom_models.py - Custom implementations of GRU, LSTM and mLSTM cells as used in representation training on UniRef50
  • data_utils.py - Convenience functions for data management.
  • formatted.txt and seqs.txt - Tutorial files.

License

Copyright 2018, 2019 Ethan Alley, Grigory Khimulya, Surojit Biswas

All the model weights are licensed under the terms of Creative Commons Attribution-NonCommercial 4.0 International License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA.

Otherwise the code in this repository is licensed under the terms of GPL v3 as specified by the gpl.txt file.

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.