Skip to content

Waller-Lab/MultiWienerNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MultiWienerNet

Deep learning for fast spatially-varying deconvolution

Setup:

Clone this project using:

git clone https://github.com/Waller-Lab/MultiWienerNet.git

Install dependencies. We provide code both in Tensorflow and in Pytorch. Tensorflow version only contains an implementation for 2D deconvolutions, whereas the Pytorch contains both 2D and 3D deconvolutions.

If using Pytorch, install the depencies as:

conda env create -f environment.yml
source activate multiwiener_pytorch

If using Tensorflow, install the depencies as:

conda env create -f environment.yml
source activate multiwiener_tf

Using pre-trained models

We provide an example of a pre-trained MutliWienerNet for fast 2D deconvolutions as well as compressive 3D deconvolutions from 2D measurements. These examples are based on data for Randoscope3D. To adapt this model to your own data, please see below.

Loading in pretrained models

The pre-trained models can be downloaded: here (pytorch) and here (tensorflow) Please download these and place them in the pytorch/saved_models and tensorflow/saved_models

Loading in data

We provide several limited examples of 2D and 3D data in /data/ You can download the full dataset that we have used for training here 2D, here 3D. You also need to download the PFSs here and add it to the /data folder.

Training for your own microscope/imaging system

Characterize your imaging system forward model

To retrain MultiWienerNet to work for your own imaging system, you first need to simulate realistic measurements from your imaging system to create sharp/blurred image pairs. If you already have a spatially-varying model for your imaging system (e.g. in Zemax), you can use this. If you do not have a model for your spatially-varying imaging system, we propose you follow the following calibration approach:

  • Scan a bead on a 8x8 grid across your microscope/imaging system's field of view. Repeat this for each depth plane of interest.
  • Run your data through our low rank code here

Create a dataset

You can simulate data using the low rank forward model as shown in the SVD notebook above or use your own field-varying forward model to simulate measurements.

Train your network

We have provided training scripts for 2D imaging in tensorflow here and for single-shot 3D imaging in pytorch here.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published