One-Network-to-Solve-Them-All

It is a Tensorflow implementation of our paper One Network to Solve Them All --- Solving Linear Inverse Problems using Deep Projection Models. If you find our paper or implementation useful in your research, please cite:

           
@article{chang2017projector,
    title={One Network to Solve Them All --- 
        Solving Linear Inverse Problems using Deep Projection Models},
    author={J. H. Rick Chang and 
        Chun-Liang Li and 
        Barnab{\'a}s P{\'o}czos and 
        B. V. K. Vijaya Kumar and 
        Aswin C. Sankaranarayanan},
    journal={arXiv preprint arXiv:1703.09912},
    year={2017}
}

Brief introduction

The goal of the proposed framework is to solve linear inverse problems of the following form:

where y is the linear measurements, e.g., a low-resolution image, A is the linear operator that generates y from an image x. In image super-resolution, A can indicate direct downsampling or box averaging. The optimization problem is solved with a constraint that the solution x lies in the natural image set X.

To solve the optimization problem, we found that in proximal algorithms like alternating direction method of multipliers (ADMM), the constraint usually appears in solving a subproblem --- projecting current estimate of x to the set X. Thereby, we propose to learn the projection operator with a deep neural net. Since the projection opereator is independent to individual linear operator A, once the network is trained, it can be used in any linear inverse problem. Since we do not have the exaxt definition of the natural image set, we use the decision boundary of a classifier to approximate the set.

There are multiple methods to achieve this approximation. For example, given a large image dataset, we can create nonimage signals by perturbing the images in the dataset and then train a classifier to differentiate the two classes. While this method is simple, the decision boundary will be loose. To get a tighter approximation, we found that during the training process, the projected images of the projection network become closer and closer to the natural image set. Thus, if we use these projected images as negative instances, we will learn a tighter decision boundary. This framework is motivated by adversarial generative net.

Once the projection network is trained, we can solve any linear inverse problem with ADMM. An illustration of the testing process is shown below.

Prerequest

The code is tested under Ubuntu 16.04 with CUDA 8.0 on Titan X Pascal and GTX 1080. We use Python 2.7 and Tensorflow 0.12.1.

We train the model on two datasets, MS-CELEB-1M dataset and ImageNet dataset. The dataset should be plased under ~/dataset. For example, we put MS-CELEB-1M dataset at ~/dataset/celeb-1m. We load the datasets via load_celeb.py and load_imagenet.py, both can be easily adapt to other datasets. In this tutorial, we take MS-CELEB-1M as an example to illustrate how to use our code. For ImageNet, the usage is alomst the same by replacing import load_celeb with import load_imagenet. Please see the comments in these files for more information.

Train a Projector

cd projector
source run_celeb.sh

The above steps train a projection network on MS-CELEB-1M dataset. Please refer to our paper for the details of parameters. The model files are saved in model directory.

Run ADMM to solve the Linear Inverse Problems

We have to preprocess the reference batch used in virtual batch normalization for testing. Note that you need to modifiy the filepath of your trained model!

cd admm
python update_popmean.py 

We then run demo script for differet linear inverse problems. Likewise, you need to modifiy the filepath of your updated model!

python paper_demo.py

In our experience, models trained for 50,000 iterations should give you the result similar to we reported in the paper. Note that you may need to adjust the value of alpha (the penalty parameter) for each task and different hyper-parameters used to train the model.

Here are some sampled result reported in the papers.

Trained models

The trained model used in the paper can be found here. A newer version of the model that uses imresize by nearest neighbor algorithm to replace upsampling/downsampling by stride is here. We found that using imresize to perform upsampling and downsampling provides more stable projectors. We have not fully explored this method, so the resulted images may look blurrier than the original model used in the paper.

To use these models, you need to modify the filepath in admm/update_popmean.py and admm/paper_demo.py. Check the comments in these files. Also note that the alpha in admm/paper_demo.py depends on the dataset, the model, and the problem itself. So like solving the traditional LASSO problems, you need to tune alpha to get nice results. In the file admm/paper_demo.py we provide the values used in the paper to solve the ms-celeb-1m dataset. They provide a good starting point for other datasets and models.

Acknowledgement

Part of our code is based on https://github.com/jazzsaxmafia/Inpainting