Code and data for paper "Deep Painterly Harmonization": https://arxiv.org/abs/1804.03189
Code and data for paper "Deep Painterly Harmonization"
This software is published for academic and non-commercial use only.
This code is based on torch. It has been tested on Ubuntu 16.04 LTS.
Dependencies: * Torch (with loadcaffe) * Matlab or Octave
Download VGG-19:
sh models/download_models.sh
Compile
cuda_utils.cu(Adjust
PREFIXand
NVCC_PREFIXin
makefilefor your machine):
make clean && make
To generate all results (in
data/) using the provided scripts, simply run
python gen_all.pyin Python and then
run('filt_cnn_artifact.m')in Matlab or Octave. The final output will be in
results/.
Note that in the paper we trained a CNN on a dataset of 80,000 paintings collected from wikiart.org, which estimates the stylization level of a given painting and adjust weights accordingly. We will release the pre-trained model in the next update. Users will need to set those weights manually if running on their new paintings for now.
Removed a few images due to copyright issue. Full set here for testing use only.
Here are some results from our algorithm (from left to right are original painting, naive composite and our output):
If you find this work useful for your research, please cite:
@article{luan2018deep, title={Deep Painterly Harmonization}, author={Luan, Fujun and Paris, Sylvain and Shechtman, Eli and Bala, Kavita}, journal={arXiv preprint arXiv:1804.03189}, year={2018} }
Feel free to contact me if there is any question (Fujun Luan [email protected]).