Giter Club home page Giter Club logo

deep_dream-master's Introduction

deep_dream

An implementation of the Deep Dream image processing algorithm which is able to process large (wallpaper-sized) images despite GPU or main memory limits. It is also able to use multiple processes to take advantage of several CPUs and/or GPUs.

This implementation of Deep Dream is able to divide the gradient ascent step into tiles if a too-large image is being processed. By default, any image larger than 512x512 will be divided into tiles no larger than 512x512. The tile seams are obscured by applying a random shift on each gradient ascent step (this also greatly improves the image quality by summing over the translation dependence inherent to the neural network architecture). Further, several tiles can be processed simultaneously on machines with more than one compute device (CPU or GPU).

  1. Usage
  2. Example
  3. CNN.dream_guided() example
  4. Models
  5. Pre-built Amazon image
  6. Requirements
  7. Python 3.5 build tips

Usage

Usage: deep_dream_cli.py [OPTIONS] IN_FILE [OUT_FILE]

  CLI interface to deep_dream.

Options:
  --cpu-workers INTEGER        The number of CPU workers to start.
  --gpus INTEGER_LIST          The CUDA device IDs to use.
  --guide-image TEXT           The guide image to use.
  --layers RE_LIST             The network layers to target.
  --max-input-size INTEGER...  Rescale the input image to fit into this size.
  --max-tile-size INTEGER      The maximum dimension of a tile.
  --min-size INTEGER           Don't use scales where the small edge of the
                               image is below this.
  --model TEXT                 The model to use. Valid values: GOOGLENET_BVLC,
                               GOOGLENET_PLACES205, GOOGLENET_PLACES365,
                               RESNET_50.
  --n INTEGER                  The number of iterations per scale.
  --per-octave INTEGER         The number of scales per octave.
  --smoothing FLOAT            The per-iteration smoothing factor. Try
                               0.02-0.1.
  --step-size FLOAT            The strength of each iteration.
  --tv-weight FLOAT            The per-scale denoising weight. Higher values
                               smooth the image less. Try 25-250.
  --help                       Show this message and exit.

Example

import deep_dream as dd
from PIL import Image

cnn = dd.CNN(dd.GOOGLENET_PLACES365, gpus=[0])
img = Image.open('kodim/img0022.jpg').resize((768, 512), Image.LANCZOS)

out = cnn.dream(img, 'inception_4a/output', min_size=64, per_octave=4, n=8, step_size=0.5, smoothing=0.02)
dd.to_image(out).save('example_med.jpg', quality=85)

out = cnn.dream(img, 'inception_4a/output', min_size=64, per_octave=4, n=12, step_size=1.2, smoothing=0.01)
dd.to_image(out).save('example_out.jpg', quality=85)

CNN.dream_guided() example

Input:

Guide:

Combined output:

Gradient ascent was performed using layers inception_(3a-b, 4a-e, 5a-b)/output. This is a reasonable set of layers for dream_guided() to work well. Note that the input and the guide do not have to be the same size; the output will be the same size as the input.

Models

Locations of pre-trained .caffemodel files (run get_models.sh to automatically download them):

  • bvlc_googlenet: tends toward visualizing abstract patterns, dogs, insects, and amorphous creatures.
  • googlenet_places205: tends toward visualizing buildings and landscapes.
  • googlenet_places365: newer than the places205-trained model, often more aesthetically pleasing output, tends toward visualizing buildings and landscapes.

Pre-built Amazon image

This AMI in us-west-2 contains deep_dream with all dependencies preinstalled and built for Python 3.5, and all models downloaded. It should be launched in a g2.2xlarge or g2.8xlarge instance. These instance types have 1 and 4 GPUs respectively. You can use all four GPUs from deep_dream_cli.py or deep_dream_test.py by specifying the parameter --gpus 0,1,2,3.

Requirements

  • Python 3.5.
  • Caffe, built against Python 3.5. (See the Python 3.5 build tips.) I would encourage you to use Caffe's nVidia GPU support if possible: it runs several times faster on even a laptop GPU (GeForce GT 750M) than on the CPU.
  • The contents of requirements.txt. (pip install -U -r requirements.txt)
    • openexrpython needs to be installed from git master instead of 1.2.0 from PyPI for optional OpenEXR export. (pip install -U git+https://github.com/jamesbowman/openexrpython)
  • Pre-trained Caffe models (run get_models.sh; see Models section).

Python 3.5 build tips

You will need protobuf 3 (currently in beta) for its Python 3 compatibility: 2.x will not work! Check out protobuf and build/install both the main protobuf package (C++/protoc) and the Python module in python/. Do this before attempting to build Caffe.

Linux (Tested on Ubuntu 16.04 LTS)

  • First see the Ubuntu 15.10/16.04 installation guide on the Caffe GitHub wiki.

  • Python 3.5 Makefile.config settings, with python3.5 installed via apt-get:

    PYTHON_INCLUDE := /usr/include/python3.5m \
            /usr/local/lib/python3.5/dist-packages/numpy/core/include
    PYTHON_LIB := /usr/lib
    PYTHON_LIBRARIES := boost_python-py35 python3.5m
  • I used openblas in this configuration. MKL is probably faster in CPU mode.

OS X (Tested on El Capitan 10.11)

  • Python 3.5 Makefile.config settings, with python3 installed through homebrew:

    PYTHON_DIR := /usr/local/opt/python3/Frameworks/Python.framework/Versions/3.5
    PYTHON_INCLUDE := $(PYTHON_DIR)/include/python3.5m \
            /usr/local/lib/python3.5/site-packages/numpy/core/include
    PYTHON_LIB := $(PYTHON_DIR)/lib
    PYTHON_LIBRARIES := boost_python3 python3.5m
  • This assumes you installed numpy with pip into the python3.5 system site-packages directory. If you're in a virtualenv this may change.

  • Leave the BLAS setting at atlas, unless you want to try MKL (faster in CPU mode). Recent OS X ships with an optimized multithreaded BLAS so there is little reason IMO to use openblas anymore.

deep_dream-master's People

Contributors

oussemarez avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.