This guide will walk you through multiple ways to setup dream-creator
on Ubuntu and Windows. If you wish to install PyTorch and dream-creator on a different operating system like MacOS, installation guides can be found here.
Note that in order to reduce their size, the pre-packaged binary releases (pip, Conda, etc...) have removed support for some older GPUs, and thus you will have to install from source in order to use these GPUs.
The pip and Conda packages ship with CUDA and cuDNN already built in, so after you have installed PyTorch with pip or Conda, you can skip to installing dream-creator.
Following the pip installation instructions here, you can install PyTorch with the following commands:
# in a terminal, run the commands
cd ~/
pip install torch torchvision
Or:
cd ~/
pip3 install torch torchvision
Now continue on to installing dream-creator to install dream-creator.
Following the Conda installation instructions here, you can install PyTorch with the following command:
conda install pytorch torchvision -c pytorch
Now continue on to installing dream-creator to install dream-creator.
If you have a CUDA-capable GPU from NVIDIA then you can
speed up dream-creator
with CUDA.
Instructions for downloading and installing the latest CUDA version on all supported operating systems, can be found here.
cuDNN is a library from NVIDIA that efficiently implements many of the operations (like convolutions and pooling) that are commonly used in deep learning.
After registering as a developer with NVIDIA, you can download cuDNN here. Make sure that you use the appropriate version of cuDNN for your version of CUDA.
Follow the download instructions on Nvidia's site to install cuDNN correctly.
Note that the cuDNN backend can only be used for GPU mode.
(Optional) Steps 1-3: Install PyTorch with support for AMD GPUs using Radeon Open Compute Stack (ROCm)
It is recommended that if you wish to use PyTorch with an AMD GPU, you install it via the official ROCm dockerfile: https://rocm.github.io/pytorch.html
- Supported AMD GPUs for the dockerfile are: Vega10 / gfx900 generation discrete graphics cards (Vega56, Vega64, or MI25).
PyTorch does not officially provide support for compilation on the host with AMD GPUs, but a user guide posted here apparently works well.
ROCm utilizes a CUDA porting tool called HIP, which automatically converts CUDA code into HIP code. HIP code can run on both AMD and Nvidia GPUs.
To install PyTorch from source on Ubuntu (Instructions may be different if you are using a different OS):
cd ~/
git clone --recursive https://github.com/pytorch/pytorch
cd pytorch
python setup.py install
cd ~/
git clone --recursive https://github.com/pytorch/vision
cd vision
python setup.py install
To check that your torch installation is working, run the command python
or python3
to enter the Python interpreter. Then type import torch
and hit enter.
You can then type print(torch.version.cuda)
and print(torch.backends.cudnn.version())
to confirm that you are using the desired versions of CUDA and cuDNN.
To quit just type exit()
or use Ctrl-D.
Now continue on to installing dream-creator to install dream-creator.
If you wish to install PyTorch on Windows From Source or via Conda, you can find instructions on the PyTorch website: https://pytorch.org/
First, you will need to download Python 3 and install it: https://www.python.org/downloads/windows/. I recommend using the executable installer for the latest version of Python 3.
Then using https://pytorch.org/, get the correct pip command, paste it into the Command Prompt (CMD) and hit enter:
pip install torch===1.6.0 torchvision===0.7.0 -f https://download.pytorch.org/whl/torch_stable.html
After installing PyTorch, download the dream-creator Github repository and extract/unzip it to the desired location.
Then copy the file path to your dream-creator folder, and paste it into the Command Prompt, with cd
in front of it and then hit enter.
In the example below, the dream-creator folder was placed on the desktop:
cd C:\Users\<User_Name>\Desktop\dream-creator-master
You can now continue on to installing dream-creator, skipping the git clone
step.
First we clone dream-creator
from GitHub:
cd ~/
git clone https://github.com/ProGamerGov/dream-creator.git
cd dream-creator
To test that dream-creator works, we'll first need to create a test dataset:
python data_tools/create_test_data.py -output_dir test_data -num_images 1200 -shape_size 200,200
Now we calculate the mean and standard deviation of our newly created test dataset:
python data_tools/calc_ms.py -data_path test_data
The calc_ms.py
script will then output something that looks like this:
-data_mean 126.7094,126.0554,126.9935 -data_sd 73.655,73.3885,73.7058
Then using the mean and standard deviation parameter inputs we generated above, we can now train a GoogleNet model using the test dataset:
python train_googlenet.py -data_path test_data -balance_classes -batch_size 96 -num_epochs 10 -data_mean 126.7094,126.0554,126.9935 -data_sd 73.655,73.3885,73.7058
If everything is working properly with the training script, then you should see output like this:
Total 2400 images, split into 2 classes
Classes:
{'ellipse': 0, 'rectangle': 1}
Model has 10,021,318 learnable parameters
Epoch 1/10
------------
train Loss: 0.9526 Acc: 0.8250
Time Elapsed 0m 23s
val Loss: 0.0319 Acc: 0.9896
Time Elapsed 0m 27s
Epoch 2/10
------------
train Loss: 0.3052 Acc: 0.9260
Time Elapsed 0m 49s
val Loss: 0.0699 Acc: 0.9646
Time Elapsed 0m 52s
Epoch 3/10
------------
train Loss: 0.2157 Acc: 0.9417
Time Elapsed 1m 15s
val Loss: 0.0142 Acc: 0.9938
Time Elapsed 1m 19s
Epoch 4/10
------------
train Loss: 1.6389 Acc: 0.5719
Time Elapsed 1m 42s
val Loss: 0.6930 Acc: 0.5125
Time Elapsed 1m 45s
Epoch 5/10
------------
train Loss: 1.3215 Acc: 0.5089
Time Elapsed 2m 8s
val Loss: 0.6986 Acc: 0.5125
Time Elapsed 2m 12s
After training has finished, we can now visualize the newly created DeepDream model's FC layers using the following command:
python vis_multi.py -model_file bvlc_out010.pth -num_iterations 200
The vis_multi.py
script should end up creating two output images, where one image has more circlelike features and the other has more squarelike features. Using more complex datasets that have more classes and images will yield far better looking results. You can find a list of image collection tools, possible sources of images, and duplicate image detection tools on the dream-creator wiki.
Finally, to visualize a single layer and channel or to DeepDream your own image, we can use the following command:
# Random Noise content image
python vis.py -model_file bvlc_out010.pth -layer mixed5a -channel 7 -num_iterations 200
# With Content image
python vis.py -model_file bvlc_out010.pth -layer mixed5a -channel 7 -content_image examples/small/fc_flowers.jpg -image_size 512,512 -num_iterations 200