You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+18-3Lines changed: 18 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -55,11 +55,20 @@ cd autoencoder_based_image_compression/kodak_tensorflow/
55
55
and H.265 are stored in the folder "eae/visualization/test/checking_reconstructing/kodak/".
56
56
57
57
## Quick start: training an autoencoder
58
-
1. First of all, ImageNet images must be downloaded. In our case, it is sufficient to download the ILSVRC2012 validation images, "ILSVRC2012_img_val.tar" (6.3 GB), see [ImageNetDownloadWebPage](http://image-net.org/download). Let's say that, in your computer, the path to "ILSVRC2012_img_val.tar" is "path/to/folder_0/ILSVRC2012_img_val.tar" and you want the unpacked images to be put into the folder "path/to/folder_1/" before the script "creating_imagenet.py" preprocesses them. The creation of the ImageNet training and validaton sets of luminance images is then done via
58
+
1. First of all, ImageNet images must be downloaded. In our case, it is sufficient to download the ILSVRC2012 validation
59
+
images, "ILSVRC2012_img_val.tar" (6.3 GB), see [ImageNetDownloadWebPage](http://image-net.org/download). Let's say that,
60
+
in your computer, the path to "ILSVRC2012_img_val.tar" is "path/to/folder_0/ILSVRC2012_img_val.tar" and you want the
61
+
unpacked images to be put into the folder "path/to/folder_1/" before the script "creating_imagenet.py" preprocesses
62
+
them. The creation of the ImageNet training and validaton sets of luminance images is then done via
2. The training of an autoencoder on the ImageNet training set is done via the command below. 1.0 is the value of the quantization bin widths at the beginning of the training. 14000.0 is the value of the coefficient weighting the distortion term and the rate term in the objective function to be minimized over the parameters of the autoencoder. The script "training_eae_imagenet.py" enables to split the entire autoencoder training into several successive parts. The last argument 0 means that "training_eae_imagenet.py" runs the first part of the entire training. For each successive part, the last argument is incremented by 1.
66
+
2. The training of an autoencoder on the ImageNet training set is done via the command below. 1.0 is the value of the
67
+
quantization bin widths at the beginning of the training. 14000.0 is the value of the coefficient weighting the
68
+
distortion term and the rate term in the objective function to be minimized over the parameters of the autoencoder.
69
+
The script "training_eae_imagenet.py" enables to split the entire autoencoder training into several successive parts.
70
+
The last argument 0 means that "training_eae_imagenet.py" runs the first part of the entire training. For each successive
71
+
part, the last argument is incremented by 1.
63
72
```sh
64
73
python training_eae_imagenet.py 1.0 14000.0 0
65
74
```
@@ -68,7 +77,10 @@ cd autoencoder_based_image_compression/kodak_tensorflow/
68
77
The documentation "documentation_kodak/documentation_code.html" describes all the functionalities of the code of the paper.
69
78
70
79
## A simple example
71
-
Another piece of code is a simple example for introducing the code of the paper. This piece of code is stored in the folder "svhn". Its documentation is in the file "documentation_svhn/documentation_code.html". If you feel comfortable with autoencoders, this piece of code can be skipped. Its purpose is to clarify the training of a rate-distortion optimized autoencoder. That is why a simple rate-distortion optimized autoencoder with very few hidden units is trained on tiny images (32x32 SVHN digits).
80
+
Another piece of code is a simple example for introducing the code of the paper. This piece of code is stored in the folder
81
+
"svhn". Its documentation is in the file "documentation_svhn/documentation_code.html". If you feel comfortable with autoencoders,
82
+
this piece of code can be skipped. Its purpose is to clarify the training of a rate-distortion optimized autoencoder. That is why
83
+
a simple rate-distortion optimized autoencoder with very few hidden units is trained on tiny images (32x32 SVHN digits).
72
84
73
85
## Citing
74
86
```
@@ -78,3 +90,6 @@ Another piece of code is a simple example for introducing the code of the paper.
0 commit comments