You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: examples/convex/classificationCIFAR10/TrainICNNOnCIFAR10Example.md
+39-43
Original file line number
Diff line number
Diff line change
@@ -1,11 +1,11 @@
1
1
2
2
# <spanstyle="color:rgb(213,80,0)">Train Fully Convex Neural Network for Image Classification</span>
3
3
4
-
This example shows how to create a fully input convex neural network and train it on CIFAR\-10 data. This example uses fully connected based convex networks, rather than the more typical convolutional networks, proven to give higher accuracy on the training and test data set. The aim of this example is to demonstrate the expressive capabilities convex constrained networks have by classifying natural images and demonstrating high accuracies on the training set. Further discussion on the expressive capabilities of convex networks for tasks including image classification can be found in \[1\].
4
+
This example shows how to create a fully input convex convolutional neural network and train it on CIFAR\-10 data\[1\].
5
5
6
6
# Prepare Data
7
7
8
-
Download the CIFAR\-10 data set \[1\]. The data set contains 60,000 images. Each image is 32\-by\-32 pixels in size and has three color channels (RGB). The size of the data set is 175 MB. Depending on your internet connection, the download process can take time.
8
+
Download the CIFAR\-10 data set \[2\]. The data set contains 60,000 images. Each image is 32\-by\-32 pixels in size and has three color channels (RGB). The size of the data set is 175 MB. Depending on your internet connection, the download process can take time.
9
9
10
10
```matlab
11
11
datadir = ".";
@@ -18,16 +18,6 @@ Load the CIFAR\-10 training and test images as 4\-D arrays. The training set con
For illustration in this example, subsample this data set evenly in each class. You can increase the number of samples by moving the slider to smaller values.
22
-
23
-
```matlab
24
-
subSampleFrequency = 10;
25
-
XTrain = XTrain(:,:,:,1:subSampleFrequency:end);
26
-
XTest = XTest(:,:,:,1:subSampleFrequency:end);
27
-
TTrain = TTrain(1:subSampleFrequency:end);
28
-
TTest = TTest(1:subSampleFrequency:end);
29
-
```
30
-
31
21
You can display a random sample of the training images using the following code.
32
22
33
23
<pre>
@@ -37,42 +27,45 @@ im = imtile(XTrain(:,:,:,idx),ThumbnailSize=[96,96]);
37
27
imshow(im)
38
28
</pre>
39
29
40
-
# Define FICNN Network Architecture
30
+
# Define FICCNN Network Architecture
41
31
42
-
Use the <samp>buildConstrainedNetwork</samp> function to create a fully input convex neural network suitable for this data set.
32
+
Use the <samp>buildConvexCNN</samp> function to create a fully input convex convolutional neural network suitable for this data set.
43
33
44
-
- The CIFAR\-10 images are 32\-by\-32 pixels. Therefore, create a fully convex network specifying the <samp>inputSize=[32 32 3]</samp>.
45
-
-Specify a vector a hidden unit sizes of decreasing value in <samp>numHiddenUnits</samp>. The final number of outputs of the network must be equal to the number of classes, which in this example is 10.
34
+
- The CIFAR\-10 images are 32\-by\-32 pixels, and belong to one of ten classes. Therefore, create a fully convex network specifying the <samp>inputSize=[32 32 3]</samp> and the <samp>numClasses=10</samp>.
35
+
-For each convolutional layer, specify the filter size in <samp>filterSize</samp>, the number of filters in <samp>numFilters</samp>, and the stride size in <samp>stride</samp>.
46
36
```matlab
47
37
inputSize = [32 32 3];
48
-
numHiddenUnits = [512 128 32 10];
38
+
numClasses = 10;
39
+
filterSize = [3; 3; 3; 3; 3; 1; 1];
40
+
numFilters = [96; 96; 192; 192; 192; 192; 10];
41
+
stride = [1; 2; 1; 2; 1; 1; 1];
49
42
```
50
43
51
-
Seed the network initialization for reproducibility.
44
+
Seed the network initialization for reproducibility.
Train for a specified number of epochs with a mini\-batch size of 256. To attain high training accuracy, you may need to train for a larger number of epochs, for example <samp>numEpochs=8000</samp>, which could take several hours.
79
+
Train for a specified number of epochs with a mini\-batch size of 256. To attain high training accuracy, you may need to train for a larger number of epochs, for example <samp>numEpochs=400</samp>, which could take several hours.
87
80
88
81
```matlab
89
-
numEpochs = 8000;
82
+
numEpochs = 400;
90
83
miniBatchSize = 256;
91
-
initialLearnRate = 0.1;
92
-
decay = 0.005;
84
+
initialLearnRate = 0.0025;
85
+
decay = eps;
93
86
lossMetric = "crossentropy";
87
+
l2Regularization = 1e-4;
94
88
```
95
89
96
90
Create a <samp>minibatchqueue</samp> object that processes and manages mini\-batches of images during training. For each mini\-batch:
@@ -103,6 +97,7 @@ Create a <samp>minibatchqueue</samp> object that processes and manages mini\-bat
The networks output has been constrained to be convex in every pixel in every colour. Even with this level of restriction, the network is able to fit reasonably well to the training data. You can see poor accuracy on the test data set but, as discussed at the start of the example, it is not anticipated that such a fully input convex network comprising of fully connected operations should generalize well to natural image classification.
To summarise, the fully input convex network is able to fit to the training data set, which is labelled natural images. The training can take a considerable amount of time owing to the weight projection to the constrained set after each gradient update, which slows down training convergence. Nevertheless, this example illustrates the flexibility and expressivity convex neural networks have to correctly classifying natural images.
199
194
200
-
# Supporting Functions
201
-
## MiniBatch Preprocessing Function
195
+
# Supporting Functions
196
+
## Mini\-Batch Preprocessing Function
202
197
203
-
The <samp>preprocessMiniBatch</samp> function preprocesses a mini\-batch of predictors and labels using the following steps:
198
+
The <samp>preprocessMiniBatch</samp> function preprocesses a mini\-batch of predictions and labels using the following steps:
204
199
205
200
1. Preprocess the images using the <samp>preprocessMiniBatchPredictors</samp> function.
206
201
2. Extract the label data from the incoming cell array and concatenate into a categorical array along the second dimension.
207
-
3. One\-hot encode the categorical labels into numeric arrays. Encoding into the first dimension produces an encoded array that matches the shape of the network output.
202
+
3. One\-hot encode the categorical labels into numeric arrays. Encoding in the first dimension produces an encoded array that matches the shape of the network output.
208
203
```matlab
209
204
function [X,T] = preprocessMiniBatch(dataX,dataT)
210
205
@@ -219,19 +214,20 @@ T = onehotencode(T,1);
219
214
220
215
end
221
216
```
222
-
## Mini\-Batch Predictors Preprocessing Function
217
+
## Mini\-Batch Predictors Preprocessing Function
223
218
224
-
The <samp>preprocessMiniBatchPredictors</samp> function preprocesses a mini\-batch of predictors by extracting the image data from the input cell array and concatenate into a numeric array. For grayscale input, concatenating over the fourth dimension adds a third dimension to each image, to use as a singleton channel dimension. You divide by 255 to normalize the pixels to <samp>[0,1]</samp> range.
219
+
The <samp>preprocessMiniBatchPredictors</samp> function preprocesses a mini\-batch of predictors by extracting the image data from the input cell array and concatenating it into a numeric array. For grayscale input, concatenating over the fourth dimension adds a third dimension to each image, to use as a singleton channel dimension. You divide by 255 to normalize the pixels to <samp>[0,1]</samp> range.
225
220
226
221
```matlab
227
222
function X = preprocessMiniBatchPredictors(dataX)
228
-
X = single(cat(4,dataX{1:end}))/255;
223
+
X = (single(cat(4,dataX{1:end}))/255); % Normalizes to [0, 1]
224
+
X = 2*X - 1; % Normalizes to [-1, 1].
229
225
end
230
226
```
231
-
# References
232
-
233
-
\[1\] Amos, Brandon, et al. Input Convex Neural Networks. arXiv:1609.07152, arXiv, 14 June 2017. arXiv.org, https://doi.org/10.48550/arXiv.1609.07152.
\[2\] Krizhevsky, Alex. "Learning multiple layers of features from tiny images." (2009). https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf
0 commit comments