|
Hi, Assume we have 99999 images of size 50*50. We want to learn features from this data and then do some classification. Sparse coding and convolution neural network(CNN) have been used. Here is the scenario:
Briefly speaking after learning 75 features and applying convolution/pooling we will have 74999250 images of size 20x20(i.e [74999250,20x20]). Now my question is that if I want to apply another layer of convolution/pooling( basically repeat all above steps 1-4), in step one for feature learning my training data set would be [74999250,20*20]?? In another words, when adding multiple layer of convolution/pooling, in fact we are increasing number input data[i.e training data] meanwhile decreasing the size of input, is it correct??? Any help appreciated. |
|
That sounds right, so in the second round you use [74999250,11x11] images to learn another 75*X features (assuming all X features are connected to all 75 in the previous layer of your network). That sounds really costly. Appreciate an expert's opinion since I am not one. |
|
Hi lnc7 check this out. Look at figure 1 think about it and it should answer your question. Hopefully :) |
|
Looking at the vast amount of parameters, maybe it would have to apply regularization to avoid using more samples than parameters. See the slides from http://trent.st/ffx/ for an easy introduction. |