-
Hi, Thanks, |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
Hi Mareeta, In my understanding, the analysis transform is implemented by a convolutional neural network. This architecture utilizes a number of learnable filters to scan the input image and computes the output feature maps by convolution. The output size of each convolutional layer(including the size of the bottleneck layer) is determined by the size of the input layer and the striding length. In other words, the CNN architecture is adaptable to any (large enough) sized input. During the training period, the input image sizes are required to be the same because you are training in batches, every image has to have the same height and widths so that it forms a 4D matrix before being fed into the network. However, images from most training sets come in all sizes. That's why cropping is necessary. You can change the patch size from 256x256 to any size you like, as long as you crop the images to the same size. In testing, we want to test the full image of each image from the test set so that results are comparable to certain standards. So there shouldn't be random cropping in the testing, since that would essentially change your test set, giving unreliable results. As mentioned, the CNN adapts to different sizes of inputs so there are no restraints on the size of test images. (Unless you want to test them in patches as well, which, empirically is faster) Lingyu |
Beta Was this translation helpful? Give feedback.
Hi Mareeta,
In my understanding, the analysis transform is implemented by a convolutional neural network. This architecture utilizes a number of learnable filters to scan the input image and computes the output feature maps by convolution. The output size of each convolutional layer(including the size of the bottleneck layer) is determined by the size of the input layer and the striding length. In other words, the CNN architecture is adaptable to any (large enough) sized input.
During the training period, the input image sizes are required to be the same because you are training in batches, every image has to have the same height and widths so that it forms a 4D matrix before being fed in…