Dumb question - at 1:35:00 ish you say that most modern architectures can handle images of various sizes.
I was under the impression that in order to feed forward an image through a network, it has to be of a consistent size. In other words, whatever architecture you have, the number of inputs has to match the number of pixels.
So if you have a CNN with 400 input nodes, then your images can only be 20x20 (or something that multiplies to be 400). Am I missing something?
How can you create one architecture that accepts images of varying sizes?