I’m very new to fastai and it’s my first personal project using it.
My goal is to locate a particular symbol on a paper sheet.
So I want to do image regression, feeding the model with images associated with coordinates of a point or maybe better a bounding box. (something like in the lesson 6 and the BIWI dataset)
Right now I’m collecting images and creating a little tool to help me fill in the coordinates of the symbol for each image.
My concern is that my images are not of the same size and not on the same orientation.
What I understood is that images should be the same size, thus be transformed and squared during the training. But I don’t understand how the coordinates will behave according to this.
In my case cropping the image is not possible but adding pads is
If my original images is of size 800x600 and the symbol is at position 700x500, if my image is resized to 256x256 for example, the coordinates will be outside, right ? Same concern if pads are added to the image
Is it automagicaly handled by the library ?
Or should I do something specific to better prepare my data ?
thank you for your insights