I want to detect signatures on photos and I cannot get good results. I have 60 labels in the train dataset and cannot anything out of it while Google AutoML does a decent job.
I tried changing ratios and scales for boxes for no avail. Any thoughts?
Also when I plot bboxes, images from dls.valid_ds are not resized - how do I get resized ones? Answer - with Resize(size, method=ResizeMethod.Squish)(img)
I think something similar happened with me last year. You can use kmeans clustering with IoU as distance metric to automate finding anchor boxes for you. This is usually done in YOLO models. Here’s one link you can refer - https://lars76.github.io/object-detection/k-means-anchor-boxes/
If I remember correctly, I used kmeans to find both anchor ratios and sizes. It worked quite well!
You may be better off here, but I still am not quite understanding asto why this would be an issue. Both v1 and v2 should be outputting the same thing. Can you make a databunch using the same base file and show what it’s outputted bounding box tensors look like? I’ve looked at this extensively and their endpoints should be the same.
I hate this issue of different styles of bounding box representation. Someone should standardise it! It gets very confusing sometimes, especially when writing code to generate anchor boxes and loss functions.
I would also recommend you to reimplement it yourself. That’s what I did as well.
Btw anyone knows why fastai uses a range of -1 to 1 for anchor boxes instead of standard 0 to 1 range.
I used the 0 to 1 range for anchor boxes in my implementation and it worked fine.