print what you have in the output, so we can help
You can use argmax as we discussed in class to get the class indexes from that.
Ah, beautiful, thanks!
I think I got what I was looking for by doing the following. At least I got a segmented image.
got it! thank!
Have a look at this topic:
what is this spec?
Here Jeremy says that we can fine-tune an already trained model, with its misclassified images and that it needs to be done for more epochs or with a higher learning rate. But what is the intuition behind the higher learning rate ?
I understand that misclassified examples are likely to be interesting but wouldn’t increasing the learning rate be harmful for a network that has already been trained ? Also, aren’t difficult examples likely to be particular cases, and retraining only on them will just overfit to those examples ?
Thanks for the reference. Unfortunatelly, the model does not train and it through an error when I try to train it:
learn.fit_one_cycle(1, 1e-2, moms=(0.8,0.7))
Maybe better actually is more epochs, instead of higher LR. I haven’t tried either - would be interesting to experiment.
One of the datasets we saw in Lesson 3 was Planet. The images in this dataset are 256x256 (called Kaggle chips) and were generated from satellite images with 6600x2200. So, one satellite image can generate several Kaggle chips.
My question: I have one (and only one) satellite image of 5500x5000 and I want to create a training with this. Is there a way I could use fastai to help me generate crops of this satellite image?
I mean, is there any transformation that crop random patches of an image?
Image generator from images instead of folders (Segmentation task)
Sorry for the late reply… spec should be: grey_im
no worries i got model to run on gray scale…
but the possible downside is m getting overiffited to Train data, diff between Tran loss and val loss is like .180 and .235
does reducing the color channels affects the ability of the model to generalize ?
with human eye it all looks clearly distinct what is object and what is background in grayscale… compared to colored image where some times both are not separable…
is this the way machine also looks ?
there was one parameter in tfms crop in old fai
that should be passed as CROP.CENTRE or CROP.RANDOM ,then it would take the sz parameter to accordingly slize the image
How can we perform multi-label classification for text in fastai ?
I am not able to figure it out !
Hello, I’m trying to segment aerial images, and sometimes due to the weather conditions the “temperature” of images may vary so some of images are more blueish, and some are more yellowish.
Does anyone knows if there is a “vision.transform” operation to vary color aspect of the image to make my model more robust to the color changes?
I was thinking of imgaug library but I don’t know how to put it in Data Block API in the best way…
Thanks in advance!
hey, what dataset is this and what are you trying to predict?
Hi, could anyone give more explanation on how to choose the threshold for f_score, fbeta, and how does accuracy_thresh work? I’m kind of confusing how the accuracy_thresh computes…
has anyone seen any datasets for domain specific law, like compliance, etc…?