Lesson 3 In-Class Discussion ✅

No, the learning rate after unfreezing is very often different.

1 Like

The idea of progressive resizing. One example: Progressive Growing of GANs for Improved Quality, Stability, and Variation | Research

https://www.fast.ai/2018/04/30/dawnbench-fastai/

Instead, we turned to a method we’d developed at fast.ai, and teach in lessons 1 & 2 of our deep learning course: progressive resizing. Variations of this technique have shown up in the academic literature before (Progressive Growing of GANs and Enhanced Deep Residual Networks) but have never to our knowledge been applied to image classification.

11 Likes

I am trying to update fastai but I get permission denied. Anything I need to add specifically? I tried conda and pip and both had permission issues.

I’ve been reading Leslie Smith’s paper and he provides some guidance on batch size with one cycle training. Basically I think go as large as you can, up to that point that you get diminishing returns.

1 Like

3e-3 === 0.003
1e-3 === 0.001

2 Likes

isn’t dice score a more relevant metric for segmentation problems?

2 Likes

What about loss function for multi label classification? Does the same loss function ( cross entropy) work for multi label classification as well?
In Kerala , I used to use binary cross entropy+sigmoid for last layer, it’s not clear how fastai takes care of this

1 Like

How can we ignore specific pixel value in an image while training ? i.e. igonire pixel value = 255

Hey @rachel I finally have 8 votes on this

:slightly_smiling_face::slightly_smiling_face::slightly_smiling_face:

8 Likes

This is the metric that was used in the paper introducing camvid, that’s why Jeremy is using it.

It almost feels like some form of data augmentation

1 Like

Given it usually takes a minute or so to train a whole epoch, how is lr_find so fast when looking at lots of different learning rates? Does it run just a few iterations for each learning rate? I looked at the documentation but still don’t quite understand how it works.

2 Likes

It does it for you. If you want to know more, you should ask on the advanced forum for now, this will be covered later in the course.

Any recommendations for making sense of cutting edge academic papers? I often see an interesting-looking paper on something I’m generally familiar with, but the jargon in academic papers can be overwhelming.

3 Likes

As per my understanding, it plots loss for different learning rates for different mini batches. That’s why it doesn’t take long

It does 100 iterations from 1e-5 to 10, growing the learning rate exponentially.

2 Likes

Progressive resizing is also talked about in the fast.ai dawnbench blog posts here and here.

11 Likes

You’d want to use BCEWithLogitsLoss in Pytorch, it’s binary cross entropy+sigmoid combined

2 Likes

Welcome to Microsoft’s Github.

6 Likes

Some tips around reading (and implementing) new papers are covered in part 2 of the course.

4 Likes