Lesson 11 discussion and wiki

Lesson resources

Papers

Notes and other resources

13 Likes

Could I get a refresher on what callbacks and hooks were?

This is the focus of the last lesson. The video is accessible in the general announcements thread.

How does lsuv work on test/val data? do we still adjust mean and std parameters on test data?

Is there a reason we use standard deviation instead of Mean Absolute Deviation?

1 Like

So just to confirm, LSUV is something you run on all the layers once at the beginning, not during training? What if your batch size is small, could you overfit to that batch?

6 Likes

possible to give an overview of lsuv again ?(high level)

1 Like

I want to hear Jeremy pronounce Imagenette in French accent!

3 Likes

It’s only the initialization of your network. Then, you train it.

4 Likes
6 Likes

Wait for it :wink:

2 Likes

Even though it’s live you can always go back in time on the video if you missed something.

2 Likes

but is it better than say Kaiming?

2 Likes

Yes, it trains better, especially for a deeper network.

8 Likes

Interesting about the size of the images affecting what works and what doesn’t. What’s the intuition behind?
Does it mean that some of the operations don’t really “scale” with size?

Probably the resolution is too low to learn good filters for full-size images?

1 Like

The network sees less (or more) details, so it can’t get to the same results.

2 Likes

you got it :slight_smile:

Oh ok - so it’s really the size of the objects, rather than the size of the images per se

Ah! got it now…thought the readjustment will done after every epoch/few iterations…thanks