Lesson 12 (2019) discussion and wiki

Lesson resources

Software requirements

  • Some nbs of this lesson also currently require pytorch-nightly if yours got de-installed see this.
  • Notebook 10c (and subsequent) requires the NVIDIA apex python library. In the environment you created for fastai go to the fastai directory and run pip install git+https://github.com/NVIDIA/apex.

Papers

Blogs

Notes and other resources

AMA about Swift at the end of class

Post your questions in this separate thread:

12 Likes

Kind of a tangential question but since Jeremy mentioned it: do you have any tips on debugging Deep Learning models?

9 Likes

Has mixup been successfully used in NLP yet?

Not that I know of, but you should definitely try :wink:

2 Likes

The Audio module won’t be covered today ???

1 Like

Last week we shared an updated schedule. The audio module will be covered in an extra session that will be livestreamed once the course ends. We had more material than will fit in the 7 weeks of the course.

Edited to add: the dates of the extra sessions have not been set yet.

8 Likes

Has anyone tried mixup and normal augmentation like rotation/zoom?

Seems like they’d be different augmentations that could be used together.

1 Like

I tried playing around with mixup on NLP embeddings this past week and from early experimenting it seems to work well (maybe someone else has spent more time on it already!)

13 Likes

We tried yes, and it’s the same as mixup without normal augmentation in our experiments.

2 Likes

Do you have an intuition for why that might be?

How broadly can we apply mixup? Could you use it on an image regression problem?

Mixup is much more powerful in terms of data augmentation, so it more or less erases everything else.

5 Likes

on mac you can just type ctrl cmd space and type the name of the letter (ie ‘gamma’)

1 Like

As long as there is a way to mixup your labels, you can try. It has been wildly experimented in single label classification, not so much in other areas.

it seems mixup forces the model to behave linearly between training classes, why is this desirable?

I’d reverse it into, why is this not desirable?

3 Likes

adversarial attacks proofing I guess

What about backprop with new loss?

Some researchers actually found it helps against adversarial attacks (for a generalized version of Mixup): here.

5 Likes