I went ahead and created my first blog on ResNet. I really wanted to understand how ResNet worked, and it didn’t feel like any of the existing blog really hit the kernel of the matter! I hope you all check it out. It’s supposed to be a very easy read.
Any ideas for improvement is welcome. I wish, for that matter, that I could understand the mathematics in better detail too, but don’t feel like I can. In particular, I wanted to understand how the skip connections actually help the gradients, and/or ultimately the classification problem. The detailed math seems to be discussed here. https://arxiv.org/abs/1603.05027
Looking good! BTW if you mention you’re doing the fast.ai course when you blog we can share you blog as an example of cool stuff our students are doing. And it can help bring more folks to the course.
I really liked your post. You might be using simple techniques - as you state in your article - but you put them to good use to tell an interesting story.
One comment though - you might want to run a spellcheck on the text and also proof read it for issues that stand out. One trick to proof reading is reading the article out loud. Also, the plot really needs a legend and then I think you can remove the code that generated it (and that doesn’t contain any info on which line color is for which batch size either).
@radek, thanks a lot for your feedback. I will take care of those-important- details you comment. I’m quite new to 2.0 stuff (believe it or not, almost zero internet footprint until recently) and also learning now Python (R programmer here), so still many things to improve. But this course, you all guys and Jeremy’s tireless example has motivated me to try participate and contribute a bit more.
I think you are off to a great start and I think you have a very nice way with words. The only reason I commented on your post - as I also state in the PM - is that I liked it a lot and I think that with minor changes you can make it even better.
I’d like to share this - as @radek says it needs a little light editing so I’ll wait until that’s done. I also suggest explaining that first LR vs loss curve (i.e. “learning rate finder”), since it’s super awesome and pretty much no-one knows about it. In explaining it, it would be good to show the LR vs batch curve as well, and link to the original paper (already linked in the lesson 1 notebook).
Great! I will add those explanations, + editing (also found learning rate finder useful and awesomely informative, what gave me the idea for the post). So,I will update and report, thanks for the feedback!
Not sure if this would be relevant to the thread, but prior to Part 1v2 starting, I was still going through the original Part 1 of the course. I decided to leverage what I’ve learned in a university assignment, and figured I would blog about it and share my final assignment report and code
@jeremy, I updated the post with the expanded info about the “learning rate finder”, referencing to the paper of Leslie N. Smith and adding labels to the plots, incorporating also the kind advice of @radek as much as I’ve been able to.
I will be AFK for some hours now (its quite late here) but will be glad to change or modify any part if you find it not precise or (I hope will not be the case), just plain wrong. Anyway your previous advice made the post more informative and solid, very happy about that!
I wrote one about the ResNet architecture - I wanted to have my intuitions clear about what ‘Residual’ really means in ResNet. I do hope I got it right
Here is the blog post . The next architecture to understand would be (no surprises) ResNeXt!
As a passing note, this course has been an amazing catalyst to make me think in multiple dimensions. And the forum feels like 7 days of awesome learning per week! Thanks @jeremy!
It’s an ELI5 like explanation to the topic. I originally wanted to wait till I could blog about an awesome idea but I decided to pull the trigger and go with the FastAI philosophy to help the learner who’d be a step behind me.