where we try to predict whether a url is a phishing one or a good one.
The authors were kind enough to share their dataset, roughly 1000000 samples for each class.
Approach:
The approach starts simple by using an LSTM plus all the goodies provided by Fastai, including fit_one_cycle, lr_find, etc.
Results
Initial results seem quite interesting, with the F1-score going from 98.76 to 99.25 in our model. Similarly all the metrics cited in the paper (AUC, accuracy, precision, recall) are improved, even without the 3-fold cross-validation:
It is very interesting how far we can go using the power of fastai and a straightforward LSTM model.
Of course more involved techniques can further improve this model. Stay tuned!
Hi everyone! I’ve created an open source web app that allows a user to take a picture using an interface similar to a native camera app. After the picture is taken, it’s sent to a fast.ai CNN model running on render. The modified server will the return a classification of the image along with related content.
There are a handful of really helpful prebuilt apps that allow a user to upload a photo (this and this), but I couldn’t find an app that allowed the user to skip the step of manually uploading it.
You can check out a demo app here that recognizes everyday objects (cars, people, trees, computers, chairs, etc).
I hope it’s helpful to someone and I welcome any feedback or pull requests that could make it more helpful or clear.
I’ve tried out SimCLR, and it seems to be a good direction to go with self-supervised learning.
Pre-training on ImageNet with SimCLR for 50 epochs and then fine-tune on ImageNette gives us 92% accuracy on the validation, whereas starting from random weights gives 79%. Starting from pre-trained weights on supervised ImageNet gives 95.5%.
I think that the first layer filters the model came up with is just mesmerizing:
I’ve trained a model on MNIST 14x14 downscaled images.
This model knows the structure of numbers on this scale.
The original 28x28 images have the same structure but just having twice the size.
If we adjust the first layer’s scale accordingly, we can use this network directly on 28x28 pictures without any fine-tuning. It is just adjusting the first layer’s dilation and stride to be twice it’s original value and we’re done, we have the same accuracy as we had on 14x14 pictures.
Or we can have first layer filters twice the size as was before and having weights of the 14x14 model’s first layer’s weights resized as if they were images.
This gives a slight accuracy drop, but training for just a little bit with a learning rate of 1e-100!!, we have back our accuracy.
Just wanted to say thanks so much for the fastai course, it was a brilliant introduction and way to get started!
I’ve been using what I learned on this course and others to explore generative algorithms. Here are some of the pictures I’ve generated so far using style transfer
Hi everyone,
as you knew, WHO suggests washing hands and not touching face would protect citizens from coronavirus infection
but not touching your face is easier said than done! you can check out this video for example.
so based on Lesson 1,2 we built a face touching classifier, Touching your face is bad for ya!
When you touch your face, a warning sound will be played. That’s it.
I’m the creator of SeeMe.ai, an AI marketplace (currently in beta).
We focus on letting you train, use and share AI models with your friends and the rest of the world. We have been a fan of fast.ai since its early days, and are happy to contribute to this wonderful community.
Once you have trained your model, you can easily deploy it and use it on the web, mobile, our Python SDK or via the API. After testing, you can share it with others so they can use your model.
We have just created a fastai v1 quick guide for you, that helps you get up and running and we would love for you to try it out and get your feedback!
Thank you for the answer. I´ve found another jawa script that worked for me just fine. The original one was generating an empty txt file in Chrome after searching the images.
I wrote an article on Medium on how to deploy your machine learning model on AWS, using Cortex. In the example I used the first course of Part I (2019), and deployed a pets classifier as a web service. Dig in and tell me what you think!
Predicting the origin of an ancient, ceramic artifact
Covid-19 has been depressing for a data guy like me. Always thinking about stats, but these ones suck. Trying to keep my mind off things, I finished up a passion project of mine: IndAIna. I love how I can combine my love for history with data science. Check out the link and stay safe, y’all! https://bit.ly/2UlJfmv
Its been a while, but I have managed to finally get round to training a model using DeOldify to re-colour Anime sketches. Got some amazing results so far!
Coming from a non-development background, last few months have been an extremely frustrating experience. I learned about version control by accidentally wiping out 3 months worth of work and observed the perils of rm -rf after virtually destroying my machine.
But going through the fastai forums and seeing people share amazing projects and positive messages kept me going.
If nothing else, this course made me have a newfound respect for software developers. Cheers guys!