This week we’ll have (time permitting) some ‘Ask Me Anything’ questions. Your question doesn’t have to be related to this lesson, although it should be something that you think Jeremy or Rachel might be reasonably well qualified to answer! e.g. Ask about deep learning, building data-oriented startups, data-driven medicine, Kaggle, MOOC development, etc…
We’ll primarily use ‘likes’ to prioritize questions, so please vote for questions you’re interested in.
Have you given any thought to doing a regular online broadcast (maybe monthly or biweekly ) to go through new papers and new learnings? I know many people would appreciate it.
Do you have any advice for people looking to transition into working as a deep learning practitioner, if they have a machine learning background, (and, say, have taken a course like this) but don’t have a PhD? I feel like I’m both constantly hearing both that companies want new people who have competence or knowledge in DL, and also seeing companies have strict PhD requirements for said positions.
How do we better utilize multi-GPU systems in Keras with either Theano or TensorFlow backend? From what I could gather you may need to split the data into batches and then train them on different GPUs and it is not straight-forward. Having a multiple GPU systems make it much much easier to iterate and try different models.
Are there any plans to formalize some of the utilities that have been made for the class and some of the models that have been made in the class to let practitioners start their work faster and make their lives easier?
It makes data parallelization very easy. Essentially that code just runs a copy of the model separately on each gpu, then concatenates the output on the CPU (so you can effectively run batch sizes that are #GPUs x larger).
Haven’t tried it on Keras 2 yet but that script works well for Keras 1 on TensorFlow.
As there are many topics covered in this course crying for a mobile app to be built, can you recommend good resources on how to deploy DL in moblle apps? More general pointers towards DL deployment in production are also highly welcome.
How do we when the model is doing as well as possible given the noisiness of the data? Is there a way to get an idea of the upper limit of performance?
For some domains, they use inter-rater agreement to get an estimate of “good enough”. Some problems, like predicting click-through rate, most of the (human) factors are unknowable and you just hope your model is well calibrated. Are there rules of thumb for getting an idea of the size of the signal and the noise?
I think one more valuable step is to teach to people how to transform their notebooks implementation into Desktop or Mobile apps.
I am planning to work on, and I think it would be worth if you organize some sessions on.
I like this idea. I started a channel with @Matthew and @sravya8 for the same reason , but it’d be nice to have something official. We can revive our old Data institute channel?