Cygwin alias
How can I create alias in Cywgin?
ssh ubuntu@34.229.220.175 -L8888:localhost:8888
You can do:
alias <your alias name>='ssh -i <path-to-aws-key> ubuntu@34.229.220.175 -L8888:localhost:8888'
Thanks!
@jeremy, @yinterian does it make sense to create alias? Public IP on AWS changes every time I restarts my p2
I am getting this error when running lesson2 :
out of memory at /opt/conda/conda-bld/pytorch_1503965122592/work/torch/lib/THC/generic/THCStorage.cu:66
I checked on some blogs and found that it might be because batch size is too big, changed the bs to 24 but still get the same error.
How should i be resolving this?
Our default batch size is 64 and runs normally on AWS and crestle. are you running in locally or on cloud?
AWS p2.xlarge instance
Have you tried restarting kernel and rerunning code?
If you have saved weights, please skip learn.fit and load to avoid training time.
I restarted the instance and it is back to normal.
@vikbehal - If you want to have the same IP Address, you should look at āElastic IPā inside āNetwork and Securityā and associate that Elastic IP to your instance.
When creating Elastic IP, make sure you choose VPC when you select āAllocate New Addressā. Once itās created, do Actions -> Associate Address. That way you have the same IP Address. If your P2 instance is in a VPC (most likely if you use P2), the elastic IP would be associated even when instance is stopped.
Elastic IP has a cost of 1cent/hour for each hour the instance you have associated with the IP is NOT running.
@jeremy With respect to the below image when Iām training with a single learning rate, where does the training loss after each epoch maps to the loss plot across iterations. Because I can see the loss after each epoch is lower that the lowest point in the loss graph
Iām not at all sure whatās happening in those graphs you show! Iāve never really plotted that with a constant learning rate - I suspect that somehow without cyclical learning rates itās not being updated correctly. Iāll try to look into itā¦
Okayā¦but a more basic doubt is the training loss calculated after each epoch, is it the loss on the whole training data using the modified weights at the end of the epoch?
Itās the loss at the end of the epoch that are printed. (To be more technical: an exponentially weighted moving average of the last few iterations.)
Ah!!! Your question inspired me to study the code more closely, and I see that whatās being printed for the training loss is not quite correct. Iāve just pushed a fix - git pull
and you should see that your chart makes more sense (hopefully!)
Thanks for the clarification and Iāll check the plot again and let you know
@jeremy : What does this code do and i get an error while executing it.
Try changing that from a 1 to a 0. Iām on my phone currently so I canāt tell you the command to check which gpu you want to use, but if you only have 1 gpu it will be 0 instead of 1.
Basically that is just telling Cuda which gpu to use.
Can you provide more context? Which Notebook is this? Are you running it in GPU? What machine G2 / P2?
Sorry I shouldnāt have left that in the notebook - itās used to select a different GPU if you have multiple GPUs. Since you have just one GPU, you should remove that line.
Ok. Thank you.