2-way sync is actually triggered every 2 minutes. So it will take anywhere from 0 -2 mins for any updates in your code and output folder to show up on Google Drive. Similarly anything uploaded on data folder, will get downloaded on machine within 2 mins. This workflow is detailed in help article in my above reply. Sync is generally an expensive operation (especially when dealing with folders containing large number of files) and hence we chose not to sync too often.
Let me know the issue this delay of upto 2 minutes is causing for you. We are open to align to what community needs.
Iām afraid , second issue is also by design, Sync back from machine to Google Drive only modifies and adds new files (not deletes). This is to make sure any accidental delete on your machine, doesnāt delete the data on cloud. Again, I am open to your feedback and suggestions on this.
On Google Colab when I try to run these fastai imports
from fastai.transforms import *
from fastai.conv_learner import *
from fastai.model import *
from fastai.dataset import *
from fastai.sgdr import *
from fastai.plots import *
@prakashguptaā¦as it is almost not practical to uppload large files in Google Driveā¦ I tried downloading the data directly in clouderizer terminal using curlwget. It succeeded but couldnot unzip. tried installing unzip using sudo but didnāt work out. Do you know a way to unzip from clouderizer terminal.
Thanks
If you download data set in clouderizer terminalā¦ is it temperory because i find it is gone next time i restart Clouderizer. It is not in data folder while i could use in prevous session. Do i need to download the data each session. Thanks
data folder is synced down (Google Drive -> Machine) code and out folder is synced up (Machine -> Google Drive)
Please see the last section of the help article above for details.
In your case, you are directly downloading to data folder on machine. This will not get persisted to Google Drive. This is by design.
In case you wish to persist this data, try downloading this in code folder. It should then sync to Google Drive.
Donāt forget to update dataset path in your notebook to accomodate this path change.
I met this problem several times and today I found a solution. just check how many free GPU RAM do you have (in my case it was 0% free left), if there is few memory left, clean your GPU RAM up and then reconnect. You will find that you have your GPU RAM 100% free!!!
To check your GPU memory
!ln -sf /opt/bin/nvidia-smi /usr/bin/nvidia-smi
!pip install gputil
!pip install psutil
!pip install humanize
import psutil
import humanize
import os
import GPUtil as GPU
GPUs = GPU.getGPUs()
# XXX: only one GPU on Colab and isnāt guaranteed
gpu = GPUs[0]
def printm():
process = psutil.Process(os.getpid())
print("Gen RAM Free: " + humanize.naturalsize( psutil.virtual_memory().available ), " I Proc size: " + humanize.naturalsize( process.memory_info().rss))
print('GPU RAM Free: {0:.0f}MB | Used: {1:.0f}MB | Util {2:3.0f}% | Total {3:.0f}MB'.format(gpu.memoryFree, gpu.memoryUsed, gpu.memoryUtil*100, gpu.memoryTotal))
printm()
Hey, from what I understood, everytime you start a notebook on colab all that setup and installation process has to happen. Which means just to start working on it you spent around an hour on just installation process. Please clarify me if Iām wrong.
Hey Shubham, itās not that bad. From my experience setting up the environment takes up to 10 minutes, can be almost fully automated and needs to be done roughly once per day (since virtual machine can live up to 12 hours).
Google Colab is a fantastic resource. It just takes time to get used to it. For studying (and some development) Iād say itās more convenient than AWS.
The biggest drawback is limited RAM (12GB) so you need to be aware of your notebookās memory consumption (and mitigate it) at all times.
Well, first you have to visualize your Colab Notebook as a cloud service. Automatically they have some packages installed, like Jupyter Notebook, but not something specific like fast.ai library.
Based on that you can use the !{command-line string} to run some code directly in the bash using only your notebook. If you need some data you can download using a command line.
Example: The fisrt lesson notebook need fast.ai library installled and the data folder dogscats. To install everything correct I use before any code this sequences of commands:
!pip install git+https://github.com/fastai/fastai.git
!wget http://files.fast.ai/data/dogscats.zip
!unzip dogscats.zip
!mkdir data
!mv dogscats data
@prakashgupta, I donāt quite understand the use of Clouderizer. I understand that it provides scripts which are imported to Google colab to create the environment required for the course but what does it happen with the three folders and the sync to Drive? I donāt understand if, say, I download the dataset for the first lesson, will it be saved to my Drive account or to Clouderizer or none of them?. Thank you for your help in the forum, I read your previous posts but I couldnāt understand them completely.