Part 2 Lesson 8 wiki

(Vijay Kumar) #293

Hope you have selected the interpreter, with fastai environment (environment.yml) available with downloaded code.
Once interpreter set you are able to navigate.

(Pierre Guillou) #294

[ EDIT ] : it works :slight_smile: The problem came from the selection of the python interpreter (ctrl+shift-p : Python: Select Interpreter).
The default path to my fastai environement is well setup in my user parameters in Visual Studio Code ("python.pythonPath": "C:\\Users\\username\\Anaconda3\\envs\\fastai\\python.exe",) but I have to select it (ctrl+shift-p) after each restart of Visual Studio Code. Any advice to avoid that ?

Hi @Vijay and @Patrick, I can not get “Go to symbol (ctr+t)” works in Visual Studio Code.

  • I’m using Windows 10 and Visual Studio Code is working.
  • I did open my fastai folder and select the python interpreter of my fastai virtual environment (I’m using an NVIDIA GPU on Windows).
  • I downloaded ctags (universal ctags and I tried as well exuberant ctags) and unzip it in my fastai folder in a folder called ctags : C:\Users\username\fastai\ctags\ctags.exe
  • I updated my Windows PATH with the path to ctags.exe
  • I updated my user parameters in Visual Studio Code with : "python.workspaceSymbols.ctagsPath": "C:\\Users\\username\\fastai\\ctags\\ctags.exe",

What else can I do ? Thanks.

(Pierre Guillou) #295

Hi, my notes on lesson 8. Hope it can help new fastai fellows :slight_smile:

(Shubham Gupta) #296

Hey guys, check out my new blog on Introduction to Object Detection. Hope you enjoy it and feel free to comment in case of any queries or suggestions.

(Ramon) #297


I’m trying to extend the bounding boxes in lesson 8 by ‘rotated bounding boxes’. I’m doing this by passing in four coordinates top-right(x,y), bottom-right(x,y, bottom-left(x,y) and top-left(x,y).

By using the same code, only the first four values get passed through the different datasets and loaders:

tfms = tfms_from_model(f_model, sz, crop_type=CropType.NO, tfm_y=tfm_y, aug_tfms=augs)
md = ImageClassifierData.from_csv(PATH, JPEGS, BB_CSV, tfms=tfms, continuous=True, bs=4)
bbox = to_np(y[1])
[194. 368. 217. 400. 0. 0. 0. 0.]

How come? I’m trying to understand but I can’t figure it out. Hints are very appreciated, thanks in advance!

(Michael) #298

Hey @ramon,

I was starting with lecture 8 today and I guess it only needs 2x x,y = 4 coordinates for the box to be sufficiently defined.

Do you use training data with rotated bounding boxes?
With the Pascal dataset I would guess that it will learn the “aligned” bounding boxes.

Best regards

(Ramon) #299

Hi @MicPie, thanks for your reply. Yes I know the bounding boxes in the lecture only use two coordinates (top-left and bottom-right). But I’m trying to extend this for a different dataset that is using rotated bounding boxes. It’s defined by 8 values, the csv looks like ImageId, y1 x1 y2 x2 y3 x3 y4 x4

(Michael) #300

I am currently watching lesson 9 and there Jeremy talks about the transformations in the “Bbox only” section in the jupyter notebook. As you can see at the transformed/rotated pictures of the woman the bounding box is not rotated and is only resized (and still aligned vertically and horizontally).

I guess to rotate the bounding box you have to adapt the class “CoordTransform”.
In addition, you have to use “tfm_y=TfmType.COORD” to transform the coordinates of the bounding box too.

However, maybe the pixel transformation can be of use with TfmType.PIXEL (see class “TfmType”, should be covered in later lessons on image segmentation).

(Ramon) #301

Hi @MicPie, Yes indeed, class "CoordTransform’ is the part of the code which is replacing the values by zero’s. That’s what I was looking for, thanks for your help. I’m trying to add a new TfmType to handle rotated boxes. Maybe in the end it’s better to use the PIXEL type, but in this way I’m getting to understand the frame more.

(Ramon) #302

Fixed it, if anyone is interested let me know :wink:

(Michael) #303

Nice work, looks very nice!

Do you also were able to train the neural network?
I would be interested if you use the coordinates of the four corners or maybe the two corners with a angle.

Best regards

(Ramon) #304

Hi @MicPie,

Not yet, but Ill continue next weekend. Ill keep you updated.
By the way, eventually I’m trying to solve the Kaggle airbus challenge (, interested in teaming up?

Kind regards,

(Thomas) #305

I am interested

(Ramon) #306

Hi @tcapelle,

Thanks for your response. Interested in the code and/or teaming up?


(David) #307

Would be interested in teaming up too!

(Ramon) #308

Great, I added you to the conversation, did you get that message?

(David) #309

…I did not…

(Thomas) #310

teaming up!
I have been playing with a kernel to detect ship presence first.


It seems there is no alias to data folder in the dl2 folder (contrary to dl1).

I just figured out how to create one on linux terminal to make the notebook work.
cd into dl2 folder, then :
“ln -s ~/data”


At the end of the pascal notebook, Jeremy unfreezes to layergroup -2. Then does lr find with the following result

From this graph, he chooses for the next iterations lr=5*10^-3

Can anyone explain to me wy he chooses such value ? (The graph doesn’t look like the usual “decrease then increase” graph…)

Thanks a lot !