Lesson 3 - BIWI Head Pose - Not able to train on Augmented Dataset

Hi all,

I am running the Lesson - 3 - BIWI Head Pose notebook.
I noticed that the notebook suddenly ends after the creation of the augmented dataset. When I tried to use this dataset to further train my model, I am getting an error.

Here is my code:

learn2 = cnn_learner(data, models.resnet34)

It is while running the lr_find() function when I am getting the following error:

LR Finder is complete, type {learner_name}.recorder.plot() to see the graph.
RuntimeError                              Traceback (most recent call last)
<ipython-input-14-19b58a336fcf> in <module>()
----> 1 learn2.lr_find()

6 frames
/usr/local/lib/python3.6/dist-packages/torch/utils/data/dataloader.py in _process_next_batch(self, batch)
    606                 raise Exception("KeyError:" + batch.exc_msg)
    607             else:
--> 608                 raise batch.exc_type(batch.exc_msg)
    609         return batch

RuntimeError: Traceback (most recent call last):
  File "/usr/local/lib/python3.6/dist-packages/torch/utils/data/_utils/worker.py", line 99, in _worker_loop
    samples = collate_fn([dataset[i] for i in batch_indices])
  File "/usr/local/lib/python3.6/dist-packages/fastai/torch_core.py", line 127, in data_collate
    return torch.utils.data.dataloader.default_collate(to_data(batch))
  File "/usr/local/lib/python3.6/dist-packages/torch/utils/data/_utils/collate.py", line 68, in default_collate
    return [default_collate(samples) for samples in transposed]
  File "/usr/local/lib/python3.6/dist-packages/torch/utils/data/_utils/collate.py", line 68, in <listcomp>
    return [default_collate(samples) for samples in transposed]
  File "/usr/local/lib/python3.6/dist-packages/torch/utils/data/_utils/collate.py", line 43, in default_collate
    return torch.stack(batch, 0, out=out)
RuntimeError: invalid argument 0: Sizes of tensors must match except in dimension 0. Got 1 and 0 in dimension 1 at /pytorch/aten/src/TH/generic/THTensor.cpp:711

I have checked the dataset creation function and the only difference between this function and the previous one is that we are using a different custom transformation function.

The original dataset code snippet:

data = (PointsItemList.from_folder(path)
        .split_by_valid_func(lambda o: o.parent.name=='13')
        .transform(get_transforms(), tfm_y=True, size=(120,160))

The later dataset creation code snippet which is giving error:

tfms = get_transforms(max_rotate=20, max_zoom=1.5, max_lighting=0.5, max_warp=0.4, p_affine=1., p_lighting=1.)

data = (PointsItemList.from_folder(path)
        .split_by_valid_func(lambda o: o.parent.name=='13')
        .transform(tfms, tfm_y=True, size=(120,160))

If someone has already faced this issue before, please reply your solutions in this thread.

Same issue. Do you have any solutions now?

I have the same Problem

Unfortunately, no. I have been busy with course completion and my own personal projects to dive further into the issue.

But what I did find out was that if we do not include max_zoom or keep it as 1, then we do not face any issues.

tfms = get_transforms(max_rotate=20, max_zoom=1, max_lighting=0.5, max_warp=0.4, p_affine=1., p_lighting=1.)

But that defeats the purpose of transformation because now we are not zooming any of the images and only applying the other transformations.

But if you really want to continue to go forward with the dataset, i think this is an OK workaround.


Refer to my reply to @Jasmine. Hope that helps as a workaround.

Thank you for your workaround @AdityaGogoi . This works for further training. Hope we will figure out why the max_zoom matters later on.

Sure @Jasmine . I will keep investigating the issue and update this thread if I get something.

I found that it may be caused by the transformed coordinates exceeding the size of the image

1 Like

@huanling I found the same thing. If ImagePoint is transformed out of bounds the .data property becomes an empty tensor([], size=(0, 2)) which causes torch.stack to complain.