FIX: F.binary_cross_entropy keeps crashing the GPU

I have a similar problem (see here: Structured Learner) but so far I could not find a good hint on what is wrong.

I checked my categorical variables and found a mistake (index was not set) that I corrected.
Also reducing the bs did not help (I also thought maybe I’m out of gpu ram and pytorch was not able to load it properly).

I use is_multi=False because I want to do 1/0 classification.

With this setup my final layer looks like this:
A.) output of learn.model:

  (outp): Linear(in_features=500, out_features=1, bias=True)
  (emb_drop): Dropout(p=0.4)
  (drops): ModuleList(
    (0): Dropout(p=0.001)
    (1): Dropout(p=0.01)

B.) output of learn.crit:
<function torch.nn.functional.nll_loss(input, target, weight=None, size_average=True, ignore_index=-100, reduce=True)>

Has somebody an idea where I should look for the error?