I have nothing to compare to, but I have a feeling there is something wrong with my GPU machine. I currently SSH into it via my Mac and run notebook from the SSH Terminal. The GPU machine is running Ubuntu 16.04, and is using an nVidia GTX 1070 w/ 8Gb RAM. 16 gb RAM, i5-6600 3.3 GHz Quad-Core.
As I was running the data sets, this is what I got:
vgg = Vgg16()
vgg.fit(batches, val_batches, nb_epoch=1)
`output_shape` argument not specified for layer lambda_3 and cannot be automatically inferred with the
Theano backend. Defaulting to output shape `(None, 3, 224, 224)` (same as input shape). If the expected
output shape is different, specify it via the `output_shape` argument.
Found 23000 images belonging to 2 classes.
Found 2000 images belonging to 2 classes.
23000/23000 [==============================] -
4724s - loss: 0.1153 - acc: 0.9683 - val_loss: 0.0472 - val_acc: 0.9835
It took 4724 seconds to run this, and that seems like a long time. Am I missing something?
I also checked the GPU machine physically, and the fan didn't even come on the GPU when it was processing this. Should this have kicked in a fan at least?
I installed the nVidia drivers and CDNN, but how can I tell if the GPU is running them?
My first run with a GPU server via SSH, so not sure what to expect. Thanks for the help out there.