This repository was archived by the owner on Dec 11, 2023. It is now read-only.

Description
I have running the training with the following command:
python -m nmt.nmt \
--src=vi --tgt=en \
--vocab_prefix=/tmp/nmt_data/vocab \
--train_prefix=/tmp/nmt_data/train \
--dev_prefix=/tmp/nmt_data/tst2012 \
--test_prefix=/tmp/nmt_data/tst2013 \
--out_dir=/tmp/nmt_model \
--num_train_steps=12000 \
--steps_per_stats=100 \
--num_layers=2 \
--num_units=128 \
--dropout=0.2 \
--metrics=bleu \
--nums_gpu=1
I have one GPU (GPU Radeon RX 580). Upon running the experimentation, I see the CPUs are fully utilized and the GPU usage remains insignificant(<5%).
I saw this in the log:
Devices visible to TensorFlow: [_DeviceAttributes(/job:localhost/replica:0/task:0/device:CPU:0, CPU, 268435456)]
Can anyone provide any pointer why GPU usage remains low?