Hi, Having done some experiments with TensorFlow on CPU I’ve installed tensorflow-gpu within Docker with nvidia-docker and the tests run nicely. Could somebody please point me to documentation or a HowTo to associate OpenNMT-tf with this set-up.
Hi @guillaumekln , It just kept on building until it filled the disk. Can opennmt-tf really occupy 824.5 GB? And I don’t understand why it should be writing into an OpenNMT subdirectory to which no reference was made in the Dockerfile. Any idea what could be happening?
Regards,
Terence
Hi,
I’m trying to install OpenNMT with Tensorflow-gpu using Docker.
My Dockerfile looks like this
FROM tensorflow/tensorflow:1.14.0-gpu
WORKDIR /root
RUN pip install OpenNMT-tf
ENTRYPOINT []
The first command docker build -t opennmt / opennmt-tf -f dockerfile . do not worry me. The second one gives me error messages and I have no idea how to get them away
WARNING: Logging before flag parsing goes to stderr.
W0829 15:45:47.053921 139708002342720 deprecation_wrapper.py:119] From /usr/local/lib/python2.7/dist-packages/opennmt/decoders/rnn_decoder.py:435: The name tf.nn.rnn_cell.RNNCell is deprecated. Please use tf.compat.v1.nn.rnn_cell.RNNCell instead.
W0829 15:45:47.517523 139708002342720 deprecation_wrapper.py:119] From /usr/local/lib/python2.7/dist-packages/opennmt/optimizers/adafactor.py:32: The name tf.train.Optimizer is deprecated. Please use tf.compat.v1.train.Optimizer instead.
W0829 15:45:47.518002 139708002342720 deprecation_wrapper.py:119] From /usr/local/lib/python2.7/dist-packages/opennmt/optimizers/multistep_adam.py:36: The name tf.train.AdamOptimizer is deprecated. Please use tf.compat.v1.train.AdamOptimizer instead.
My first attempt with Opennmt Lua with GPU and docker worked fine. Was probably easier too
Hi, It’s not quite a direct answer, but have you tried nmtwizard/opennmt-tf? I find that works nicely. Here’s my command for inference via a server on a laptop without a CPU:
docker run -p 5000:5000 -v e:\tf_models\serving:/tf_models/serving nmtwizard/opennmt-tf --model %1 --model_storage /tf_models/serving serve --host 0.0.0.0 --port 5000
Hi, thanks for the tip. I tried it first without mounting directories, but it does not work.
The easy way: sudo nvidia-docker run nmtwizard/opennmt-tf
I get some warnings and this error message: entrypoint.py: error: too few arguments.
Hi, Sorry: the script I gave you was for running on my Windows laptop without a GPU. The Linux GPU command is:
CUDA_VISIBLE_DEVICES=0 nvidia-docker run -p 5000:5000 -v$PWD:/home/miguel nmtwizard/opennmt-tf --model ned2eng_0104 --model_storage /home/miguel/tf_experiments/ned2eng_tf/serving serve --host 0.0.0.0 --port 5000
Both are running as I write this. You need to make sure everything is adapted to your local system.