Replies: 13 comments
-
>>> markmp |
Beta Was this translation helpful? Give feedback.
-
>>> lissyx |
Beta Was this translation helpful? Give feedback.
-
>>> lissyx |
Beta Was this translation helpful? Give feedback.
-
>>> lissyx |
Beta Was this translation helpful? Give feedback.
-
>>> lissyx |
Beta Was this translation helpful? Give feedback.
-
>>> lissyx |
Beta Was this translation helpful? Give feedback.
-
>>> markmp |
Beta Was this translation helpful? Give feedback.
-
>>> markmp |
Beta Was this translation helpful? Give feedback.
-
>>> markmp |
Beta Was this translation helpful? Give feedback.
-
>>> markmp |
Beta Was this translation helpful? Give feedback.
-
>>> lissyx |
Beta Was this translation helpful? Give feedback.
-
>>> markmp |
Beta Was this translation helpful? Give feedback.
-
>>> lissyx |
Beta Was this translation helpful? Give feedback.
-
>>> markmp
[March 7, 2018, 5:37am]
Hi,
After much wrangling, I was able to get DeepSpeech working on the new
Amazon V100 Nvidia instances (p3.2xlarge). The inference seems quite
slow though - I seem to get .95x - 1.25x real-time in order to do
inference on V100...i.e. about 2 seconds for inference on a 2 second
audio clip. This is a top card, and this seems much slower than what
others seem to be reporting (closer to 0.3x-0.4x).
For comparison, the CPU seems to take 2x to 2.5x to do the same
inference. Really surprised the V100 isn't performing better and I'm
wondering if I'm doing something suboptimal.
Getting the CPU inference going was fine - but the GPU inference was
very frustrating because of mismatches between tensorflow, deepspeech,
cuda, and cudnn. In the end, the only config I could get running was: slash
pip install slash 'tensorflow-gpu==1.5.0' slash
pip install deepspeech-gpu (the pypi package doesn't work - so I used
the artifact here:
https://tools.taskcluster.net/index/project.deepspeech.deepspeech.native_client.master/gpu) slash
manually install Cuda 9.0 slash
manually install cudnn 7.0.5
Is anyone getting faster performance on the amazon v100? Or is .95x the
best I can hope for?
thx
[This is an archived TTS discussion thread from discourse.mozilla.org/t/inference-time-on-v100-seems-slow]
Beta Was this translation helpful? Give feedback.
All reactions