Skip to content

/root/fairseq/examples/speech_recognition/w2l_decoder.py:41: UserWarning: wav2letter python bindings are required to use this functionality. #2

@youssefavx

Description

@youssefavx

Hi! Thank you so much for setting up this docker, I've been looking for so long for some sort of simple way to test this model, just put input audio in and get text back and it's ridiculous how complex it all is to set up. This whole thing should've been packaged into a pip install line! so thank you for your work!

After running the installation example:

docker build -t wav2vec2 -f wav2letter.Dockerfile .
docker run -d -it --rm -v $PWD/data:/root/data --name w2v2 wav2vec2
docker exec -it w2v2 bash

Everything went smoothly up til this line:

python examples/wav2vec/recognize.py --wav_path /root/data/temp.wav --w2v_path /root/data/wav2vec_small_10m.pt --target_dict_path /root/data/dict.ltr.txt

Which I modified to this (because it seemed that the wav file was not found:

#I first cd'd back to the root, and did:

python3 fairseq/examples/wav2vec/recognize.py --wav_path /root/data/temp.wav --w2v_path /root/data/wav2vec2_vox_960h.pt --target_dict_path /root/data/dict.ltr.txt

Now I get this error:

/root/fairseq/examples/speech_recognition/w2l_decoder.py:41: UserWarning: wav2letter python bindings are required to use this functionality. Please install from https://github.com/facebookresearch/wav2letter/wiki/Python-bindings
  "wav2letter python bindings are required to use this functionality. Please install from https://github.com/facebookresearch/wav2letter/wiki/Python-bindings"

I also get this error right above the previous error even though I have CUDA 9 or 10 I believe:

/usr/local/lib/python3.6/dist-packages/torch/cuda/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx (Triggered internally at  /pytorch/c10/cuda/CUDAFunctions.cpp:100.)
  return torch._C._cuda_getDeviceCount() > 0

I do remember that to get cuda working on a mac with Pytorch you do have to build Pytorch from source which also seems like another 10 thousand steps :(. Let alone integrating that into docker somehow which I don't understand.

Anyway, would appreciate any help whatsoever to get this working. Hopefully this thing can run without CUDA. At the moment I don't want to fine tune the model, simply test.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions