-
Notifications
You must be signed in to change notification settings - Fork 90
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not working on any other video other than Kennedy or Mona #24
Comments
I'm having the same problem as you. Did you find a solution? @Tauseefahmed1451 @saifhassan |
anyone found ? @oykuparlakk |
same here @saifhassan any recommendations? |
I have this problem too |
Unfortunately I couldn't find it @gutscdav000 @AIhasArrived @start-life |
it's sufficient uncomment (and slightly modify) the last line of the command that creates the hd video is the commented
I'm not able yet to fix passing global variable, but if you uncomment and modify with values it should work properly. E.g. with
|
@lbdave94 I used your command: |
@tranthuhoai3786 I think you just have to change the |
I have followed all the instructions and downloaded every model and placed it at defined places. Now when I run the sh file, it works if the name is kennedy and the audio is also the given one. Although if I only change the audio, it seems to process everything but does not store the output video. Lastly, if I change the video (my videos are also real humans with clear faces and no obstructions, looking into the camera) it fails at extracting frames. Also, torchvision gives a warning about deprecated parameter
This is the entire output
(lipsync) (base) root@07cb68bb-cd8a-4674-8c71-08d02fdf1aaf:/opt/trainml/Wav2Lip-HD(main)# bash run_final.shhey... /opt/trainml/Wav2Lip-HD hey... /opt/trainml/Wav2Lip-HD hey... /opt/trainml/Wav2Lip-HD Using cuda for inference. Traceback (most recent call last): File "/opt/trainml/Wav2Lip-HD/inference.py", line 359, in <module> main() File "/opt/trainml/Wav2Lip-HD/inference.py", line 263, in main mel = audio.melspectrogram(wav) File "/opt/trainml/Wav2Lip-HD/audio.py", line 47, in melspectrogram S = _amp_to_db(_linear_to_mel(np.abs(D))) - hp.ref_level_db File "/opt/trainml/Wav2Lip-HD/audio.py", line 95, in _linear_to_mel _mel_basis = _build_mel_basis() File "/opt/trainml/Wav2Lip-HD/audio.py", line 100, in _build_mel_basis return librosa.filters.mel(hp.sample_rate, hp.n_fft, n_mels=hp.num_mels, TypeError: mel() takes 0 positional arguments but 2 positional arguments (and 3 keyword-only arguments) were given Frames extracted and stored at frames_wav2lip/cot /opt/trainml/lipsync/lib/python3.10/site-packages/torchvision/transforms/functional_tensor.py:5: UserWarning: The torchvision.transforms.functional_tensor module is deprecated in 0.15 and will be **removed in 0.17**. Please don't rely on it. You probably just need to use APIs in torchvision.transforms.functional or in torchvision.transforms.v2.functional. warnings.warn( /opt/trainml/lipsync/lib/python3.10/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead. warnings.warn( /opt/trainml/lipsync/lib/python3.10/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or
Nonefor 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing
weights=None. warnings.warn(msg)
The text was updated successfully, but these errors were encountered: