You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Guessed Channel Layout for Input Stream #0.0 : mono
Input #0, wav, from '/tmp/gradio/178d11976f94d691ff57dd90a34c7603b4309ea3/acknowledgement_english.wav':
Duration: 00:00:16.57, bitrate: 768 kb/s
Stream #0:0: Audio: pcm_s16le ([1][0][0][0] / 0x0001), 48000 Hz, mono, s16, 768 kb/s
Stream mapping:
Stream #0:0 -> #0:0 (pcm_s16le (native) -> pcm_s16le (native))
Press [q] to stop, [?] for help
-async is forwarded to lavfi similarly to -af aresample=async=1:min_hard_comp=0.100000:first_pts=0.
Output #0, wav, to 'tmp/lipsynced_result_20240306024455/lipsynced_result_20240306024455_16K.wav':
Metadata:
ISFT : Lavf58.76.100
Stream #0:0: Audio: pcm_s16le ([1][0][0][0] / 0x0001), 16000 Hz, mono, s16, 256 kb/s
Metadata:
encoder : Lavc58.134.100 pcm_s16le
size= 518kB time=00:00:16.57 bitrate= 256.1kbits/s speed= 905x
video:0kB audio:518kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: 0.014706%
Some weights of Wav2Vec2Model were not initialized from the model checkpoint at jonatasgrosman/wav2vec2-large-xlsr-53-english and are newly initialized: ['wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original0', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original1']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
2024-03-06 02:45:04.494899: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-03-06 02:45:04.494954: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-03-06 02:45:04.496171: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-03-06 02:45:05.725115: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
/usr/local/lib/python3.10/dist-packages/torch/nn/functional.py:4296: UserWarning: Default grid_sample and affine_grid behavior has changed to align_corners=False since 1.3.0. Please specify align_corners=True if the old behavior is desired. See the documentation of grid_sample for details.
warnings.warn(
Traceback (most recent call last):
File "/content/drive/MyDrive/dreamtalk/inference_for_demo_video.py", line 230, in
render_video(
File "/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "/content/drive/MyDrive/dreamtalk/generators/utils.py", line 112, in render_video
torchvision.io.write_video(silent_video_path, transformed_imgs.cpu(), fps)
File "/usr/local/lib/python3.10/dist-packages/torchvision/io/video.py", line 134, in write_video
container.mux(packet)
File "av/container/output.pyx", line 211, in av.container.output.OutputContainer.mux
File "av/container/output.pyx", line 217, in av.container.output.OutputContainer.mux_one
File "av/container/output.pyx", line 172, in av.container.output.OutputContainer.start_encoding
File "av/error.pyx", line 336, in av.error.err_check
av.error.FileNotFoundError: [Errno 2] No such file or directory
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/gradio/queueing.py", line 495, in call_prediction
output = await route_utils.call_process_api(
File "/usr/local/lib/python3.10/dist-packages/gradio/route_utils.py", line 232, in call_process_api
output = await app.get_blocks().process_api(
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1561, in process_api
result = await self.call_function(
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1179, in call_function
prediction = await anyio.to_thread.run_sync(
File "/usr/local/lib/python3.10/dist-packages/anyio/to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 807, in run
result = context.run(func, *args)
File "/usr/local/lib/python3.10/dist-packages/gradio/utils.py", line 695, in wrapper
response = f(*args, **kwargs)
File "/content/drive/MyDrive/dreamtalk/app.py", line 47, in infer
execute_command(command)
File "/content/drive/MyDrive/dreamtalk/app.py", line 27, in execute_command
subprocess.run(command, check=True)
File "/usr/lib/python3.10/subprocess.py", line 526, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['python', 'inference_for_demo_video.py', '--wav_path=/tmp/gradio/178d11976f94d691ff57dd90a34c7603b4309ea3/acknowledgement_english.wav', '--style_clip_path=data/style_clip/3DMM/M030_front_neutral_level1_001.mat', '--pose_path=data/pose/RichardShelby_front_neutral_level1_001.mat', '--image_path=/tmp/gradio/077f22810adc22e2aaf724be2c6e65713593a23d/cut_img.png', '--cfg_scale=1.0', '--max_gen_len=30', '--output_name=lipsynced_result_20240306024455']' returned non-zero exit status 1.
The text was updated successfully, but these errors were encountered:
Guessed Channel Layout for Input Stream #0.0 : mono
Input #0, wav, from '/tmp/gradio/178d11976f94d691ff57dd90a34c7603b4309ea3/acknowledgement_english.wav':
Duration: 00:00:16.57, bitrate: 768 kb/s
Stream #0:0: Audio: pcm_s16le ([1][0][0][0] / 0x0001), 48000 Hz, mono, s16, 768 kb/s
Stream mapping:
Stream #0:0 -> #0:0 (pcm_s16le (native) -> pcm_s16le (native))
Press [q] to stop, [?] for help
-async is forwarded to lavfi similarly to -af aresample=async=1:min_hard_comp=0.100000:first_pts=0.
Output #0, wav, to 'tmp/lipsynced_result_20240306024455/lipsynced_result_20240306024455_16K.wav':
Metadata:
ISFT : Lavf58.76.100
Stream #0:0: Audio: pcm_s16le ([1][0][0][0] / 0x0001), 16000 Hz, mono, s16, 256 kb/s
Metadata:
encoder : Lavc58.134.100 pcm_s16le
size= 518kB time=00:00:16.57 bitrate= 256.1kbits/s speed= 905x
video:0kB audio:518kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: 0.014706%
Some weights of Wav2Vec2Model were not initialized from the model checkpoint at jonatasgrosman/wav2vec2-large-xlsr-53-english and are newly initialized: ['wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original0', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original1']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
2024-03-06 02:45:04.494899: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-03-06 02:45:04.494954: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-03-06 02:45:04.496171: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-03-06 02:45:05.725115: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
/usr/local/lib/python3.10/dist-packages/torch/nn/functional.py:4296: UserWarning: Default grid_sample and affine_grid behavior has changed to align_corners=False since 1.3.0. Please specify align_corners=True if the old behavior is desired. See the documentation of grid_sample for details.
warnings.warn(
Traceback (most recent call last):
File "/content/drive/MyDrive/dreamtalk/inference_for_demo_video.py", line 230, in
render_video(
File "/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "/content/drive/MyDrive/dreamtalk/generators/utils.py", line 112, in render_video
torchvision.io.write_video(silent_video_path, transformed_imgs.cpu(), fps)
File "/usr/local/lib/python3.10/dist-packages/torchvision/io/video.py", line 134, in write_video
container.mux(packet)
File "av/container/output.pyx", line 211, in av.container.output.OutputContainer.mux
File "av/container/output.pyx", line 217, in av.container.output.OutputContainer.mux_one
File "av/container/output.pyx", line 172, in av.container.output.OutputContainer.start_encoding
File "av/error.pyx", line 336, in av.error.err_check
av.error.FileNotFoundError: [Errno 2] No such file or directory
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/gradio/queueing.py", line 495, in call_prediction
output = await route_utils.call_process_api(
File "/usr/local/lib/python3.10/dist-packages/gradio/route_utils.py", line 232, in call_process_api
output = await app.get_blocks().process_api(
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1561, in process_api
result = await self.call_function(
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1179, in call_function
prediction = await anyio.to_thread.run_sync(
File "/usr/local/lib/python3.10/dist-packages/anyio/to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 807, in run
result = context.run(func, *args)
File "/usr/local/lib/python3.10/dist-packages/gradio/utils.py", line 695, in wrapper
response = f(*args, **kwargs)
File "/content/drive/MyDrive/dreamtalk/app.py", line 47, in infer
execute_command(command)
File "/content/drive/MyDrive/dreamtalk/app.py", line 27, in execute_command
subprocess.run(command, check=True)
File "/usr/lib/python3.10/subprocess.py", line 526, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['python', 'inference_for_demo_video.py', '--wav_path=/tmp/gradio/178d11976f94d691ff57dd90a34c7603b4309ea3/acknowledgement_english.wav', '--style_clip_path=data/style_clip/3DMM/M030_front_neutral_level1_001.mat', '--pose_path=data/pose/RichardShelby_front_neutral_level1_001.mat', '--image_path=/tmp/gradio/077f22810adc22e2aaf724be2c6e65713593a23d/cut_img.png', '--cfg_scale=1.0', '--max_gen_len=30', '--output_name=lipsynced_result_20240306024455']' returned non-zero exit status 1.
The text was updated successfully, but these errors were encountered: