You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
load models models/Whisper/faster-whisper
Error transcribing file on line parallel_for failed: cudaErrorNoKernelImageForDevice: no kernel image is available for execution on the device
/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/torch/cuda/memory.py:303: FutureWarning: torch.cuda.reset_max_memory_allocated now calls torch.cuda.reset_peak_memory_stats, which resets /all/ peak memory stats.
warnings.warn(
Traceback (most recent call last):
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/queueing.py", line 501, in call_prediction
output = await route_utils.call_process_api(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/route_utils.py", line 253, in call_process_api
output = await app.get_blocks().process_api(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/blocks.py", line 1704, in process_api
data = await anyio.to_thread.run_sync(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync
return await get_async_backend().run_sync_in_worker_thread(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2144, in run_sync_in_worker_thread
return await future
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 851, in run
result = context.run(func, *args)
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/blocks.py", line 1460, in postprocess_data
self.validate_outputs(fn_index, predictions) # type: ignore
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/blocks.py", line 1434, in validate_outputs
raise ValueError(
ValueError: An event handler (transcribe_file) didn't receive enough output values (needed: 2, received: 1).
The text was updated successfully, but these errors were encountered:
load models models/Whisper/faster-whisper
Error transcribing file on line parallel_for failed: cudaErrorNoKernelImageForDevice: no kernel image is available for execution on the device
/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/torch/cuda/memory.py:303: FutureWarning: torch.cuda.reset_max_memory_allocated now calls torch.cuda.reset_peak_memory_stats, which resets /all/ peak memory stats.
warnings.warn(
Traceback (most recent call last):
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/queueing.py", line 501, in call_prediction
output = await route_utils.call_process_api(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/route_utils.py", line 253, in call_process_api
output = await app.get_blocks().process_api(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/blocks.py", line 1704, in process_api
data = await anyio.to_thread.run_sync(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync
return await get_async_backend().run_sync_in_worker_thread(
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2144, in run_sync_in_worker_thread
return await future
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 851, in run
result = context.run(func, *args)
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/blocks.py", line 1460, in postprocess_data
self.validate_outputs(fn_index, predictions) # type: ignore
File "/home/ai/other/miniconda3/envs/whisper-webui/lib/python3.10/site-packages/gradio/blocks.py", line 1434, in validate_outputs
raise ValueError(
ValueError: An event handler (transcribe_file) didn't receive enough output values (needed: 2, received: 1).
The text was updated successfully, but these errors were encountered: