Skip to content

[BUG] 980ti: cudaErrorNoKernelImageForDevice #57

@kgd192

Description

@kgd192

Is there an existing issue for this?

  • I have searched the existing issues

Current Behavior

Unable to Transscripe audio with the following Error

RuntimeError: parallel_for failed: cudaErrorNoKernelImageForDevice: no kernel image is available for execution on the device

See #50

Expected Behavior

Analyze Audio Using GPU

Steps To Reproduce

Rund Docker Container link to HA and try to use Voice Assistant

Environment

- OS: Debian 12
- How docker service was installed: apt

CPU architecture

x86-64

Docker creation

services:
  faster-whisper:                                                                             image: lscr.io/linuxserver/faster-whisper:gpu-legacy                                      container_name: faster-whisper
    environment:
      - PUID=1004
      - PGID=1005                                                                               - TZ=Etc/UTC                                                                              - WHISPER_MODEL=small.en
      - LOCAL_ONLY= #optional
      - WHISPER_BEAM=20 #optional
      - WHISPER_LANG=en #optional
    volumes:
      - ./faster-whisper/data:/config
    ports:
      - 10300:10300
    restart: unless-stopped
    network_mode: host
    runtime: nvidia
    deploy:
      resources:
        reservations:
          devices:
            - driver: nvidia
              count: all
              capabilities:
                - gpu
                - utility
                - compute

Container logs

migrations] started
[migrations] no migrations found
───────────────────────────────────────

      ██╗     ███████╗██╗ ██████╗
      ██║     ██╔════╝██║██╔═══██╗
      ██║     ███████╗██║██║   ██║
      ██║     ╚════██║██║██║   ██║
      ███████╗███████║██║╚██████╔╝
      ╚══════╝╚══════╝╚═╝ ╚═════╝

   Brought to you by linuxserver.io
───────────────────────────────────────

To support LSIO projects visit:
https://www.linuxserver.io/donate/

───────────────────────────────────────
GID/UID
───────────────────────────────────────

User UID:    1004
User GID:    1005
───────────────────────────────────────
Linuxserver.io version: v3.0.2-ls21
Build-date: 2026-01-04T06:55:30+00:00
───────────────────────────────────────

[custom-init] No custom files found, skipping...
INFO:httpx:HTTP Request: GET https://huggingface.co/api/models/Systran/faster-whisper-small.en/revision/main "HTTP/1.1 200 OK"
[2026-01-07 20:29:30.524] [ctranslate2] [thread 211] [warning] The compute type inferred from the saved model is float16, but the target device or backend do not support efficient float16 computation. The model weights have been automatically converted to use the float32 compute type instead.
INFO:__main__:Ready
Connection to localhost (127.0.0.1) 10300 port [tcp/*] succeeded!
[ls.io-init] done.
INFO:faster_whisper:Processing audio with duration 00:02.600
ERROR:asyncio:Task exception was never retrieved
future: <Task finished name='wyoming event handler' coro=<AsyncEventHandler.run() done, defined at /lsiopy/lib/python3.12/site-packages/wyoming/server.py:31> exception=RuntimeError('parallel_for failed: cudaErrorNoKernelImageForDevice: no kernel image is available for execution on the device')>
Traceback (most recent call last):
  File "/lsiopy/lib/python3.12/site-packages/wyoming/server.py", line 41, in run
    if not (await self.handle_event(event)):
            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/lsiopy/lib/python3.12/site-packages/wyoming_faster_whisper/faster_whisper_handler.py", line 77, in handle_event
    text = " ".join(segment.text for segment in segments)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/lsiopy/lib/python3.12/site-packages/wyoming_faster_whisper/faster_whisper_handler.py", line 77, in <genexpr>
    text = " ".join(segment.text for segment in segments)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/lsiopy/lib/python3.12/site-packages/faster_whisper/transcribe.py", line 1190, in generate_segments
    encoder_output = self.encode(segment)
                     ^^^^^^^^^^^^^^^^^^^^
  File "/lsiopy/lib/python3.12/site-packages/faster_whisper/transcribe.py", line 1400, in encode
    return self.model.encode(features, to_cpu=to_cpu)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: parallel_for failed: cudaErrorNoKernelImageForDevice: no kernel image is available for execution on the device

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    Status

    Issues

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions