diff --git a/docs/source/models/supported_models.md b/docs/source/models/supported_models.md index e498efc22086..7145bcf2d5f5 100644 --- a/docs/source/models/supported_models.md +++ b/docs/source/models/supported_models.md @@ -938,6 +938,26 @@ The following table lists those that are tested in vLLM. * ✅︎ ::: +#### Transcription (`--task transcription`) + +Speech2Text models trained specifically for Automatic Speech Recognition. + +:::{list-table} +:widths: 25 25 25 5 5 +:header-rows: 1 + +- * Architecture + * Models + * Example HF Models + * [LoRA](#lora-adapter) + * [PP](#distributed-serving) +- * `Whisper` + * Whisper-based + * `openai/whisper-large-v3-turbo` + * 🚧 + * 🚧 +::: + _________________ ## Model Support Policy diff --git a/vllm/entrypoints/llm.py b/vllm/entrypoints/llm.py index 73593f0c6f0a..40b7a529ebfb 100644 --- a/vllm/entrypoints/llm.py +++ b/vllm/entrypoints/llm.py @@ -421,7 +421,7 @@ def generate( instead pass them via the ``inputs`` parameter. """ runner_type = self.llm_engine.model_config.runner_type - if runner_type != "generate": + if runner_type not in ["generate", "transcription"]: messages = [ "LLM.generate() is only supported for (conditional) generation " "models (XForCausalLM, XForConditionalGeneration).",