|
12 | 12 | from torch.cuda import device_count
|
13 | 13 |
|
14 | 14 | from tensorrt_llm import LLM as PyTorchLLM
|
15 |
| -from tensorrt_llm._torch.auto_deploy.llm import LLM as AutoDeployLLM |
16 | 15 | from tensorrt_llm import MultimodalEncoder
|
17 | 16 | from tensorrt_llm._tensorrt_engine import LLM
|
| 17 | +from tensorrt_llm._torch.auto_deploy.llm import LLM as AutoDeployLLM |
18 | 18 | from tensorrt_llm._utils import mpi_rank
|
19 | 19 | from tensorrt_llm.executor.utils import LlmLauncherEnvs
|
20 | 20 | from tensorrt_llm.llmapi import (BuildConfig, CapacitySchedulerPolicy,
|
@@ -211,10 +211,13 @@ def launch_mm_encoder_server(
|
211 | 211 | default="localhost",
|
212 | 212 | help="Hostname of the server.")
|
213 | 213 | @click.option("--port", type=int, default=8000, help="Port of the server.")
|
214 |
| -@click.option("--backend", |
215 |
| - type=click.Choice(["pytorch", "trt", "_autodeploy"]), |
216 |
| - default="pytorch", |
217 |
| - help="Set to 'pytorch' for pytorch path and '_autodeploy' for autodeploy path. Default is pytorch path.") |
| 214 | +@click.option( |
| 215 | + "--backend", |
| 216 | + type=click.Choice(["pytorch", "trt", "_autodeploy"]), |
| 217 | + default="pytorch", |
| 218 | + help= |
| 219 | + "Set to 'pytorch' for pytorch path and '_autodeploy' for autodeploy path. Default is pytorch path." |
| 220 | +) |
218 | 221 | @click.option('--log_level',
|
219 | 222 | type=click.Choice(severity_map.keys()),
|
220 | 223 | default='info',
|
|
0 commit comments