Skip to content

Commit faa0ec2

Browse files
committed
fix: multi-GPU gather responses
Signed-off-by: Olya Kozlova <[email protected]>
1 parent 3902648 commit faa0ec2

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

tensorrt_llm/_torch/pyexecutor/py_executor.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1451,7 +1451,7 @@ def _enqueue_responses(self, responses: List[Tuple[int, LlmResponse]]):
14511451
if responses_list is not None:
14521452
for resp in responses_list:
14531453
if resp is not None:
1454-
gather_responses.append(resp)
1454+
gather_responses.extend(resp)
14551455
responses = gather_responses
14561456
logger.debug(
14571457
f'after gather, rank = {self.dist.rank}, responses = {responses}')

0 commit comments

Comments
 (0)