@@ -277,6 +277,7 @@ def create_rsp(id, finished: bool = False):
277
277
return tllm .Response (request_id = 0 , result = result , client_id = 0 )
278
278
279
279
280
+ @pytest .mark .skip (reason = "https://nvbugs/5477359" )
280
281
def test_GenerationResultBase ():
281
282
sampling_params = SamplingParams (max_tokens = 4 )
282
283
result = GenerationResultBase (
@@ -291,6 +292,7 @@ def test_GenerationResultBase():
291
292
assert result ._done
292
293
293
294
295
+ @pytest .mark .skip (reason = "https://nvbugs/5477359" )
294
296
def test_GenerationResult ():
295
297
request = GenerationRequest (prompt_token_ids = [12 , 23 , 34 ],
296
298
sampling_params = SamplingParams (max_tokens = 4 ))
@@ -303,6 +305,7 @@ def test_GenerationResult():
303
305
assert result ._done
304
306
305
307
308
+ @pytest .mark .skip (reason = "https://nvbugs/5477359" )
306
309
def test_DetokenizedGenerationResultBase ():
307
310
sampling_params = SamplingParams (max_tokens = 4 )
308
311
model_path = llm_models_root () / "llama-models/llama-7b-hf"
@@ -434,6 +437,7 @@ def ResponsePostprocessWorker_worker_task(pull_pipe_addr, push_pipe_addr,
434
437
worker .start ()
435
438
436
439
440
+ @pytest .mark .skip (reason = "https://nvbugs/5477369" )
437
441
def test_ResponsePostprocessWorker ():
438
442
439
443
input_pipe = ZeroMqQueue (is_server = True )
0 commit comments