Skip to content

Commit f873c7d

Browse files
crazydemoLarryXFly
authored andcommitted
[None][test] Clean cache for certain easily hang cases (NVIDIA#8619)
Signed-off-by: Ivy Zhang <25222398+crazydemo@users.noreply.github.com> Co-authored-by: Larry Xu <197874197+LarryXFly@users.noreply.github.com> Signed-off-by: Mike Iovine <6158008+mikeiovine@users.noreply.github.com>
1 parent 4cd87b1 commit f873c7d

File tree

5 files changed

+6
-5
lines changed

5 files changed

+6
-5
lines changed

tests/integration/defs/conftest.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2679,6 +2679,7 @@ def torch_empty_cache() -> None:
26792679
Manually empty the torch CUDA cache before each test, to reduce risk of OOM errors.
26802680
"""
26812681
if torch.cuda.is_available():
2682+
gc.collect()
26822683
torch.cuda.empty_cache()
26832684
gc.collect()
26842685

tests/integration/defs/test_e2e.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2273,7 +2273,7 @@ def test_ptp_quickstart_advanced_deepseek_r1_w4afp8_8gpus(
22732273

22742274
@pytest.mark.skip_less_device_memory(80000)
22752275
@pytest.mark.parametrize("model_name,model_path,gpu_count", [
2276-
("Llama3.1-70B-BF16", "llama-3.1-model/Meta-Llama-3.1-70B", 2),
2276+
("Llama3.1-70B-BF16", "llama-3.1-model/Meta-Llama-3.1-70B", 8),
22772277
("Mixtral-8x7B-BF16", "Mixtral-8x7B-v0.1", 8),
22782278
pytest.param('Llama3.1-70B-FP8',
22792279
'llama-3.1-model/Llama-3.1-70B-Instruct-FP8',
@@ -2304,7 +2304,7 @@ def test_ptp_quickstart_advanced_multi_gpus(llm_root, llm_venv, model_name,
23042304
pytest.skip(f"Not enough GPUs for {model_name}")
23052305
example_root = Path(os.path.join(llm_root, "examples", "llm-api"))
23062306
mapping = {
2307-
"Llama3.1-70B-BF16": 91.0,
2307+
"Llama3.1-70B-BF16": 24.6,
23082308
"Mixtral-8x7B-BF16": 16.5,
23092309
"Llama3.1-70B-FP8": 58.5,
23102310
"Llama3.1-405B-FP8": 63.2,

tests/integration/test_lists/qa/llm_function_core.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -657,7 +657,7 @@ test_e2e.py::test_ptp_quickstart_advanced[Llama3.1-8B-NVFP4-nvfp4-quantized/Meta
657657
test_e2e.py::test_ptp_quickstart_advanced[Llama3.2-11B-BF16-llama-3.2-models/Llama-3.2-11B-Vision]
658658
test_e2e.py::test_ptp_quickstart_advanced[Qwen3-30B-A3B-Qwen3/Qwen3-30B-A3B]
659659
test_e2e.py::test_ptp_quickstart_advanced_ngram[Llama-3.1-8B-Instruct-llama-3.1-model/Llama-3.1-8B-Instruct]
660-
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-BF16-llama-3.1-model/Meta-Llama-3.1-70B-2]
660+
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-BF16-llama-3.1-model/Meta-Llama-3.1-70B-8]
661661
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-FP8-llama-3.1-model/Llama-3.1-70B-Instruct-FP8-2]
662662
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-405B-FP8-llama-3.1-model/Llama-3.1-405B-Instruct-FP8-8]
663663
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Mixtral-8x7B-BF16-Mixtral-8x7B-v0.1-8]

tests/integration/test_lists/qa/llm_function_core_sanity.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -233,7 +233,7 @@ test_e2e.py::test_openai_consistent_chat
233233
test_e2e.py::test_openai_multi_chat_example
234234
test_e2e.py::test_ptp_quickstart
235235
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-405B-FP8-llama-3.1-model/Llama-3.1-405B-Instruct-FP8-8]
236-
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-BF16-llama-3.1-model/Meta-Llama-3.1-70B-2]
236+
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-BF16-llama-3.1-model/Meta-Llama-3.1-70B-8]
237237
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-FP8-llama-3.1-model/Llama-3.1-70B-Instruct-FP8-2]
238238
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Mixtral-8x7B-BF16-Mixtral-8x7B-v0.1-8]
239239
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Mixtral-8x7B-NVFP4-nvfp4-quantized/Mixtral-8x7B-Instruct-v0.1-8]

tests/integration/test_lists/qa/llm_function_nim.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -427,7 +427,7 @@ test_e2e.py::test_ptp_quickstart_advanced[Llama3.1-8B-BF16-llama-3.1-model/Meta-
427427
test_e2e.py::test_ptp_quickstart_advanced[Llama3.1-8B-FP8-llama-3.1-model/Llama-3.1-8B-Instruct-FP8]
428428
test_e2e.py::test_ptp_quickstart_advanced[Llama3.1-8B-NVFP4-nvfp4-quantized/Meta-Llama-3.1-8B]
429429
test_e2e.py::test_ptp_quickstart_advanced[Qwen3-30B-A3B-Qwen3/Qwen3-30B-A3B]
430-
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-BF16-llama-3.1-model/Meta-Llama-3.1-70B-2]
430+
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-BF16-llama-3.1-model/Meta-Llama-3.1-70B-8]
431431
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-70B-FP8-llama-3.1-model/Llama-3.1-70B-Instruct-FP8-2]
432432
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Llama3.1-405B-FP8-llama-3.1-model/Llama-3.1-405B-Instruct-FP8-8]
433433
test_e2e.py::test_ptp_quickstart_advanced_multi_gpus[Mixtral-8x7B-BF16-Mixtral-8x7B-v0.1-8]

0 commit comments

Comments
 (0)