Skip to content

Commit eff7d18

Browse files
committed
waive a ray test
Signed-off-by: Yan Chunwei <[email protected]>
1 parent dff32c1 commit eff7d18

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

tests/unittest/llmapi/test_llm_pytorch.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -360,6 +360,7 @@ def _check_llama_7b_multi_lora_evict_load_new_adapters(
360360

361361

362362
@skip_gpu_memory_less_than_40gb
363+
@skip_ray # https://nvbugs/5682551
363364
def test_llama_7b_multi_lora_evict_and_reload_lora_gpu_cache():
364365
"""Test eviction and re-loading a previously evicted adapter from the LoRA GPU cache, within a single
365366
llm.generate call, that's repeated twice.
@@ -454,6 +455,7 @@ def test_llama_7b_peft_cache_config_affects_peft_cache_size():
454455
cuda_graph_config=None)
455456

456457

458+
@skip_ray # https://nvbugs/5682551
457459
@skip_gpu_memory_less_than_40gb
458460
def test_llama_7b_lora_config_overrides_peft_cache_config():
459461
"""Tests that cache size args in lora_config LLM arg override the cache size

0 commit comments

Comments
 (0)