Skip to content

Commit

Permalink
mark llama_inference test as nightly (#677)
Browse files Browse the repository at this point in the history
  • Loading branch information
kamalrajkannan78 authored Nov 13, 2024
1 parent ec8ee2c commit 1f38fd4
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion forge/test/mlir/llama/test_llama_inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@
from test.mlir.llama.utils.utils import load_model


@pytest.mark.push
@pytest.mark.nightly
@pytest.mark.xfail()
@pytest.mark.parametrize("model_path", ["openlm-research/open_llama_3b", "meta-llama/Llama-3.2-1B"])
def test_llama_inference(model_path):
Expand Down
2 changes: 1 addition & 1 deletion forge/test/mlir/llama/tests/test_llama_prefil.py
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ def decode_on_cpu(model, tokenizer, input_ids, hidden_states, max_new_tokens):

@pytest.mark.parametrize("model_path", ["openlm-research/open_llama_3b", "meta-llama/Llama-3.2-1B"])
@pytest.mark.xfail()
@pytest.mark.push
@pytest.mark.nightly
def test_llama_prefil_on_device_decode_on_cpu(model_path):
"""
This function tests the inference of the Llama models split into two parts:
Expand Down

0 comments on commit 1f38fd4

Please sign in to comment.