diff --git a/examples/models/llama/runner/generation.py b/examples/models/llama/runner/generation.py index 67506990e1..159bc5f501 100644 --- a/examples/models/llama/runner/generation.py +++ b/examples/models/llama/runner/generation.py @@ -125,7 +125,7 @@ def text_completion( echo (bool, optional): Flag indicating whether to include prompt tokens in the generated output. Defaults to False. Returns: - CompletionPrediction: Completion prediction, which contains the generated text completion. + Generated list of tokens. Note: This method generates text completion for the provided prompt, employing nucleus sampling to introduce controlled randomness.