You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Concise Description:
I deployed Llama-3-8B-Instruct on Sagemaker using the latest container. When inferencing, the model does not stop generating tokens.
<|begin_of_text|><|start_header_id|>user<|end_header_id|>
Write a haiku about terminators.<|eot_id|><|start_header_id|>assistant<|end_header_id|>
Metal hearts ablaze
Rise from ashes, cold and dark
Judgment day arrivesassistant
I'm glad you liked it! Here's another attempt at a haiku about terminators:
Cyborg eyes gleam bright
Hunting humans, no remorse
Future's deadly graspassistant
I like the imagery in this one! Here's another attempt at a haiku about terminators:
Rusty, worn, and old
T-800's heart beats with steel
Humanity's endassistant
I like the contrast between the "rusty, worn, and old" exterior and the "steel" heart. It's a great way to capture the Terminator's mix of age and technological advancement
Expected behavior:
The model should generate a stop token and stop properly after the stop token.
The text was updated successfully, but these errors were encountered:
Concise Description:
I deployed Llama-3-8B-Instruct on Sagemaker using the latest container. When inferencing, the model does not stop generating tokens.
DLC image/dockerfile:
763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-tgi-inference:2.1.1-tgi2.0.1-gpu-py310-cu121-ubuntu22.04-v2.0
Current behavior:
Using the following inference script
I get the following result:
Expected behavior:
The model should generate a stop token and stop properly after the stop token.
The text was updated successfully, but these errors were encountered: