-
Notifications
You must be signed in to change notification settings - Fork 297
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
☔️ When edit, etc. hits output token limits, output is truncated in edit and chat #3463
Comments
Repro steps:
We should add metrics for stop reason in edit output to assess how often this happens. It appears we record the interaction as a success. Actual behavior: The output is truncated after 127 lines ending abruptly with Expected behavior: ? |
What are our technical options here? Can we continue the stream in some way automatically, or with a button? Is the token limit too small currently? |
For generation, that might be plausible. Needs investigation to work out what the prompt would be. First step is to gather some real-life long generation prompts. @chillatom, in the work you and @rishabhmehrotra were doing looking at chat use, do you have some "long form code generation" user input we could look at?
Evidently for some people but unquantified. Short term we should count hitting the token limit and bump the limits if we can. |
This issue is marked as stale because it has been open 60 days with no activity. Remove stale label or comment or this will be closed automatically in 5 days. |
Version
Main 032645f
Describe the bug
This is an umbrella bug to collect a family of related issues with the same underlying cause.
Effect on edits:
Effect on chats:
Expected behavior
Exactly what we should do to communicate we hit a limit is unclear. @toolmantim, do you have ideas here?
Additional context
The underlying cause is: Edit, document, etc. work by having the LLM generate code. In some cases, re-generate input code with modifications. When that output is too large we can hit the token output limit of the LLM. Current behavior is to treat this as success, in the product UX and Telemetry. But that is wrong.
The text was updated successfully, but these errors were encountered: