Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

VS Code: LLM request cancellations are not transmitted reliably to Cody Gateway #3614

Closed
valerybugakov opened this issue Mar 29, 2024 · 2 comments
Assignees

Comments

@valerybugakov
Copy link
Member

  • LLM request cancellations are not transmitted reliably to Cody Gateway, leading to token inference continuing until the maximum limit is reached. It significantly increases the load on our inference providers, increasing latency and spending. By the latest estimate, about 2/3 of inferenced tokens are "overhead" tokens.
  • The Google Sheet with pricing breakdown and approximate potential savings.
  • Slack thread with more context.
Copy link

This issue is marked as stale because it has been open 60 days with no activity. Remove stale label or comment or this will be closed automatically in 5 days.

Copy link

This issue is marked as stale because it has been open 60 days with no activity. Remove stale label or comment or this will be closed automatically in 5 days.

@github-actions github-actions bot added the Stale label Aug 31, 2024
@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Sep 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant