Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature Request: Adjustable response timeout per GenAI connector #166561

Open
cp-elastic opened this issue Sep 15, 2023 · 7 comments
Open

Feature Request: Adjustable response timeout per GenAI connector #166561

cp-elastic opened this issue Sep 15, 2023 · 7 comments
Labels
Team:ResponseOps Label for the ResponseOps team (formerly the Cases and Alerting teams)

Comments

@cp-elastic
Copy link

Describe the feature:
The default GenAI response timeout appears to be around 60 seconds. This timeout should be adjustable per connector to account for varying models and responsiveness of the API. Ideally this would be an additional field to set during the connector configuration workflow in Kibana.

Describe a specific use case for the feature:
In this specific case, I am self-hosting a large language model with an OpenAI conformant API for development purposes on a bare metal server with 24 CPU cores and 96 GB RAM. The model typically sends a response within 2 minutes, which is obviously beyond the default timeout. As adoption of LLMs and capabilities of the AI Assistant expand, this will help organizations with privacy concerns that are hosting their own LLMs on commodity hardware.

@botelastic botelastic bot added the needs-team Issues missing a team label label Sep 15, 2023
@jughosta jughosta added the Team:ResponseOps Label for the ResponseOps team (formerly the Cases and Alerting teams) label Sep 28, 2023
@elasticmachine
Copy link
Contributor

Pinging @elastic/response-ops (Team:ResponseOps)

@botelastic botelastic bot removed the needs-team Issues missing a team label label Sep 28, 2023
@cp-elastic
Copy link
Author

I've tested the new changes in Kibana 8.11.0, and I'm running into less timeout errors now when using my local LLM. I'd still like to see this be configurable, if possible, so that users can adjust timeouts per OpenAI connector to account for different models and LLM settings.

@securix-rog
Copy link

For testing different local models and optimize the performance of them, I would really appreciate this feature!

@Danouchka
Copy link

+1

1 similar comment
@pathoge
Copy link

pathoge commented Aug 14, 2024

+1

@cnasikas
Copy link
Member

cnasikas commented Sep 3, 2024

@dgieselaar
Copy link
Member

@cnasikas will defer to others, timeouts are not really an issue for the Observability AI Assistant because we use streaming.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Team:ResponseOps Label for the ResponseOps team (formerly the Cases and Alerting teams)
Projects
None yet
Development

No branches or pull requests

8 participants