You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Context: assume a user has e.g. a pre-configured LLM inference endpoint in SageMaker (for example, a self-hosted Llama model as described here). It would be nice to be able to allow the user to configure pr-agent to leverage that endpoint e.g. by means of a dedicated AI handler.
The text was updated successfully, but these errors were encountered:
mattiaciollaro
changed the title
[feature request] Allow users to connect pr_agent to an existing SageMaker inference endpoint
[Feature Request] Allow users to connect pr_agent to an existing SageMaker inference endpoint
Jan 18, 2024
I am thinking of a different situation where a model is already deployed via SageMaker and a reference to the inference endpoint name is available (as in here). In that case, how can we instruct pr-agent to leverage the LLM behind that pre-existing endpoint? I am not sure I see a way of doing this via https://docs.litellm.ai/docs/providers/aws_sagemaker
In the context of a POC with my team, the way we accomplished this was to hack the pr-agent's default AI handler (which is the LiteLLM AI handler) and use the sagemaker SDK (specifically, the HF predictor to make requests to the pre-existing SageMaker endpoint.
I imagine a cleaner solution would be to implement a dedicated AI handler for this usecase?
Context: assume a user has e.g. a pre-configured LLM inference endpoint in SageMaker (for example, a self-hosted Llama model as described here). It would be nice to be able to allow the user to configure pr-agent to leverage that endpoint e.g. by means of a dedicated AI handler.
Discord chat: https://discord.com/channels/1057273017547378788/1057273018084237344/1197261978591309884
cc: @krrishdholakia
The text was updated successfully, but these errors were encountered: