From 4a5439e598417fbc93535a2105001ec76f5562c7 Mon Sep 17 00:00:00 2001 From: Ikko Eltociear Ashimine Date: Fri, 27 Dec 2024 06:13:38 +0900 Subject: [PATCH] chore: update _extract.py (#70) reponse -> response --- src/raglite/_extract.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/raglite/_extract.py b/src/raglite/_extract.py index c902e68..a13ad2d 100644 --- a/src/raglite/_extract.py +++ b/src/raglite/_extract.py @@ -41,7 +41,7 @@ class MyNameResponse(BaseModel): system_prompt = getattr(return_type, "system_prompt", "").strip() if not llm_supports_response_format or config.llm.startswith("llama-cpp-python"): system_prompt += f"\n\nFormat your response according to this JSON schema:\n{return_type.model_json_schema()!s}" - # Constrain the reponse format to the JSON schema if it's supported by the LLM [1]. Strict mode + # Constrain the response format to the JSON schema if it's supported by the LLM [1]. Strict mode # is disabled by default because it only supports a subset of JSON schema features [2]. # [1] https://docs.litellm.ai/docs/completion/json_mode # [2] https://platform.openai.com/docs/guides/structured-outputs#some-type-specific-keywords-are-not-yet-supported