Skip to content

Commit

Permalink
Register missing engines
Browse files Browse the repository at this point in the history
  • Loading branch information
ProbablyFaiz committed Jun 26, 2024
1 parent 7a9e65d commit bf473de
Showing 1 changed file with 5 additions and 1 deletion.
6 changes: 5 additions & 1 deletion rl/llm/engines.py
Original file line number Diff line number Diff line change
Expand Up @@ -270,7 +270,9 @@ def generate(self, prompt: ChatInput) -> InferenceOutput:
generation_config={
"temperature": self.llm_config.temperature,
"max_output_tokens": self.llm_config.max_new_tokens,
"response_mime_type": "application/json" if self.llm_config.json_output else "text/plain",
"response_mime_type": "application/json"
if self.llm_config.json_output
else "text/plain",
},
system_instruction=system_message,
safety_settings={
Expand Down Expand Up @@ -825,13 +827,15 @@ def _wrap_output(self, req_output) -> InferenceOutput:
e.NAME: e
for e in (
VLLMEngine,
AsyncVLLMEngine,
WorkerVLLMEngine,
OpenAIEngine,
TogetherEngine,
GroqEngine,
AnthropicEngine,
ModalEngine,
GeminiEngine,
ManualEditEngine,
)
}

Expand Down

0 comments on commit bf473de

Please sign in to comment.