From e04f4b239527431a07d20ba54134ae1bf5a9a093 Mon Sep 17 00:00:00 2001 From: Faiz Surani Date: Wed, 2 Oct 2024 10:06:08 -0700 Subject: [PATCH] Fix closeout error in vllm --- rl/llm/engines/local.py | 1 - 1 file changed, 1 deletion(-) diff --git a/rl/llm/engines/local.py b/rl/llm/engines/local.py index 04b9136..f18dc69 100644 --- a/rl/llm/engines/local.py +++ b/rl/llm/engines/local.py @@ -392,7 +392,6 @@ def __exit__(self, exc_type, exc_value, traceback): del self.vllm gc.collect() torch.cuda.empty_cache() - torch.distributed.destroy_process_group() LOGGER.info("VLLM model unloaded.") def generate(self, prompt: InferenceInput) -> InferenceOutput: