From d6547707327031410adf0219db8d61e36a79f1e8 Mon Sep 17 00:00:00 2001 From: takatost Date: Fri, 17 Nov 2023 14:07:36 +0800 Subject: [PATCH] feat: supports for new version of openllm (#1554) --- api/core/third_party/langchain/llms/openllm.py | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/api/core/third_party/langchain/llms/openllm.py b/api/core/third_party/langchain/llms/openllm.py index 6151fe3f1c..d83f54da6e 100644 --- a/api/core/third_party/langchain/llms/openllm.py +++ b/api/core/third_party/langchain/llms/openllm.py @@ -51,7 +51,8 @@ class OpenLLM(LLM): ) -> str: params = { "prompt": prompt, - "llm_config": self.llm_kwargs + "llm_config": self.llm_kwargs, + "stop": stop, } headers = {"Content-Type": "application/json"} @@ -65,11 +66,11 @@ class OpenLLM(LLM): raise ValueError(f"OpenLLM HTTP {response.status_code} error: {response.text}") json_response = response.json() - completion = json_response["responses"][0] + completion = json_response["outputs"][0]['text'] completion = completion.lstrip(prompt) - if stop is not None: - completion = enforce_stop_tokens(completion, stop) + # if stop is not None: + # completion = enforce_stop_tokens(completion, stop) return completion