| @@ -51,7 +51,8 @@ class OpenLLM(LLM): | |||
| ) -> str: | |||
| params = { | |||
| "prompt": prompt, | |||
| "llm_config": self.llm_kwargs | |||
| "llm_config": self.llm_kwargs, | |||
| "stop": stop, | |||
| } | |||
| headers = {"Content-Type": "application/json"} | |||
| @@ -65,11 +66,11 @@ class OpenLLM(LLM): | |||
| raise ValueError(f"OpenLLM HTTP {response.status_code} error: {response.text}") | |||
| json_response = response.json() | |||
| completion = json_response["responses"][0] | |||
| completion = json_response["outputs"][0]['text'] | |||
| completion = completion.lstrip(prompt) | |||
| if stop is not None: | |||
| completion = enforce_stop_tokens(completion, stop) | |||
| # if stop is not None: | |||
| # completion = enforce_stop_tokens(completion, stop) | |||
| return completion | |||