Przeglądaj źródła

Fix/app runner typo (#2661)

tags/0.5.8
nan jiang 1 rok temu
rodzic
commit
607f9eda35
No account linked to committer's email address

+ 1
- 1
api/core/app_runner/app_runner.py Wyświetl plik



return rest_tokens return rest_tokens


def recale_llm_max_tokens(self, model_config: ModelConfigEntity,
def recalc_llm_max_tokens(self, model_config: ModelConfigEntity,
prompt_messages: list[PromptMessage]): prompt_messages: list[PromptMessage]):
# recalc max_tokens if sum(prompt_token + max_tokens) over model token limit # recalc max_tokens if sum(prompt_token + max_tokens) over model token limit
model_type_instance = model_config.provider_model_bundle.model_type_instance model_type_instance = model_config.provider_model_bundle.model_type_instance

+ 1
- 1
api/core/app_runner/basic_app_runner.py Wyświetl plik

return return


# Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit # Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit
self.recale_llm_max_tokens(
self.recalc_llm_max_tokens(
model_config=app_orchestration_config.model_config, model_config=app_orchestration_config.model_config,
prompt_messages=prompt_messages prompt_messages=prompt_messages
) )

+ 1
- 1
api/core/features/assistant_cot_runner.py Wyświetl plik

) )


# recale llm max tokens # recale llm max tokens
self.recale_llm_max_tokens(self.model_config, prompt_messages)
self.recalc_llm_max_tokens(self.model_config, prompt_messages)
# invoke model # invoke model
chunks: Generator[LLMResultChunk, None, None] = model_instance.invoke_llm( chunks: Generator[LLMResultChunk, None, None] = model_instance.invoke_llm(
prompt_messages=prompt_messages, prompt_messages=prompt_messages,

+ 1
- 1
api/core/features/assistant_fc_runner.py Wyświetl plik

) )


# recale llm max tokens # recale llm max tokens
self.recale_llm_max_tokens(self.model_config, prompt_messages)
self.recalc_llm_max_tokens(self.model_config, prompt_messages)
# invoke model # invoke model
chunks: Union[Generator[LLMResultChunk, None, None], LLMResult] = model_instance.invoke_llm( chunks: Union[Generator[LLMResultChunk, None, None], LLMResult] = model_instance.invoke_llm(
prompt_messages=prompt_messages, prompt_messages=prompt_messages,

Ładowanie…
Anuluj
Zapisz