Преглед изворни кода

feat: optimize error raise (#820)

tags/0.3.13
takatost пре 2 година
родитељ
комит
1bd0a76a20
No account linked to committer's email address

+ 1
- 1
api/core/callback_handler/llm_callback_handler.py Прегледај датотеку

) )
self.conversation_message_task.save_message(llm_message=self.llm_message, by_stopped=True) self.conversation_message_task.save_message(llm_message=self.llm_message, by_stopped=True)
else: else:
logging.exception(error)
logging.debug("on_llm_error: %s", error)

+ 9
- 4
api/core/generator/llm_generator.py Прегледај датотеку



from langchain.schema import OutputParserException from langchain.schema import OutputParserException


from core.model_providers.error import LLMError
from core.model_providers.model_factory import ModelFactory from core.model_providers.model_factory import ModelFactory
from core.model_providers.models.entity.message import PromptMessage, MessageType from core.model_providers.models.entity.message import PromptMessage, MessageType
from core.model_providers.models.entity.model_params import ModelKwargs from core.model_providers.models.entity.model_params import ModelKwargs
try: try:
output = model_instance.run(prompts) output = model_instance.run(prompts)
questions = output_parser.parse(output.content) questions = output_parser.parse(output.content)
except Exception:
logging.exception("Error generating suggested questions after answer")
except LLMError:
questions = []
except Exception as e:
logging.exception(e)
questions = [] questions = []


return questions return questions
try: try:
output = model_instance.run(prompts) output = model_instance.run(prompts)
rule_config = output_parser.parse(output.content) rule_config = output_parser.parse(output.content)
except LLMError as e:
raise e
except OutputParserException: except OutputParserException:
raise ValueError('Please give a valid input for intended audience or hoping to solve problems.') raise ValueError('Please give a valid input for intended audience or hoping to solve problems.')
except Exception:
logging.exception("Error generating prompt")
except Exception as e:
logging.exception(e)
rule_config = { rule_config = {
"prompt": "", "prompt": "",
"variables": [], "variables": [],

+ 5
- 1
api/core/model_providers/providers/azure_openai_provider.py Прегледај датотеку

if obfuscated: if obfuscated:
credentials['openai_api_key'] = encrypter.obfuscated_token(credentials['openai_api_key']) credentials['openai_api_key'] = encrypter.obfuscated_token(credentials['openai_api_key'])


return credentials
return {
'openai_api_base': credentials['openai_api_base'],
'openai_api_key': credentials['openai_api_key'],
'base_model_name': credentials['base_model_name']
}
else: else:
if hosted_model_providers.azure_openai: if hosted_model_providers.azure_openai:
return { return {

+ 5
- 2
api/tasks/generate_conversation_summary_task.py Прегледај датотеку

from werkzeug.exceptions import NotFound from werkzeug.exceptions import NotFound


from core.generator.llm_generator import LLMGenerator from core.generator.llm_generator import LLMGenerator
from core.model_providers.error import LLMError
from extensions.ext_database import db from extensions.ext_database import db
from models.model import Conversation, Message from models.model import Conversation, Message




end_at = time.perf_counter() end_at = time.perf_counter()
logging.info(click.style('Conversation summary generated: {} latency: {}'.format(conversation_id, end_at - start_at), fg='green')) logging.info(click.style('Conversation summary generated: {} latency: {}'.format(conversation_id, end_at - start_at), fg='green'))
except Exception:
logging.exception("generate conversation summary failed")
except LLMError:
pass
except Exception as e:
logging.exception(e)

Loading…
Откажи
Сачувај