| @@ -312,39 +312,39 @@ class AzureOpenAILargeLanguageModel(_CommonAzureOpenAI, LargeLanguageModel): | |||
| if user: | |||
| extra_model_kwargs["user"] = user | |||
| # clear illegal prompt messages | |||
| prompt_messages = self._clear_illegal_prompt_messages(model, prompt_messages) | |||
| block_as_stream = False | |||
| if model.startswith("o1"): | |||
| if stream: | |||
| block_as_stream = True | |||
| stream = False | |||
| if "stream_options" in extra_model_kwargs: | |||
| del extra_model_kwargs["stream_options"] | |||
| if "stop" in extra_model_kwargs: | |||
| del extra_model_kwargs["stop"] | |||
| # chat model | |||
| response = client.chat.completions.create( | |||
| messages=[self._convert_prompt_message_to_dict(m) for m in prompt_messages], | |||
| model=model, | |||
| stream=stream, | |||
| **model_parameters, | |||
| **extra_model_kwargs, | |||
| ) | |||
| # clear illegal prompt messages | |||
| prompt_messages = self._clear_illegal_prompt_messages(model, prompt_messages) | |||
| block_as_stream = False | |||
| if model.startswith("o1"): | |||
| if stream: | |||
| return self._handle_chat_generate_stream_response(model, credentials, response, prompt_messages, tools) | |||
| block_as_stream = True | |||
| stream = False | |||
| if "stream_options" in extra_model_kwargs: | |||
| del extra_model_kwargs["stream_options"] | |||
| if "stop" in extra_model_kwargs: | |||
| del extra_model_kwargs["stop"] | |||
| # chat model | |||
| response = client.chat.completions.create( | |||
| messages=[self._convert_prompt_message_to_dict(m) for m in prompt_messages], | |||
| model=model, | |||
| stream=stream, | |||
| **model_parameters, | |||
| **extra_model_kwargs, | |||
| ) | |||
| if stream: | |||
| return self._handle_chat_generate_stream_response(model, credentials, response, prompt_messages, tools) | |||
| block_result = self._handle_chat_generate_response(model, credentials, response, prompt_messages, tools) | |||
| block_result = self._handle_chat_generate_response(model, credentials, response, prompt_messages, tools) | |||
| if block_as_stream: | |||
| return self._handle_chat_block_as_stream_response(block_result, prompt_messages, stop) | |||
| if block_as_stream: | |||
| return self._handle_chat_block_as_stream_response(block_result, prompt_messages, stop) | |||
| return block_result | |||
| return block_result | |||
| def _handle_chat_block_as_stream_response( | |||
| self, | |||