| @@ -0,0 +1,5 @@ | |||
| - open-mistral-7b | |||
| - open-mixtral-8x7b | |||
| - mistral-small-latest | |||
| - mistral-medium-latest | |||
| - mistral-large-latest | |||
| @@ -0,0 +1,31 @@ | |||
| from collections.abc import Generator | |||
| from typing import Optional, Union | |||
| from core.model_runtime.entities.llm_entities import LLMResult | |||
| from core.model_runtime.entities.message_entities import PromptMessage, PromptMessageTool | |||
| from core.model_runtime.model_providers.openai_api_compatible.llm.llm import OAIAPICompatLargeLanguageModel | |||
| class MistralAILargeLanguageModel(OAIAPICompatLargeLanguageModel): | |||
| def _invoke(self, model: str, credentials: dict, | |||
| prompt_messages: list[PromptMessage], model_parameters: dict, | |||
| tools: Optional[list[PromptMessageTool]] = None, stop: Optional[list[str]] = None, | |||
| stream: bool = True, user: Optional[str] = None) \ | |||
| -> Union[LLMResult, Generator]: | |||
| self._add_custom_parameters(credentials) | |||
| # mistral dose not support user/stop arguments | |||
| stop = [] | |||
| user = None | |||
| return super()._invoke(model, credentials, prompt_messages, model_parameters, tools, stop, stream, user) | |||
| def validate_credentials(self, model: str, credentials: dict) -> None: | |||
| self._add_custom_parameters(credentials) | |||
| super().validate_credentials(model, credentials) | |||
| @staticmethod | |||
| def _add_custom_parameters(credentials: dict) -> None: | |||
| credentials['mode'] = 'chat' | |||
| credentials['endpoint_url'] = 'https://api.mistral.ai/v1' | |||
| @@ -0,0 +1,50 @@ | |||
| model: mistral-large-latest | |||
| label: | |||
| zh_Hans: mistral-large-latest | |||
| en_US: mistral-large-latest | |||
| model_type: llm | |||
| features: | |||
| - agent-thought | |||
| model_properties: | |||
| context_size: 32000 | |||
| parameter_rules: | |||
| - name: temperature | |||
| use_template: temperature | |||
| default: 0.7 | |||
| min: 0 | |||
| max: 1 | |||
| - name: top_p | |||
| use_template: top_p | |||
| default: 1 | |||
| min: 0 | |||
| max: 1 | |||
| - name: max_tokens | |||
| use_template: max_tokens | |||
| default: 1024 | |||
| min: 1 | |||
| max: 8000 | |||
| - name: safe_prompt | |||
| defulat: false | |||
| type: boolean | |||
| help: | |||
| en_US: Whether to inject a safety prompt before all conversations. | |||
| zh_Hans: 是否开启提示词审查 | |||
| label: | |||
| en_US: SafePrompt | |||
| zh_Hans: 提示词审查 | |||
| - name: random_seed | |||
| type: int | |||
| help: | |||
| en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
| zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
| label: | |||
| en_US: RandomSeed | |||
| zh_Hans: 随机数种子 | |||
| default: 0 | |||
| min: 0 | |||
| max: 2147483647 | |||
| pricing: | |||
| input: '0.008' | |||
| output: '0.024' | |||
| unit: '0.001' | |||
| currency: USD | |||
| @@ -0,0 +1,50 @@ | |||
| model: mistral-medium-latest | |||
| label: | |||
| zh_Hans: mistral-medium-latest | |||
| en_US: mistral-medium-latest | |||
| model_type: llm | |||
| features: | |||
| - agent-thought | |||
| model_properties: | |||
| context_size: 32000 | |||
| parameter_rules: | |||
| - name: temperature | |||
| use_template: temperature | |||
| default: 0.7 | |||
| min: 0 | |||
| max: 1 | |||
| - name: top_p | |||
| use_template: top_p | |||
| default: 1 | |||
| min: 0 | |||
| max: 1 | |||
| - name: max_tokens | |||
| use_template: max_tokens | |||
| default: 1024 | |||
| min: 1 | |||
| max: 8000 | |||
| - name: safe_prompt | |||
| defulat: false | |||
| type: boolean | |||
| help: | |||
| en_US: Whether to inject a safety prompt before all conversations. | |||
| zh_Hans: 是否开启提示词审查 | |||
| label: | |||
| en_US: SafePrompt | |||
| zh_Hans: 提示词审查 | |||
| - name: random_seed | |||
| type: int | |||
| help: | |||
| en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
| zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
| label: | |||
| en_US: RandomSeed | |||
| zh_Hans: 随机数种子 | |||
| default: 0 | |||
| min: 0 | |||
| max: 2147483647 | |||
| pricing: | |||
| input: '0.0027' | |||
| output: '0.0081' | |||
| unit: '0.001' | |||
| currency: USD | |||
| @@ -0,0 +1,50 @@ | |||
| model: mistral-small-latest | |||
| label: | |||
| zh_Hans: mistral-small-latest | |||
| en_US: mistral-small-latest | |||
| model_type: llm | |||
| features: | |||
| - agent-thought | |||
| model_properties: | |||
| context_size: 32000 | |||
| parameter_rules: | |||
| - name: temperature | |||
| use_template: temperature | |||
| default: 0.7 | |||
| min: 0 | |||
| max: 1 | |||
| - name: top_p | |||
| use_template: top_p | |||
| default: 1 | |||
| min: 0 | |||
| max: 1 | |||
| - name: max_tokens | |||
| use_template: max_tokens | |||
| default: 1024 | |||
| min: 1 | |||
| max: 8000 | |||
| - name: safe_prompt | |||
| defulat: false | |||
| type: boolean | |||
| help: | |||
| en_US: Whether to inject a safety prompt before all conversations. | |||
| zh_Hans: 是否开启提示词审查 | |||
| label: | |||
| en_US: SafePrompt | |||
| zh_Hans: 提示词审查 | |||
| - name: random_seed | |||
| type: int | |||
| help: | |||
| en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
| zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
| label: | |||
| en_US: RandomSeed | |||
| zh_Hans: 随机数种子 | |||
| default: 0 | |||
| min: 0 | |||
| max: 2147483647 | |||
| pricing: | |||
| input: '0.002' | |||
| output: '0.006' | |||
| unit: '0.001' | |||
| currency: USD | |||
| @@ -0,0 +1,50 @@ | |||
| model: open-mistral-7b | |||
| label: | |||
| zh_Hans: open-mistral-7b | |||
| en_US: open-mistral-7b | |||
| model_type: llm | |||
| features: | |||
| - agent-thought | |||
| model_properties: | |||
| context_size: 8000 | |||
| parameter_rules: | |||
| - name: temperature | |||
| use_template: temperature | |||
| default: 0.7 | |||
| min: 0 | |||
| max: 1 | |||
| - name: top_p | |||
| use_template: top_p | |||
| default: 1 | |||
| min: 0 | |||
| max: 1 | |||
| - name: max_tokens | |||
| use_template: max_tokens | |||
| default: 1024 | |||
| min: 1 | |||
| max: 2048 | |||
| - name: safe_prompt | |||
| defulat: false | |||
| type: boolean | |||
| help: | |||
| en_US: Whether to inject a safety prompt before all conversations. | |||
| zh_Hans: 是否开启提示词审查 | |||
| label: | |||
| en_US: SafePrompt | |||
| zh_Hans: 提示词审查 | |||
| - name: random_seed | |||
| type: int | |||
| help: | |||
| en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
| zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
| label: | |||
| en_US: RandomSeed | |||
| zh_Hans: 随机数种子 | |||
| default: 0 | |||
| min: 0 | |||
| max: 2147483647 | |||
| pricing: | |||
| input: '0.00025' | |||
| output: '0.00025' | |||
| unit: '0.001' | |||
| currency: USD | |||
| @@ -0,0 +1,50 @@ | |||
| model: open-mixtral-8x7b | |||
| label: | |||
| zh_Hans: open-mixtral-8x7b | |||
| en_US: open-mixtral-8x7b | |||
| model_type: llm | |||
| features: | |||
| - agent-thought | |||
| model_properties: | |||
| context_size: 32000 | |||
| parameter_rules: | |||
| - name: temperature | |||
| use_template: temperature | |||
| default: 0.7 | |||
| min: 0 | |||
| max: 1 | |||
| - name: top_p | |||
| use_template: top_p | |||
| default: 1 | |||
| min: 0 | |||
| max: 1 | |||
| - name: max_tokens | |||
| use_template: max_tokens | |||
| default: 1024 | |||
| min: 1 | |||
| max: 8000 | |||
| - name: safe_prompt | |||
| defulat: false | |||
| type: boolean | |||
| help: | |||
| en_US: Whether to inject a safety prompt before all conversations. | |||
| zh_Hans: 是否开启提示词审查 | |||
| label: | |||
| en_US: SafePrompt | |||
| zh_Hans: 提示词审查 | |||
| - name: random_seed | |||
| type: int | |||
| help: | |||
| en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
| zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
| label: | |||
| en_US: RandomSeed | |||
| zh_Hans: 随机数种子 | |||
| default: 0 | |||
| min: 0 | |||
| max: 2147483647 | |||
| pricing: | |||
| input: '0.0007' | |||
| output: '0.0007' | |||
| unit: '0.001' | |||
| currency: USD | |||
| @@ -0,0 +1,30 @@ | |||
| import logging | |||
| from core.model_runtime.entities.model_entities import ModelType | |||
| from core.model_runtime.errors.validate import CredentialsValidateFailedError | |||
| from core.model_runtime.model_providers.__base.model_provider import ModelProvider | |||
| logger = logging.getLogger(__name__) | |||
| class MistralAIProvider(ModelProvider): | |||
| def validate_provider_credentials(self, credentials: dict) -> None: | |||
| """ | |||
| Validate provider credentials | |||
| if validate failed, raise exception | |||
| :param credentials: provider credentials, credentials form defined in `provider_credential_schema`. | |||
| """ | |||
| try: | |||
| model_instance = self.get_model_instance(ModelType.LLM) | |||
| model_instance.validate_credentials( | |||
| model='open-mistral-7b', | |||
| credentials=credentials | |||
| ) | |||
| except CredentialsValidateFailedError as ex: | |||
| raise ex | |||
| except Exception as ex: | |||
| logger.exception(f'{self.get_provider_schema().provider} credentials validate failed') | |||
| raise ex | |||
| @@ -0,0 +1,31 @@ | |||
| provider: mistralai | |||
| label: | |||
| en_US: MistralAI | |||
| description: | |||
| en_US: Models provided by MistralAI, such as open-mistral-7b and mistral-large-latest. | |||
| zh_Hans: MistralAI 提供的模型,例如 open-mistral-7b 和 mistral-large-latest。 | |||
| icon_small: | |||
| en_US: icon_s_en.png | |||
| icon_large: | |||
| en_US: icon_l_en.png | |||
| background: "#FFFFFF" | |||
| help: | |||
| title: | |||
| en_US: Get your API Key from MistralAI | |||
| zh_Hans: 从 MistralAI 获取 API Key | |||
| url: | |||
| en_US: https://console.mistral.ai/api-keys/ | |||
| supported_model_types: | |||
| - llm | |||
| configurate_methods: | |||
| - predefined-model | |||
| provider_credential_schema: | |||
| credential_form_schemas: | |||
| - variable: api_key | |||
| label: | |||
| en_US: API Key | |||
| type: secret-input | |||
| required: true | |||
| placeholder: | |||
| zh_Hans: 在此输入您的 API Key | |||
| en_US: Enter your API Key | |||