@@ -0,0 +1,5 @@ | |||
- open-mistral-7b | |||
- open-mixtral-8x7b | |||
- mistral-small-latest | |||
- mistral-medium-latest | |||
- mistral-large-latest |
@@ -0,0 +1,31 @@ | |||
from collections.abc import Generator | |||
from typing import Optional, Union | |||
from core.model_runtime.entities.llm_entities import LLMResult | |||
from core.model_runtime.entities.message_entities import PromptMessage, PromptMessageTool | |||
from core.model_runtime.model_providers.openai_api_compatible.llm.llm import OAIAPICompatLargeLanguageModel | |||
class MistralAILargeLanguageModel(OAIAPICompatLargeLanguageModel): | |||
def _invoke(self, model: str, credentials: dict, | |||
prompt_messages: list[PromptMessage], model_parameters: dict, | |||
tools: Optional[list[PromptMessageTool]] = None, stop: Optional[list[str]] = None, | |||
stream: bool = True, user: Optional[str] = None) \ | |||
-> Union[LLMResult, Generator]: | |||
self._add_custom_parameters(credentials) | |||
# mistral dose not support user/stop arguments | |||
stop = [] | |||
user = None | |||
return super()._invoke(model, credentials, prompt_messages, model_parameters, tools, stop, stream, user) | |||
def validate_credentials(self, model: str, credentials: dict) -> None: | |||
self._add_custom_parameters(credentials) | |||
super().validate_credentials(model, credentials) | |||
@staticmethod | |||
def _add_custom_parameters(credentials: dict) -> None: | |||
credentials['mode'] = 'chat' | |||
credentials['endpoint_url'] = 'https://api.mistral.ai/v1' |
@@ -0,0 +1,50 @@ | |||
model: mistral-large-latest | |||
label: | |||
zh_Hans: mistral-large-latest | |||
en_US: mistral-large-latest | |||
model_type: llm | |||
features: | |||
- agent-thought | |||
model_properties: | |||
context_size: 32000 | |||
parameter_rules: | |||
- name: temperature | |||
use_template: temperature | |||
default: 0.7 | |||
min: 0 | |||
max: 1 | |||
- name: top_p | |||
use_template: top_p | |||
default: 1 | |||
min: 0 | |||
max: 1 | |||
- name: max_tokens | |||
use_template: max_tokens | |||
default: 1024 | |||
min: 1 | |||
max: 8000 | |||
- name: safe_prompt | |||
defulat: false | |||
type: boolean | |||
help: | |||
en_US: Whether to inject a safety prompt before all conversations. | |||
zh_Hans: 是否开启提示词审查 | |||
label: | |||
en_US: SafePrompt | |||
zh_Hans: 提示词审查 | |||
- name: random_seed | |||
type: int | |||
help: | |||
en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
label: | |||
en_US: RandomSeed | |||
zh_Hans: 随机数种子 | |||
default: 0 | |||
min: 0 | |||
max: 2147483647 | |||
pricing: | |||
input: '0.008' | |||
output: '0.024' | |||
unit: '0.001' | |||
currency: USD |
@@ -0,0 +1,50 @@ | |||
model: mistral-medium-latest | |||
label: | |||
zh_Hans: mistral-medium-latest | |||
en_US: mistral-medium-latest | |||
model_type: llm | |||
features: | |||
- agent-thought | |||
model_properties: | |||
context_size: 32000 | |||
parameter_rules: | |||
- name: temperature | |||
use_template: temperature | |||
default: 0.7 | |||
min: 0 | |||
max: 1 | |||
- name: top_p | |||
use_template: top_p | |||
default: 1 | |||
min: 0 | |||
max: 1 | |||
- name: max_tokens | |||
use_template: max_tokens | |||
default: 1024 | |||
min: 1 | |||
max: 8000 | |||
- name: safe_prompt | |||
defulat: false | |||
type: boolean | |||
help: | |||
en_US: Whether to inject a safety prompt before all conversations. | |||
zh_Hans: 是否开启提示词审查 | |||
label: | |||
en_US: SafePrompt | |||
zh_Hans: 提示词审查 | |||
- name: random_seed | |||
type: int | |||
help: | |||
en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
label: | |||
en_US: RandomSeed | |||
zh_Hans: 随机数种子 | |||
default: 0 | |||
min: 0 | |||
max: 2147483647 | |||
pricing: | |||
input: '0.0027' | |||
output: '0.0081' | |||
unit: '0.001' | |||
currency: USD |
@@ -0,0 +1,50 @@ | |||
model: mistral-small-latest | |||
label: | |||
zh_Hans: mistral-small-latest | |||
en_US: mistral-small-latest | |||
model_type: llm | |||
features: | |||
- agent-thought | |||
model_properties: | |||
context_size: 32000 | |||
parameter_rules: | |||
- name: temperature | |||
use_template: temperature | |||
default: 0.7 | |||
min: 0 | |||
max: 1 | |||
- name: top_p | |||
use_template: top_p | |||
default: 1 | |||
min: 0 | |||
max: 1 | |||
- name: max_tokens | |||
use_template: max_tokens | |||
default: 1024 | |||
min: 1 | |||
max: 8000 | |||
- name: safe_prompt | |||
defulat: false | |||
type: boolean | |||
help: | |||
en_US: Whether to inject a safety prompt before all conversations. | |||
zh_Hans: 是否开启提示词审查 | |||
label: | |||
en_US: SafePrompt | |||
zh_Hans: 提示词审查 | |||
- name: random_seed | |||
type: int | |||
help: | |||
en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
label: | |||
en_US: RandomSeed | |||
zh_Hans: 随机数种子 | |||
default: 0 | |||
min: 0 | |||
max: 2147483647 | |||
pricing: | |||
input: '0.002' | |||
output: '0.006' | |||
unit: '0.001' | |||
currency: USD |
@@ -0,0 +1,50 @@ | |||
model: open-mistral-7b | |||
label: | |||
zh_Hans: open-mistral-7b | |||
en_US: open-mistral-7b | |||
model_type: llm | |||
features: | |||
- agent-thought | |||
model_properties: | |||
context_size: 8000 | |||
parameter_rules: | |||
- name: temperature | |||
use_template: temperature | |||
default: 0.7 | |||
min: 0 | |||
max: 1 | |||
- name: top_p | |||
use_template: top_p | |||
default: 1 | |||
min: 0 | |||
max: 1 | |||
- name: max_tokens | |||
use_template: max_tokens | |||
default: 1024 | |||
min: 1 | |||
max: 2048 | |||
- name: safe_prompt | |||
defulat: false | |||
type: boolean | |||
help: | |||
en_US: Whether to inject a safety prompt before all conversations. | |||
zh_Hans: 是否开启提示词审查 | |||
label: | |||
en_US: SafePrompt | |||
zh_Hans: 提示词审查 | |||
- name: random_seed | |||
type: int | |||
help: | |||
en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
label: | |||
en_US: RandomSeed | |||
zh_Hans: 随机数种子 | |||
default: 0 | |||
min: 0 | |||
max: 2147483647 | |||
pricing: | |||
input: '0.00025' | |||
output: '0.00025' | |||
unit: '0.001' | |||
currency: USD |
@@ -0,0 +1,50 @@ | |||
model: open-mixtral-8x7b | |||
label: | |||
zh_Hans: open-mixtral-8x7b | |||
en_US: open-mixtral-8x7b | |||
model_type: llm | |||
features: | |||
- agent-thought | |||
model_properties: | |||
context_size: 32000 | |||
parameter_rules: | |||
- name: temperature | |||
use_template: temperature | |||
default: 0.7 | |||
min: 0 | |||
max: 1 | |||
- name: top_p | |||
use_template: top_p | |||
default: 1 | |||
min: 0 | |||
max: 1 | |||
- name: max_tokens | |||
use_template: max_tokens | |||
default: 1024 | |||
min: 1 | |||
max: 8000 | |||
- name: safe_prompt | |||
defulat: false | |||
type: boolean | |||
help: | |||
en_US: Whether to inject a safety prompt before all conversations. | |||
zh_Hans: 是否开启提示词审查 | |||
label: | |||
en_US: SafePrompt | |||
zh_Hans: 提示词审查 | |||
- name: random_seed | |||
type: int | |||
help: | |||
en_US: The seed to use for random sampling. If set, different calls will generate deterministic results. | |||
zh_Hans: 当开启随机数种子以后,你可以通过指定一个固定的种子来使得回答结果更加稳定 | |||
label: | |||
en_US: RandomSeed | |||
zh_Hans: 随机数种子 | |||
default: 0 | |||
min: 0 | |||
max: 2147483647 | |||
pricing: | |||
input: '0.0007' | |||
output: '0.0007' | |||
unit: '0.001' | |||
currency: USD |
@@ -0,0 +1,30 @@ | |||
import logging | |||
from core.model_runtime.entities.model_entities import ModelType | |||
from core.model_runtime.errors.validate import CredentialsValidateFailedError | |||
from core.model_runtime.model_providers.__base.model_provider import ModelProvider | |||
logger = logging.getLogger(__name__) | |||
class MistralAIProvider(ModelProvider): | |||
def validate_provider_credentials(self, credentials: dict) -> None: | |||
""" | |||
Validate provider credentials | |||
if validate failed, raise exception | |||
:param credentials: provider credentials, credentials form defined in `provider_credential_schema`. | |||
""" | |||
try: | |||
model_instance = self.get_model_instance(ModelType.LLM) | |||
model_instance.validate_credentials( | |||
model='open-mistral-7b', | |||
credentials=credentials | |||
) | |||
except CredentialsValidateFailedError as ex: | |||
raise ex | |||
except Exception as ex: | |||
logger.exception(f'{self.get_provider_schema().provider} credentials validate failed') | |||
raise ex |
@@ -0,0 +1,31 @@ | |||
provider: mistralai | |||
label: | |||
en_US: MistralAI | |||
description: | |||
en_US: Models provided by MistralAI, such as open-mistral-7b and mistral-large-latest. | |||
zh_Hans: MistralAI 提供的模型,例如 open-mistral-7b 和 mistral-large-latest。 | |||
icon_small: | |||
en_US: icon_s_en.png | |||
icon_large: | |||
en_US: icon_l_en.png | |||
background: "#FFFFFF" | |||
help: | |||
title: | |||
en_US: Get your API Key from MistralAI | |||
zh_Hans: 从 MistralAI 获取 API Key | |||
url: | |||
en_US: https://console.mistral.ai/api-keys/ | |||
supported_model_types: | |||
- llm | |||
configurate_methods: | |||
- predefined-model | |||
provider_credential_schema: | |||
credential_form_schemas: | |||
- variable: api_key | |||
label: | |||
en_US: API Key | |||
type: secret-input | |||
required: true | |||
placeholder: | |||
zh_Hans: 在此输入您的 API Key | |||
en_US: Enter your API Key |