|
|
|
@@ -0,0 +1,53 @@ |
|
|
|
model: deepseek-ai/DeepSeek-V3 |
|
|
|
label: |
|
|
|
en_US: deepseek-ai/DeepSeek-V3 |
|
|
|
model_type: llm |
|
|
|
features: |
|
|
|
- agent-thought |
|
|
|
- tool-call |
|
|
|
- stream-tool-call |
|
|
|
model_properties: |
|
|
|
mode: chat |
|
|
|
context_size: 64000 |
|
|
|
parameter_rules: |
|
|
|
- name: temperature |
|
|
|
use_template: temperature |
|
|
|
- name: max_tokens |
|
|
|
use_template: max_tokens |
|
|
|
type: int |
|
|
|
default: 512 |
|
|
|
min: 1 |
|
|
|
max: 4096 |
|
|
|
help: |
|
|
|
zh_Hans: 指定生成结果长度的上限。如果生成结果截断,可以调大该参数。 |
|
|
|
en_US: Specifies the upper limit on the length of generated results. If the generated results are truncated, you can increase this parameter. |
|
|
|
- name: top_p |
|
|
|
use_template: top_p |
|
|
|
- name: top_k |
|
|
|
label: |
|
|
|
zh_Hans: 取样数量 |
|
|
|
en_US: Top k |
|
|
|
type: int |
|
|
|
help: |
|
|
|
zh_Hans: 仅从每个后续标记的前 K 个选项中采样。 |
|
|
|
en_US: Only sample from the top K options for each subsequent token. |
|
|
|
required: false |
|
|
|
- name: frequency_penalty |
|
|
|
use_template: frequency_penalty |
|
|
|
- name: response_format |
|
|
|
label: |
|
|
|
zh_Hans: 回复格式 |
|
|
|
en_US: Response Format |
|
|
|
type: string |
|
|
|
help: |
|
|
|
zh_Hans: 指定模型必须输出的格式 |
|
|
|
en_US: specifying the format that the model must output |
|
|
|
required: false |
|
|
|
options: |
|
|
|
- text |
|
|
|
- json_object |
|
|
|
pricing: |
|
|
|
input: "1" |
|
|
|
output: "2" |
|
|
|
unit: "0.000001" |
|
|
|
currency: RMB |