| @@ -26,6 +26,11 @@ class ZhipuAIProvider(BaseModelProvider): | |||
| def _get_fixed_model_list(self, model_type: ModelType) -> list[dict]: | |||
| if model_type == ModelType.TEXT_GENERATION: | |||
| return [ | |||
| { | |||
| 'id': 'chatglm_turbo', | |||
| 'name': 'chatglm_turbo', | |||
| 'mode': ModelMode.CHAT.value, | |||
| }, | |||
| { | |||
| 'id': 'chatglm_pro', | |||
| 'name': 'chatglm_pro', | |||
| @@ -11,6 +11,12 @@ | |||
| }, | |||
| "model_flexibility": "fixed", | |||
| "price_config": { | |||
| "chatglm_turbo": { | |||
| "prompt": "0.005", | |||
| "completion": "0.005", | |||
| "unit": "0.001", | |||
| "currency": "RMB" | |||
| }, | |||
| "chatglm_pro": { | |||
| "prompt": "0.01", | |||
| "completion": "0.01", | |||
| @@ -96,7 +96,7 @@ class ZhipuAIChatLLM(BaseChatModel): | |||
| return True | |||
| client: Any = None #: :meta private: | |||
| model: str = "chatglm_lite" | |||
| model: str = "chatglm_turbo" | |||
| """Model name to use.""" | |||
| temperature: float = 0.95 | |||
| """A non-negative float that tunes the degree of randomness in generation.""" | |||