Spaces:
Runtime error
Runtime error
| # ========= Copyright 2023-2024 @ CAMEL-AI.org. All Rights Reserved. ========= | |
| # Licensed under the Apache License, Version 2.0 (the "License"); | |
| # you may not use this file except in compliance with the License. | |
| # You may obtain a copy of the License at | |
| # | |
| # http://www.apache.org/licenses/LICENSE-2.0 | |
| # | |
| # Unless required by applicable law or agreed to in writing, software | |
| # distributed under the License is distributed on an "AS IS" BASIS, | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
| # See the License for the specific language governing permissions and | |
| # limitations under the License. | |
| # ========= Copyright 2023-2024 @ CAMEL-AI.org. All Rights Reserved. ========= | |
| from __future__ import annotations | |
| from typing import Any, Dict, Optional, Union | |
| from pydantic import field_validator | |
| from camel.configs.base_config import BaseConfig | |
| class MistralConfig(BaseConfig): | |
| r"""Defines the parameters for generating chat completions using the | |
| Mistral API. | |
| reference: https://github.com/mistralai/client-python/blob/9d238f88c41689821d7b08570f13b43426f97fd6/src/mistralai/client.py#L195 | |
| #TODO: Support stream mode | |
| Args: | |
| temperature (Optional[float], optional): temperature the temperature | |
| to use for sampling, e.g. 0.5. | |
| top_p (Optional[float], optional): the cumulative probability of | |
| tokens to generate, e.g. 0.9. Defaults to None. | |
| max_tokens (Optional[int], optional): the maximum number of tokens to | |
| generate, e.g. 100. Defaults to None. | |
| stop (Optional[Union[str,list[str]]]): Stop generation if this token | |
| is detected. Or if one of these tokens is detected when providing | |
| a string list. | |
| random_seed (Optional[int], optional): the random seed to use for | |
| sampling, e.g. 42. Defaults to None. | |
| safe_prompt (bool, optional): whether to use safe prompt, e.g. true. | |
| Defaults to False. | |
| response_format (Union[Dict[str, str], ResponseFormat): format of the | |
| response. | |
| tool_choice (str, optional): Controls which (if | |
| any) tool is called by the model. :obj:`"none"` means the model | |
| will not call any tool and instead generates a message. | |
| :obj:`"auto"` means the model can pick between generating a | |
| message or calling one or more tools. :obj:`"any"` means the | |
| model must call one or more tools. :obj:`"auto"` is the default | |
| value. | |
| """ | |
| temperature: Optional[float] = None | |
| top_p: Optional[float] = None | |
| max_tokens: Optional[int] = None | |
| stop: Optional[Union[str, list[str]]] = None | |
| random_seed: Optional[int] = None | |
| safe_prompt: bool = False | |
| response_format: Optional[Union[Dict[str, str], Any]] = None | |
| tool_choice: Optional[str] = "auto" | |
| def fields_type_checking(cls, response_format): | |
| if response_format and not isinstance(response_format, dict): | |
| from mistralai.models import ResponseFormat | |
| if not isinstance(response_format, ResponseFormat): | |
| raise ValueError( | |
| f"The tool {response_format} should be an instance " | |
| "of `mistralai.models.ResponseFormat`." | |
| ) | |
| return response_format | |
| MISTRAL_API_PARAMS = {param for param in MistralConfig().model_fields.keys()} | |