126 lines
4.3 KiB
Python
126 lines
4.3 KiB
Python
|
|
"""
|
||
|
|
Translates from OpenAI's `/v1/chat/completions` to DeepSeek's `/v1/chat/completions`
|
||
|
|
"""
|
||
|
|
|
||
|
|
from typing import Any, Coroutine, List, Literal, Optional, Tuple, Union, overload
|
||
|
|
|
||
|
|
from litellm.litellm_core_utils.prompt_templates.common_utils import (
|
||
|
|
handle_messages_with_content_list_to_str_conversion,
|
||
|
|
)
|
||
|
|
from litellm.secret_managers.main import get_secret_str
|
||
|
|
from litellm.types.llms.openai import AllMessageValues
|
||
|
|
|
||
|
|
from ...openai.chat.gpt_transformation import OpenAIGPTConfig
|
||
|
|
|
||
|
|
|
||
|
|
class DeepSeekChatConfig(OpenAIGPTConfig):
|
||
|
|
def get_supported_openai_params(self, model: str) -> list:
|
||
|
|
"""
|
||
|
|
DeepSeek reasoner models support thinking parameter.
|
||
|
|
"""
|
||
|
|
params = super().get_supported_openai_params(model)
|
||
|
|
params.extend(["thinking", "reasoning_effort"])
|
||
|
|
return params
|
||
|
|
|
||
|
|
def map_openai_params(
|
||
|
|
self,
|
||
|
|
non_default_params: dict,
|
||
|
|
optional_params: dict,
|
||
|
|
model: str,
|
||
|
|
drop_params: bool,
|
||
|
|
) -> dict:
|
||
|
|
"""
|
||
|
|
Map OpenAI params to DeepSeek params.
|
||
|
|
|
||
|
|
Handles `thinking` and `reasoning_effort` parameters for DeepSeek reasoner models.
|
||
|
|
DeepSeek only supports `{"type": "enabled"}` - no budget_tokens like Anthropic.
|
||
|
|
|
||
|
|
Reference: https://api-docs.deepseek.com/guides/thinking_mode
|
||
|
|
"""
|
||
|
|
# Let parent handle standard params first
|
||
|
|
optional_params = super().map_openai_params(
|
||
|
|
non_default_params, optional_params, model, drop_params
|
||
|
|
)
|
||
|
|
|
||
|
|
# Pop thinking/reasoning_effort from optional_params first (parent may have added them)
|
||
|
|
# Then re-add only if valid for DeepSeek
|
||
|
|
thinking_value = optional_params.pop("thinking", None)
|
||
|
|
reasoning_effort = optional_params.pop("reasoning_effort", None)
|
||
|
|
|
||
|
|
# Handle thinking parameter - only accept {"type": "enabled"}
|
||
|
|
if thinking_value is not None:
|
||
|
|
if (
|
||
|
|
isinstance(thinking_value, dict)
|
||
|
|
and thinking_value.get("type") == "enabled"
|
||
|
|
):
|
||
|
|
# DeepSeek only accepts {"type": "enabled"}, ignore budget_tokens
|
||
|
|
optional_params["thinking"] = {"type": "enabled"}
|
||
|
|
|
||
|
|
# Handle reasoning_effort - map to thinking enabled
|
||
|
|
elif reasoning_effort is not None and reasoning_effort != "none":
|
||
|
|
optional_params["thinking"] = {"type": "enabled"}
|
||
|
|
|
||
|
|
return optional_params
|
||
|
|
|
||
|
|
@overload
|
||
|
|
def _transform_messages(
|
||
|
|
self, messages: List[AllMessageValues], model: str, is_async: Literal[True]
|
||
|
|
) -> Coroutine[Any, Any, List[AllMessageValues]]:
|
||
|
|
...
|
||
|
|
|
||
|
|
@overload
|
||
|
|
def _transform_messages(
|
||
|
|
self,
|
||
|
|
messages: List[AllMessageValues],
|
||
|
|
model: str,
|
||
|
|
is_async: Literal[False] = False,
|
||
|
|
) -> List[AllMessageValues]:
|
||
|
|
...
|
||
|
|
|
||
|
|
def _transform_messages(
|
||
|
|
self, messages: List[AllMessageValues], model: str, is_async: bool = False
|
||
|
|
) -> Union[List[AllMessageValues], Coroutine[Any, Any, List[AllMessageValues]]]:
|
||
|
|
"""
|
||
|
|
DeepSeek does not support content in list format.
|
||
|
|
"""
|
||
|
|
messages = handle_messages_with_content_list_to_str_conversion(messages)
|
||
|
|
if is_async:
|
||
|
|
return super()._transform_messages(
|
||
|
|
messages=messages, model=model, is_async=True
|
||
|
|
)
|
||
|
|
else:
|
||
|
|
return super()._transform_messages(
|
||
|
|
messages=messages, model=model, is_async=False
|
||
|
|
)
|
||
|
|
|
||
|
|
def _get_openai_compatible_provider_info(
|
||
|
|
self, api_base: Optional[str], api_key: Optional[str]
|
||
|
|
) -> Tuple[Optional[str], Optional[str]]:
|
||
|
|
api_base = (
|
||
|
|
api_base
|
||
|
|
or get_secret_str("DEEPSEEK_API_BASE")
|
||
|
|
or "https://api.deepseek.com/beta"
|
||
|
|
) # type: ignore
|
||
|
|
dynamic_api_key = api_key or get_secret_str("DEEPSEEK_API_KEY")
|
||
|
|
return api_base, dynamic_api_key
|
||
|
|
|
||
|
|
def get_complete_url(
|
||
|
|
self,
|
||
|
|
api_base: Optional[str],
|
||
|
|
api_key: Optional[str],
|
||
|
|
model: str,
|
||
|
|
optional_params: dict,
|
||
|
|
litellm_params: dict,
|
||
|
|
stream: Optional[bool] = None,
|
||
|
|
) -> str:
|
||
|
|
"""
|
||
|
|
If api_base is not provided, use the default DeepSeek /chat/completions endpoint.
|
||
|
|
"""
|
||
|
|
if not api_base:
|
||
|
|
api_base = "https://api.deepseek.com/beta"
|
||
|
|
|
||
|
|
if not api_base.endswith("/chat/completions"):
|
||
|
|
api_base = f"{api_base}/chat/completions"
|
||
|
|
|
||
|
|
return api_base
|