197 lines
7.0 KiB
Python
197 lines
7.0 KiB
Python
|
|
"""
|
||
|
|
Translates from OpenAI's `/v1/chat/completions` to Moonshot AI's `/v1/chat/completions`
|
||
|
|
"""
|
||
|
|
|
||
|
|
from typing import Any, Coroutine, List, Literal, Optional, Tuple, Union, overload
|
||
|
|
|
||
|
|
from litellm.litellm_core_utils.prompt_templates.common_utils import (
|
||
|
|
handle_messages_with_content_list_to_str_conversion,
|
||
|
|
)
|
||
|
|
from litellm.secret_managers.main import get_secret_str
|
||
|
|
from litellm.types.llms.openai import AllMessageValues
|
||
|
|
|
||
|
|
from ...openai.chat.gpt_transformation import OpenAIGPTConfig
|
||
|
|
|
||
|
|
|
||
|
|
class MoonshotChatConfig(OpenAIGPTConfig):
|
||
|
|
@overload
|
||
|
|
def _transform_messages(
|
||
|
|
self, messages: List[AllMessageValues], model: str, is_async: Literal[True]
|
||
|
|
) -> Coroutine[Any, Any, List[AllMessageValues]]:
|
||
|
|
...
|
||
|
|
|
||
|
|
@overload
|
||
|
|
def _transform_messages(
|
||
|
|
self,
|
||
|
|
messages: List[AllMessageValues],
|
||
|
|
model: str,
|
||
|
|
is_async: Literal[False] = False,
|
||
|
|
) -> List[AllMessageValues]:
|
||
|
|
...
|
||
|
|
|
||
|
|
def _transform_messages(
|
||
|
|
self, messages: List[AllMessageValues], model: str, is_async: bool = False
|
||
|
|
) -> Union[List[AllMessageValues], Coroutine[Any, Any, List[AllMessageValues]]]:
|
||
|
|
"""
|
||
|
|
Moonshot text-only models don't support content in list format.
|
||
|
|
Multimodal models (kimi-k2.5, kimi-latest, etc.) accept the
|
||
|
|
standard OpenAI content array with non-text blocks (image_url,
|
||
|
|
input_audio, video_url, file, etc.).
|
||
|
|
|
||
|
|
If any message contains a non-text content part, skip flattening
|
||
|
|
so the multimodal payload is preserved.
|
||
|
|
"""
|
||
|
|
has_non_text = False
|
||
|
|
for m in messages:
|
||
|
|
_content = m.get("content")
|
||
|
|
if _content and isinstance(_content, list):
|
||
|
|
if any(c.get("type") != "text" for c in _content):
|
||
|
|
has_non_text = True
|
||
|
|
break
|
||
|
|
|
||
|
|
if not has_non_text:
|
||
|
|
messages = handle_messages_with_content_list_to_str_conversion(messages)
|
||
|
|
|
||
|
|
if is_async:
|
||
|
|
return super()._transform_messages(
|
||
|
|
messages=messages, model=model, is_async=True
|
||
|
|
)
|
||
|
|
else:
|
||
|
|
return super()._transform_messages(
|
||
|
|
messages=messages, model=model, is_async=False
|
||
|
|
)
|
||
|
|
|
||
|
|
def _get_openai_compatible_provider_info(
|
||
|
|
self, api_base: Optional[str], api_key: Optional[str]
|
||
|
|
) -> Tuple[Optional[str], Optional[str]]:
|
||
|
|
api_base = (
|
||
|
|
api_base
|
||
|
|
or get_secret_str("MOONSHOT_API_BASE")
|
||
|
|
or "https://api.moonshot.ai/v1"
|
||
|
|
) # type: ignore
|
||
|
|
dynamic_api_key = api_key or get_secret_str("MOONSHOT_API_KEY")
|
||
|
|
return api_base, dynamic_api_key
|
||
|
|
|
||
|
|
def get_complete_url(
|
||
|
|
self,
|
||
|
|
api_base: Optional[str],
|
||
|
|
api_key: Optional[str],
|
||
|
|
model: str,
|
||
|
|
optional_params: dict,
|
||
|
|
litellm_params: dict,
|
||
|
|
stream: Optional[bool] = None,
|
||
|
|
) -> str:
|
||
|
|
"""
|
||
|
|
If api_base is not provided, use the default Moonshot AI /chat/completions endpoint.
|
||
|
|
"""
|
||
|
|
if not api_base:
|
||
|
|
api_base = "https://api.moonshot.ai/v1"
|
||
|
|
|
||
|
|
if not api_base.endswith("/chat/completions"):
|
||
|
|
api_base = f"{api_base}/chat/completions"
|
||
|
|
|
||
|
|
return api_base
|
||
|
|
|
||
|
|
def get_supported_openai_params(self, model: str) -> list:
|
||
|
|
"""
|
||
|
|
Get the supported OpenAI params for Moonshot AI models
|
||
|
|
|
||
|
|
Moonshot AI limitations:
|
||
|
|
- functions parameter is not supported (use tools instead)
|
||
|
|
- tool_choice doesn't support "required" value
|
||
|
|
- kimi-thinking-preview doesn't support tool calls at all
|
||
|
|
"""
|
||
|
|
excluded_params: List[str] = ["functions"]
|
||
|
|
|
||
|
|
# kimi-thinking-preview has additional limitations
|
||
|
|
if "kimi-thinking-preview" in model:
|
||
|
|
excluded_params.extend(["tools", "tool_choice"])
|
||
|
|
|
||
|
|
base_openai_params = super().get_supported_openai_params(model=model)
|
||
|
|
final_params: List[str] = []
|
||
|
|
for param in base_openai_params:
|
||
|
|
if param not in excluded_params:
|
||
|
|
final_params.append(param)
|
||
|
|
|
||
|
|
return final_params
|
||
|
|
|
||
|
|
def map_openai_params(
|
||
|
|
self,
|
||
|
|
non_default_params: dict,
|
||
|
|
optional_params: dict,
|
||
|
|
model: str,
|
||
|
|
drop_params: bool,
|
||
|
|
) -> dict:
|
||
|
|
"""
|
||
|
|
Map OpenAI parameters to Moonshot AI parameters
|
||
|
|
|
||
|
|
Handles Moonshot AI specific limitations:
|
||
|
|
- tool_choice doesn't support "required" value
|
||
|
|
- Temperature <0.3 limitation for n>1
|
||
|
|
"""
|
||
|
|
supported_openai_params = self.get_supported_openai_params(model)
|
||
|
|
for param, value in non_default_params.items():
|
||
|
|
if param == "max_completion_tokens":
|
||
|
|
optional_params["max_tokens"] = value
|
||
|
|
elif param in supported_openai_params:
|
||
|
|
optional_params[param] = value
|
||
|
|
|
||
|
|
##########################################
|
||
|
|
# temperature limitations
|
||
|
|
# 1. `temperature` on KIMI API is [0, 1] but OpenAI is [0, 2]
|
||
|
|
# 2. If temperature < 0.3 and n > 1, KIMI will raise an exception.
|
||
|
|
# If we enter this condition, we set the temperature to 0.3 as suggested by Moonshot AI
|
||
|
|
##########################################
|
||
|
|
if "temperature" in optional_params:
|
||
|
|
if optional_params["temperature"] > 1:
|
||
|
|
optional_params["temperature"] = 1
|
||
|
|
if optional_params["temperature"] < 0.3 and optional_params.get("n", 1) > 1:
|
||
|
|
optional_params["temperature"] = 0.3
|
||
|
|
return optional_params
|
||
|
|
|
||
|
|
def transform_request(
|
||
|
|
self,
|
||
|
|
model: str,
|
||
|
|
messages: List[AllMessageValues],
|
||
|
|
optional_params: dict,
|
||
|
|
litellm_params: dict,
|
||
|
|
headers: dict,
|
||
|
|
) -> dict:
|
||
|
|
"""
|
||
|
|
Transform the overall request to be sent to the API.
|
||
|
|
Returns:
|
||
|
|
dict: The transformed request. Sent as the body of the API call.
|
||
|
|
"""
|
||
|
|
# Add tool_choice="required" message if needed
|
||
|
|
if optional_params.get("tool_choice", None) == "required":
|
||
|
|
messages = self._add_tool_choice_required_message(
|
||
|
|
messages=messages,
|
||
|
|
optional_params=optional_params,
|
||
|
|
)
|
||
|
|
|
||
|
|
# Call parent transform_request which handles _transform_messages
|
||
|
|
return super().transform_request(
|
||
|
|
model=model,
|
||
|
|
messages=messages,
|
||
|
|
optional_params=optional_params,
|
||
|
|
litellm_params=litellm_params,
|
||
|
|
headers=headers,
|
||
|
|
)
|
||
|
|
|
||
|
|
def _add_tool_choice_required_message(
|
||
|
|
self, messages: List[AllMessageValues], optional_params: dict
|
||
|
|
) -> List[AllMessageValues]:
|
||
|
|
"""
|
||
|
|
Add a message to the messages list to indicate that the tool choice is required.
|
||
|
|
|
||
|
|
https://platform.moonshot.ai/docs/guide/migrating-from-openai-to-kimi#about-tool_choice
|
||
|
|
"""
|
||
|
|
messages.append(
|
||
|
|
{
|
||
|
|
"role": "user",
|
||
|
|
"content": "Please select a tool to handle the current issue.", # Usually, the Kimi large language model understands the intention to invoke a tool and selects one for invocation
|
||
|
|
}
|
||
|
|
)
|
||
|
|
optional_params.pop("tool_choice")
|
||
|
|
return messages
|