Added LiteLLM to the stack

This commit is contained in:
2025-08-18 09:40:50 +00:00
parent 0648c1968c
commit d220b04e32
2682 changed files with 533609 additions and 1 deletions

View File

@@ -0,0 +1,479 @@
import os
import sys
from unittest.mock import MagicMock, patch
import pytest
sys.path.insert(
0, os.path.abspath("../..")
) # Adds the parent directory to the system path
from litellm.llms.mistral.chat.transformation import MistralConfig
@pytest.mark.asyncio
async def test_mistral_chat_transformation():
mistral_config = MistralConfig()
result = mistral_config._transform_messages(
**{
"messages": [
{
"content": [
{"type": "text", "text": "Here is a representation of text"},
{
"type": "image_url",
"image_url": "https://images.pexels.com/photos/13268478/pexels-photo-13268478.jpeg",
},
],
"role": "user",
}
],
"model": "mistral-medium-latest",
"is_async": True,
}
)
class TestMistralReasoningSupport:
"""Test suite for Mistral Magistral reasoning functionality."""
def test_get_supported_openai_params_magistral_model(self):
"""Test that magistral models support reasoning parameters."""
mistral_config = MistralConfig()
# Test magistral model supports reasoning parameters
supported_params = mistral_config.get_supported_openai_params("mistral/magistral-medium-2506")
assert "reasoning_effort" in supported_params
assert "thinking" in supported_params
# Test non-magistral model doesn't include reasoning parameters
supported_params_normal = mistral_config.get_supported_openai_params("mistral/mistral-large-latest")
assert "reasoning_effort" not in supported_params_normal
assert "thinking" not in supported_params_normal
def test_map_openai_params_reasoning_effort(self):
"""Test that reasoning_effort parameter is properly mapped for magistral models."""
mistral_config = MistralConfig()
# Test reasoning_effort mapping for magistral model
optional_params = {}
result = mistral_config.map_openai_params(
non_default_params={"reasoning_effort": "low"},
optional_params=optional_params,
model="mistral/magistral-medium-2506",
drop_params=False,
)
assert result.get("_add_reasoning_prompt") is True
# Test reasoning_effort ignored for non-magistral model
optional_params_normal = {}
result_normal = mistral_config.map_openai_params(
non_default_params={"reasoning_effort": "low"},
optional_params=optional_params_normal,
model="mistral/mistral-large-latest",
drop_params=False,
)
assert "_add_reasoning_prompt" not in result_normal
def test_map_openai_params_thinking(self):
"""Test that thinking parameter is properly mapped for magistral models."""
mistral_config = MistralConfig()
# Test thinking mapping for magistral model
optional_params = {}
result = mistral_config.map_openai_params(
non_default_params={"thinking": {"budget": 1000}},
optional_params=optional_params,
model="mistral/magistral-small-2506",
drop_params=False,
)
assert result.get("_add_reasoning_prompt") is True
def test_get_mistral_reasoning_system_prompt(self):
"""Test that the reasoning system prompt is properly formatted."""
prompt = MistralConfig._get_mistral_reasoning_system_prompt()
assert isinstance(prompt, str)
assert len(prompt) > 50 # Ensure it's not empty
def test_add_reasoning_system_prompt_no_existing_system_message(self):
"""Test adding reasoning system prompt when no system message exists."""
mistral_config = MistralConfig()
messages = [
{"role": "user", "content": "What is 2+2?"}
]
optional_params = {"_add_reasoning_prompt": True}
result = mistral_config._add_reasoning_system_prompt_if_needed(messages, optional_params)
# Should add a new system message at the beginning
assert len(result) == 2
assert result[0]["role"] == "system"
assert "<think>" in result[0]["content"]
assert result[1]["role"] == "user"
assert result[1]["content"] == "What is 2+2?"
# Should remove the internal flag
assert "_add_reasoning_prompt" not in optional_params
def test_add_reasoning_system_prompt_with_existing_system_message(self):
"""Test adding reasoning system prompt when system message already exists."""
mistral_config = MistralConfig()
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "What is 2+2?"}
]
optional_params = {"_add_reasoning_prompt": True}
result = mistral_config._add_reasoning_system_prompt_if_needed(messages, optional_params)
# Should modify existing system message
assert len(result) == 2
assert result[0]["role"] == "system"
assert "<think>" in result[0]["content"]
assert "You are a helpful assistant." in result[0]["content"]
assert result[1]["role"] == "user"
# Should remove the internal flag
assert "_add_reasoning_prompt" not in optional_params
def test_add_reasoning_system_prompt_with_existing_list_content(self):
"""Test adding reasoning system prompt when system message has list content."""
mistral_config = MistralConfig()
messages = [
{
"role": "system",
"content": [
{"type": "text", "text": "You are a helpful assistant."},
{"type": "text", "text": "You always provide detailed explanations."}
]
},
{"role": "user", "content": "What is 2+2?"}
]
optional_params = {"_add_reasoning_prompt": True}
result = mistral_config._add_reasoning_system_prompt_if_needed(messages, optional_params)
# Should modify existing system message preserving list format
assert len(result) == 2
assert result[0]["role"] == "system"
assert isinstance(result[0]["content"], list)
# First item should be the reasoning prompt
assert result[0]["content"][0]["type"] == "text"
assert "<think>" in result[0]["content"][0]["text"]
# Original content should be preserved
assert "You are a helpful assistant." in result[0]["content"][1]["text"]
assert "You always provide detailed explanations." in result[0]["content"][2]["text"]
assert result[1]["role"] == "user"
# Should remove the internal flag
assert "_add_reasoning_prompt" not in optional_params
def test_add_reasoning_system_prompt_preserves_content_types(self):
"""Test that reasoning prompt preserves original content types (string vs list)."""
mistral_config = MistralConfig()
# Test with string content
string_messages = [
{"role": "system", "content": "You are helpful."},
{"role": "user", "content": "Hello"}
]
string_params = {"_add_reasoning_prompt": True}
string_result = mistral_config._add_reasoning_system_prompt_if_needed(string_messages, string_params)
assert isinstance(string_result[0]["content"], str)
assert "<think>" in string_result[0]["content"]
assert "You are helpful." in string_result[0]["content"]
# Test with list content
list_messages = [
{
"role": "system",
"content": [{"type": "text", "text": "You are helpful."}]
},
{"role": "user", "content": "Hello"}
]
list_params = {"_add_reasoning_prompt": True}
list_result = mistral_config._add_reasoning_system_prompt_if_needed(list_messages, list_params)
assert isinstance(list_result[0]["content"], list)
assert list_result[0]["content"][0]["type"] == "text"
assert "<think>" in list_result[0]["content"][0]["text"]
assert "You are helpful." in list_result[0]["content"][1]["text"]
def test_add_reasoning_system_prompt_no_flag(self):
"""Test that no modification happens when _add_reasoning_prompt flag is not set."""
mistral_config = MistralConfig()
messages = [
{"role": "user", "content": "What is 2+2?"}
]
optional_params = {}
result = mistral_config._add_reasoning_system_prompt_if_needed(messages, optional_params)
# Should return messages unchanged
assert result == messages
assert len(result) == 1
def test_transform_request_magistral_with_reasoning(self):
"""Test transform_request method for magistral model with reasoning."""
mistral_config = MistralConfig()
messages = [
{"role": "user", "content": "What is 15 * 7?"}
]
optional_params = {"_add_reasoning_prompt": True}
result = mistral_config.transform_request(
model="mistral/magistral-medium-2506",
messages=messages,
optional_params=optional_params,
litellm_params={},
headers={}
)
# Should have added system message
assert len(result["messages"]) == 2
assert result["messages"][0]["role"] == "system"
assert "<think>" in result["messages"][0]["content"]
assert result["messages"][1]["role"] == "user"
# Should remove internal flag from optional_params
assert "_add_reasoning_prompt" not in result
def test_transform_request_magistral_without_reasoning(self):
"""Test transform_request method for magistral model without reasoning."""
mistral_config = MistralConfig()
messages = [
{"role": "user", "content": "What is 15 * 7?"}
]
optional_params = {}
result = mistral_config.transform_request(
model="mistral/magistral-medium-2506",
messages=messages,
optional_params=optional_params,
litellm_params={},
headers={}
)
# Should not modify messages
assert len(result["messages"]) == 1
assert result["messages"][0]["role"] == "user"
def test_transform_request_non_magistral_with_reasoning_params(self):
"""Test that non-magistral models ignore reasoning parameters."""
mistral_config = MistralConfig()
messages = [
{"role": "user", "content": "What is 15 * 7?"}
]
optional_params = {"_add_reasoning_prompt": True}
result = mistral_config.transform_request(
model="mistral/mistral-large-latest",
messages=messages,
optional_params=optional_params,
litellm_params={},
headers={}
)
# Should not add system message for non-magistral models
assert len(result["messages"]) == 1
assert result["messages"][0]["role"] == "user"
def test_case_insensitive_magistral_detection(self):
"""Test that magistral model detection is case-insensitive."""
mistral_config = MistralConfig()
# Test various case combinations
models_to_test = [
"mistral/Magistral-medium-2506",
"mistral/MAGISTRAL-MEDIUM-2506",
"mistral/magistral-SMALL-2506",
"MaGiStRaL-medium-2506"
]
for model in models_to_test:
supported_params = mistral_config.get_supported_openai_params(model)
assert "reasoning_effort" in supported_params, f"Failed for model: {model}"
def test_end_to_end_reasoning_workflow(self):
"""Test the complete workflow from parameter to system prompt injection."""
mistral_config = MistralConfig()
# Step 1: Map parameters
optional_params = {}
mapped_params = mistral_config.map_openai_params(
non_default_params={"reasoning_effort": "high", "temperature": 0.7},
optional_params=optional_params,
model="mistral/magistral-medium-2506",
drop_params=False,
)
assert mapped_params.get("_add_reasoning_prompt") is True
assert mapped_params.get("temperature") == 0.7
# Step 2: Transform request
messages = [
{"role": "user", "content": "Solve for x: 2x + 5 = 13"}
]
result = mistral_config.transform_request(
model="mistral/magistral-medium-2506",
messages=messages,
optional_params=mapped_params,
litellm_params={},
headers={}
)
# Verify final result
assert len(result["messages"]) == 2
assert result["messages"][0]["role"] == "system"
assert "<think>" in result["messages"][0]["content"]
assert result["messages"][1]["role"] == "user"
assert result["messages"][1]["content"] == "Solve for x: 2x + 5 = 13"
assert result.get("temperature") == 0.7
assert "_add_reasoning_prompt" not in result
class TestMistralNameHandling:
"""Test suite for Mistral name handling in messages."""
def test_handle_name_in_message_tool_role_empty_name_removes_name(self):
"""Test that empty name is removed for tool messages."""
# Test with empty string
tool_message = {"role": "tool", "content": "Function result", "name": ""}
result = MistralConfig._handle_name_in_message(tool_message)
assert "name" not in result
assert result["role"] == "tool"
assert result["content"] == "Function result"
def test_handle_name_in_message_tool_role_valid_name_keeps_name(self):
"""Test that valid name is kept for tool messages."""
# Test with normal function name
tool_message = {"role": "tool", "content": "Function result", "name": "get_weather"}
result = MistralConfig._handle_name_in_message(tool_message)
assert "name" in result
assert result["name"] == "get_weather"
assert result["role"] == "tool"
assert result["content"] == "Function result"
def test_handle_name_in_message_no_name_field(self):
"""Test that messages without name field are unchanged."""
# Test with user role
user_message = {"role": "user", "content": "Hello"}
result = MistralConfig._handle_name_in_message(user_message)
assert "name" not in result
assert result["role"] == "user"
assert result["content"] == "Hello"
class TestMistralParallelToolCalls:
"""Test suite for Mistral parallel tool calls functionality."""
def test_get_supported_openai_params_includes_parallel_tool_calls(self):
"""Test that parallel_tool_calls is in supported parameters."""
mistral_config = MistralConfig()
supported_params = mistral_config.get_supported_openai_params("mistral/mistral-large-latest")
assert "parallel_tool_calls" in supported_params
def test_transform_request_preserves_parallel_tool_calls(self):
"""Test that transform_request preserves parallel_tool_calls parameter."""
mistral_config = MistralConfig()
messages = [
{"role": "user", "content": "What's the weather like?"}
]
optional_params = {"parallel_tool_calls": True}
result = mistral_config.transform_request(
model="mistral/mistral-large-latest",
messages=messages,
optional_params=optional_params,
litellm_params={},
headers={}
)
assert result.get("parallel_tool_calls") is True
assert len(result["messages"]) == 1
assert result["messages"][0]["role"] == "user"
class TestMistralEmptyContentHandling:
"""Test suite for Mistral empty content response handling functionality."""
def test_handle_empty_content_response_converts_empty_string_to_none(self):
"""Test that empty string content is converted to None."""
response_data = {
"choices": [
{
"message": {
"content": "",
"role": "assistant"
},
"finish_reason": "stop"
}
]
}
result = MistralConfig._handle_empty_content_response(response_data)
assert result["choices"][0]["message"]["content"] is None
def test_handle_empty_content_response_preserves_actual_content(self):
"""Test that actual content is preserved unchanged."""
response_data = {
"choices": [
{
"message": {
"content": "Hello, how can I help you?",
"role": "assistant"
},
"finish_reason": "stop"
}
]
}
result = MistralConfig._handle_empty_content_response(response_data)
assert result["choices"][0]["message"]["content"] == "Hello, how can I help you?"
def test_handle_empty_content_response_handles_multiple_choices(self):
"""Test that only the first choice is processed for empty content."""
response_data = {
"choices": [
{
"message": {
"content": "",
"role": "assistant"
},
"finish_reason": "stop"
},
{
"message": {
"content": "",
"role": "assistant"
},
"finish_reason": "stop"
}
]
}
result = MistralConfig._handle_empty_content_response(response_data)
# Only first choice should be converted to None
assert result["choices"][0]["message"]["content"] is None
# Second choice should remain as empty string
assert result["choices"][1]["message"]["content"] is None

View File

@@ -0,0 +1,160 @@
import pytest
import litellm
@pytest.fixture(autouse=True)
def add_mistral_api_key_to_env(monkeypatch):
"""Add Mistral API key to environment for testing."""
monkeypatch.setenv("MISTRAL_API_KEY", "fake-mistral-api-key-12345")
@pytest.fixture
def mistral_api_response():
"""Mock response data for Mistral API calls."""
return {
"id": "chatcmpl-mistral-123",
"object": "chat.completion",
"created": 1677652288,
"model": "mistral-medium-latest",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello from Mistral! How can I help you today?",
},
"finish_reason": "stop",
}
],
"usage": {"prompt_tokens": 10, "completion_tokens": 15, "total_tokens": 25},
}
@pytest.fixture
def mistral_api_response_with_empty_content():
"""Mock response data for Mistral API calls with empty content that should be converted to None."""
return {
"id": "chatcmpl-mistral-123",
"object": "chat.completion",
"created": 1677652288,
"model": "mistral-medium-latest",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "", # Empty string that should be converted to None
},
"finish_reason": "stop",
}
],
"usage": {"prompt_tokens": 10, "completion_tokens": 0, "total_tokens": 10},
}
@pytest.mark.parametrize("sync_mode", [True, False])
@pytest.mark.asyncio
async def test_mistral_basic_completion(sync_mode, respx_mock, mistral_api_response):
"""Test basic Mistral completion functionality."""
litellm.disable_aiohttp_transport = True
model = "mistral/mistral-medium-latest"
messages = [{"role": "user", "content": "Hello, how are you?"}]
# Mock the Mistral API endpoint
respx_mock.post("https://api.mistral.ai/v1/chat/completions").respond(
json=mistral_api_response
)
if sync_mode:
response = litellm.completion(model=model, messages=messages)
else:
response = await litellm.acompletion(model=model, messages=messages)
# Verify response
assert response.choices[0].message.content == "Hello from Mistral! How can I help you today?"
assert response.model == "mistral-medium-latest"
assert response.usage.total_tokens == 25
@pytest.mark.parametrize("sync_mode", [True, False])
@pytest.mark.asyncio
async def test_mistral_transform_response_empty_content_conversion(sync_mode, respx_mock, mistral_api_response_with_empty_content):
"""
Test that Mistral's transform_response method is being called by verifying
the specific behavior of converting empty string content to None.
This test verifies that the _handle_empty_content_response method in
MistralConfig.transform_response is being applied.
"""
litellm.disable_aiohttp_transport = True
model = "mistral/mistral-medium-latest"
messages = [{"role": "user", "content": "Generate an empty response"}]
# Mock the Mistral API endpoint with empty content
respx_mock.post("https://api.mistral.ai/v1/chat/completions").respond(
json=mistral_api_response_with_empty_content
)
if sync_mode:
response = litellm.completion(model=model, messages=messages)
else:
response = await litellm.acompletion(model=model, messages=messages)
# Verify that the transform_response method was called by checking that
# empty string content was converted to None (Mistral-specific behavior)
assert response.choices[0].message.content is None
assert response.model == "mistral-medium-latest"
assert response.usage.total_tokens == 10
@pytest.mark.parametrize("sync_mode", [True, False])
@pytest.mark.asyncio
async def test_mistral_transform_request_name_field_removal(sync_mode, respx_mock, mistral_api_response):
"""
Test that Mistral's transform_request method is being called by verifying
the specific behavior of removing the 'name' field from non-tool messages.
This test verifies that the _handle_name_in_message method in
MistralConfig._transform_messages is being applied.
"""
litellm.disable_aiohttp_transport = True
model = "mistral/mistral-medium-latest"
# Include a message with 'name' field that should be removed for non-tool messages
messages = [
{"role": "user", "content": "Hello", "name": "should_be_removed"},
{"role": "assistant", "content": "Hi there!"},
{"role": "user", "content": "How are you?"}
]
# Mock the Mistral API endpoint
respx_mock.post("https://api.mistral.ai/v1/chat/completions").respond(
json=mistral_api_response
)
if sync_mode:
response = litellm.completion(model=model, messages=messages)
else:
response = await litellm.acompletion(model=model, messages=messages)
# Verify the response works (if transform_request wasn't called, the API would reject the request)
assert response.choices[0].message.content == "Hello from Mistral! How can I help you today?"
assert response.model == "mistral-medium-latest"
# Verify that the request was made (if transform_request failed, this would fail)
assert len(respx_mock.calls) == 1
# Get the actual request that was made
request = respx_mock.calls[0].request
import json
request_data = json.loads(request.content.decode('utf-8'))
# Verify that the 'name' field was removed from the user message
# (Mistral API only supports 'name' in tool messages)
user_message = request_data["messages"][0]
assert user_message["role"] == "user"
assert user_message["content"] == "Hello"
assert "name" not in user_message # The 'name' field should have been removed