Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion packages/ai-providers/server-ai-langchain/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -138,7 +138,7 @@ provider = await LangChainProvider.create(config)
async def invoke():
return await provider.invoke_model(messages)

response = await config.tracker.track_metrics_of(
response = await config.tracker.track_metrics_of_async(
invoke,
lambda r: r.metrics
)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,9 @@
create_langchain_model,
get_ai_metrics_from_response,
get_ai_usage_from_response,
get_tool_calls_from_response,
map_provider,
sum_token_usage_from_messages,
)
from ldai_langchain.langchain_model_runner import LangChainModelRunner
from ldai_langchain.langchain_runner_factory import LangChainRunnerFactory
Expand All @@ -18,5 +20,7 @@
'create_langchain_model',
'get_ai_metrics_from_response',
'get_ai_usage_from_response',
'get_tool_calls_from_response',
'map_provider',
'sum_token_usage_from_messages',
]
Original file line number Diff line number Diff line change
Expand Up @@ -115,3 +115,41 @@ def get_ai_metrics_from_response(response: Any) -> LDAIMetrics:
:return: LDAIMetrics with success status and token usage
"""
return LDAIMetrics(success=True, usage=get_ai_usage_from_response(response))


def get_tool_calls_from_response(response: Any) -> List[str]:
"""
Get tool call names from a LangChain provider response.

:param response: The response from the LangChain model
:return: List of tool names in order, or empty list if none
"""
names: List[str] = []
if hasattr(response, 'tool_calls') and isinstance(response.tool_calls, list):
for tc in response.tool_calls:
n = tc.get('name')
if n:
names.append(str(n))
return names


def sum_token_usage_from_messages(messages: List[Any]) -> Optional[TokenUsage]:
"""
Sum token usage across LangChain messages using get_ai_usage_from_response per message.

:param messages: List of message objects (e.g. from a graph state)
:return: Aggregated TokenUsage, or None if no usage on any message
"""
in_sum = 0
out_sum = 0
total_sum = 0
for m in messages:
u = get_ai_usage_from_response(m)
if u is None:
continue
in_sum += u.input
out_sum += u.output
total_sum += u.total
if in_sum == 0 and out_sum == 0 and total_sum == 0:
return None
return TokenUsage(total=total_sum, input=in_sum, output=out_sum)
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,15 @@

from ldai import LDMessage

from ldai_langchain import LangChainModelRunner, LangChainRunnerFactory, convert_messages_to_langchain, get_ai_metrics_from_response, map_provider
from ldai_langchain import (
LangChainModelRunner,
LangChainRunnerFactory,
convert_messages_to_langchain,
get_ai_metrics_from_response,
get_tool_calls_from_response,
map_provider,
sum_token_usage_from_messages,
)


class TestConvertMessages:
Expand Down Expand Up @@ -237,6 +245,82 @@ async def test_returns_success_false_when_structured_model_invocation_throws_err
assert result.metrics.usage is None


class TestGetToolCallsFromResponse:
"""Tests for get_tool_calls_from_response."""

def test_returns_tool_call_names_in_order(self):
"""Should return tool call names from response.tool_calls."""
mock_response = MagicMock()
mock_response.tool_calls = [
{'name': 'search', 'args': {}},
{'name': 'calculator', 'args': {}},
]
assert get_tool_calls_from_response(mock_response) == ['search', 'calculator']

def test_returns_empty_list_when_tool_calls_is_empty(self):
"""Should return empty list when tool_calls is an empty list."""
mock_response = MagicMock()
mock_response.tool_calls = []
assert get_tool_calls_from_response(mock_response) == []

def test_returns_empty_list_when_no_tool_calls_attribute(self):
"""Should return empty list when response has no tool_calls attribute."""
mock_response = MagicMock(spec=[])
assert get_tool_calls_from_response(mock_response) == []

def test_returns_empty_list_when_tool_calls_is_not_a_list(self):
"""Should return empty list when tool_calls is not a list."""
mock_response = MagicMock()
mock_response.tool_calls = 'not-a-list'
assert get_tool_calls_from_response(mock_response) == []

def test_skips_tool_calls_without_name(self):
"""Should skip tool calls that have no name."""
mock_response = MagicMock()
mock_response.tool_calls = [{'args': {}}, {'name': 'search', 'args': {}}]
assert get_tool_calls_from_response(mock_response) == ['search']


class TestSumTokenUsageFromMessages:
"""Tests for sum_token_usage_from_messages."""

def test_sums_usage_across_messages(self):
"""Should sum token usage from all messages."""
msg1 = AIMessage(content='a')
msg1.usage_metadata = {'total_tokens': 10, 'input_tokens': 6, 'output_tokens': 4}
msg2 = AIMessage(content='b')
msg2.usage_metadata = {'total_tokens': 20, 'input_tokens': 12, 'output_tokens': 8}

result = sum_token_usage_from_messages([msg1, msg2])

assert result is not None
assert result.total == 30
assert result.input == 18
assert result.output == 12

def test_returns_none_when_no_usage_on_any_message(self):
"""Should return None when no message has usage metadata."""
msg = AIMessage(content='hello')
assert sum_token_usage_from_messages([msg]) is None

def test_returns_none_for_empty_list(self):
"""Should return None for an empty message list."""
assert sum_token_usage_from_messages([]) is None

def test_skips_messages_without_usage(self):
"""Should skip messages that have no usage and sum the rest."""
msg1 = AIMessage(content='a')
msg2 = AIMessage(content='b')
msg2.usage_metadata = {'total_tokens': 5, 'input_tokens': 3, 'output_tokens': 2}

result = sum_token_usage_from_messages([msg1, msg2])

assert result is not None
assert result.total == 5
assert result.input == 3
assert result.output == 2


class TestGetLlm:
"""Tests for LangChainModelRunner.get_llm."""

Expand Down
6 changes: 3 additions & 3 deletions packages/sdk/server-ai/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -150,8 +150,8 @@ async def main():
# Create LangChain model from configuration
llm = await LangChainProvider.create_langchain_model(ai_config)

# Use with tracking
response = await ai_config.tracker.track_metrics_of(
# Use with tracking (sync invoke)
response = ai_config.tracker.track_metrics_of(
lambda: llm.invoke(messages),
lambda result: LangChainProvider.get_ai_metrics_from_response(result)
)
Expand Down Expand Up @@ -190,7 +190,7 @@ async def main():
temperature=ai_config.model.get_parameter('temperature') if ai_config.model else 0.5,
)

result = await ai_config.tracker.track_metrics_of(
result = await ai_config.tracker.track_metrics_of_async(
call_custom_provider,
map_custom_provider_metrics
)
Expand Down
2 changes: 1 addition & 1 deletion packages/sdk/server-ai/src/ldai/judge/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@ async def evaluate(
messages = self._construct_evaluation_messages(input_text, output_text)
assert self._evaluation_response_structure is not None

response = await self._ai_config_tracker.track_metrics_of(
response = await self._ai_config_tracker.track_metrics_of_async(
lambda: self._model_runner.invoke_structured_model(messages, self._evaluation_response_structure),
lambda result: result.metrics,
)
Expand Down
2 changes: 1 addition & 1 deletion packages/sdk/server-ai/src/ldai/managed_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ async def invoke(self, prompt: str) -> ModelResponse:
config_messages = self._ai_config.messages or []
all_messages = config_messages + self._messages

response = await self._tracker.track_metrics_of(
response = await self._tracker.track_metrics_of_async(
lambda: self._model_runner.invoke_model(all_messages),
lambda result: result.metrics,
)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -77,7 +77,7 @@ def _with_fallback(
continue
result = fn(provider_factory)
if result is not None:
log.debug(f"Successfully created capability using provider '{provider_type}'")
log.debug(f"Successfully invoked create function with provider '{provider_type}'")
return result
except Exception as exc:
log.warning(f"Provider '{provider_type}' failed: {exc}")
Expand Down
Loading
Loading