diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml
index 4d0e2d9..e2f90fb 100644
--- a/.github/workflows/ci.yml
+++ b/.github/workflows/ci.yml
@@ -21,10 +21,10 @@ jobs:
runs-on: ${{ github.repository == 'stainless-sdks/warp-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }}
if: (github.event_name == 'push' || github.event.pull_request.head.repo.fork) && (github.event_name != 'push' || github.event.head_commit.message != 'codegen metadata')
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
- name: Install uv
- uses: astral-sh/setup-uv@v5
+ uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2
with:
version: '0.10.2'
@@ -43,10 +43,10 @@ jobs:
id-token: write
runs-on: ${{ github.repository == 'stainless-sdks/warp-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }}
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
- name: Install uv
- uses: astral-sh/setup-uv@v5
+ uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2
with:
version: '0.10.2'
@@ -61,7 +61,7 @@ jobs:
github.repository == 'stainless-sdks/warp-api-python' &&
!startsWith(github.ref, 'refs/heads/stl/')
id: github-oidc
- uses: actions/github-script@v8
+ uses: actions/github-script@ed597411d8f924073f98dfc5c65a23a2325f34cd # v8.0.0
with:
script: core.setOutput('github_token', await core.getIDToken());
@@ -81,10 +81,10 @@ jobs:
runs-on: ${{ github.repository == 'stainless-sdks/warp-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }}
if: github.event_name == 'push' || github.event.pull_request.head.repo.fork
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
- name: Install uv
- uses: astral-sh/setup-uv@v5
+ uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2
with:
version: '0.10.2'
diff --git a/.github/workflows/publish-pypi.yml b/.github/workflows/publish-pypi.yml
index 74160e1..bae8377 100644
--- a/.github/workflows/publish-pypi.yml
+++ b/.github/workflows/publish-pypi.yml
@@ -17,10 +17,10 @@ jobs:
id-token: write
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
- name: Install uv
- uses: astral-sh/setup-uv@v5
+ uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2
with:
version: '0.9.13'
diff --git a/.github/workflows/release-doctor.yml b/.github/workflows/release-doctor.yml
index bb0047d..2f3f96e 100644
--- a/.github/workflows/release-doctor.yml
+++ b/.github/workflows/release-doctor.yml
@@ -12,7 +12,7 @@ jobs:
if: github.repository == 'warpdotdev/oz-sdk-python' && (github.event_name == 'push' || github.event_name == 'workflow_dispatch' || startsWith(github.head_ref, 'release-please') || github.head_ref == 'next')
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2
- name: Check release environment
run: |
diff --git a/.release-please-manifest.json b/.release-please-manifest.json
index a713055..d52d2b9 100644
--- a/.release-please-manifest.json
+++ b/.release-please-manifest.json
@@ -1,3 +1,3 @@
{
- ".": "0.12.0"
+ ".": "0.13.0"
}
\ No newline at end of file
diff --git a/.stats.yml b/.stats.yml
index 2cb0b3b..8e0d8bf 100644
--- a/.stats.yml
+++ b/.stats.yml
@@ -1,4 +1,4 @@
-configured_endpoints: 22
-openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/warp-bnavetta/warp-api-1fecc5f5d6ee664d804b81bd1aa6eec4d3f170ffa788d214fead4f7e95ab9d4e.yml
-openapi_spec_hash: 82990b03bd5a93e45bfc79db56ae7fc0
-config_hash: f52e7636f248f25c4ea0b086e7326816
+configured_endpoints: 23
+openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/warp-bnavetta/warp-api-b719dd35d7850ee303cdebf54fa3dfddb492a6f578344c2060cfae013b61541c.yml
+openapi_spec_hash: 4c21e0d940ef5fc42767be5380571c5d
+config_hash: 236823a4936c76818117c16aa5c188df
diff --git a/CHANGELOG.md b/CHANGELOG.md
index 4431ffa..43d4c99 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,5 +1,30 @@
# Changelog
+## 0.13.0 (2026-05-13)
+
+Full Changelog: [v0.12.0...v0.13.0](https://github.com/warpdotdev/oz-sdk-python/compare/v0.12.0...v0.13.0)
+
+### Features
+
+* Add default harness selection for agents ([8cd06f6](https://github.com/warpdotdev/oz-sdk-python/commit/8cd06f6fcb48444e3861ff130d2fb4fa820e3aa7))
+* Add per-agent AWS Bedrock OIDC inference role (backend) ([e4aadec](https://github.com/warpdotdev/oz-sdk-python/commit/e4aadec916b2bea7e05e0e263b307449002e7f8b))
+* **agents:** add prompt property to agent identity data model ([520c835](https://github.com/warpdotdev/oz-sdk-python/commit/520c8350b5b7850242adc082fe83a7ae666e6da6))
+* **api:** api update ([0976362](https://github.com/warpdotdev/oz-sdk-python/commit/09763624c5e172aa28fa43ed18bc0e391c04331d))
+* **api:** api update ([96d640c](https://github.com/warpdotdev/oz-sdk-python/commit/96d640cd68290e1a915fca51a8629cf9dab4d81d))
+* **api:** api update ([2730eea](https://github.com/warpdotdev/oz-sdk-python/commit/2730eea19d91df998de5ef1b5da4989670a5889a))
+* **api:** api update ([af81ef3](https://github.com/warpdotdev/oz-sdk-python/commit/af81ef3a2d48f378209cf69a3074997cb02b1b6c))
+* **api:** api update ([99b2d31](https://github.com/warpdotdev/oz-sdk-python/commit/99b2d31ba1a4c2c3d79fc11c5eb8d611c55613b8))
+* Codex auth: API key support. ([cb84f0c](https://github.com/warpdotdev/oz-sdk-python/commit/cb84f0c6c9625d8c0657c728d5fadae065eb37c9))
+* **internal/types:** support eagerly validating pydantic iterators ([a588f0e](https://github.com/warpdotdev/oz-sdk-python/commit/a588f0e5ba27503659b5abf01d0ebc01f652950d))
+* **memory:** agent identity memory store attachments — API layer ([94b5348](https://github.com/warpdotdev/oz-sdk-python/commit/94b5348152c6b0bfb03b0d3887366c4a65e397fb))
+* **memory:** wire memory stores into run pipeline and add listing endpoint ([6bb74c2](https://github.com/warpdotdev/oz-sdk-python/commit/6bb74c2b695cd268fe8466fc6099f082370ba54e))
+* Retrieve memories in third party harnesses ([7689e12](https://github.com/warpdotdev/oz-sdk-python/commit/7689e121d6f22efad3d81828721f8ed900b9cd28))
+
+
+### Bug Fixes
+
+* **client:** add missing f-string prefix in file type error message ([17a8e5b](https://github.com/warpdotdev/oz-sdk-python/commit/17a8e5bf17b882a440a067fee1569caf679f8b55))
+
## 0.12.0 (2026-05-07)
Full Changelog: [v0.11.0...v0.12.0](https://github.com/warpdotdev/oz-sdk-python/compare/v0.11.0...v0.12.0)
diff --git a/api.md b/api.md
index bd33474..e82443c 100644
--- a/api.md
+++ b/api.md
@@ -94,6 +94,7 @@ Methods:
- client.agent.agent.update(uid, \*\*params) -> AgentResponse
- client.agent.agent.list() -> ListAgentIdentitiesResponse
- client.agent.agent.delete(uid) -> None
+- client.agent.agent.get(uid) -> AgentResponse
## Sessions
diff --git a/pyproject.toml b/pyproject.toml
index 367b68a..533a8e4 100644
--- a/pyproject.toml
+++ b/pyproject.toml
@@ -1,6 +1,6 @@
[project]
name = "oz-agent-sdk"
-version = "0.12.0"
+version = "0.13.0"
description = "The official Python library for the oz-api API"
dynamic = ["readme"]
license = "Apache-2.0"
diff --git a/src/oz_agent_sdk/_files.py b/src/oz_agent_sdk/_files.py
index 0fdce17..76da9e0 100644
--- a/src/oz_agent_sdk/_files.py
+++ b/src/oz_agent_sdk/_files.py
@@ -99,7 +99,7 @@ async def async_to_httpx_files(files: RequestFiles | None) -> HttpxRequestFiles
elif is_sequence_t(files):
files = [(key, await _async_transform_file(file)) for key, file in files]
else:
- raise TypeError("Unexpected file type input {type(files)}, expected mapping or sequence")
+ raise TypeError(f"Unexpected file type input {type(files)}, expected mapping or sequence")
return files
diff --git a/src/oz_agent_sdk/_models.py b/src/oz_agent_sdk/_models.py
index e22dd2a..69f41a6 100644
--- a/src/oz_agent_sdk/_models.py
+++ b/src/oz_agent_sdk/_models.py
@@ -25,7 +25,9 @@
ClassVar,
Protocol,
Required,
+ Annotated,
ParamSpec,
+ TypeAlias,
TypedDict,
TypeGuard,
final,
@@ -79,7 +81,15 @@
from ._constants import RAW_RESPONSE_HEADER
if TYPE_CHECKING:
+ from pydantic import GetCoreSchemaHandler, ValidatorFunctionWrapHandler
+ from pydantic_core import CoreSchema, core_schema
from pydantic_core.core_schema import ModelField, ModelSchema, LiteralSchema, ModelFieldsSchema
+else:
+ try:
+ from pydantic_core import CoreSchema, core_schema
+ except ImportError:
+ CoreSchema = None
+ core_schema = None
__all__ = ["BaseModel", "GenericModel"]
@@ -396,6 +406,76 @@ def model_dump_json(
)
+class _EagerIterable(list[_T], Generic[_T]):
+ """
+ Accepts any Iterable[T] input (including generators), consumes it
+ eagerly, and validates all items upfront.
+
+ Validation preserves the original container type where possible
+ (e.g. a set[T] stays a set[T]). Serialization (model_dump / JSON)
+ always emits a list — round-tripping through model_dump() will not
+ restore the original container type.
+ """
+
+ @classmethod
+ def __get_pydantic_core_schema__(
+ cls,
+ source_type: Any,
+ handler: GetCoreSchemaHandler,
+ ) -> CoreSchema:
+ (item_type,) = get_args(source_type) or (Any,)
+ item_schema: CoreSchema = handler.generate_schema(item_type)
+ list_of_items_schema: CoreSchema = core_schema.list_schema(item_schema)
+
+ return core_schema.no_info_wrap_validator_function(
+ cls._validate,
+ list_of_items_schema,
+ serialization=core_schema.plain_serializer_function_ser_schema(
+ cls._serialize,
+ info_arg=False,
+ ),
+ )
+
+ @staticmethod
+ def _validate(v: Iterable[_T], handler: "ValidatorFunctionWrapHandler") -> Any:
+ original_type: type[Any] = type(v)
+
+ # Normalize to list so list_schema can validate each item
+ if isinstance(v, list):
+ items: list[_T] = v
+ else:
+ try:
+ items = list(v)
+ except TypeError as e:
+ raise TypeError("Value is not iterable") from e
+
+ # Validate items against the inner schema
+ validated: list[_T] = handler(items)
+
+ # Reconstruct original container type
+ if original_type is list:
+ return validated
+ # str(list) produces the list's repr, not a string built from items,
+ # so skip reconstruction for str and its subclasses.
+ if issubclass(original_type, str):
+ return validated
+ try:
+ return original_type(validated)
+ except (TypeError, ValueError):
+ # If the type cannot be reconstructed, just return the validated list
+ return validated
+
+ @staticmethod
+ def _serialize(v: Iterable[_T]) -> list[_T]:
+ """Always serialize as a list so Pydantic's JSON encoder is happy."""
+ if isinstance(v, list):
+ return v
+ return list(v)
+
+
+EagerIterable: TypeAlias = Annotated[Iterable[_T], _EagerIterable]
+
+
def _construct_field(value: object, field: FieldInfo, key: str) -> object:
if value is None:
return field_get_default(field)
diff --git a/src/oz_agent_sdk/_version.py b/src/oz_agent_sdk/_version.py
index b96fedf..1762ab1 100644
--- a/src/oz_agent_sdk/_version.py
+++ b/src/oz_agent_sdk/_version.py
@@ -1,4 +1,4 @@
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
__title__ = "oz_agent_sdk"
-__version__ = "0.12.0" # x-release-please-version
+__version__ = "0.13.0" # x-release-please-version
diff --git a/src/oz_agent_sdk/resources/agent/agent_.py b/src/oz_agent_sdk/resources/agent/agent_.py
index ab7835e..447bc5d 100644
--- a/src/oz_agent_sdk/resources/agent/agent_.py
+++ b/src/oz_agent_sdk/resources/agent/agent_.py
@@ -50,7 +50,11 @@ def create(
self,
*,
name: str,
+ base_model: Optional[str] | Omit = omit,
description: Optional[str] | Omit = omit,
+ inference_providers: agent_create_params.InferenceProviders | Omit = omit,
+ memory_stores: Iterable[agent_create_params.MemoryStore] | Omit = omit,
+ prompt: Optional[str] | Omit = omit,
secrets: Iterable[agent_create_params.Secret] | Omit = omit,
skills: SequenceNotStr[str] | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
@@ -68,8 +72,18 @@ def create(
Args:
name: A name for the agent
+ base_model: Optional base model for runs executed by this agent.
+
description: Optional description of the agent
+ inference_providers: Inference provider settings used for LLM calls.
+
+ memory_stores: Optional list of memory stores to attach to the agent. Each store must be
+ team-owned by the same team as the agent. Duplicate UIDs within a single request
+ are rejected.
+
+ prompt: Optional base prompt for this agent
+
secrets: Optional list of secrets associated with the agent. Duplicate names within a
single request are rejected. Each entry is unioned into the run-time secret
scope when the agent executes.
@@ -94,7 +108,11 @@ def create(
body=maybe_transform(
{
"name": name,
+ "base_model": base_model,
"description": description,
+ "inference_providers": inference_providers,
+ "memory_stores": memory_stores,
+ "prompt": prompt,
"secrets": secrets,
"skills": skills,
},
@@ -110,8 +128,12 @@ def update(
self,
uid: str,
*,
+ base_model: Optional[str] | Omit = omit,
description: Optional[str] | Omit = omit,
+ inference_providers: Optional[agent_update_params.InferenceProviders] | Omit = omit,
+ memory_stores: Optional[Iterable[agent_update_params.MemoryStore]] | Omit = omit,
name: str | Omit = omit,
+ prompt: Optional[str] | Omit = omit,
secrets: Optional[Iterable[agent_update_params.Secret]] | Omit = omit,
skills: Optional[SequenceNotStr[str]] | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
@@ -124,13 +146,24 @@ def update(
"""Update an existing agent.
Args:
- description: Replacement description.
+ base_model: Replacement base model.
+
+ Omit or pass `null` to leave unchanged, or pass an empty
+ string to clear.
- Omit or pass `null` to leave unchanged, or use an empty
+ description: Replacement description. Omit or pass `null` to leave unchanged, or use an empty
value to clear.
+ inference_providers: Inference provider settings used for LLM calls.
+
+ memory_stores: Replacement list of memory stores. Omit to leave unchanged, pass an empty array
+ to clear, or pass a non-empty array to replace.
+
name: The new name for the agent
+ prompt: Replacement prompt. Omit or pass `null` to leave unchanged, or use an empty
+ value to clear.
+
secrets: Replacement list of secrets. Omit to leave unchanged, pass an empty array to
clear, or pass a non-empty array to replace. Duplicate names are rejected.
@@ -151,8 +184,12 @@ def update(
path_template("/agent/identities/{uid}", uid=uid),
body=maybe_transform(
{
+ "base_model": base_model,
"description": description,
+ "inference_providers": inference_providers,
+ "memory_stores": memory_stores,
"name": name,
+ "prompt": prompt,
"secrets": secrets,
"skills": skills,
},
@@ -222,6 +259,42 @@ def delete(
cast_to=NoneType,
)
+ def get(
+ self,
+ uid: str,
+ *,
+ # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
+ # The extra values given here take precedence over values defined on the client or passed to this method.
+ extra_headers: Headers | None = None,
+ extra_query: Query | None = None,
+ extra_body: Body | None = None,
+ timeout: float | httpx.Timeout | None | NotGiven = not_given,
+ ) -> AgentResponse:
+ """Retrieve a single agent by its unique identifier.
+
+ The response includes an
+ `available` flag indicating whether the agent is within the team's plan limit
+ and may be used for runs.
+
+ Args:
+ extra_headers: Send extra headers
+
+ extra_query: Add additional query parameters to the request
+
+ extra_body: Add additional JSON properties to the request
+
+ timeout: Override the client-level default timeout for this request, in seconds
+ """
+ if not uid:
+ raise ValueError(f"Expected a non-empty value for `uid` but received {uid!r}")
+ return self._get(
+ path_template("/agent/identities/{uid}", uid=uid),
+ options=make_request_options(
+ extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout
+ ),
+ cast_to=AgentResponse,
+ )
+
class AsyncAgentResource(AsyncAPIResource):
"""Operations for running and managing cloud agents"""
@@ -249,7 +322,11 @@ async def create(
self,
*,
name: str,
+ base_model: Optional[str] | Omit = omit,
description: Optional[str] | Omit = omit,
+ inference_providers: agent_create_params.InferenceProviders | Omit = omit,
+ memory_stores: Iterable[agent_create_params.MemoryStore] | Omit = omit,
+ prompt: Optional[str] | Omit = omit,
secrets: Iterable[agent_create_params.Secret] | Omit = omit,
skills: SequenceNotStr[str] | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
@@ -267,8 +344,18 @@ async def create(
Args:
name: A name for the agent
+ base_model: Optional base model for runs executed by this agent.
+
description: Optional description of the agent
+ inference_providers: Inference provider settings used for LLM calls.
+
+ memory_stores: Optional list of memory stores to attach to the agent. Each store must be
+ team-owned by the same team as the agent. Duplicate UIDs within a single request
+ are rejected.
+
+ prompt: Optional base prompt for this agent
+
secrets: Optional list of secrets associated with the agent. Duplicate names within a
single request are rejected. Each entry is unioned into the run-time secret
scope when the agent executes.
@@ -293,7 +380,11 @@ async def create(
body=await async_maybe_transform(
{
"name": name,
+ "base_model": base_model,
"description": description,
+ "inference_providers": inference_providers,
+ "memory_stores": memory_stores,
+ "prompt": prompt,
"secrets": secrets,
"skills": skills,
},
@@ -309,8 +400,12 @@ async def update(
self,
uid: str,
*,
+ base_model: Optional[str] | Omit = omit,
description: Optional[str] | Omit = omit,
+ inference_providers: Optional[agent_update_params.InferenceProviders] | Omit = omit,
+ memory_stores: Optional[Iterable[agent_update_params.MemoryStore]] | Omit = omit,
name: str | Omit = omit,
+ prompt: Optional[str] | Omit = omit,
secrets: Optional[Iterable[agent_update_params.Secret]] | Omit = omit,
skills: Optional[SequenceNotStr[str]] | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
@@ -323,13 +418,24 @@ async def update(
"""Update an existing agent.
Args:
- description: Replacement description.
+ base_model: Replacement base model.
+
+ Omit or pass `null` to leave unchanged, or pass an empty
+ string to clear.
- Omit or pass `null` to leave unchanged, or use an empty
+ description: Replacement description. Omit or pass `null` to leave unchanged, or use an empty
value to clear.
+ inference_providers: Inference provider settings used for LLM calls.
+
+ memory_stores: Replacement list of memory stores. Omit to leave unchanged, pass an empty array
+ to clear, or pass a non-empty array to replace.
+
name: The new name for the agent
+ prompt: Replacement prompt. Omit or pass `null` to leave unchanged, or use an empty
+ value to clear.
+
secrets: Replacement list of secrets. Omit to leave unchanged, pass an empty array to
clear, or pass a non-empty array to replace. Duplicate names are rejected.
@@ -350,8 +456,12 @@ async def update(
path_template("/agent/identities/{uid}", uid=uid),
body=await async_maybe_transform(
{
+ "base_model": base_model,
"description": description,
+ "inference_providers": inference_providers,
+ "memory_stores": memory_stores,
"name": name,
+ "prompt": prompt,
"secrets": secrets,
"skills": skills,
},
@@ -421,6 +531,42 @@ async def delete(
cast_to=NoneType,
)
+ async def get(
+ self,
+ uid: str,
+ *,
+ # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
+ # The extra values given here take precedence over values defined on the client or passed to this method.
+ extra_headers: Headers | None = None,
+ extra_query: Query | None = None,
+ extra_body: Body | None = None,
+ timeout: float | httpx.Timeout | None | NotGiven = not_given,
+ ) -> AgentResponse:
+ """Retrieve a single agent by its unique identifier.
+
+ The response includes an
+ `available` flag indicating whether the agent is within the team's plan limit
+ and may be used for runs.
+
+ Args:
+ extra_headers: Send extra headers
+
+ extra_query: Add additional query parameters to the request
+
+ extra_body: Add additional JSON properties to the request
+
+ timeout: Override the client-level default timeout for this request, in seconds
+ """
+ if not uid:
+ raise ValueError(f"Expected a non-empty value for `uid` but received {uid!r}")
+ return await self._get(
+ path_template("/agent/identities/{uid}", uid=uid),
+ options=make_request_options(
+ extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout
+ ),
+ cast_to=AgentResponse,
+ )
+
class AgentResourceWithRawResponse:
def __init__(self, agent: AgentResource) -> None:
@@ -438,6 +584,9 @@ def __init__(self, agent: AgentResource) -> None:
self.delete = to_raw_response_wrapper(
agent.delete,
)
+ self.get = to_raw_response_wrapper(
+ agent.get,
+ )
class AsyncAgentResourceWithRawResponse:
@@ -456,6 +605,9 @@ def __init__(self, agent: AsyncAgentResource) -> None:
self.delete = async_to_raw_response_wrapper(
agent.delete,
)
+ self.get = async_to_raw_response_wrapper(
+ agent.get,
+ )
class AgentResourceWithStreamingResponse:
@@ -474,6 +626,9 @@ def __init__(self, agent: AgentResource) -> None:
self.delete = to_streamed_response_wrapper(
agent.delete,
)
+ self.get = to_streamed_response_wrapper(
+ agent.get,
+ )
class AsyncAgentResourceWithStreamingResponse:
@@ -492,3 +647,6 @@ def __init__(self, agent: AsyncAgentResource) -> None:
self.delete = async_to_streamed_response_wrapper(
agent.delete,
)
+ self.get = async_to_streamed_response_wrapper(
+ agent.get,
+ )
diff --git a/src/oz_agent_sdk/types/agent/agent_create_params.py b/src/oz_agent_sdk/types/agent/agent_create_params.py
index 5f31e19..db351bc 100644
--- a/src/oz_agent_sdk/types/agent/agent_create_params.py
+++ b/src/oz_agent_sdk/types/agent/agent_create_params.py
@@ -3,20 +3,36 @@
from __future__ import annotations
from typing import Iterable, Optional
-from typing_extensions import Required, TypedDict
+from typing_extensions import Literal, Required, TypedDict
from ..._types import SequenceNotStr
-__all__ = ["AgentCreateParams", "Secret"]
+__all__ = ["AgentCreateParams", "InferenceProviders", "InferenceProvidersAws", "MemoryStore", "Secret"]
class AgentCreateParams(TypedDict, total=False):
name: Required[str]
"""A name for the agent"""
+ base_model: Optional[str]
+ """Optional base model for runs executed by this agent."""
+
description: Optional[str]
"""Optional description of the agent"""
+ inference_providers: InferenceProviders
+ """Inference provider settings used for LLM calls."""
+
+ memory_stores: Iterable[MemoryStore]
+ """
+ Optional list of memory stores to attach to the agent. Each store must be
+ team-owned by the same team as the agent. Duplicate UIDs within a single request
+ are rejected.
+ """
+
+ prompt: Optional[str]
+ """Optional base prompt for this agent"""
+
secrets: Iterable[Secret]
"""
Optional list of secrets associated with the agent. Duplicate names within a
@@ -34,6 +50,39 @@ class AgentCreateParams(TypedDict, total=False):
"""
+class InferenceProvidersAws(TypedDict, total=False):
+ """
+ Configures AWS Bedrock as the LLM inference provider for this
+ agent or run.
+ """
+
+ disabled: bool
+ """If true, opt out of Bedrock at this layer."""
+
+ role_arn: str
+ """IAM role ARN to assume when calling Bedrock."""
+
+
+class InferenceProviders(TypedDict, total=False):
+ """Inference provider settings used for LLM calls."""
+
+ aws: InferenceProvidersAws
+ """Configures AWS Bedrock as the LLM inference provider for this agent or run."""
+
+
+class MemoryStore(TypedDict, total=False):
+ """Reference to a memory store to attach to an agent."""
+
+ access: Required[Literal["read_write", "read_only"]]
+ """Access level for the store."""
+
+ instructions: Required[str]
+ """Instructions for how the agent should use this memory store. Must not be empty."""
+
+ uid: Required[str]
+ """UID of the memory store."""
+
+
class Secret(TypedDict, total=False):
"""Reference to a managed secret by name."""
diff --git a/src/oz_agent_sdk/types/agent/agent_response.py b/src/oz_agent_sdk/types/agent/agent_response.py
index 8b46c85..fac7f4a 100644
--- a/src/oz_agent_sdk/types/agent/agent_response.py
+++ b/src/oz_agent_sdk/types/agent/agent_response.py
@@ -2,10 +2,24 @@
from typing import List, Optional
from datetime import datetime
+from typing_extensions import Literal
from ..._models import BaseModel
-__all__ = ["AgentResponse", "Secret"]
+__all__ = ["AgentResponse", "MemoryStore", "Secret", "InferenceProviders", "InferenceProvidersAws"]
+
+
+class MemoryStore(BaseModel):
+ """Reference to a memory store to attach to an agent."""
+
+ access: Literal["read_write", "read_only"]
+ """Access level for the store."""
+
+ instructions: str
+ """Instructions for how the agent should use this memory store. Must not be empty."""
+
+ uid: str
+ """UID of the memory store."""
class Secret(BaseModel):
@@ -15,6 +29,26 @@ class Secret(BaseModel):
"""Name of the managed secret."""
+class InferenceProvidersAws(BaseModel):
+ """
+ Configures AWS Bedrock as the LLM inference provider for this
+ agent or run.
+ """
+
+ disabled: Optional[bool] = None
+ """If true, opt out of Bedrock at this layer."""
+
+ role_arn: Optional[str] = None
+ """IAM role ARN to assume when calling Bedrock."""
+
+
+class InferenceProviders(BaseModel):
+ """Inference provider settings used for LLM calls."""
+
+ aws: Optional[InferenceProvidersAws] = None
+ """Configures AWS Bedrock as the LLM inference provider for this agent or run."""
+
+
class AgentResponse(BaseModel):
available: bool
"""Whether this agent is within the team's plan limit and can be used for runs"""
@@ -22,6 +56,12 @@ class AgentResponse(BaseModel):
created_at: datetime
"""When the agent was created (RFC3339)"""
+ memory_stores: List[MemoryStore]
+ """
+ Memory stores attached to this agent. Always present; empty when no stores are
+ attached.
+ """
+
name: str
"""Name of the agent"""
@@ -37,5 +77,21 @@ class AgentResponse(BaseModel):
uid: str
"""Unique identifier for the agent"""
+ base_model: Optional[str] = None
+ """Base model for runs executed by this agent.
+
+ The precedence order for model resolution is:
+
+ 1. The model specified on the run itself
+ 2. The agent's base model
+ 3. The team's default model
+ """
+
description: Optional[str] = None
"""Optional description of the agent"""
+
+ inference_providers: Optional[InferenceProviders] = None
+ """Inference provider settings used for LLM calls."""
+
+ prompt: Optional[str] = None
+ """Optional base prompt for this agent"""
diff --git a/src/oz_agent_sdk/types/agent/agent_update_params.py b/src/oz_agent_sdk/types/agent/agent_update_params.py
index d3cc68a..29278da 100644
--- a/src/oz_agent_sdk/types/agent/agent_update_params.py
+++ b/src/oz_agent_sdk/types/agent/agent_update_params.py
@@ -3,23 +3,45 @@
from __future__ import annotations
from typing import Iterable, Optional
-from typing_extensions import Required, TypedDict
+from typing_extensions import Literal, Required, TypedDict
from ..._types import SequenceNotStr
-__all__ = ["AgentUpdateParams", "Secret"]
+__all__ = ["AgentUpdateParams", "InferenceProviders", "InferenceProvidersAws", "MemoryStore", "Secret"]
class AgentUpdateParams(TypedDict, total=False):
+ base_model: Optional[str]
+ """Replacement base model.
+
+ Omit or pass `null` to leave unchanged, or pass an empty string to clear.
+ """
+
description: Optional[str]
"""Replacement description.
Omit or pass `null` to leave unchanged, or use an empty value to clear.
"""
+ inference_providers: Optional[InferenceProviders]
+ """Inference provider settings used for LLM calls."""
+
+ memory_stores: Optional[Iterable[MemoryStore]]
+ """Replacement list of memory stores.
+
+ Omit to leave unchanged, pass an empty array to clear, or pass a non-empty array
+ to replace.
+ """
+
name: str
"""The new name for the agent"""
+ prompt: Optional[str]
+ """Replacement prompt.
+
+ Omit or pass `null` to leave unchanged, or use an empty value to clear.
+ """
+
secrets: Optional[Iterable[Secret]]
"""Replacement list of secrets.
@@ -35,6 +57,39 @@ class AgentUpdateParams(TypedDict, total=False):
"""
+class InferenceProvidersAws(TypedDict, total=False):
+ """
+ Configures AWS Bedrock as the LLM inference provider for this
+ agent or run.
+ """
+
+ disabled: bool
+ """If true, opt out of Bedrock at this layer."""
+
+ role_arn: str
+ """IAM role ARN to assume when calling Bedrock."""
+
+
+class InferenceProviders(TypedDict, total=False):
+ """Inference provider settings used for LLM calls."""
+
+ aws: InferenceProvidersAws
+ """Configures AWS Bedrock as the LLM inference provider for this agent or run."""
+
+
+class MemoryStore(TypedDict, total=False):
+ """Reference to a memory store to attach to an agent."""
+
+ access: Required[Literal["read_write", "read_only"]]
+ """Access level for the store."""
+
+ instructions: Required[str]
+ """Instructions for how the agent should use this memory store. Must not be empty."""
+
+ uid: Required[str]
+ """UID of the memory store."""
+
+
class Secret(TypedDict, total=False):
"""Reference to a managed secret by name."""
diff --git a/src/oz_agent_sdk/types/agent/run_item.py b/src/oz_agent_sdk/types/agent/run_item.py
index 5b802d8..4293ce2 100644
--- a/src/oz_agent_sdk/types/agent/run_item.py
+++ b/src/oz_agent_sdk/types/agent/run_item.py
@@ -44,6 +44,9 @@ class RequestUsage(BaseModel):
inference_cost: Optional[float] = None
"""Cost of LLM inference for the run"""
+ platform_cost: Optional[float] = None
+ """Cost of platform usage for the run"""
+
class Schedule(BaseModel):
"""
diff --git a/src/oz_agent_sdk/types/ambient_agent_config.py b/src/oz_agent_sdk/types/ambient_agent_config.py
index 148931d..2d5b76a 100644
--- a/src/oz_agent_sdk/types/ambient_agent_config.py
+++ b/src/oz_agent_sdk/types/ambient_agent_config.py
@@ -1,6 +1,6 @@
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
-from typing import Dict, Optional
+from typing import Dict, List, Optional
from typing_extensions import Literal
from pydantic import Field as FieldInfo
@@ -8,7 +8,15 @@
from .._models import BaseModel
from .mcp_server_config import McpServerConfig
-__all__ = ["AmbientAgentConfig", "Harness", "HarnessAuthSecrets", "SessionSharing"]
+__all__ = [
+ "AmbientAgentConfig",
+ "Harness",
+ "HarnessAuthSecrets",
+ "InferenceProviders",
+ "InferenceProvidersAws",
+ "MemoryStore",
+ "SessionSharing",
+]
class Harness(BaseModel):
@@ -41,6 +49,39 @@ class HarnessAuthSecrets(BaseModel):
"""
+class InferenceProvidersAws(BaseModel):
+ """
+ Configures AWS Bedrock as the LLM inference provider for this
+ agent or run.
+ """
+
+ disabled: Optional[bool] = None
+ """If true, opt out of Bedrock at this layer."""
+
+ role_arn: Optional[str] = None
+ """IAM role ARN to assume when calling Bedrock."""
+
+
+class InferenceProviders(BaseModel):
+ """Inference provider settings used for LLM calls."""
+
+ aws: Optional[InferenceProvidersAws] = None
+ """Configures AWS Bedrock as the LLM inference provider for this agent or run."""
+
+
+class MemoryStore(BaseModel):
+ """Reference to a memory store to attach to an agent."""
+
+ access: Literal["read_write", "read_only"]
+ """Access level for the store."""
+
+ instructions: str
+ """Instructions for how the agent should use this memory store. Must not be empty."""
+
+ uid: str
+ """UID of the memory store."""
+
+
class SessionSharing(BaseModel):
"""
Configures sharing behavior for the run's shared session.
@@ -97,9 +138,15 @@ class AmbientAgentConfig(BaseModel):
floor(max_instance_runtime_seconds / 60) for your billing tier).
"""
+ inference_providers: Optional[InferenceProviders] = None
+ """Inference provider settings used for LLM calls."""
+
mcp_servers: Optional[Dict[str, McpServerConfig]] = None
"""Map of MCP server configurations by name"""
+ memory_stores: Optional[List[MemoryStore]] = None
+ """Memory stores to attach to this run."""
+
api_model_id: Optional[str] = FieldInfo(alias="model_id", default=None)
"""LLM model to use (uses team default if not specified)"""
diff --git a/src/oz_agent_sdk/types/ambient_agent_config_param.py b/src/oz_agent_sdk/types/ambient_agent_config_param.py
index f4c6ee4..4e662df 100644
--- a/src/oz_agent_sdk/types/ambient_agent_config_param.py
+++ b/src/oz_agent_sdk/types/ambient_agent_config_param.py
@@ -2,12 +2,20 @@
from __future__ import annotations
-from typing import Dict
-from typing_extensions import Literal, TypedDict
+from typing import Dict, Iterable
+from typing_extensions import Literal, Required, TypedDict
from .mcp_server_config_param import McpServerConfigParam
-__all__ = ["AmbientAgentConfigParam", "Harness", "HarnessAuthSecrets", "SessionSharing"]
+__all__ = [
+ "AmbientAgentConfigParam",
+ "Harness",
+ "HarnessAuthSecrets",
+ "InferenceProviders",
+ "InferenceProvidersAws",
+ "MemoryStore",
+ "SessionSharing",
+]
class Harness(TypedDict, total=False):
@@ -40,6 +48,39 @@ class HarnessAuthSecrets(TypedDict, total=False):
"""
+class InferenceProvidersAws(TypedDict, total=False):
+ """
+ Configures AWS Bedrock as the LLM inference provider for this
+ agent or run.
+ """
+
+ disabled: bool
+ """If true, opt out of Bedrock at this layer."""
+
+ role_arn: str
+ """IAM role ARN to assume when calling Bedrock."""
+
+
+class InferenceProviders(TypedDict, total=False):
+ """Inference provider settings used for LLM calls."""
+
+ aws: InferenceProvidersAws
+ """Configures AWS Bedrock as the LLM inference provider for this agent or run."""
+
+
+class MemoryStore(TypedDict, total=False):
+ """Reference to a memory store to attach to an agent."""
+
+ access: Required[Literal["read_write", "read_only"]]
+ """Access level for the store."""
+
+ instructions: Required[str]
+ """Instructions for how the agent should use this memory store. Must not be empty."""
+
+ uid: Required[str]
+ """UID of the memory store."""
+
+
class SessionSharing(TypedDict, total=False):
"""
Configures sharing behavior for the run's shared session.
@@ -96,9 +137,15 @@ class AmbientAgentConfigParam(TypedDict, total=False):
floor(max_instance_runtime_seconds / 60) for your billing tier).
"""
+ inference_providers: InferenceProviders
+ """Inference provider settings used for LLM calls."""
+
mcp_servers: Dict[str, McpServerConfigParam]
"""Map of MCP server configurations by name"""
+ memory_stores: Iterable[MemoryStore]
+ """Memory stores to attach to this run."""
+
model_id: str
"""LLM model to use (uses team default if not specified)"""
diff --git a/tests/api_resources/agent/test_agent_.py b/tests/api_resources/agent/test_agent_.py
index a23b05e..0b91ac0 100644
--- a/tests/api_resources/agent/test_agent_.py
+++ b/tests/api_resources/agent/test_agent_.py
@@ -33,7 +33,22 @@ def test_method_create(self, client: OzAPI) -> None:
def test_method_create_with_all_params(self, client: OzAPI) -> None:
agent = client.agent.agent.create(
name="name",
+ base_model="base_model",
description="description",
+ inference_providers={
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
+ memory_stores=[
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
+ prompt="prompt",
secrets=[{"name": "name"}],
skills=["string"],
)
@@ -78,8 +93,23 @@ def test_method_update(self, client: OzAPI) -> None:
def test_method_update_with_all_params(self, client: OzAPI) -> None:
agent = client.agent.agent.update(
uid="uid",
+ base_model="base_model",
description="description",
+ inference_providers={
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
+ memory_stores=[
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
name="name",
+ prompt="prompt",
secrets=[{"name": "name"}],
skills=["string"],
)
@@ -189,6 +219,48 @@ def test_path_params_delete(self, client: OzAPI) -> None:
"",
)
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ def test_method_get(self, client: OzAPI) -> None:
+ agent = client.agent.agent.get(
+ "uid",
+ )
+ assert_matches_type(AgentResponse, agent, path=["response"])
+
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ def test_raw_response_get(self, client: OzAPI) -> None:
+ response = client.agent.agent.with_raw_response.get(
+ "uid",
+ )
+
+ assert response.is_closed is True
+ assert response.http_request.headers.get("X-Stainless-Lang") == "python"
+ agent = response.parse()
+ assert_matches_type(AgentResponse, agent, path=["response"])
+
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ def test_streaming_response_get(self, client: OzAPI) -> None:
+ with client.agent.agent.with_streaming_response.get(
+ "uid",
+ ) as response:
+ assert not response.is_closed
+ assert response.http_request.headers.get("X-Stainless-Lang") == "python"
+
+ agent = response.parse()
+ assert_matches_type(AgentResponse, agent, path=["response"])
+
+ assert cast(Any, response.is_closed) is True
+
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ def test_path_params_get(self, client: OzAPI) -> None:
+ with pytest.raises(ValueError, match=r"Expected a non-empty value for `uid` but received ''"):
+ client.agent.agent.with_raw_response.get(
+ "",
+ )
+
class TestAsyncAgent:
parametrize = pytest.mark.parametrize(
@@ -208,7 +280,22 @@ async def test_method_create(self, async_client: AsyncOzAPI) -> None:
async def test_method_create_with_all_params(self, async_client: AsyncOzAPI) -> None:
agent = await async_client.agent.agent.create(
name="name",
+ base_model="base_model",
description="description",
+ inference_providers={
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
+ memory_stores=[
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
+ prompt="prompt",
secrets=[{"name": "name"}],
skills=["string"],
)
@@ -253,8 +340,23 @@ async def test_method_update(self, async_client: AsyncOzAPI) -> None:
async def test_method_update_with_all_params(self, async_client: AsyncOzAPI) -> None:
agent = await async_client.agent.agent.update(
uid="uid",
+ base_model="base_model",
description="description",
+ inference_providers={
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
+ memory_stores=[
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
name="name",
+ prompt="prompt",
secrets=[{"name": "name"}],
skills=["string"],
)
@@ -363,3 +465,45 @@ async def test_path_params_delete(self, async_client: AsyncOzAPI) -> None:
await async_client.agent.agent.with_raw_response.delete(
"",
)
+
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ async def test_method_get(self, async_client: AsyncOzAPI) -> None:
+ agent = await async_client.agent.agent.get(
+ "uid",
+ )
+ assert_matches_type(AgentResponse, agent, path=["response"])
+
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ async def test_raw_response_get(self, async_client: AsyncOzAPI) -> None:
+ response = await async_client.agent.agent.with_raw_response.get(
+ "uid",
+ )
+
+ assert response.is_closed is True
+ assert response.http_request.headers.get("X-Stainless-Lang") == "python"
+ agent = await response.parse()
+ assert_matches_type(AgentResponse, agent, path=["response"])
+
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ async def test_streaming_response_get(self, async_client: AsyncOzAPI) -> None:
+ async with async_client.agent.agent.with_streaming_response.get(
+ "uid",
+ ) as response:
+ assert not response.is_closed
+ assert response.http_request.headers.get("X-Stainless-Lang") == "python"
+
+ agent = await response.parse()
+ assert_matches_type(AgentResponse, agent, path=["response"])
+
+ assert cast(Any, response.is_closed) is True
+
+ @pytest.mark.skip(reason="Mock server tests are disabled")
+ @parametrize
+ async def test_path_params_get(self, async_client: AsyncOzAPI) -> None:
+ with pytest.raises(ValueError, match=r"Expected a non-empty value for `uid` but received ''"):
+ await async_client.agent.agent.with_raw_response.get(
+ "",
+ )
diff --git a/tests/api_resources/agent/test_schedules.py b/tests/api_resources/agent/test_schedules.py
index 0bf7d09..38f62c2 100644
--- a/tests/api_resources/agent/test_schedules.py
+++ b/tests/api_resources/agent/test_schedules.py
@@ -43,6 +43,12 @@ def test_method_create_with_all_params(self, client: OzAPI) -> None:
"harness": {"type": "oz"},
"harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"},
"idle_timeout_minutes": 1,
+ "inference_providers": {
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
"mcp_servers": {
"foo": {
"args": ["string"],
@@ -53,6 +59,13 @@ def test_method_create_with_all_params(self, client: OzAPI) -> None:
"warp_id": "warp_id",
}
},
+ "memory_stores": [
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
"model_id": "model_id",
"name": "name",
"session_sharing": {"public_access": "VIEWER"},
@@ -163,6 +176,12 @@ def test_method_update_with_all_params(self, client: OzAPI) -> None:
"harness": {"type": "oz"},
"harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"},
"idle_timeout_minutes": 1,
+ "inference_providers": {
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
"mcp_servers": {
"foo": {
"args": ["string"],
@@ -173,6 +192,13 @@ def test_method_update_with_all_params(self, client: OzAPI) -> None:
"warp_id": "warp_id",
}
},
+ "memory_stores": [
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
"model_id": "model_id",
"name": "name",
"session_sharing": {"public_access": "VIEWER"},
@@ -410,6 +436,12 @@ async def test_method_create_with_all_params(self, async_client: AsyncOzAPI) ->
"harness": {"type": "oz"},
"harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"},
"idle_timeout_minutes": 1,
+ "inference_providers": {
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
"mcp_servers": {
"foo": {
"args": ["string"],
@@ -420,6 +452,13 @@ async def test_method_create_with_all_params(self, async_client: AsyncOzAPI) ->
"warp_id": "warp_id",
}
},
+ "memory_stores": [
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
"model_id": "model_id",
"name": "name",
"session_sharing": {"public_access": "VIEWER"},
@@ -530,6 +569,12 @@ async def test_method_update_with_all_params(self, async_client: AsyncOzAPI) ->
"harness": {"type": "oz"},
"harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"},
"idle_timeout_minutes": 1,
+ "inference_providers": {
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
"mcp_servers": {
"foo": {
"args": ["string"],
@@ -540,6 +585,13 @@ async def test_method_update_with_all_params(self, async_client: AsyncOzAPI) ->
"warp_id": "warp_id",
}
},
+ "memory_stores": [
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
"model_id": "model_id",
"name": "name",
"session_sharing": {"public_access": "VIEWER"},
diff --git a/tests/api_resources/test_agent.py b/tests/api_resources/test_agent.py
index a10a54e..980b5b2 100644
--- a/tests/api_resources/test_agent.py
+++ b/tests/api_resources/test_agent.py
@@ -164,6 +164,12 @@ def test_method_run_with_all_params(self, client: OzAPI) -> None:
"harness": {"type": "oz"},
"harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"},
"idle_timeout_minutes": 1,
+ "inference_providers": {
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
"mcp_servers": {
"foo": {
"args": ["string"],
@@ -174,6 +180,13 @@ def test_method_run_with_all_params(self, client: OzAPI) -> None:
"warp_id": "warp_id",
}
},
+ "memory_stores": [
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
"model_id": "model_id",
"name": "name",
"session_sharing": {"public_access": "VIEWER"},
@@ -361,6 +374,12 @@ async def test_method_run_with_all_params(self, async_client: AsyncOzAPI) -> Non
"harness": {"type": "oz"},
"harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"},
"idle_timeout_minutes": 1,
+ "inference_providers": {
+ "aws": {
+ "disabled": True,
+ "role_arn": "role_arn",
+ }
+ },
"mcp_servers": {
"foo": {
"args": ["string"],
@@ -371,6 +390,13 @@ async def test_method_run_with_all_params(self, async_client: AsyncOzAPI) -> Non
"warp_id": "warp_id",
}
},
+ "memory_stores": [
+ {
+ "access": "read_write",
+ "instructions": "instructions",
+ "uid": "uid",
+ }
+ ],
"model_id": "model_id",
"name": "name",
"session_sharing": {"public_access": "VIEWER"},
diff --git a/tests/test_models.py b/tests/test_models.py
index c0b225d..ae6ec09 100644
--- a/tests/test_models.py
+++ b/tests/test_models.py
@@ -1,7 +1,8 @@
import json
-from typing import TYPE_CHECKING, Any, Dict, List, Union, Optional, cast
+from typing import TYPE_CHECKING, Any, Dict, List, Union, Iterable, Optional, cast
from datetime import datetime, timezone
-from typing_extensions import Literal, Annotated, TypeAliasType
+from collections import deque
+from typing_extensions import Literal, Annotated, TypedDict, TypeAliasType
import pytest
import pydantic
@@ -9,7 +10,7 @@
from oz_agent_sdk._utils import PropertyInfo
from oz_agent_sdk._compat import PYDANTIC_V1, parse_obj, model_dump, model_json
-from oz_agent_sdk._models import DISCRIMINATOR_CACHE, BaseModel, construct_type
+from oz_agent_sdk._models import DISCRIMINATOR_CACHE, BaseModel, EagerIterable, construct_type
class BasicModel(BaseModel):
@@ -961,3 +962,56 @@ def __getattr__(self, attr: str) -> Item: ...
assert model.a.prop == 1
assert isinstance(model.a, Item)
assert model.other == "foo"
+
+
+# NOTE: Workaround for Pydantic Iterable behavior.
+# Iterable fields are replaced with a ValidatorIterator and may be consumed
+# during serialization, which can cause subsequent dumps to return empty data.
+# See: https://github.com/pydantic/pydantic/issues/9541
+@pytest.mark.parametrize(
+ "data, expected_validated",
+ [
+ ([1, 2, 3], [1, 2, 3]),
+ ((1, 2, 3), (1, 2, 3)),
+ (set([1, 2, 3]), set([1, 2, 3])),
+ (iter([1, 2, 3]), [1, 2, 3]),
+ ([], []),
+ ((x for x in [1, 2, 3]), [1, 2, 3]),
+ (map(lambda x: x, [1, 2, 3]), [1, 2, 3]),
+ (frozenset([1, 2, 3]), frozenset([1, 2, 3])),
+ (deque([1, 2, 3]), deque([1, 2, 3])),
+ ],
+ ids=["list", "tuple", "set", "iterator", "empty", "generator", "map", "frozenset", "deque"],
+)
+@pytest.mark.skipif(PYDANTIC_V1, reason="this is only supported in pydantic v2")
+def test_iterable_construction(data: Iterable[int], expected_validated: Iterable[int]) -> None:
+ class TypeWithIterable(TypedDict):
+ items: EagerIterable[int]
+
+ class Model(BaseModel):
+ data: TypeWithIterable
+
+ m = Model.model_validate({"data": {"items": data}})
+ assert m.data["items"] == expected_validated
+
+ # Verify repeated dumps don't lose data (the original bug)
+ assert m.model_dump()["data"]["items"] == list(expected_validated)
+ assert m.model_dump()["data"]["items"] == list(expected_validated)
+
+
+@pytest.mark.skipif(PYDANTIC_V1, reason="this is only supported in pydantic v2")
+def test_iterable_construction_str_falls_back_to_list() -> None:
+ # str is iterable (over chars), but str(list_of_chars) produces the list's repr
+ # rather than reconstructing a string from items. We special-case str to fall
+ # back to list instead of attempting reconstruction.
+ class TypeWithIterable(TypedDict):
+ items: EagerIterable[str]
+
+ class Model(BaseModel):
+ data: TypeWithIterable
+
+ m = Model.model_validate({"data": {"items": "hello"}})
+
+ # falls back to list of chars rather than calling str(["h", "e", "l", "l", "o"])
+ assert m.data["items"] == ["h", "e", "l", "l", "o"]
+ assert m.model_dump()["data"]["items"] == ["h", "e", "l", "l", "o"]