Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .release-please-manifest.json
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
{
".": "0.1.0-alpha.30"
".": "0.1.0-alpha.31"
}
2 changes: 1 addition & 1 deletion .stats.yml
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
configured_endpoints: 55
openapi_spec_hash: c894ce3fb9db92c69816f06896e30067
openapi_spec_hash: d555131ff52c78852f82acda3348224b
config_hash: 48c3812186c899cdef23cc8de76bd2aa
15 changes: 15 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,20 @@
# Changelog

## 0.1.0-alpha.31 (2025-10-20)

Full Changelog: [v0.1.0-alpha.30...v0.1.0-alpha.31](https://github.com/cleanlab/codex-python/compare/v0.1.0-alpha.30...v0.1.0-alpha.31)

### Features

* **api:** api update ([766e331](https://github.com/cleanlab/codex-python/commit/766e3314ad8a72522fa29804b3edc328bab3b4af))
* **api:** api update ([377dc6b](https://github.com/cleanlab/codex-python/commit/377dc6b4dad3ba898b9fda4fe305d4c25b1ee781))
* **api:** api update ([321392c](https://github.com/cleanlab/codex-python/commit/321392cfd9ac98fe1ad0b06d21f26441e70c90fc))


### Chores

* bump `httpx-aiohttp` version to 0.1.9 ([666f855](https://github.com/cleanlab/codex-python/commit/666f855d5532589f5ffbe0b53a5de65355ec33aa))

## 0.1.0-alpha.30 (2025-10-14)

Full Changelog: [v0.1.0-alpha.29...v0.1.0-alpha.30](https://github.com/cleanlab/codex-python/compare/v0.1.0-alpha.29...v0.1.0-alpha.30)
Expand Down
4 changes: 2 additions & 2 deletions pyproject.toml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
[project]
name = "codex-sdk"
version = "0.1.0-alpha.30"
version = "0.1.0-alpha.31"
description = "Internal SDK used within cleanlab-codex package. Refer to https://pypi.org/project/cleanlab-codex/ instead."
dynamic = ["readme"]
license = "MIT"
Expand Down Expand Up @@ -39,7 +39,7 @@ Homepage = "https://github.com/cleanlab/codex-python"
Repository = "https://github.com/cleanlab/codex-python"

[project.optional-dependencies]
aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.8"]
aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.9"]

[tool.rye]
managed = true
Expand Down
2 changes: 1 addition & 1 deletion requirements-dev.lock
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ httpx==0.28.1
# via codex-sdk
# via httpx-aiohttp
# via respx
httpx-aiohttp==0.1.8
httpx-aiohttp==0.1.9
# via codex-sdk
idna==3.4
# via anyio
Expand Down
2 changes: 1 addition & 1 deletion requirements.lock
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@ httpcore==1.0.9
httpx==0.28.1
# via codex-sdk
# via httpx-aiohttp
httpx-aiohttp==0.1.8
httpx-aiohttp==0.1.9
# via codex-sdk
idna==3.4
# via anyio
Expand Down
2 changes: 1 addition & 1 deletion src/codex/_version.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.

__title__ = "codex"
__version__ = "0.1.0-alpha.30" # x-release-please-version
__version__ = "0.1.0-alpha.31" # x-release-please-version
17 changes: 2 additions & 15 deletions src/codex/types/project_detect_response.py
Original file line number Diff line number Diff line change
@@ -1,20 +1,10 @@
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.

from typing import Dict, List, Optional
from typing import Dict, Optional

from .._models import BaseModel

__all__ = ["ProjectDetectResponse", "DeterministicGuardrailsResults", "EvalScores"]


class DeterministicGuardrailsResults(BaseModel):
guardrail_name: str

should_guardrail: bool

fallback_message: Optional[str] = None

matches: Optional[List[str]] = None
__all__ = ["ProjectDetectResponse", "EvalScores"]


class EvalScores(BaseModel):
Expand All @@ -32,9 +22,6 @@ class EvalScores(BaseModel):


class ProjectDetectResponse(BaseModel):
deterministic_guardrails_results: Optional[Dict[str, DeterministicGuardrailsResults]] = None
"""Results from deterministic guardrails applied to the response."""

escalated_to_sme: bool
"""
True if the question should be escalated to Codex for an SME to review, False
Expand Down
30 changes: 30 additions & 0 deletions src/codex/types/project_list_response.py
Original file line number Diff line number Diff line change
Expand Up @@ -101,6 +101,12 @@ class ProjectConfigEvalConfigCustomEvals(BaseModel):


class ProjectConfigEvalConfigDefaultEvalsContextSufficiency(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -142,6 +148,12 @@ class ProjectConfigEvalConfigDefaultEvalsContextSufficiency(BaseModel):


class ProjectConfigEvalConfigDefaultEvalsQueryEase(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -183,6 +195,12 @@ class ProjectConfigEvalConfigDefaultEvalsQueryEase(BaseModel):


class ProjectConfigEvalConfigDefaultEvalsResponseGroundedness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -224,6 +242,12 @@ class ProjectConfigEvalConfigDefaultEvalsResponseGroundedness(BaseModel):


class ProjectConfigEvalConfigDefaultEvalsResponseHelpfulness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -265,6 +289,12 @@ class ProjectConfigEvalConfigDefaultEvalsResponseHelpfulness(BaseModel):


class ProjectConfigEvalConfigDefaultEvalsTrustworthiness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down
30 changes: 30 additions & 0 deletions src/codex/types/project_retrieve_response.py
Original file line number Diff line number Diff line change
Expand Up @@ -99,6 +99,12 @@ class ConfigEvalConfigCustomEvals(BaseModel):


class ConfigEvalConfigDefaultEvalsContextSufficiency(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -140,6 +146,12 @@ class ConfigEvalConfigDefaultEvalsContextSufficiency(BaseModel):


class ConfigEvalConfigDefaultEvalsQueryEase(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -181,6 +193,12 @@ class ConfigEvalConfigDefaultEvalsQueryEase(BaseModel):


class ConfigEvalConfigDefaultEvalsResponseGroundedness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -222,6 +240,12 @@ class ConfigEvalConfigDefaultEvalsResponseGroundedness(BaseModel):


class ConfigEvalConfigDefaultEvalsResponseHelpfulness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -263,6 +287,12 @@ class ConfigEvalConfigDefaultEvalsResponseHelpfulness(BaseModel):


class ConfigEvalConfigDefaultEvalsTrustworthiness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down
30 changes: 30 additions & 0 deletions src/codex/types/project_return_schema.py
Original file line number Diff line number Diff line change
Expand Up @@ -99,6 +99,12 @@ class ConfigEvalConfigCustomEvals(BaseModel):


class ConfigEvalConfigDefaultEvalsContextSufficiency(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -140,6 +146,12 @@ class ConfigEvalConfigDefaultEvalsContextSufficiency(BaseModel):


class ConfigEvalConfigDefaultEvalsQueryEase(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -181,6 +193,12 @@ class ConfigEvalConfigDefaultEvalsQueryEase(BaseModel):


class ConfigEvalConfigDefaultEvalsResponseGroundedness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -222,6 +240,12 @@ class ConfigEvalConfigDefaultEvalsResponseGroundedness(BaseModel):


class ConfigEvalConfigDefaultEvalsResponseHelpfulness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down Expand Up @@ -263,6 +287,12 @@ class ConfigEvalConfigDefaultEvalsResponseHelpfulness(BaseModel):


class ConfigEvalConfigDefaultEvalsTrustworthiness(BaseModel):
display_name: str
"""Human-friendly name for display.

For default evals, use standardized labels from DEFAULT_EVAL_ISSUE_TYPE_LABELS.
"""

eval_key: str
"""
Unique key for eval metric - currently maps to the TrustworthyRAG name property
Expand Down
3 changes: 3 additions & 0 deletions src/codex/types/projects/query_log_list_by_group_response.py
Original file line number Diff line number Diff line change
Expand Up @@ -472,6 +472,9 @@ class QueryLogsByGroupQueryLog(BaseModel):
primary_eval_issue_score: Optional[float] = None
"""Score of the primary eval issue"""

served_remediation_id: Optional[str] = None
"""ID of the remediation that was served if cache hit, otherwise None."""

tools: Optional[List[QueryLogsByGroupQueryLogTool]] = None
"""Tools to use for the LLM call.

Expand Down
3 changes: 3 additions & 0 deletions src/codex/types/projects/query_log_list_groups_response.py
Original file line number Diff line number Diff line change
Expand Up @@ -467,6 +467,9 @@ class QueryLogListGroupsResponse(BaseModel):
primary_eval_issue_score: Optional[float] = None
"""Score of the primary eval issue"""

served_remediation_id: Optional[str] = None
"""ID of the remediation that was served if cache hit, otherwise None."""

tools: Optional[List[Tool]] = None
"""Tools to use for the LLM call.

Expand Down
3 changes: 3 additions & 0 deletions src/codex/types/projects/query_log_list_response.py
Original file line number Diff line number Diff line change
Expand Up @@ -452,6 +452,9 @@ class QueryLogListResponse(BaseModel):
primary_eval_issue_score: Optional[float] = None
"""Score of the primary eval issue"""

served_remediation_id: Optional[str] = None
"""ID of the remediation that was served if cache hit, otherwise None."""

tools: Optional[List[Tool]] = None
"""Tools to use for the LLM call.

Expand Down
3 changes: 3 additions & 0 deletions src/codex/types/projects/query_log_retrieve_response.py
Original file line number Diff line number Diff line change
Expand Up @@ -459,6 +459,9 @@ class QueryLogRetrieveResponse(BaseModel):
primary_eval_issue_score: Optional[float] = None
"""Score of the primary eval issue"""

served_remediation_id: Optional[str] = None
"""ID of the remediation that was served if cache hit, otherwise None."""

tools: Optional[List[Tool]] = None
"""Tools to use for the LLM call.

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -459,6 +459,9 @@ class QueryLog(BaseModel):
primary_eval_issue_score: Optional[float] = None
"""Score of the primary eval issue"""

served_remediation_id: Optional[str] = None
"""ID of the remediation that was served if cache hit, otherwise None."""

tools: Optional[List[QueryLogTool]] = None
"""Tools to use for the LLM call.

Expand Down