From 93a15f90ac7512bce7c78109d9c350af61f20c63 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 11 Feb 2026 05:29:51 +0000 Subject: [PATCH 1/4] fix(api): typo in docs --- .stats.yml | 4 +-- src/isaacus/resources/enrichments.py | 28 ++++++------------- src/isaacus/types/enrichment_create_params.py | 14 +++------- src/isaacus/types/enrichment_response.py | 2 +- src/isaacus/types/ilgs/v1/document.py | 2 +- tests/api_resources/test_enrichments.py | 4 +-- 6 files changed, 18 insertions(+), 36 deletions(-) diff --git a/.stats.yml b/.stats.yml index 8b31efe..fc2373c 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 5 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/isaacus%2Fisaacus-854d1a74fd0240b79b6a7902200adf22b85d5cb67710abe7c0177b4f8801157f.yml -openapi_spec_hash: 9a141dbe42dfb83a674e69441888776f +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/isaacus%2Fisaacus-5033f9217ca384041df97464049b6ed07fd6d6dc47d76db59246fa67edec9ee5.yml +openapi_spec_hash: ed015407e5771e22d0b48721f6f88e11 config_hash: 9040e7359f066240ad536041fb2c5185 diff --git a/src/isaacus/resources/enrichments.py b/src/isaacus/resources/enrichments.py index dba9061..7d1752c 100644 --- a/src/isaacus/resources/enrichments.py +++ b/src/isaacus/resources/enrichments.py @@ -49,7 +49,7 @@ def create( *, model: Literal["kanon-2-enricher"], texts: Union[SequenceNotStr[str], str], - overflow_strategy: Optional[Literal["auto", "drop_end", "chunk"]] | Omit = omit, + overflow_strategy: Optional[Literal["auto", "drop_end"]] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. # The extra values given here take precedence over values defined on the client or passed to this method. extra_headers: Headers | None = None, @@ -71,15 +71,9 @@ def create( overflow_strategy: The strategy for handling content exceeding the model's maximum input length. - `auto`, which is the recommended setting, currently behaves the same as `chunk`, - which intelligently breaks the input up into smaller chunks and then stitches - the results back together into a single prediction. In the future `auto` may - implement even more sophisticated strategies for handling long contexts such as - leveraging chunk overlap and/or a specialized stitching model. - - `chunk` breaks the input up into smaller chunks that fit within the model's - context window and then intelligently merges the results into a single - prediction at the cost of a minor accuracy drop. + `auto` currently behaves the same as `drop_end`, dropping excess tokens from the + end of input. In the future, `auto` may implement more sophisticated strategies + such as chunking and context-aware stitching. `drop_end` drops tokens from the end of input exceeding the model's maximum input length. @@ -137,7 +131,7 @@ async def create( *, model: Literal["kanon-2-enricher"], texts: Union[SequenceNotStr[str], str], - overflow_strategy: Optional[Literal["auto", "drop_end", "chunk"]] | Omit = omit, + overflow_strategy: Optional[Literal["auto", "drop_end"]] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. # The extra values given here take precedence over values defined on the client or passed to this method. extra_headers: Headers | None = None, @@ -159,15 +153,9 @@ async def create( overflow_strategy: The strategy for handling content exceeding the model's maximum input length. - `auto`, which is the recommended setting, currently behaves the same as `chunk`, - which intelligently breaks the input up into smaller chunks and then stitches - the results back together into a single prediction. In the future `auto` may - implement even more sophisticated strategies for handling long contexts such as - leveraging chunk overlap and/or a specialized stitching model. - - `chunk` breaks the input up into smaller chunks that fit within the model's - context window and then intelligently merges the results into a single - prediction at the cost of a minor accuracy drop. + `auto` currently behaves the same as `drop_end`, dropping excess tokens from the + end of input. In the future, `auto` may implement more sophisticated strategies + such as chunking and context-aware stitching. `drop_end` drops tokens from the end of input exceeding the model's maximum input length. diff --git a/src/isaacus/types/enrichment_create_params.py b/src/isaacus/types/enrichment_create_params.py index 0cf1194..4c412e5 100644 --- a/src/isaacus/types/enrichment_create_params.py +++ b/src/isaacus/types/enrichment_create_params.py @@ -25,18 +25,12 @@ class EnrichmentCreateParams(TypedDict, total=False): No more than 8 texts can be enriched in a single request. """ - overflow_strategy: Optional[Literal["auto", "drop_end", "chunk"]] + overflow_strategy: Optional[Literal["auto", "drop_end"]] """The strategy for handling content exceeding the model's maximum input length. - `auto`, which is the recommended setting, currently behaves the same as `chunk`, - which intelligently breaks the input up into smaller chunks and then stitches - the results back together into a single prediction. In the future `auto` may - implement even more sophisticated strategies for handling long contexts such as - leveraging chunk overlap and/or a specialized stitching model. - - `chunk` breaks the input up into smaller chunks that fit within the model's - context window and then intelligently merges the results into a single - prediction at the cost of a minor accuracy drop. + `auto` currently behaves the same as `drop_end`, dropping excess tokens from the + end of input. In the future, `auto` may implement more sophisticated strategies + such as chunking and context-aware stitching. `drop_end` drops tokens from the end of input exceeding the model's maximum input length. diff --git a/src/isaacus/types/enrichment_response.py b/src/isaacus/types/enrichment_response.py index 41e8e38..13013bf 100644 --- a/src/isaacus/types/enrichment_response.py +++ b/src/isaacus/types/enrichment_response.py @@ -20,7 +20,7 @@ class Result(BaseModel): document: Document """ The document enriched into version 1.0.0 of the Isaacus Legal Graph Schema - (IGLS). + (ILGS). All spans in an enriched document graph are indexed into the Unicode code point space of a source document. diff --git a/src/isaacus/types/ilgs/v1/document.py b/src/isaacus/types/ilgs/v1/document.py index dbd1845..808eae2 100644 --- a/src/isaacus/types/ilgs/v1/document.py +++ b/src/isaacus/types/ilgs/v1/document.py @@ -23,7 +23,7 @@ class Document(BaseModel): """ - The document enriched into version 1.0.0 of the Isaacus Legal Graph Schema (IGLS). + The document enriched into version 1.0.0 of the Isaacus Legal Graph Schema (ILGS). All spans in an enriched document graph are indexed into the Unicode code point space of a source document. diff --git a/tests/api_resources/test_enrichments.py b/tests/api_resources/test_enrichments.py index eac10a9..0960d04 100644 --- a/tests/api_resources/test_enrichments.py +++ b/tests/api_resources/test_enrichments.py @@ -32,7 +32,7 @@ def test_method_create_with_all_params(self, client: Isaacus) -> None: enrichment = client.enrichments.create( model="kanon-2-enricher", texts=['1.5 You (the "User") agree to be bound by these Terms.'], - overflow_strategy="auto", + overflow_strategy=None, ) assert_matches_type(EnrichmentResponse, enrichment, path=["response"]) @@ -85,7 +85,7 @@ async def test_method_create_with_all_params(self, async_client: AsyncIsaacus) - enrichment = await async_client.enrichments.create( model="kanon-2-enricher", texts=['1.5 You (the "User") agree to be bound by these Terms.'], - overflow_strategy="auto", + overflow_strategy=None, ) assert_matches_type(EnrichmentResponse, enrichment, path=["response"]) From 1dfc52c3c6c495de8969f37f42ad6010e7bea03e Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 11 Feb 2026 05:31:00 +0000 Subject: [PATCH 2/4] fix(api): typo in docs --- .stats.yml | 4 +-- src/isaacus/resources/enrichments.py | 28 +++++++++++++------ src/isaacus/types/enrichment_create_params.py | 14 +++++++--- tests/api_resources/test_enrichments.py | 4 +-- 4 files changed, 34 insertions(+), 16 deletions(-) diff --git a/.stats.yml b/.stats.yml index fc2373c..0df1ed2 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 5 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/isaacus%2Fisaacus-5033f9217ca384041df97464049b6ed07fd6d6dc47d76db59246fa67edec9ee5.yml -openapi_spec_hash: ed015407e5771e22d0b48721f6f88e11 +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/isaacus%2Fisaacus-84e454eb280ff5e816abbb592ea590979e180df5c49af8cbe9bec81efb847f81.yml +openapi_spec_hash: 9eaef1b127c6dc6ff23ea1367d42459a config_hash: 9040e7359f066240ad536041fb2c5185 diff --git a/src/isaacus/resources/enrichments.py b/src/isaacus/resources/enrichments.py index 7d1752c..dba9061 100644 --- a/src/isaacus/resources/enrichments.py +++ b/src/isaacus/resources/enrichments.py @@ -49,7 +49,7 @@ def create( *, model: Literal["kanon-2-enricher"], texts: Union[SequenceNotStr[str], str], - overflow_strategy: Optional[Literal["auto", "drop_end"]] | Omit = omit, + overflow_strategy: Optional[Literal["auto", "drop_end", "chunk"]] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. # The extra values given here take precedence over values defined on the client or passed to this method. extra_headers: Headers | None = None, @@ -71,9 +71,15 @@ def create( overflow_strategy: The strategy for handling content exceeding the model's maximum input length. - `auto` currently behaves the same as `drop_end`, dropping excess tokens from the - end of input. In the future, `auto` may implement more sophisticated strategies - such as chunking and context-aware stitching. + `auto`, which is the recommended setting, currently behaves the same as `chunk`, + which intelligently breaks the input up into smaller chunks and then stitches + the results back together into a single prediction. In the future `auto` may + implement even more sophisticated strategies for handling long contexts such as + leveraging chunk overlap and/or a specialized stitching model. + + `chunk` breaks the input up into smaller chunks that fit within the model's + context window and then intelligently merges the results into a single + prediction at the cost of a minor accuracy drop. `drop_end` drops tokens from the end of input exceeding the model's maximum input length. @@ -131,7 +137,7 @@ async def create( *, model: Literal["kanon-2-enricher"], texts: Union[SequenceNotStr[str], str], - overflow_strategy: Optional[Literal["auto", "drop_end"]] | Omit = omit, + overflow_strategy: Optional[Literal["auto", "drop_end", "chunk"]] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. # The extra values given here take precedence over values defined on the client or passed to this method. extra_headers: Headers | None = None, @@ -153,9 +159,15 @@ async def create( overflow_strategy: The strategy for handling content exceeding the model's maximum input length. - `auto` currently behaves the same as `drop_end`, dropping excess tokens from the - end of input. In the future, `auto` may implement more sophisticated strategies - such as chunking and context-aware stitching. + `auto`, which is the recommended setting, currently behaves the same as `chunk`, + which intelligently breaks the input up into smaller chunks and then stitches + the results back together into a single prediction. In the future `auto` may + implement even more sophisticated strategies for handling long contexts such as + leveraging chunk overlap and/or a specialized stitching model. + + `chunk` breaks the input up into smaller chunks that fit within the model's + context window and then intelligently merges the results into a single + prediction at the cost of a minor accuracy drop. `drop_end` drops tokens from the end of input exceeding the model's maximum input length. diff --git a/src/isaacus/types/enrichment_create_params.py b/src/isaacus/types/enrichment_create_params.py index 4c412e5..0cf1194 100644 --- a/src/isaacus/types/enrichment_create_params.py +++ b/src/isaacus/types/enrichment_create_params.py @@ -25,12 +25,18 @@ class EnrichmentCreateParams(TypedDict, total=False): No more than 8 texts can be enriched in a single request. """ - overflow_strategy: Optional[Literal["auto", "drop_end"]] + overflow_strategy: Optional[Literal["auto", "drop_end", "chunk"]] """The strategy for handling content exceeding the model's maximum input length. - `auto` currently behaves the same as `drop_end`, dropping excess tokens from the - end of input. In the future, `auto` may implement more sophisticated strategies - such as chunking and context-aware stitching. + `auto`, which is the recommended setting, currently behaves the same as `chunk`, + which intelligently breaks the input up into smaller chunks and then stitches + the results back together into a single prediction. In the future `auto` may + implement even more sophisticated strategies for handling long contexts such as + leveraging chunk overlap and/or a specialized stitching model. + + `chunk` breaks the input up into smaller chunks that fit within the model's + context window and then intelligently merges the results into a single + prediction at the cost of a minor accuracy drop. `drop_end` drops tokens from the end of input exceeding the model's maximum input length. diff --git a/tests/api_resources/test_enrichments.py b/tests/api_resources/test_enrichments.py index 0960d04..eac10a9 100644 --- a/tests/api_resources/test_enrichments.py +++ b/tests/api_resources/test_enrichments.py @@ -32,7 +32,7 @@ def test_method_create_with_all_params(self, client: Isaacus) -> None: enrichment = client.enrichments.create( model="kanon-2-enricher", texts=['1.5 You (the "User") agree to be bound by these Terms.'], - overflow_strategy=None, + overflow_strategy="auto", ) assert_matches_type(EnrichmentResponse, enrichment, path=["response"]) @@ -85,7 +85,7 @@ async def test_method_create_with_all_params(self, async_client: AsyncIsaacus) - enrichment = await async_client.enrichments.create( model="kanon-2-enricher", texts=['1.5 You (the "User") agree to be bound by these Terms.'], - overflow_strategy=None, + overflow_strategy="auto", ) assert_matches_type(EnrichmentResponse, enrichment, path=["response"]) From 2e34941b8dfcf1397f94e20a7a52b92f6a0ed35f Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Thu, 12 Feb 2026 02:11:08 +0000 Subject: [PATCH 3/4] feat(api): revoke ILGS ID consecutive integer promise --- .stats.yml | 4 ++-- src/isaacus/types/ilgs/v1/date.py | 5 +---- src/isaacus/types/ilgs/v1/external_document.py | 3 +-- src/isaacus/types/ilgs/v1/location.py | 10 ++-------- src/isaacus/types/ilgs/v1/person.py | 15 +++------------ src/isaacus/types/ilgs/v1/quote.py | 15 +++------------ src/isaacus/types/ilgs/v1/segment.py | 10 ++-------- src/isaacus/types/ilgs/v1/term.py | 5 +---- 8 files changed, 15 insertions(+), 52 deletions(-) diff --git a/.stats.yml b/.stats.yml index 0df1ed2..046016b 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 5 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/isaacus%2Fisaacus-84e454eb280ff5e816abbb592ea590979e180df5c49af8cbe9bec81efb847f81.yml -openapi_spec_hash: 9eaef1b127c6dc6ff23ea1367d42459a +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/isaacus%2Fisaacus-daf4baba7c77b0df59c77b13dd57c56da4900b5b3f95f81a8346ae9762419f95.yml +openapi_spec_hash: 346769e832e482cea4496f92d56dd8b6 config_hash: 9040e7359f066240ad536041fb2c5185 diff --git a/src/isaacus/types/ilgs/v1/date.py b/src/isaacus/types/ilgs/v1/date.py index 6aaa370..4ca4bff 100644 --- a/src/isaacus/types/ilgs/v1/date.py +++ b/src/isaacus/types/ilgs/v1/date.py @@ -58,10 +58,7 @@ class Date(BaseModel): """ person: Optional[str] = None - """ - A unique identifier for a legal person in the format `per:{index}` where - `{index}` is a non-negative incrementing integer starting from zero. - """ + """A unique identifier for a legal person in the format `per:{identifier}`.""" mentions: List[Span] """ diff --git a/src/isaacus/types/ilgs/v1/external_document.py b/src/isaacus/types/ilgs/v1/external_document.py index 8f2b135..f2833e3 100644 --- a/src/isaacus/types/ilgs/v1/external_document.py +++ b/src/isaacus/types/ilgs/v1/external_document.py @@ -14,8 +14,7 @@ class ExternalDocument(BaseModel): id: str """ - The unique identifier of the external document in the format `exd:{index}` where - `{index}` is a non-negative incrementing integer starting from zero. + The unique identifier of the external document in the format `exd:{identifier}`. """ name: Span diff --git a/src/isaacus/types/ilgs/v1/location.py b/src/isaacus/types/ilgs/v1/location.py index facd492..b84bb94 100644 --- a/src/isaacus/types/ilgs/v1/location.py +++ b/src/isaacus/types/ilgs/v1/location.py @@ -13,10 +13,7 @@ class Location(BaseModel): """A location identified within a document.""" id: str - """ - The unique identifier of the location in the format `loc:{index}` where - `{index}` is a non-negative incrementing integer starting from zero. - """ + """The unique identifier of the location in the format `loc:{identifier}`.""" name: Span """A zero-based, half-open span into the Unicode code point space of input text. @@ -41,10 +38,7 @@ class Location(BaseModel): """ parent: Optional[str] = None - """ - A unique identifier for a location in the format `loc:{index}` where `{index}` - is a non-negative incrementing integer starting from zero. - """ + """A unique identifier for a location in the format `loc:{identifier}`.""" children: List[str] """ diff --git a/src/isaacus/types/ilgs/v1/person.py b/src/isaacus/types/ilgs/v1/person.py index 38c9fdc..0b58fdb 100644 --- a/src/isaacus/types/ilgs/v1/person.py +++ b/src/isaacus/types/ilgs/v1/person.py @@ -13,10 +13,7 @@ class Person(BaseModel): """A legal person identified in a document.""" id: str - """ - The unique identifier of the person in the format `per:{index}` where `{index}` - is a non-negative incrementing integer starting from zero. - """ + """The unique identifier of the person in the format `per:{identifier}`.""" name: Span """A zero-based, half-open span into the Unicode code point space of input text. @@ -177,10 +174,7 @@ class Person(BaseModel): """ parent: Optional[str] = None - """ - A unique identifier for a legal person in the format `per:{index}` where - `{index}` is a non-negative incrementing integer starting from zero. - """ + """A unique identifier for a legal person in the format `per:{identifier}`.""" children: List[str] """ @@ -189,10 +183,7 @@ class Person(BaseModel): """ residence: Optional[str] = None - """ - A unique identifier for a location in the format `loc:{index}` where `{index}` - is a non-negative incrementing integer starting from zero. - """ + """A unique identifier for a location in the format `loc:{identifier}`.""" mentions: List[Span] """ diff --git a/src/isaacus/types/ilgs/v1/quote.py b/src/isaacus/types/ilgs/v1/quote.py index 3bcaad9..82c27c6 100644 --- a/src/isaacus/types/ilgs/v1/quote.py +++ b/src/isaacus/types/ilgs/v1/quote.py @@ -12,22 +12,13 @@ class Quote(BaseModel): """A quotation within a document.""" source_segment: Optional[str] = None - """ - A unique identifier for a segment in the format `seg:{index}` where `{index}` is - a non-negative incrementing integer starting from zero. - """ + """A unique identifier for a segment in the format `seg:{identifier}`.""" source_document: Optional[str] = None - """ - A unique identifier for an external document in the format `exd:{index}` where - `{index}` is a non-negative incrementing integer starting from zero. - """ + """A unique identifier for an external document in the format `exd:{identifier}`.""" source_person: Optional[str] = None - """ - A unique identifier for a legal person in the format `per:{index}` where - `{index}` is a non-negative incrementing integer starting from zero. - """ + """A unique identifier for a legal person in the format `per:{identifier}`.""" amending: bool """ diff --git a/src/isaacus/types/ilgs/v1/segment.py b/src/isaacus/types/ilgs/v1/segment.py index 3ca21d2..4814fcd 100644 --- a/src/isaacus/types/ilgs/v1/segment.py +++ b/src/isaacus/types/ilgs/v1/segment.py @@ -15,10 +15,7 @@ class Segment(BaseModel): """ id: str - """ - The unique identifier of the segment in the format `seg:{index}` where `{index}` - is a non-negative incrementing integer starting from zero. - """ + """The unique identifier of the segment in the format `seg:{identifier}`.""" kind: Literal["container", "unit", "item", "figure"] """ @@ -183,10 +180,7 @@ class Segment(BaseModel): """ parent: Optional[str] = None - """ - A unique identifier for a segment in the format `seg:{index}` where `{index}` is - a non-negative incrementing integer starting from zero. - """ + """A unique identifier for a segment in the format `seg:{identifier}`.""" children: List[str] """ diff --git a/src/isaacus/types/ilgs/v1/term.py b/src/isaacus/types/ilgs/v1/term.py index 7b34272..0ead336 100644 --- a/src/isaacus/types/ilgs/v1/term.py +++ b/src/isaacus/types/ilgs/v1/term.py @@ -12,10 +12,7 @@ class Term(BaseModel): """A term assigned a definite meaning within a document.""" id: str - """ - The unique identifier of the term in the format `term:{index}` where `{index}` - is a non-negative incrementing integer starting from zero. - """ + """The unique identifier of the term in the format `term:{identifier}`.""" name: Span """A zero-based, half-open span into the Unicode code point space of input text. From c6b88f0117ab4b99029129ea299f80d14f037b07 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Thu, 12 Feb 2026 02:11:45 +0000 Subject: [PATCH 4/4] release: 0.19.0 --- .release-please-manifest.json | 2 +- CHANGELOG.md | 14 ++++++++++++++ pyproject.toml | 2 +- src/isaacus/_version.py | 2 +- 4 files changed, 17 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index d661066..e756293 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.18.1" + ".": "0.19.0" } \ No newline at end of file diff --git a/CHANGELOG.md b/CHANGELOG.md index b80f11a..6d79f0d 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,19 @@ # Changelog +## 0.19.0 (2026-02-12) + +Full Changelog: [v0.18.1...v0.19.0](https://github.com/isaacus-dev/isaacus-python/compare/v0.18.1...v0.19.0) + +### Features + +* **api:** revoke ILGS ID consecutive integer promise ([2e34941](https://github.com/isaacus-dev/isaacus-python/commit/2e34941b8dfcf1397f94e20a7a52b92f6a0ed35f)) + + +### Bug Fixes + +* **api:** typo in docs ([1dfc52c](https://github.com/isaacus-dev/isaacus-python/commit/1dfc52c3c6c495de8969f37f42ad6010e7bea03e)) +* **api:** typo in docs ([93a15f9](https://github.com/isaacus-dev/isaacus-python/commit/93a15f90ac7512bce7c78109d9c350af61f20c63)) + ## 0.18.1 (2026-02-11) Full Changelog: [v0.18.0...v0.18.1](https://github.com/isaacus-dev/isaacus-python/compare/v0.18.0...v0.18.1) diff --git a/pyproject.toml b/pyproject.toml index b866199..601caa8 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "isaacus" -version = "0.18.1" +version = "0.19.0" description = "The official Python library for the isaacus API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/isaacus/_version.py b/src/isaacus/_version.py index 61ceebe..5682338 100644 --- a/src/isaacus/_version.py +++ b/src/isaacus/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "isaacus" -__version__ = "0.18.1" # x-release-please-version +__version__ = "0.19.0" # x-release-please-version