diff --git a/.devcontainer/Dockerfile b/.devcontainer/Dockerfile old mode 100755 new mode 100644 diff --git a/.devcontainer/devcontainer.json b/.devcontainer/devcontainer.json old mode 100755 new mode 100644 diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml old mode 100755 new mode 100644 diff --git a/.github/workflows/publish-pypi.yml b/.github/workflows/publish-pypi.yml old mode 100755 new mode 100644 diff --git a/.github/workflows/release-doctor.yml b/.github/workflows/release-doctor.yml old mode 100755 new mode 100644 diff --git a/.gitignore b/.gitignore old mode 100755 new mode 100644 diff --git a/.python-version b/.python-version old mode 100755 new mode 100644 diff --git a/.release-please-manifest.json b/.release-please-manifest.json old mode 100755 new mode 100644 index cdc55cc..abc263a --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.1.0-alpha.13" + ".": "0.1.0-alpha.14" } diff --git a/.stats.yml b/.stats.yml old mode 100755 new mode 100644 index 112acec..0222f21 --- a/.stats.yml +++ b/.stats.yml @@ -1,2 +1,2 @@ configured_endpoints: 22 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/prompt-foundry%2Fprompt-foundry-sdk-84eb1e848fd165a965df78740df71651a8b413f023e5fdb6eb1acb9cce08e006.yml +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/prompt-foundry%2Fprompt-foundry-sdk-d232bc0a3409f271bb778f567e5caeb18bea44e721c3d82a78f475f9ca3faa2b.yml diff --git a/Brewfile b/Brewfile old mode 100755 new mode 100644 diff --git a/CHANGELOG.md b/CHANGELOG.md index 56a3cf7..2911ca0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,51 @@ # Changelog +## 0.1.0-alpha.14 (2024-09-14) + +Full Changelog: [v0.1.0-alpha.13...v0.1.0-alpha.14](https://github.com/prompt-foundry/python-sdk/compare/v0.1.0-alpha.13...v0.1.0-alpha.14) + +### Features + +* **api:** OpenAPI spec update via Stainless API ([#271](https://github.com/prompt-foundry/python-sdk/issues/271)) ([4c6751a](https://github.com/prompt-foundry/python-sdk/commit/4c6751aa300a930de5bebc9bcd0a51fe8d86534a)) +* **api:** OpenAPI spec update via Stainless API ([#273](https://github.com/prompt-foundry/python-sdk/issues/273)) ([e9d3e2d](https://github.com/prompt-foundry/python-sdk/commit/e9d3e2d5888bde33ca1b572ff9eb151f5d156e49)) +* **api:** OpenAPI spec update via Stainless API ([#274](https://github.com/prompt-foundry/python-sdk/issues/274)) ([080663e](https://github.com/prompt-foundry/python-sdk/commit/080663ea3f50326a5d842045aba98bc4f8f1a510)) +* **api:** OpenAPI spec update via Stainless API ([#275](https://github.com/prompt-foundry/python-sdk/issues/275)) ([b19a6f9](https://github.com/prompt-foundry/python-sdk/commit/b19a6f9741b92092bd8821dd2270afce492fc55f)) +* **api:** OpenAPI spec update via Stainless API ([#276](https://github.com/prompt-foundry/python-sdk/issues/276)) ([c9c3357](https://github.com/prompt-foundry/python-sdk/commit/c9c3357d3472d7ad1a5599b5e61331adbda3ec45)) +* **api:** OpenAPI spec update via Stainless API ([#277](https://github.com/prompt-foundry/python-sdk/issues/277)) ([1f274a8](https://github.com/prompt-foundry/python-sdk/commit/1f274a86f96e2fd7e2e5c60b85e4d4dfe1e3c734)) +* **api:** OpenAPI spec update via Stainless API ([#278](https://github.com/prompt-foundry/python-sdk/issues/278)) ([56cacb5](https://github.com/prompt-foundry/python-sdk/commit/56cacb5ad53fa814163afa31e4b247abee281818)) +* **api:** OpenAPI spec update via Stainless API ([#279](https://github.com/prompt-foundry/python-sdk/issues/279)) ([f9f4520](https://github.com/prompt-foundry/python-sdk/commit/f9f452070e8ac4149bc9f6dd19c1033bacd5e909)) +* **api:** OpenAPI spec update via Stainless API ([#280](https://github.com/prompt-foundry/python-sdk/issues/280)) ([18ffd3c](https://github.com/prompt-foundry/python-sdk/commit/18ffd3cfbc584090777cfdbf47d1772ed48b1884)) +* **api:** OpenAPI spec update via Stainless API ([#281](https://github.com/prompt-foundry/python-sdk/issues/281)) ([af39ea5](https://github.com/prompt-foundry/python-sdk/commit/af39ea571f490b98433da4ccc7a432cef4858b8d)) +* **api:** OpenAPI spec update via Stainless API ([#282](https://github.com/prompt-foundry/python-sdk/issues/282)) ([7df3d07](https://github.com/prompt-foundry/python-sdk/commit/7df3d073c549d78f97c87213e1d16258ebe4f01c)) +* **api:** OpenAPI spec update via Stainless API ([#283](https://github.com/prompt-foundry/python-sdk/issues/283)) ([fcd9e6f](https://github.com/prompt-foundry/python-sdk/commit/fcd9e6fa3ab06ee59402393fe3f153e70a851014)) +* **api:** OpenAPI spec update via Stainless API ([#284](https://github.com/prompt-foundry/python-sdk/issues/284)) ([4e5ce8c](https://github.com/prompt-foundry/python-sdk/commit/4e5ce8c858497127ba6389f319bdff61b61af97b)) +* **api:** OpenAPI spec update via Stainless API ([#285](https://github.com/prompt-foundry/python-sdk/issues/285)) ([e5105fa](https://github.com/prompt-foundry/python-sdk/commit/e5105fa21ad9859ccb8bdea5f11db74d82401606)) +* **api:** OpenAPI spec update via Stainless API ([#286](https://github.com/prompt-foundry/python-sdk/issues/286)) ([7a8d128](https://github.com/prompt-foundry/python-sdk/commit/7a8d128961aaef728df795aafd693fb1f256d647)) +* **api:** OpenAPI spec update via Stainless API ([#287](https://github.com/prompt-foundry/python-sdk/issues/287)) ([3f07e39](https://github.com/prompt-foundry/python-sdk/commit/3f07e39f446c97a84d037b81a6b178bb2176d0b8)) +* **api:** OpenAPI spec update via Stainless API ([#288](https://github.com/prompt-foundry/python-sdk/issues/288)) ([9e7569a](https://github.com/prompt-foundry/python-sdk/commit/9e7569ad26ab4fd61e3aa84e70293d869673c3c6)) +* **api:** OpenAPI spec update via Stainless API ([#289](https://github.com/prompt-foundry/python-sdk/issues/289)) ([9614fa0](https://github.com/prompt-foundry/python-sdk/commit/9614fa0e9c9b28cab752193d3c2c34b4e138f57f)) +* **api:** OpenAPI spec update via Stainless API ([#290](https://github.com/prompt-foundry/python-sdk/issues/290)) ([793f960](https://github.com/prompt-foundry/python-sdk/commit/793f9600798a3b656e1069da9471eff70d70ac10)) +* **api:** OpenAPI spec update via Stainless API ([#291](https://github.com/prompt-foundry/python-sdk/issues/291)) ([2c4cab2](https://github.com/prompt-foundry/python-sdk/commit/2c4cab2527d40beff0daec2f12d9d444f831dd3c)) +* **api:** OpenAPI spec update via Stainless API ([#292](https://github.com/prompt-foundry/python-sdk/issues/292)) ([244ea00](https://github.com/prompt-foundry/python-sdk/commit/244ea00a86a35f3eb68cf2fd7f0e1c6d79054e5b)) +* **api:** OpenAPI spec update via Stainless API ([#293](https://github.com/prompt-foundry/python-sdk/issues/293)) ([8998b8d](https://github.com/prompt-foundry/python-sdk/commit/8998b8dae1460bd0892ce00d3eba5683d0f46aa0)) +* **api:** OpenAPI spec update via Stainless API ([#294](https://github.com/prompt-foundry/python-sdk/issues/294)) ([473d442](https://github.com/prompt-foundry/python-sdk/commit/473d442b72510d41da63512e4146b9c02f655b1d)) +* **api:** OpenAPI spec update via Stainless API ([#295](https://github.com/prompt-foundry/python-sdk/issues/295)) ([3137813](https://github.com/prompt-foundry/python-sdk/commit/3137813996092309100a794ff085ffdbd491c587)) +* **api:** OpenAPI spec update via Stainless API ([#296](https://github.com/prompt-foundry/python-sdk/issues/296)) ([686cce8](https://github.com/prompt-foundry/python-sdk/commit/686cce833e31ec43864f6fb59b8f8d23a18b6046)) +* **api:** OpenAPI spec update via Stainless API ([#297](https://github.com/prompt-foundry/python-sdk/issues/297)) ([8af54a0](https://github.com/prompt-foundry/python-sdk/commit/8af54a04f733b4f926b227168e220b109237ecd4)) +* **api:** OpenAPI spec update via Stainless API ([#299](https://github.com/prompt-foundry/python-sdk/issues/299)) ([0da70b8](https://github.com/prompt-foundry/python-sdk/commit/0da70b849dd098ce919c2c4761e6fa95b5bb40a5)) + + +### Chores + +* **internal:** codegen related update ([#300](https://github.com/prompt-foundry/python-sdk/issues/300)) ([68f8e74](https://github.com/prompt-foundry/python-sdk/commit/68f8e74e3ea4213fce959184ae503da561c07229)) +* pyproject.toml formatting changes ([#298](https://github.com/prompt-foundry/python-sdk/issues/298)) ([0a92048](https://github.com/prompt-foundry/python-sdk/commit/0a92048e7bbf562ed88d6940163e78083774c106)) + + +### Documentation + +* **readme:** add section on determining installed version ([#301](https://github.com/prompt-foundry/python-sdk/issues/301)) ([229a803](https://github.com/prompt-foundry/python-sdk/commit/229a803d0777cb72a9c34c86f314e002fcd4ea2d)) +* update CONTRIBUTING.md ([#302](https://github.com/prompt-foundry/python-sdk/issues/302)) ([c205f0e](https://github.com/prompt-foundry/python-sdk/commit/c205f0ea97d9faf98d205d2da14108d67b06d154)) + ## 0.1.0-alpha.13 (2024-08-20) Full Changelog: [v0.1.0-alpha.12...v0.1.0-alpha.13](https://github.com/prompt-foundry/python-sdk/compare/v0.1.0-alpha.12...v0.1.0-alpha.13) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md old mode 100755 new mode 100644 index 1e7789b..1066e58 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -31,13 +31,13 @@ $ pip install -r requirements-dev.lock ## Modifying/Adding code -Most of the SDK is generated code, and any modified code will be overridden on the next generation. The -`src/prompt_foundry_python_sdk/lib/` and `examples/` directories are exceptions and will never be overridden. +Most of the SDK is generated code. Modifications to code will be persisted between generations, but may +result in merge conflicts between manual patches and changes from the generator. The generator will never +modify the contents of the `src/prompt_foundry_python_sdk/lib/` and `examples/` directories. ## Adding and running examples -All files in the `examples/` directory are not modified by the Stainless generator and can be freely edited or -added to. +All files in the `examples/` directory are not modified by the generator and can be freely edited or added to. ```bash # add an example to examples/.py diff --git a/LICENSE b/LICENSE old mode 100755 new mode 100644 diff --git a/README.md b/README.md old mode 100755 new mode 100644 index ec8465e..58a1928 --- a/README.md +++ b/README.md @@ -456,6 +456,17 @@ We take backwards-compatibility seriously and work hard to ensure you can rely o We are keen for your feedback; please open an [issue](https://www.github.com/prompt-foundry/python-sdk/issues) with questions, bugs, or suggestions. +### Determining the installed version + +If you've upgraded to the latest version but aren't seeing any new features you were expecting then your python environment is likely still using an older version. + +You can determine the version that is being used at runtime with: + +```py +import prompt_foundry_python_sdk +print(prompt_foundry_python_sdk.__version__) +``` + ## Requirements Python 3.7 or higher. diff --git a/SECURITY.md b/SECURITY.md old mode 100755 new mode 100644 diff --git a/api.md b/api.md old mode 100755 new mode 100644 index 0ca765e..1571049 --- a/api.md +++ b/api.md @@ -16,10 +16,10 @@ Types: ```python from prompt_foundry_python_sdk.types import ( - Parameters, PromptConfiguration, PromptListResponse, PromptDeleteResponse, + PromptGetParametersResponse, ) ``` @@ -30,7 +30,7 @@ Methods: - client.prompts.list() -> PromptListResponse - client.prompts.delete(id) -> PromptDeleteResponse - client.prompts.get(id) -> PromptConfiguration -- client.prompts.get_parameters(id, \*\*params) -> Parameters +- client.prompts.get_parameters(id, \*\*params) -> PromptGetParametersResponse # Tools diff --git a/bin/check-release-environment b/bin/check-release-environment old mode 100755 new mode 100644 diff --git a/bin/publish-pypi b/bin/publish-pypi old mode 100755 new mode 100644 diff --git a/examples/.keep b/examples/.keep old mode 100755 new mode 100644 diff --git a/mypy.ini b/mypy.ini old mode 100755 new mode 100644 diff --git a/noxfile.py b/noxfile.py old mode 100755 new mode 100644 diff --git a/pyproject.toml b/pyproject.toml old mode 100755 new mode 100644 index 97f4cd1..9dfce02 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "prompt_foundry_python_sdk" -version = "0.1.0-alpha.13" +version = "0.1.0-alpha.14" description = "The prompt engineering, prompt management, and prompt evaluation tool for Python " keywords = ["llm evaluation", "prompt management", "prompt testing", "prompt engineering"] dynamic = ["readme"] @@ -16,7 +16,6 @@ dependencies = [ "distro>=1.7.0, <2", "sniffio", "cached-property; python_version < '3.8'", - ] requires-python = ">= 3.7" classifiers = [ @@ -37,8 +36,6 @@ classifiers = [ "License :: OSI Approved :: Apache Software License" ] - - [project.urls] Homepage = "https://https://www.promptfoundry.ai" Repository = "https://github.com/prompt-foundry/python-sdk" @@ -61,7 +58,6 @@ dev-dependencies = [ "dirty-equals>=0.6.0", "importlib-metadata>=6.7.0", "rich>=13.7.1", - ] [tool.rye.scripts] diff --git a/release-please-config.json b/release-please-config.json old mode 100755 new mode 100644 diff --git a/requirements-dev.lock b/requirements-dev.lock old mode 100755 new mode 100644 diff --git a/requirements.lock b/requirements.lock old mode 100755 new mode 100644 diff --git a/scripts/utils/ruffen-docs.py b/scripts/utils/ruffen-docs.py old mode 100755 new mode 100644 diff --git a/src/prompt-foundry-python-sdk/lib/.keep b/src/prompt-foundry-python-sdk/lib/.keep old mode 100755 new mode 100644 diff --git a/src/prompt-foundry-sdk/lib/.keep b/src/prompt-foundry-sdk/lib/.keep old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/__init__.py b/src/prompt_foundry_python_sdk/__init__.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_base_client.py b/src/prompt_foundry_python_sdk/_base_client.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_client.py b/src/prompt_foundry_python_sdk/_client.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_compat.py b/src/prompt_foundry_python_sdk/_compat.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_constants.py b/src/prompt_foundry_python_sdk/_constants.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_exceptions.py b/src/prompt_foundry_python_sdk/_exceptions.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_files.py b/src/prompt_foundry_python_sdk/_files.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_models.py b/src/prompt_foundry_python_sdk/_models.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_qs.py b/src/prompt_foundry_python_sdk/_qs.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_resource.py b/src/prompt_foundry_python_sdk/_resource.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_response.py b/src/prompt_foundry_python_sdk/_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_streaming.py b/src/prompt_foundry_python_sdk/_streaming.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_types.py b/src/prompt_foundry_python_sdk/_types.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/__init__.py b/src/prompt_foundry_python_sdk/_utils/__init__.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_logs.py b/src/prompt_foundry_python_sdk/_utils/_logs.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_proxy.py b/src/prompt_foundry_python_sdk/_utils/_proxy.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_reflection.py b/src/prompt_foundry_python_sdk/_utils/_reflection.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_streams.py b/src/prompt_foundry_python_sdk/_utils/_streams.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_sync.py b/src/prompt_foundry_python_sdk/_utils/_sync.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_transform.py b/src/prompt_foundry_python_sdk/_utils/_transform.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_typing.py b/src/prompt_foundry_python_sdk/_utils/_typing.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_utils/_utils.py b/src/prompt_foundry_python_sdk/_utils/_utils.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/_version.py b/src/prompt_foundry_python_sdk/_version.py old mode 100755 new mode 100644 index 69495c2..0418302 --- a/src/prompt_foundry_python_sdk/_version.py +++ b/src/prompt_foundry_python_sdk/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "prompt_foundry_python_sdk" -__version__ = "0.1.0-alpha.13" # x-release-please-version +__version__ = "0.1.0-alpha.14" # x-release-please-version diff --git a/src/prompt_foundry_python_sdk/lib/.keep b/src/prompt_foundry_python_sdk/lib/.keep old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/py.typed b/src/prompt_foundry_python_sdk/py.typed old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/resources/__init__.py b/src/prompt_foundry_python_sdk/resources/__init__.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/resources/completion.py b/src/prompt_foundry_python_sdk/resources/completion.py old mode 100755 new mode 100644 index 908b994..41a1ba1 --- a/src/prompt_foundry_python_sdk/resources/completion.py +++ b/src/prompt_foundry_python_sdk/resources/completion.py @@ -29,10 +29,21 @@ class CompletionResource(SyncAPIResource): @cached_property def with_raw_response(self) -> CompletionResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return CompletionResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> CompletionResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return CompletionResourceWithStreamingResponse(self) def create( @@ -98,10 +109,21 @@ def create( class AsyncCompletionResource(AsyncAPIResource): @cached_property def with_raw_response(self) -> AsyncCompletionResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return AsyncCompletionResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> AsyncCompletionResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return AsyncCompletionResourceWithStreamingResponse(self) async def create( diff --git a/src/prompt_foundry_python_sdk/resources/evaluation_assertions.py b/src/prompt_foundry_python_sdk/resources/evaluation_assertions.py old mode 100755 new mode 100644 index 909e4ba..fb46e8d --- a/src/prompt_foundry_python_sdk/resources/evaluation_assertions.py +++ b/src/prompt_foundry_python_sdk/resources/evaluation_assertions.py @@ -37,10 +37,21 @@ class EvaluationAssertionsResource(SyncAPIResource): @cached_property def with_raw_response(self) -> EvaluationAssertionsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return EvaluationAssertionsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> EvaluationAssertionsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return EvaluationAssertionsResourceWithStreamingResponse(self) @overload @@ -914,10 +925,21 @@ def get( class AsyncEvaluationAssertionsResource(AsyncAPIResource): @cached_property def with_raw_response(self) -> AsyncEvaluationAssertionsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return AsyncEvaluationAssertionsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> AsyncEvaluationAssertionsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return AsyncEvaluationAssertionsResourceWithStreamingResponse(self) @overload diff --git a/src/prompt_foundry_python_sdk/resources/evaluations.py b/src/prompt_foundry_python_sdk/resources/evaluations.py old mode 100755 new mode 100644 index 8eaf74b..833d793 --- a/src/prompt_foundry_python_sdk/resources/evaluations.py +++ b/src/prompt_foundry_python_sdk/resources/evaluations.py @@ -31,10 +31,21 @@ class EvaluationsResource(SyncAPIResource): @cached_property def with_raw_response(self) -> EvaluationsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return EvaluationsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> EvaluationsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return EvaluationsResourceWithStreamingResponse(self) def create( @@ -235,10 +246,21 @@ def get( class AsyncEvaluationsResource(AsyncAPIResource): @cached_property def with_raw_response(self) -> AsyncEvaluationsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return AsyncEvaluationsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> AsyncEvaluationsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return AsyncEvaluationsResourceWithStreamingResponse(self) async def create( diff --git a/src/prompt_foundry_python_sdk/resources/prompts.py b/src/prompt_foundry_python_sdk/resources/prompts.py old mode 100755 new mode 100644 index 49ea9fb..d1b9842 --- a/src/prompt_foundry_python_sdk/resources/prompts.py +++ b/src/prompt_foundry_python_sdk/resources/prompts.py @@ -21,10 +21,10 @@ async_to_streamed_response_wrapper, ) from .._base_client import make_request_options -from ..types.parameters import Parameters from ..types.prompt_configuration import PromptConfiguration from ..types.prompt_list_response import PromptListResponse from ..types.prompt_delete_response import PromptDeleteResponse +from ..types.prompt_get_parameters_response import PromptGetParametersResponse __all__ = ["PromptsResource", "AsyncPromptsResource"] @@ -32,10 +32,21 @@ class PromptsResource(SyncAPIResource): @cached_property def with_raw_response(self) -> PromptsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return PromptsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> PromptsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return PromptsResourceWithStreamingResponse(self) def create( @@ -228,7 +239,7 @@ def get_parameters( extra_query: Query | None = None, extra_body: Body | None = None, timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN, - ) -> Parameters: + ) -> PromptGetParametersResponse: """ Fetches the configured model parameters and messages rendered with the provided variables mapped to the set LLM provider. This endpoint abstracts the need to @@ -257,7 +268,7 @@ def get_parameters( if not id: raise ValueError(f"Expected a non-empty value for `id` but received {id!r}") return cast( - Parameters, + PromptGetParametersResponse, self._post( f"/sdk/v1/prompts/{id}", body=maybe_transform( @@ -272,7 +283,9 @@ def get_parameters( options=make_request_options( extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout ), - cast_to=cast(Any, Parameters), # Union types cannot be passed in as arguments in the type system + cast_to=cast( + Any, PromptGetParametersResponse + ), # Union types cannot be passed in as arguments in the type system ), ) @@ -280,10 +293,21 @@ def get_parameters( class AsyncPromptsResource(AsyncAPIResource): @cached_property def with_raw_response(self) -> AsyncPromptsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return AsyncPromptsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> AsyncPromptsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return AsyncPromptsResourceWithStreamingResponse(self) async def create( @@ -476,7 +500,7 @@ async def get_parameters( extra_query: Query | None = None, extra_body: Body | None = None, timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN, - ) -> Parameters: + ) -> PromptGetParametersResponse: """ Fetches the configured model parameters and messages rendered with the provided variables mapped to the set LLM provider. This endpoint abstracts the need to @@ -505,7 +529,7 @@ async def get_parameters( if not id: raise ValueError(f"Expected a non-empty value for `id` but received {id!r}") return cast( - Parameters, + PromptGetParametersResponse, await self._post( f"/sdk/v1/prompts/{id}", body=await async_maybe_transform( @@ -520,7 +544,9 @@ async def get_parameters( options=make_request_options( extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout ), - cast_to=cast(Any, Parameters), # Union types cannot be passed in as arguments in the type system + cast_to=cast( + Any, PromptGetParametersResponse + ), # Union types cannot be passed in as arguments in the type system ), ) diff --git a/src/prompt_foundry_python_sdk/resources/tools.py b/src/prompt_foundry_python_sdk/resources/tools.py old mode 100755 new mode 100644 index 3461db4..d4917e1 --- a/src/prompt_foundry_python_sdk/resources/tools.py +++ b/src/prompt_foundry_python_sdk/resources/tools.py @@ -31,10 +31,21 @@ class ToolsResource(SyncAPIResource): @cached_property def with_raw_response(self) -> ToolsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return ToolsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> ToolsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return ToolsResourceWithStreamingResponse(self) def create( @@ -231,10 +242,21 @@ def get( class AsyncToolsResource(AsyncAPIResource): @cached_property def with_raw_response(self) -> AsyncToolsResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return the + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#accessing-raw-response-data-eg-headers + """ return AsyncToolsResourceWithRawResponse(self) @cached_property def with_streaming_response(self) -> AsyncToolsResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/prompt-foundry/python-sdk#with_streaming_response + """ return AsyncToolsResourceWithStreamingResponse(self) async def create( diff --git a/src/prompt_foundry_python_sdk/types/__init__.py b/src/prompt_foundry_python_sdk/types/__init__.py old mode 100755 new mode 100644 index 0b3fb27..dbf8263 --- a/src/prompt_foundry_python_sdk/types/__init__.py +++ b/src/prompt_foundry_python_sdk/types/__init__.py @@ -4,7 +4,6 @@ from .tool import Tool as Tool from .evaluation import Evaluation as Evaluation -from .parameters import Parameters as Parameters from .tool_create_params import ToolCreateParams as ToolCreateParams from .tool_list_response import ToolListResponse as ToolListResponse from .tool_update_params import ToolUpdateParams as ToolUpdateParams @@ -22,6 +21,7 @@ from .completion_create_response import CompletionCreateResponse as CompletionCreateResponse from .evaluation_delete_response import EvaluationDeleteResponse as EvaluationDeleteResponse from .prompt_get_parameters_params import PromptGetParametersParams as PromptGetParametersParams +from .prompt_get_parameters_response import PromptGetParametersResponse as PromptGetParametersResponse from .evaluation_assertion_list_params import EvaluationAssertionListParams as EvaluationAssertionListParams from .evaluation_assertion_create_params import EvaluationAssertionCreateParams as EvaluationAssertionCreateParams from .evaluation_assertion_list_response import EvaluationAssertionListResponse as EvaluationAssertionListResponse diff --git a/src/prompt_foundry_python_sdk/types/completion_create_params.py b/src/prompt_foundry_python_sdk/types/completion_create_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/completion_create_response.py b/src/prompt_foundry_python_sdk/types/completion_create_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation.py b/src/prompt_foundry_python_sdk/types/evaluation.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_assertion.py b/src/prompt_foundry_python_sdk/types/evaluation_assertion.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_assertion_create_params.py b/src/prompt_foundry_python_sdk/types/evaluation_assertion_create_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_assertion_delete_response.py b/src/prompt_foundry_python_sdk/types/evaluation_assertion_delete_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_assertion_list_params.py b/src/prompt_foundry_python_sdk/types/evaluation_assertion_list_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_assertion_list_response.py b/src/prompt_foundry_python_sdk/types/evaluation_assertion_list_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_assertion_update_params.py b/src/prompt_foundry_python_sdk/types/evaluation_assertion_update_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_create_params.py b/src/prompt_foundry_python_sdk/types/evaluation_create_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_delete_response.py b/src/prompt_foundry_python_sdk/types/evaluation_delete_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_list_response.py b/src/prompt_foundry_python_sdk/types/evaluation_list_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/evaluation_update_params.py b/src/prompt_foundry_python_sdk/types/evaluation_update_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/parameters.py b/src/prompt_foundry_python_sdk/types/parameters.py deleted file mode 100755 index 803b0c7..0000000 --- a/src/prompt_foundry_python_sdk/types/parameters.py +++ /dev/null @@ -1,428 +0,0 @@ -# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. - -from typing import TYPE_CHECKING, Dict, List, Union, Optional -from typing_extensions import Literal, TypeAlias - -from .._models import BaseModel - -__all__ = [ - "Parameters", - "AnthropicModelParameters", - "AnthropicModelParametersParameters", - "AnthropicModelParametersParametersMessage", - "AnthropicModelParametersParametersMessageContentUnionMember1", - "AnthropicModelParametersParametersMessageContentUnionMember1TextBlockParam", - "AnthropicModelParametersParametersMessageContentUnionMember1ImageBlockParam", - "AnthropicModelParametersParametersMessageContentUnionMember1ImageBlockParamSource", - "AnthropicModelParametersParametersMessageContentUnionMember1ToolUseBlockParam", - "AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParam", - "AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1", - "AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1TextBlockParam", - "AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParam", - "AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParamSource", - "AnthropicModelParametersParametersMetadata", - "AnthropicModelParametersParametersToolChoice", - "AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceAuto", - "AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceAny", - "AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceTool", - "AnthropicModelParametersParametersTool", - "AnthropicModelParametersParametersToolInputSchema", - "OpenAIModelParameters", - "OpenAIModelParametersParameters", - "OpenAIModelParametersParametersMessage", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestSystemMessage", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessage", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartText", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImage", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImageImageURL", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessage", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageFunctionCall", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCall", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCallFunction", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestToolMessage", - "OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestFunctionMessage", - "OpenAIModelParametersParametersResponseFormat", - "OpenAIModelParametersParametersStreamOptions", - "OpenAIModelParametersParametersToolChoice", - "OpenAIModelParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoice", - "OpenAIModelParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoiceFunction", - "OpenAIModelParametersParametersTool", - "OpenAIModelParametersParametersToolFunction", -] - - -class AnthropicModelParametersParametersMessageContentUnionMember1TextBlockParam(BaseModel): - text: str - - type: Literal["text"] - - -class AnthropicModelParametersParametersMessageContentUnionMember1ImageBlockParamSource(BaseModel): - data: str - - media_type: Literal["image/jpeg", "image/png", "image/gif", "image/webp"] - - type: Literal["base64"] - - -class AnthropicModelParametersParametersMessageContentUnionMember1ImageBlockParam(BaseModel): - source: AnthropicModelParametersParametersMessageContentUnionMember1ImageBlockParamSource - - type: Literal["image"] - - -class AnthropicModelParametersParametersMessageContentUnionMember1ToolUseBlockParam(BaseModel): - id: str - - input: Dict[str, str] - - name: str - - type: Literal["tool_use"] - - -class AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1TextBlockParam( - BaseModel -): - text: str - - type: Literal["text"] - - -class AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParamSource( - BaseModel -): - data: str - - media_type: Literal["image/jpeg", "image/png", "image/gif", "image/webp"] - - type: Literal["base64"] - - -class AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParam( - BaseModel -): - source: AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParamSource - - type: Literal["image"] - - -AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1: TypeAlias = Union[ - AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1TextBlockParam, - AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParam, -] - - -class AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParam(BaseModel): - tool_use_id: str - - type: Literal["tool_result"] - - content: Union[ - str, - List[AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1], - None, - ] = None - - is_error: Optional[bool] = None - - -AnthropicModelParametersParametersMessageContentUnionMember1: TypeAlias = Union[ - AnthropicModelParametersParametersMessageContentUnionMember1TextBlockParam, - AnthropicModelParametersParametersMessageContentUnionMember1ImageBlockParam, - AnthropicModelParametersParametersMessageContentUnionMember1ToolUseBlockParam, - AnthropicModelParametersParametersMessageContentUnionMember1ToolResultBlockParam, -] - - -class AnthropicModelParametersParametersMessage(BaseModel): - content: Union[str, List[AnthropicModelParametersParametersMessageContentUnionMember1]] - - role: Literal["user", "assistant"] - - -class AnthropicModelParametersParametersMetadata(BaseModel): - user_id: Optional[str] = None - - -class AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceAuto(BaseModel): - type: Literal["auto"] - - -class AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceAny(BaseModel): - type: Literal["any"] - - -class AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceTool(BaseModel): - name: str - - type: Literal["tool"] - - -AnthropicModelParametersParametersToolChoice: TypeAlias = Union[ - AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceAuto, - AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceAny, - AnthropicModelParametersParametersToolChoiceMessageCreateParamsToolChoiceTool, -] - - -class AnthropicModelParametersParametersToolInputSchema(BaseModel): - type: Literal["object"] - - properties: Optional[object] = None - - if TYPE_CHECKING: - # Stub to indicate that arbitrary properties are accepted. - # To access properties that are not valid identifiers you can use `getattr`, e.g. - # `getattr(obj, '$type')` - def __getattr__(self, attr: str) -> Optional[object]: ... - - -class AnthropicModelParametersParametersTool(BaseModel): - input_schema: AnthropicModelParametersParametersToolInputSchema - - name: str - - description: Optional[str] = None - - -class AnthropicModelParametersParameters(BaseModel): - max_tokens: float - - messages: List[AnthropicModelParametersParametersMessage] - - model: Union[ - str, - Literal[ - "claude-3-5-sonnet-20240620", - "claude-3-opus-20240229", - "claude-3-sonnet-20240229", - "claude-3-haiku-20240307", - ], - ] - - metadata: Optional[AnthropicModelParametersParametersMetadata] = None - - stop_sequences: Optional[List[str]] = None - - stream: Optional[bool] = None - - system: Optional[str] = None - - temperature: Optional[float] = None - - tool_choice: Optional[AnthropicModelParametersParametersToolChoice] = None - - tools: Optional[List[AnthropicModelParametersParametersTool]] = None - - top_k: Optional[float] = None - - top_p: Optional[float] = None - - -class AnthropicModelParameters(BaseModel): - name: str - - parameters: AnthropicModelParametersParameters - - provider: Literal["anthropic"] - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestSystemMessage(BaseModel): - content: str - - role: Literal["system"] - - name: Optional[str] = None - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartText( - BaseModel -): - text: str - - type: Literal["text"] - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImageImageURL( - BaseModel -): - url: str - - detail: Optional[Literal["auto", "low", "high"]] = None - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImage( - BaseModel -): - image_url: OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImageImageURL - - type: Literal["image_url"] - - -OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1: TypeAlias = Union[ - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartText, - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImage, -] - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessage(BaseModel): - content: Union[ - str, List[OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1] - ] - - role: Literal["user"] - - name: Optional[str] = None - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageFunctionCall(BaseModel): - arguments: str - - name: str - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCallFunction(BaseModel): - arguments: str - - name: str - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCall(BaseModel): - id: str - - function: OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCallFunction - - type: Literal["function"] - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessage(BaseModel): - role: Literal["assistant"] - - content: Optional[str] = None - - function_call: Optional[ - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageFunctionCall - ] = None - - name: Optional[str] = None - - tool_calls: Optional[ - List[OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCall] - ] = None - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestToolMessage(BaseModel): - content: str - - role: Literal["tool"] - - tool_call_id: str - - -class OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestFunctionMessage(BaseModel): - content: Optional[str] = None - - name: str - - role: Literal["function"] - - -OpenAIModelParametersParametersMessage: TypeAlias = Union[ - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestSystemMessage, - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestUserMessage, - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestAssistantMessage, - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestToolMessage, - OpenAIModelParametersParametersMessageOpenAIChatCompletionRequestFunctionMessage, -] - - -class OpenAIModelParametersParametersResponseFormat(BaseModel): - type: Optional[Literal["text", "json_object"]] = None - - -class OpenAIModelParametersParametersStreamOptions(BaseModel): - include_usage: bool - - -class OpenAIModelParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoiceFunction(BaseModel): - name: str - - -class OpenAIModelParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoice(BaseModel): - function: OpenAIModelParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoiceFunction - - type: Literal["function"] - - -OpenAIModelParametersParametersToolChoice: TypeAlias = Union[ - Literal["none", "auto", "required"], OpenAIModelParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoice -] - - -class OpenAIModelParametersParametersToolFunction(BaseModel): - name: str - - description: Optional[str] = None - - parameters: Optional[Dict[str, Optional[object]]] = None - - -class OpenAIModelParametersParametersTool(BaseModel): - function: OpenAIModelParametersParametersToolFunction - - type: Literal["function"] - - -class OpenAIModelParametersParameters(BaseModel): - messages: List[OpenAIModelParametersParametersMessage] - - model: str - - frequency_penalty: Optional[float] = None - - logit_bias: Optional[Dict[str, int]] = None - - logprobs: Optional[bool] = None - - max_tokens: Optional[int] = None - - n: Optional[int] = None - - parallel_tool_calls: Optional[bool] = None - - presence_penalty: Optional[float] = None - - response_format: Optional[OpenAIModelParametersParametersResponseFormat] = None - - seed: Optional[int] = None - - stop: Union[str, List[str], None] = None - - stream: Optional[bool] = None - - stream_options: Optional[OpenAIModelParametersParametersStreamOptions] = None - - temperature: Optional[float] = None - - tool_choice: Optional[OpenAIModelParametersParametersToolChoice] = None - - tools: Optional[List[OpenAIModelParametersParametersTool]] = None - - top_logprobs: Optional[int] = None - - top_p: Optional[float] = None - - user: Optional[str] = None - - -class OpenAIModelParameters(BaseModel): - name: str - - parameters: OpenAIModelParametersParameters - - provider: Literal["openai"] - - -Parameters: TypeAlias = Union[AnthropicModelParameters, OpenAIModelParameters] diff --git a/src/prompt_foundry_python_sdk/types/prompt_configuration.py b/src/prompt_foundry_python_sdk/types/prompt_configuration.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/prompt_create_params.py b/src/prompt_foundry_python_sdk/types/prompt_create_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/prompt_delete_response.py b/src/prompt_foundry_python_sdk/types/prompt_delete_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/prompt_get_parameters_params.py b/src/prompt_foundry_python_sdk/types/prompt_get_parameters_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/prompt_get_parameters_response.py b/src/prompt_foundry_python_sdk/types/prompt_get_parameters_response.py new file mode 100644 index 0000000..5873f85 --- /dev/null +++ b/src/prompt_foundry_python_sdk/types/prompt_get_parameters_response.py @@ -0,0 +1,470 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import TYPE_CHECKING, Dict, List, Union, Optional +from typing_extensions import Literal, Annotated, TypeAlias + +from pydantic import Field as FieldInfo + +from .._utils import PropertyInfo +from .._models import BaseModel + +__all__ = [ + "PromptGetParametersResponse", + "AnthropicModelNonStreamingParameters", + "AnthropicModelNonStreamingParametersParameters", + "AnthropicModelNonStreamingParametersParametersMessage", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1TextBlockParam", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ImageBlockParam", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ImageBlockParamSource", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolUseBlockParam", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParam", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1TextBlockParam", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParam", + "AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParamSource", + "AnthropicModelNonStreamingParametersParametersMetadata", + "AnthropicModelNonStreamingParametersParametersToolChoice", + "AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceAuto", + "AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceAny", + "AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceTool", + "AnthropicModelNonStreamingParametersParametersTool", + "AnthropicModelNonStreamingParametersParametersToolInputSchema", + "OpenAIModelNonStreamingParameters", + "OpenAIModelNonStreamingParametersParameters", + "OpenAIModelNonStreamingParametersParametersMessage", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestSystemMessage", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessage", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartText", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImage", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImageImageURL", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessage", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageFunctionCall", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCall", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCallFunction", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestToolMessage", + "OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestFunctionMessage", + "OpenAIModelNonStreamingParametersParametersResponseFormat", + "OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatText", + "OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonObject", + "OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonSchema", + "OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonSchemaJsonSchema", + "OpenAIModelNonStreamingParametersParametersToolChoice", + "OpenAIModelNonStreamingParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoice", + "OpenAIModelNonStreamingParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoiceFunction", + "OpenAIModelNonStreamingParametersParametersTool", + "OpenAIModelNonStreamingParametersParametersToolFunction", +] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1TextBlockParam(BaseModel): + text: str + + type: Literal["text"] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ImageBlockParamSource(BaseModel): + data: str + + media_type: Literal["image/jpeg", "image/png", "image/gif", "image/webp"] + + type: Literal["base64"] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ImageBlockParam(BaseModel): + source: AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ImageBlockParamSource + + type: Literal["image"] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolUseBlockParam(BaseModel): + id: str + + input: Dict[str, str] + + name: str + + type: Literal["tool_use"] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1TextBlockParam( + BaseModel +): + text: str + + type: Literal["text"] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParamSource( + BaseModel +): + data: str + + media_type: Literal["image/jpeg", "image/png", "image/gif", "image/webp"] + + type: Literal["base64"] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParam( + BaseModel +): + source: AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParamSource + + type: Literal["image"] + + +AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1: TypeAlias = Union[ + AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1TextBlockParam, + AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1ImageBlockParam, +] + + +class AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParam(BaseModel): + tool_use_id: str + + type: Literal["tool_result"] + + content: Union[ + str, + List[ + AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParamContentUnionMember1 + ], + None, + ] = None + + is_error: Optional[bool] = None + + +AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1: TypeAlias = Union[ + AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1TextBlockParam, + AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ImageBlockParam, + AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolUseBlockParam, + AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1ToolResultBlockParam, +] + + +class AnthropicModelNonStreamingParametersParametersMessage(BaseModel): + content: Union[str, List[AnthropicModelNonStreamingParametersParametersMessageContentUnionMember1]] + + role: Literal["user", "assistant"] + + +class AnthropicModelNonStreamingParametersParametersMetadata(BaseModel): + user_id: Optional[str] = None + + +class AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceAuto(BaseModel): + type: Literal["auto"] + + +class AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceAny(BaseModel): + type: Literal["any"] + + +class AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceTool(BaseModel): + name: str + + type: Literal["tool"] + + +AnthropicModelNonStreamingParametersParametersToolChoice: TypeAlias = Union[ + AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceAuto, + AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceAny, + AnthropicModelNonStreamingParametersParametersToolChoiceMessageCreateParamsToolChoiceTool, +] + + +class AnthropicModelNonStreamingParametersParametersToolInputSchema(BaseModel): + type: Literal["object"] + + properties: Optional[object] = None + + if TYPE_CHECKING: + # Stub to indicate that arbitrary properties are accepted. + # To access properties that are not valid identifiers you can use `getattr`, e.g. + # `getattr(obj, '$type')` + def __getattr__(self, attr: str) -> Optional[object]: ... + + +class AnthropicModelNonStreamingParametersParametersTool(BaseModel): + input_schema: AnthropicModelNonStreamingParametersParametersToolInputSchema + + name: str + + description: Optional[str] = None + + +class AnthropicModelNonStreamingParametersParameters(BaseModel): + max_tokens: float + + messages: List[AnthropicModelNonStreamingParametersParametersMessage] + + model: Union[ + str, + Literal[ + "claude-3-5-sonnet-20240620", + "claude-3-opus-20240229", + "claude-3-sonnet-20240229", + "claude-3-haiku-20240307", + ], + ] + + stream: Literal[False] + + metadata: Optional[AnthropicModelNonStreamingParametersParametersMetadata] = None + + stop_sequences: Optional[List[str]] = None + + system: Optional[str] = None + + temperature: Optional[float] = None + + tool_choice: Optional[AnthropicModelNonStreamingParametersParametersToolChoice] = None + + tools: Optional[List[AnthropicModelNonStreamingParametersParametersTool]] = None + + top_k: Optional[float] = None + + top_p: Optional[float] = None + + +class AnthropicModelNonStreamingParameters(BaseModel): + name: str + + parameters: AnthropicModelNonStreamingParametersParameters + + provider: Literal["anthropic"] + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestSystemMessage(BaseModel): + content: str + + role: Literal["system"] + + name: Optional[str] = None + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartText( + BaseModel +): + text: str + + type: Literal["text"] + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImageImageURL( + BaseModel +): + url: str + + detail: Optional[Literal["auto", "low", "high"]] = None + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImage( + BaseModel +): + image_url: OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImageImageURL + + type: Literal["image_url"] + + +OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1: TypeAlias = Union[ + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartText, + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1OpenAIChatCompletionRequestMessageContentPartImage, +] + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessage(BaseModel): + content: Union[ + str, + List[ + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessageContentUnionMember1 + ], + ] + + role: Literal["user"] + + name: Optional[str] = None + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageFunctionCall( + BaseModel +): + arguments: str + + name: str + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCallFunction( + BaseModel +): + arguments: str + + name: str + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCall(BaseModel): + id: str + + function: ( + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCallFunction + ) + + type: Literal["function"] + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessage(BaseModel): + role: Literal["assistant"] + + content: Optional[str] = None + + function_call: Optional[ + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageFunctionCall + ] = None + + name: Optional[str] = None + + tool_calls: Optional[ + List[OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessageToolCall] + ] = None + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestToolMessage(BaseModel): + content: str + + role: Literal["tool"] + + tool_call_id: str + + +class OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestFunctionMessage(BaseModel): + content: Optional[str] = None + + name: str + + role: Literal["function"] + + +OpenAIModelNonStreamingParametersParametersMessage: TypeAlias = Union[ + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestSystemMessage, + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestUserMessage, + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestAssistantMessage, + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestToolMessage, + OpenAIModelNonStreamingParametersParametersMessageOpenAIChatCompletionRequestFunctionMessage, +] + + +class OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatText(BaseModel): + type: Literal["text"] + + +class OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonObject(BaseModel): + type: Literal["json_object"] + + +class OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonSchemaJsonSchema(BaseModel): + name: str + + strict: Optional[bool] = None + + description: Optional[str] = None + + schema_: Optional[Dict[str, Optional[object]]] = FieldInfo(alias="schema", default=None) + + +class OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonSchema(BaseModel): + json_schema: OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonSchemaJsonSchema + + type: Literal["json_schema"] + + +OpenAIModelNonStreamingParametersParametersResponseFormat: TypeAlias = Annotated[ + Union[ + OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatText, + OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonObject, + OpenAIModelNonStreamingParametersParametersResponseFormatOpenAIResponseFormatJsonSchema, + ], + PropertyInfo(discriminator="type"), +] + + +class OpenAIModelNonStreamingParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoiceFunction(BaseModel): + name: str + + +class OpenAIModelNonStreamingParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoice(BaseModel): + function: OpenAIModelNonStreamingParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoiceFunction + + type: Literal["function"] + + +OpenAIModelNonStreamingParametersParametersToolChoice: TypeAlias = Union[ + Literal["none", "auto", "required"], + OpenAIModelNonStreamingParametersParametersToolChoiceOpenAIChatCompletionNamedToolChoice, +] + + +class OpenAIModelNonStreamingParametersParametersToolFunction(BaseModel): + name: str + + description: Optional[str] = None + + parameters: Optional[Dict[str, Optional[object]]] = None + + +class OpenAIModelNonStreamingParametersParametersTool(BaseModel): + function: OpenAIModelNonStreamingParametersParametersToolFunction + + type: Literal["function"] + + +class OpenAIModelNonStreamingParametersParameters(BaseModel): + messages: List[OpenAIModelNonStreamingParametersParametersMessage] + + model: str + + response_format: OpenAIModelNonStreamingParametersParametersResponseFormat + + stream: Literal[False] + + frequency_penalty: Optional[float] = None + + logit_bias: Optional[Dict[str, int]] = None + + logprobs: Optional[bool] = None + + max_tokens: Optional[int] = None + + n: Optional[int] = None + + parallel_tool_calls: Optional[bool] = None + + presence_penalty: Optional[float] = None + + seed: Optional[int] = None + + stop: Union[str, List[str], None] = None + + temperature: Optional[float] = None + + tool_choice: Optional[OpenAIModelNonStreamingParametersParametersToolChoice] = None + + tools: Optional[List[OpenAIModelNonStreamingParametersParametersTool]] = None + + top_logprobs: Optional[int] = None + + top_p: Optional[float] = None + + user: Optional[str] = None + + +class OpenAIModelNonStreamingParameters(BaseModel): + name: str + + parameters: OpenAIModelNonStreamingParametersParameters + + provider: Literal["openai"] + + +PromptGetParametersResponse: TypeAlias = Union[AnthropicModelNonStreamingParameters, OpenAIModelNonStreamingParameters] diff --git a/src/prompt_foundry_python_sdk/types/prompt_list_response.py b/src/prompt_foundry_python_sdk/types/prompt_list_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/prompt_update_params.py b/src/prompt_foundry_python_sdk/types/prompt_update_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/tool.py b/src/prompt_foundry_python_sdk/types/tool.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/tool_create_params.py b/src/prompt_foundry_python_sdk/types/tool_create_params.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/tool_delete_response.py b/src/prompt_foundry_python_sdk/types/tool_delete_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/tool_list_response.py b/src/prompt_foundry_python_sdk/types/tool_list_response.py old mode 100755 new mode 100644 diff --git a/src/prompt_foundry_python_sdk/types/tool_update_params.py b/src/prompt_foundry_python_sdk/types/tool_update_params.py old mode 100755 new mode 100644 diff --git a/tests/__init__.py b/tests/__init__.py old mode 100755 new mode 100644 diff --git a/tests/api_resources/__init__.py b/tests/api_resources/__init__.py old mode 100755 new mode 100644 diff --git a/tests/api_resources/test_completion.py b/tests/api_resources/test_completion.py old mode 100755 new mode 100644 diff --git a/tests/api_resources/test_evaluation_assertions.py b/tests/api_resources/test_evaluation_assertions.py old mode 100755 new mode 100644 diff --git a/tests/api_resources/test_evaluations.py b/tests/api_resources/test_evaluations.py old mode 100755 new mode 100644 diff --git a/tests/api_resources/test_prompts.py b/tests/api_resources/test_prompts.py old mode 100755 new mode 100644 index ffd64fb..fd27b0f --- a/tests/api_resources/test_prompts.py +++ b/tests/api_resources/test_prompts.py @@ -10,10 +10,10 @@ from tests.utils import assert_matches_type from prompt_foundry_python_sdk import PromptFoundry, AsyncPromptFoundry from prompt_foundry_python_sdk.types import ( - Parameters, PromptListResponse, PromptConfiguration, PromptDeleteResponse, + PromptGetParametersResponse, ) base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") @@ -705,7 +705,7 @@ def test_method_get_parameters(self, client: PromptFoundry) -> None: prompt = client.prompts.get_parameters( id="1212121", ) - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) @parametrize def test_method_get_parameters_with_all_params(self, client: PromptFoundry) -> None: @@ -820,7 +820,7 @@ def test_method_get_parameters_with_all_params(self, client: PromptFoundry) -> N user="user", variables={"foo": "string"}, ) - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) @parametrize def test_raw_response_get_parameters(self, client: PromptFoundry) -> None: @@ -831,7 +831,7 @@ def test_raw_response_get_parameters(self, client: PromptFoundry) -> None: assert response.is_closed is True assert response.http_request.headers.get("X-Stainless-Lang") == "python" prompt = response.parse() - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) @parametrize def test_streaming_response_get_parameters(self, client: PromptFoundry) -> None: @@ -842,7 +842,7 @@ def test_streaming_response_get_parameters(self, client: PromptFoundry) -> None: assert response.http_request.headers.get("X-Stainless-Lang") == "python" prompt = response.parse() - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) assert cast(Any, response.is_closed) is True @@ -1540,7 +1540,7 @@ async def test_method_get_parameters(self, async_client: AsyncPromptFoundry) -> prompt = await async_client.prompts.get_parameters( id="1212121", ) - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) @parametrize async def test_method_get_parameters_with_all_params(self, async_client: AsyncPromptFoundry) -> None: @@ -1655,7 +1655,7 @@ async def test_method_get_parameters_with_all_params(self, async_client: AsyncPr user="user", variables={"foo": "string"}, ) - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) @parametrize async def test_raw_response_get_parameters(self, async_client: AsyncPromptFoundry) -> None: @@ -1666,7 +1666,7 @@ async def test_raw_response_get_parameters(self, async_client: AsyncPromptFoundr assert response.is_closed is True assert response.http_request.headers.get("X-Stainless-Lang") == "python" prompt = await response.parse() - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) @parametrize async def test_streaming_response_get_parameters(self, async_client: AsyncPromptFoundry) -> None: @@ -1677,7 +1677,7 @@ async def test_streaming_response_get_parameters(self, async_client: AsyncPrompt assert response.http_request.headers.get("X-Stainless-Lang") == "python" prompt = await response.parse() - assert_matches_type(Parameters, prompt, path=["response"]) + assert_matches_type(PromptGetParametersResponse, prompt, path=["response"]) assert cast(Any, response.is_closed) is True diff --git a/tests/api_resources/test_tools.py b/tests/api_resources/test_tools.py old mode 100755 new mode 100644 diff --git a/tests/conftest.py b/tests/conftest.py old mode 100755 new mode 100644 diff --git a/tests/sample_file.txt b/tests/sample_file.txt old mode 100755 new mode 100644 diff --git a/tests/test_client.py b/tests/test_client.py old mode 100755 new mode 100644 diff --git a/tests/test_deepcopy.py b/tests/test_deepcopy.py old mode 100755 new mode 100644 diff --git a/tests/test_extract_files.py b/tests/test_extract_files.py old mode 100755 new mode 100644 diff --git a/tests/test_files.py b/tests/test_files.py old mode 100755 new mode 100644 diff --git a/tests/test_models.py b/tests/test_models.py old mode 100755 new mode 100644 diff --git a/tests/test_qs.py b/tests/test_qs.py old mode 100755 new mode 100644 diff --git a/tests/test_required_args.py b/tests/test_required_args.py old mode 100755 new mode 100644 diff --git a/tests/test_response.py b/tests/test_response.py old mode 100755 new mode 100644 diff --git a/tests/test_streaming.py b/tests/test_streaming.py old mode 100755 new mode 100644 diff --git a/tests/test_transform.py b/tests/test_transform.py old mode 100755 new mode 100644 diff --git a/tests/test_utils/test_proxy.py b/tests/test_utils/test_proxy.py old mode 100755 new mode 100644 diff --git a/tests/test_utils/test_typing.py b/tests/test_utils/test_typing.py old mode 100755 new mode 100644 diff --git a/tests/utils.py b/tests/utils.py old mode 100755 new mode 100644