From 14c9edb88cc0cd83cd40aacd9b8f7dc4e32c1354 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Hugues=20de=20Saxc=C3=A9?= Date: Wed, 12 Jun 2024 17:48:58 +0200 Subject: [PATCH 1/3] fix: use promptType and switch llm steps to be run --- literalai/api/generation_helpers.py | 7 ++++++- literalai/api/gql.py | 4 ++-- tests/e2e/test_e2e.py | 8 ++++---- 3 files changed, 12 insertions(+), 7 deletions(-) diff --git a/literalai/api/generation_helpers.py b/literalai/api/generation_helpers.py index 33cb7ee..8952907 100644 --- a/literalai/api/generation_helpers.py +++ b/literalai/api/generation_helpers.py @@ -49,7 +49,12 @@ def create_generation_helper(generation: Union[ChatGeneration, CompletionGenerat variables = {"generation": generation.to_dict()} def process_response(response): - return BaseGeneration.from_dict(response["data"]["createGeneration"]) + generation_as_step = response["data"]["createGeneration"] + deprecated_generation = { + "id": generation_as_step["id"], + "type": generation_as_step["promptType"], + } + return BaseGeneration.from_dict(deprecated_generation) description = "create generation" diff --git a/literalai/api/gql.py b/literalai/api/gql.py index ea0b677..bbfdf64 100644 --- a/literalai/api/gql.py +++ b/literalai/api/gql.py @@ -766,8 +766,8 @@ CREATE_GENERATION = """ mutation CreateGeneration($generation: GenerationPayloadInput!) { createGeneration(generation: $generation) { - id, - type + id + promptType } } """ diff --git a/tests/e2e/test_e2e.py b/tests/e2e/test_e2e.py index c81742b..2a34fc4 100644 --- a/tests/e2e/test_e2e.py +++ b/tests/e2e/test_e2e.py @@ -313,13 +313,13 @@ async def assert_delete(thread_id: str, step_id: str): @async_client.thread def thread_decorated(): - @async_client.step(name="foo", type="llm", tags=["to_score"]) + @async_client.step(name="foo", type="run", tags=["to_score"]) def step_decorated(): t = async_client.get_current_thread() s = async_client.get_current_step() assert s is not None assert s.name == "foo" - assert s.type == "llm" + assert s.type == "run" assert s.tags == ["to_score"] return t.id, s.id @@ -330,13 +330,13 @@ def step_decorated(): @async_client.thread async def a_thread_decorated(): - @async_client.step(name="foo", type="llm") + @async_client.step(name="foo", type="run") async def a_step_decorated(): t = async_client.get_current_thread() s = async_client.get_current_step() assert s is not None assert s.name == "foo" - assert s.type == "llm" + assert s.type == "run" return t.id, s.id return await a_step_decorated() From 205f4bbc64535c09929aaeca338af1b2f11ee367 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Hugues=20de=20Saxc=C3=A9?= Date: Thu, 13 Jun 2024 16:33:47 +0200 Subject: [PATCH 2/3] fix: keep using llm instead of run --- tests/e2e/test_e2e.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/tests/e2e/test_e2e.py b/tests/e2e/test_e2e.py index 2a34fc4..c81742b 100644 --- a/tests/e2e/test_e2e.py +++ b/tests/e2e/test_e2e.py @@ -313,13 +313,13 @@ async def assert_delete(thread_id: str, step_id: str): @async_client.thread def thread_decorated(): - @async_client.step(name="foo", type="run", tags=["to_score"]) + @async_client.step(name="foo", type="llm", tags=["to_score"]) def step_decorated(): t = async_client.get_current_thread() s = async_client.get_current_step() assert s is not None assert s.name == "foo" - assert s.type == "run" + assert s.type == "llm" assert s.tags == ["to_score"] return t.id, s.id @@ -330,13 +330,13 @@ def step_decorated(): @async_client.thread async def a_thread_decorated(): - @async_client.step(name="foo", type="run") + @async_client.step(name="foo", type="llm") async def a_step_decorated(): t = async_client.get_current_thread() s = async_client.get_current_step() assert s is not None assert s.name == "foo" - assert s.type == "run" + assert s.type == "llm" return t.id, s.id return await a_step_decorated() From cbecfaa493b02881ffb550b67422c4b60ad62a8f Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Hugues=20de=20Saxc=C3=A9?= Date: Fri, 14 Jun 2024 11:35:28 +0200 Subject: [PATCH 3/3] fix: revert sdk changes --- literalai/api/generation_helpers.py | 7 +------ literalai/api/gql.py | 2 +- 2 files changed, 2 insertions(+), 7 deletions(-) diff --git a/literalai/api/generation_helpers.py b/literalai/api/generation_helpers.py index 8952907..33cb7ee 100644 --- a/literalai/api/generation_helpers.py +++ b/literalai/api/generation_helpers.py @@ -49,12 +49,7 @@ def create_generation_helper(generation: Union[ChatGeneration, CompletionGenerat variables = {"generation": generation.to_dict()} def process_response(response): - generation_as_step = response["data"]["createGeneration"] - deprecated_generation = { - "id": generation_as_step["id"], - "type": generation_as_step["promptType"], - } - return BaseGeneration.from_dict(deprecated_generation) + return BaseGeneration.from_dict(response["data"]["createGeneration"]) description = "create generation" diff --git a/literalai/api/gql.py b/literalai/api/gql.py index bbfdf64..f65bd08 100644 --- a/literalai/api/gql.py +++ b/literalai/api/gql.py @@ -767,7 +767,7 @@ mutation CreateGeneration($generation: GenerationPayloadInput!) { createGeneration(generation: $generation) { id - promptType + type } } """