Skip to content

Commit 154fec5

Browse files
authored
Clean up dataclasses reprs (#1812)
1 parent ed2a8ad commit 154fec5

File tree

17 files changed

+175
-355
lines changed

17 files changed

+175
-355
lines changed

docs/agents.md

Lines changed: 14 additions & 51 deletions
Original file line numberDiff line numberDiff line change
@@ -135,30 +135,21 @@ async def main():
135135
UserPromptPart(
136136
content='What is the capital of France?',
137137
timestamp=datetime.datetime(...),
138-
part_kind='user-prompt',
139138
)
140-
],
141-
instructions=None,
142-
kind='request',
139+
]
143140
)
144141
),
145142
CallToolsNode(
146143
model_response=ModelResponse(
147-
parts=[TextPart(content='Paris', part_kind='text')],
144+
parts=[TextPart(content='Paris')],
148145
usage=Usage(
149-
requests=1,
150-
request_tokens=56,
151-
response_tokens=1,
152-
total_tokens=57,
153-
details=None,
146+
requests=1, request_tokens=56, response_tokens=1, total_tokens=57
154147
),
155148
model_name='gpt-4o',
156149
timestamp=datetime.datetime(...),
157-
kind='response',
158-
vendor_id=None,
159150
)
160151
),
161-
End(data=FinalResult(output='Paris', tool_name=None, tool_call_id=None)),
152+
End(data=FinalResult(output='Paris')),
162153
]
163154
"""
164155
print(agent_run.result.output)
@@ -207,30 +198,24 @@ async def main():
207198
UserPromptPart(
208199
content='What is the capital of France?',
209200
timestamp=datetime.datetime(...),
210-
part_kind='user-prompt',
211201
)
212-
],
213-
instructions=None,
214-
kind='request',
202+
]
215203
)
216204
),
217205
CallToolsNode(
218206
model_response=ModelResponse(
219-
parts=[TextPart(content='Paris', part_kind='text')],
207+
parts=[TextPart(content='Paris')],
220208
usage=Usage(
221209
requests=1,
222210
request_tokens=56,
223211
response_tokens=1,
224212
total_tokens=57,
225-
details=None,
226213
),
227214
model_name='gpt-4o',
228215
timestamp=datetime.datetime(...),
229-
kind='response',
230-
vendor_id=None,
231216
)
232217
),
233-
End(data=FinalResult(output='Paris', tool_name=None, tool_call_id=None)),
218+
End(data=FinalResult(output='Paris')),
234219
]
235220
"""
236221
```
@@ -370,7 +355,7 @@ if __name__ == '__main__':
370355
[
371356
'=== UserPromptNode: What will the weather be like in Paris on Tuesday? ===',
372357
'=== ModelRequestNode: streaming partial request tokens ===',
373-
"[Request] Starting part 0: ToolCallPart(tool_name='weather_forecast', args=None, tool_call_id='0001', part_kind='tool-call')",
358+
"[Request] Starting part 0: ToolCallPart(tool_name='weather_forecast', tool_call_id='0001')",
374359
'[Request] Part 0 args_delta={"location":"Pa',
375360
'[Request] Part 0 args_delta=ris","forecast_',
376361
'[Request] Part 0 args_delta=date":"2030-01-',
@@ -379,7 +364,7 @@ if __name__ == '__main__':
379364
'[Tools] The LLM calls tool=\'weather_forecast\' with args={"location":"Paris","forecast_date":"2030-01-01"} (tool_call_id=\'0001\')',
380365
"[Tools] Tool call '0001' returned => The forecast in Paris on 2030-01-01 is 24°C and sunny.",
381366
'=== ModelRequestNode: streaming partial request tokens ===',
382-
"[Request] Starting part 0: TextPart(content='It will be ', part_kind='text')",
367+
"[Request] Starting part 0: TextPart(content='It will be ')",
383368
'[Result] The model produced a final output (tool_name=None)',
384369
"[Request] Part 0 text delta: 'warm and sunny '",
385370
"[Request] Part 0 text delta: 'in Paris on '",
@@ -417,9 +402,7 @@ result_sync = agent.run_sync(
417402
print(result_sync.output)
418403
#> Rome
419404
print(result_sync.usage())
420-
"""
421-
Usage(requests=1, request_tokens=62, response_tokens=1, total_tokens=63, details=None)
422-
"""
405+
#> Usage(requests=1, request_tokens=62, response_tokens=1, total_tokens=63)
423406

424407
try:
425408
result_sync = agent.run_sync(
@@ -831,32 +814,22 @@ with capture_run_messages() as messages: # (2)!
831814
UserPromptPart(
832815
content='Please get me the volume of a box with size 6.',
833816
timestamp=datetime.datetime(...),
834-
part_kind='user-prompt',
835817
)
836-
],
837-
instructions=None,
838-
kind='request',
818+
]
839819
),
840820
ModelResponse(
841821
parts=[
842822
ToolCallPart(
843823
tool_name='calc_volume',
844824
args={'size': 6},
845825
tool_call_id='pyd_ai_tool_call_id',
846-
part_kind='tool-call',
847826
)
848827
],
849828
usage=Usage(
850-
requests=1,
851-
request_tokens=62,
852-
response_tokens=4,
853-
total_tokens=66,
854-
details=None,
829+
requests=1, request_tokens=62, response_tokens=4, total_tokens=66
855830
),
856831
model_name='gpt-4o',
857832
timestamp=datetime.datetime(...),
858-
kind='response',
859-
vendor_id=None,
860833
),
861834
ModelRequest(
862835
parts=[
@@ -865,32 +838,22 @@ with capture_run_messages() as messages: # (2)!
865838
tool_name='calc_volume',
866839
tool_call_id='pyd_ai_tool_call_id',
867840
timestamp=datetime.datetime(...),
868-
part_kind='retry-prompt',
869841
)
870-
],
871-
instructions=None,
872-
kind='request',
842+
]
873843
),
874844
ModelResponse(
875845
parts=[
876846
ToolCallPart(
877847
tool_name='calc_volume',
878848
args={'size': 6},
879849
tool_call_id='pyd_ai_tool_call_id',
880-
part_kind='tool-call',
881850
)
882851
],
883852
usage=Usage(
884-
requests=1,
885-
request_tokens=72,
886-
response_tokens=8,
887-
total_tokens=80,
888-
details=None,
853+
requests=1, request_tokens=72, response_tokens=8, total_tokens=80
889854
),
890855
model_name='gpt-4o',
891856
timestamp=datetime.datetime(...),
892-
kind='response',
893-
vendor_id=None,
894857
),
895858
]
896859
"""

docs/api/models/function.md

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -28,11 +28,8 @@ async def model_function(
2828
UserPromptPart(
2929
content='Testing my agent...',
3030
timestamp=datetime.datetime(...),
31-
part_kind='user-prompt',
3231
)
33-
],
34-
instructions=None,
35-
kind='request',
32+
]
3633
)
3734
]
3835
"""

docs/direct.md

Lines changed: 2 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -27,9 +27,7 @@ model_response = model_request_sync(
2727
print(model_response.parts[0].content)
2828
#> Paris
2929
print(model_response.usage)
30-
"""
31-
Usage(requests=1, request_tokens=56, response_tokens=1, total_tokens=57, details=None)
32-
"""
30+
#> Usage(requests=1, request_tokens=56, response_tokens=1, total_tokens=57)
3331
```
3432

3533
_(This example is complete, it can be run "as is")_
@@ -82,20 +80,11 @@ async def main():
8280
tool_name='divide',
8381
args={'numerator': '123', 'denominator': '456'},
8482
tool_call_id='pyd_ai_2e0e396768a14fe482df90a29a78dc7b',
85-
part_kind='tool-call',
8683
)
8784
],
88-
usage=Usage(
89-
requests=1,
90-
request_tokens=55,
91-
response_tokens=7,
92-
total_tokens=62,
93-
details=None,
94-
),
85+
usage=Usage(requests=1, request_tokens=55, response_tokens=7, total_tokens=62),
9586
model_name='gpt-4.1-nano',
9687
timestamp=datetime.datetime(...),
97-
kind='response',
98-
vendor_id=None,
9988
)
10089
"""
10190
```

0 commit comments

Comments
 (0)