v2.2.73
What's Changed
- feat: add context CLI command with --url and --goal parameters by @MervinPraison in #1026
- fix: implement real-time streaming for Agent.start() method by @MervinPraison in #1027
- fix: implement real-time streaming for Agent.start() method by @MervinPraison in #1028
- fix: enhance Gemini streaming robustness with graceful JSON parsing error handling by @MervinPraison in #1029
- fix: bypass display_generation for OpenAI streaming to enable raw chunk output by @MervinPraison in #1030
- fix: eliminate streaming pause caused by telemetry tracking by @github-actions[bot] in #1032
- Fix litellm deprecation warnings for issue #1033 by @github-actions[bot] in #1034
- fix: correct tool call argument parsing in streaming mode by @MervinPraison in #1037
- feat: Add comprehensive performance monitoring system by @MervinPraison in #1038
- Fix: Comprehensive LiteLLM deprecation warning suppression by @MervinPraison in #1039
- PR #1038: Monitoring examples by @MervinPraison in #1040
- PR #1039: Logging by @MervinPraison in #1041
- fix: ensure display_generating is called when verbose=True regardless of streaming mode by @MervinPraison in #1042
- fix: enhance LiteLLM streaming error handling for JSON parsing errors (Issue #1043) by @MervinPraison in #1044
- fix: correct display_generating logic to only show when stream=False AND verbose=True by @MervinPraison in #1045
- fix: implement proper streaming fallback logic for JSON parsing errors by @MervinPraison in #1046
- fix: resolve display_generating issue by ensuring stream parameter is correctly passed by @MervinPraison in #1047
- PR #1046: Changes from Claude by @MervinPraison in #1048
- Fix: Add display_generating support for OpenAI non-streaming mode by @MervinPraison in #1049
- fix: Remove display_generating when stream=false to prevent streaming-like behavior by @MervinPraison in #1050
New Contributors
- @github-actions[bot] made their first contribution in #1032
Full Changelog: v2.2.72...v2.2.73