Skip to content

Conversation

@kausmeows
Copy link
Contributor

@kausmeows kausmeows commented May 26, 2025

Changelog

New Features:

  • Claude File Upload: we can now upload a file to Anthropic directly and then use it as an input to an agent.
  • Claude 4 code execution tool: allows Claude to execute Python code in a secure, sandboxed environment.
  • Prompt caching with Anthropic models: allows resuming from specific prefixes in your prompts. This approach significantly reduces processing time and costs for repetitive tasks or prompts with consistent elements.
  • Vercel v0 Model: Added support for new Vercel v0 models and cookbook examples.
  • Qdrant Hybrid Search support
  • Markdown Knowledge Base: Added native support for Markdown-based knowledge bases.
  • AI/ML API platform integration: introduces integration with AI/ML API, a platform providing AI/ML models. AI/ML API provides 300+ AI models including Deepseek, Gemini, ChatGPT. The models run at enterprise-grade rate limits and uptimes
  • Update pydantic and dataclass in function handling: Now supports class as input to a function

Improvements:

  • Timeout handling for API calls in ExaTools class:
    • Timeout functionality to Exa API calls to prevent indefinite hanging of search operations. The implementation uses Python's concurrent.futures module to enforce timeouts on all Exa API operations (search, get contents, find similar, and answer generation).
    • This change addresses the issue where Exa search functions would hang indefinitely, causing potential service disruptions and resource leaks.
  • Fetch messages from last n sessions:
    • A tool for the agent, something like get_previous_session_messages(number_of_sessions: int) that returns a list of messages that the agent can analyse
    • Switch on with search_previous_sessions_history
  • Redis Expiration: Added expire key to set TTL on Redis keys.
  •  Add anthropic cache write to agent session metrics: adds cache_creation_input_tokens to agent session metrics, to allow for tracking Anthropic cache write statistics

Bug Fixes:

  • HF custom embedder:
  • Add role_map for OpenAIChat: This allows certain models that don’t adhere to OpenAI’s role mapping to be used vir OpenAILike.
  • Use content hash as id in upsert: in pgvector use reproducible content_hash in upsert as id
  • Insert in vector db passes only last chunk meta_data: insert in vector db passes only last chunk meta_data. issue link- https://discord.com/channels/965734768803192842/1219054452221153463/1376631140047130649
  • Remove argument sanitization: a safer way to do this that won't break arguments that shouldn't be sanitized
  • Handle async tools when running async agents on playground: Fixes a regression where using Agents with async tools (e.g. MCP tools) was breaking in the Playground.

@kausmeows kausmeows requested a review from a team as a code owner May 26, 2025 13:22
@kausmeows kausmeows changed the title update chore: Release 1.5.5 May 26, 2025
@dirkbrnd dirkbrnd merged commit 95bad1a into main May 27, 2025
11 checks passed
@dirkbrnd dirkbrnd deleted the release-1.5.5 branch May 27, 2025 10:21
Mustafa-Esoofally pushed a commit that referenced this pull request Jun 3, 2025
# Changelog

## New Features:

- **Claude File Upload:** we can now upload a file to Anthropic directly
and then use it as an input to an agent.
- **Claude 4 code execution tool:** allows Claude to execute Python code
in a secure, sandboxed environment.
- **Prompt caching with Anthropic models:** allows resuming from
specific prefixes in your prompts. This approach significantly reduces
processing time and costs for repetitive tasks or prompts with
consistent elements.
- **Vercel v0 Model:** Added support for new Vercel v0 models and
cookbook examples.
- **Qdrant Hybrid Search support**
- **Markdown Knowledge Base**: Added native support for Markdown-based
knowledge bases.
- **AI/ML API platform integration**: introduces integration with [AI/ML
API](https://aimlapi.com/models/?utm_source=agno&utm_medium=github&utm_campaign=integration),
a platform providing AI/ML models. AI/ML API provides 300+ AI models
including Deepseek, Gemini, ChatGPT. The models run at enterprise-grade
rate limits and uptimes
- **Update pydantic and dataclass in function handling**: Now supports
class as input to a function

## Improvements:

- **Timeout handling for API calls in ExaTools class:**
- Timeout functionality to Exa API calls to prevent indefinite hanging
of search operations. The implementation uses
Python's `concurrent.futures` module to enforce timeouts on all Exa API
operations (search, get contents, find similar, and answer generation).
- This change addresses the issue where Exa search functions would hang
indefinitely, causing potential service disruptions and resource leaks.
- **Fetch messages from last n sessions:**
- A tool for the agent, something
like `get_previous_session_messages(number_of_sessions: int)` that
returns a list of messages that the agent can analyse
    - Switch on with `search_previous_sessions_history`
- **Redis Expiration**: Added expire key to set TTL on Redis keys.
-  **Add anthropic cache write to agent session metrics**: adds
cache_creation_input_tokens to agent session metrics, to allow for
tracking Anthropic cache write statistics

## Bug Fixes:

- **HF custom embedder:**
- It turns out that HF has changed some things on their API and they've
deprecated .post on their
InferenceClient()- https://discuss.huggingface.co/t/getting-error-attributeerror-inferenceclient-object-has-no-attribute-post/156682
- Also we can no longer use- `id: str =
"jinaai/jina-embeddings-v2-base-code"` as default, because these models
are no longer provided by the `HF Inference API`. Changed the default
to- `id: str = "intfloat/multilingual-e5-large"`
- **Add `role_map` for `OpenAIChat`:** This allows certain models that
don’t adhere to OpenAI’s role mapping to be used vir `OpenAILike`.
- **Use content hash as id in upsert: in pgvector** use reproducible
content_hash in upsert as id
- **Insert in vector db passes only last chunk meta_data:** insert in
vector db passes only last chunk meta_data. issue link-
https://discord.com/channels/965734768803192842/1219054452221153463/1376631140047130649
- **Remove argument sanitization**: a safer way to do this that won't
break arguments that shouldn't be sanitized
- **Handle async tools when running async agents on playground**: Fixes
a regression where using Agents with async tools (e.g. MCP tools) was
breaking in the Playground.

---------

Co-authored-by: Dirk Brand <dirkbrnd@gmail.com>
Mustafa-Esoofally pushed a commit that referenced this pull request Jun 4, 2025
# Changelog

## New Features:

- **Claude File Upload:** we can now upload a file to Anthropic directly
and then use it as an input to an agent.
- **Claude 4 code execution tool:** allows Claude to execute Python code
in a secure, sandboxed environment.
- **Prompt caching with Anthropic models:** allows resuming from
specific prefixes in your prompts. This approach significantly reduces
processing time and costs for repetitive tasks or prompts with
consistent elements.
- **Vercel v0 Model:** Added support for new Vercel v0 models and
cookbook examples.
- **Qdrant Hybrid Search support**
- **Markdown Knowledge Base**: Added native support for Markdown-based
knowledge bases.
- **AI/ML API platform integration**: introduces integration with [AI/ML
API](https://aimlapi.com/models/?utm_source=agno&utm_medium=github&utm_campaign=integration),
a platform providing AI/ML models. AI/ML API provides 300+ AI models
including Deepseek, Gemini, ChatGPT. The models run at enterprise-grade
rate limits and uptimes
- **Update pydantic and dataclass in function handling**: Now supports
class as input to a function

## Improvements:

- **Timeout handling for API calls in ExaTools class:**
- Timeout functionality to Exa API calls to prevent indefinite hanging
of search operations. The implementation uses
Python's `concurrent.futures` module to enforce timeouts on all Exa API
operations (search, get contents, find similar, and answer generation).
- This change addresses the issue where Exa search functions would hang
indefinitely, causing potential service disruptions and resource leaks.
- **Fetch messages from last n sessions:**
- A tool for the agent, something
like `get_previous_session_messages(number_of_sessions: int)` that
returns a list of messages that the agent can analyse
    - Switch on with `search_previous_sessions_history`
- **Redis Expiration**: Added expire key to set TTL on Redis keys.
-  **Add anthropic cache write to agent session metrics**: adds
cache_creation_input_tokens to agent session metrics, to allow for
tracking Anthropic cache write statistics

## Bug Fixes:

- **HF custom embedder:**
- It turns out that HF has changed some things on their API and they've
deprecated .post on their
InferenceClient()- https://discuss.huggingface.co/t/getting-error-attributeerror-inferenceclient-object-has-no-attribute-post/156682
- Also we can no longer use- `id: str =
"jinaai/jina-embeddings-v2-base-code"` as default, because these models
are no longer provided by the `HF Inference API`. Changed the default
to- `id: str = "intfloat/multilingual-e5-large"`
- **Add `role_map` for `OpenAIChat`:** This allows certain models that
don’t adhere to OpenAI’s role mapping to be used vir `OpenAILike`.
- **Use content hash as id in upsert: in pgvector** use reproducible
content_hash in upsert as id
- **Insert in vector db passes only last chunk meta_data:** insert in
vector db passes only last chunk meta_data. issue link-
https://discord.com/channels/965734768803192842/1219054452221153463/1376631140047130649
- **Remove argument sanitization**: a safer way to do this that won't
break arguments that shouldn't be sanitized
- **Handle async tools when running async agents on playground**: Fixes
a regression where using Agents with async tools (e.g. MCP tools) was
breaking in the Playground.

---------

Co-authored-by: Dirk Brand <dirkbrnd@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants