Evaluation Metrics for Retrieval should have flexible comparison attributes to allow for consistent evaluation across different chunking strategies #9331
Labels
P2
Medium priority, add to the next sprint if no P1 available
Is your feature request related to a problem? Please describe.
Doing evaluations with a given ground truth set of documents where the user put manual effort into curating it. The ground truth is based on retrieving the correct page from a document (joined id: (file_id, page_number)).
In all Document evaluators, doc.content is used for comparison.
haystack/haystack/components/evaluators/document_mrr.py
Line 73 in e3f9da1
Describe the solution you'd like
I would like to define how the comparison should be done via a comparison field:
The text was updated successfully, but these errors were encountered: