Skip to content

Commit f7bac39

Browse files
mongodbenBen Perlmutter
andauthored
(EAI-895): Remove AnswerRelevancy tracing metric (#686)
remove AnswerRelevancy Co-authored-by: Ben Perlmutter <mongodben@mongodb.com>
1 parent 7b09e21 commit f7bac39

File tree

2 files changed

+3
-28
lines changed

2 files changed

+3
-28
lines changed

packages/chatbot-server-mongodb-public/src/tracing/getLlmAsAJudgeScores.test.ts

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,6 @@ jest.mock("autoevals", () => ({
88
ContextRelevancy: jest.fn().mockResolvedValue({
99
score: 0.8,
1010
}),
11-
AnswerRelevancy: jest.fn().mockResolvedValue({
12-
score: 0.8,
13-
}),
1411
}));
1512

1613
afterEach(() => {
@@ -100,7 +97,6 @@ describe("getLlmAsAJudgeScores", () => {
10097
const scores = await getLlmAsAJudgeScores(fakeBaseConfig, willJudge);
10198
expect(scores).toEqual({
10299
ContextRelevancy: 0.8,
103-
AnswerRelevancy: 0.8,
104100
Faithfulness: 0.8,
105101
});
106102
});

packages/chatbot-server-mongodb-public/src/tracing/getLlmAsAJudgeScores.ts

Lines changed: 3 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -31,13 +31,7 @@ const makeEvaluateWithLlmAsAJudge = (
3131
openAiConfig: LlmAsAJudge["openAiConfig"]
3232
) =>
3333
wrapTraced(
34-
async function ({
35-
input,
36-
output,
37-
context,
38-
judgeEmbeddingModel,
39-
judgeModel,
40-
}: ScorerArgs) {
34+
async function ({ input, output, context, judgeModel }: ScorerArgs) {
4135
return Promise.all([
4236
traced(
4337
async () =>
@@ -52,20 +46,6 @@ const makeEvaluateWithLlmAsAJudge = (
5246
name: "Faithfulness",
5347
}
5448
),
55-
traced(
56-
async () =>
57-
AnswerRelevancy({
58-
input,
59-
output,
60-
context,
61-
model: judgeModel,
62-
embeddingModel: judgeEmbeddingModel,
63-
...openAiConfig,
64-
}),
65-
{
66-
name: "AnswerRelevancy",
67-
}
68-
),
6949
traced(
7050
async () =>
7151
ContextRelevancy({
@@ -113,19 +93,18 @@ export async function getLlmAsAJudgeScores(
11393

11494
const evaluateWithLlmAsAJudge = makeEvaluateWithLlmAsAJudge(openAiConfig);
11595

116-
const [faithfulness, answerRelevancy, contextRelevancy] = context
96+
const [faithfulness, contextRelevancy] = context
11797
? await evaluateWithLlmAsAJudge({
11898
input,
11999
output,
120100
context,
121101
judgeModel,
122102
judgeEmbeddingModel,
123103
})
124-
: [nullScore, nullScore, nullScore];
104+
: [nullScore, nullScore];
125105

126106
return {
127107
Faithfulness: faithfulness.score,
128-
AnswerRelevancy: answerRelevancy.score,
129108
ContextRelevancy: contextRelevancy.score,
130109
};
131110
}

0 commit comments

Comments
 (0)