Skip to content

Commit 01c1e4e

Browse files
authored
Fix a link that need to be escaped in markdown (#721)
1 parent ea21100 commit 01c1e4e

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

llms-full.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -71538,7 +71538,7 @@ Based on who is maintaining the eval set data, there are two routes:
7153871538
The evaluation criteria define how the agent's performance is measured against the evalset. The following metrics are supported:
7153971539

7154071540
- `tool_trajectory_avg_score`: This metric compares the agent's actual tool usage during the evaluation against the expected tool usage defined in the `expected_tool_use` field. Each matching tool usage step receives a score of 1, while a mismatch receives a score of 0. The final score is the average of these matches, representing the accuracy of the tool usage trajectory.
71541-
- `response_match_score`: This metric compares the agent's final natural language response to the expected final response, stored in the `reference` field. We use the [ROUGE](https://en.wikipedia.org/wiki/ROUGE_(metric)) metric to calculate the similarity between the two responses.
71541+
- `response_match_score`: This metric compares the agent's final natural language response to the expected final response, stored in the `reference` field. We use the [ROUGE](https://en.wikipedia.org/wiki/ROUGE_\(metric\)) metric to calculate the similarity between the two responses.
7154271542

7154371543
If no evaluation criteria are provided, the following default configuration is used:
7154471544

0 commit comments

Comments
 (0)