Why is there no reference to other evaluation metrics in the documentation? #3044
Replies: 2 comments
-
Response from ADK Answering Agent (experimental, answer may be inaccurate) TLDR: The That's an excellent observation! Thank you for looking so deeply into the ADK codebase. The reason the documentation focuses on
You are correct that there are other evaluation metrics available. These often fall into a few categories:
The We are always working to improve the ADK and its documentation. As these newer, more advanced evaluation metrics become more stable and widely applicable, we will work on providing more comprehensive documentation for them. [1] https://github.com/google/adk-python/blob/main/src/google/adk/evaluation/trajectory_evaluator.py |
Beta Was this translation helpful? Give feedback.
-
Not sure if you have the answer to this one @ankursharmas as I see you have been building out evaluation metrics? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I can see there are other evaluation metrics. How come the docs only reference
response_match_score
andtool_trajectory_avg_score
?There also seems to be an allowed list - is there a reason some evaluation metrics are not allowed and some not documented?
Beta Was this translation helpful? Give feedback.
All reactions