Replies: 1 comment
-
|
On user experience, I would add: I think these should be the KPIs to measure agent performance:
Some questions we should solve:
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Opening this discussion to start drafting what an evaluation framework for agentic analytics should be.
Problem
When deploying an agentic analytics solution teams want to be sure the agentic solution does not hallucinate when answering questions. Hallucination can be of multiple factors, it can hallucinates tables or columns names, but it does not limit to hallucination there is also a information retrieval problem.
As a an admin deploying nao I want to make sure before deploying to my stakeholders that the agentic loop on a determined set of questions, situations or scenario is able to limit hallucinations and retrieve the right context to answer the problem.
We then want to create a tool that gives the capabilities to users to describe problems inputs and outputs and then be able to run these problems against the agentic loop of multiple models and compare the performances.
Eventually the users can see the impact of adding this or this context on the performance of the output (eg. should I add a preview of the tables or not).
Beta Was this translation helpful? Give feedback.
All reactions