Skip to content

Commit 0c7140c

Browse files
VasiliyRadElva
andauthored
fix(broken_link): Broken link to agenteval_cq_math.ipynb (#2677)
GitHub notebook URL returns HTTP 404. Bug: file-7a8d1ff7 Fixed by Elva. Co-authored-by: Elva <elva@ag2.ai>
1 parent faa187b commit 0c7140c

1 file changed

Lines changed: 2 additions & 2 deletions

File tree

  • website/docs/_blogs/2023-11-20-AgentEval

website/docs/_blogs/2023-11-20-AgentEval/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ tags: [Evaluation]
1414
**TL;DR:**
1515
* As a developer of an LLM-powered application, how can you assess the utility it brings to end users while helping them with their tasks?
1616
* To shed light on the question above, we introduce `AgentEval` — the first version of the framework to assess the utility of any LLM-powered application crafted to assist users in specific tasks. AgentEval aims to simplify the evaluation process by automatically proposing a set of criteria tailored to the unique purpose of your application. This allows for a comprehensive assessment, quantifying the utility of your application against the suggested criteria.
17-
* We demonstrate how `AgentEval` work using [math problems dataset](https://docs.ag2.ai/latest/docs/blog/2023/06/28/MathChat) as an example in the [following notebook](https://github.com/ag2ai/ag2/blob/main/notebook/agenteval_cq_math.ipynb). Any feedback would be useful for future development. Please contact us on our [Discord](https://discord.gg/pAbnFJrkgZ).
17+
* We demonstrate how `AgentEval` work using [math problems dataset](https://docs.ag2.ai/latest/docs/blog/2023/06/28/MathChat) as an example in the [following notebook](https://github.com/ag2ai/ag2/blob/main/notebook/agentchat_agenteval_cq_math.ipynb). Any feedback would be useful for future development. Please contact us on our [Discord](https://discord.gg/pAbnFJrkgZ).
1818

1919

2020

@@ -56,7 +56,7 @@ critic = autogen.AssistantAgent(
5656
)
5757
```
5858

59-
Next, the critic is given successful and failed examples of the task execution; then, it is able to return a list of criteria (Fig. 1). For reference, use the [following notebook](https://github.com/ag2ai/ag2/blob/main/notebook/agenteval_cq_math.ipynb).
59+
Next, the critic is given successful and failed examples of the task execution; then, it is able to return a list of criteria (Fig. 1). For reference, use the [following notebook](https://github.com/ag2ai/ag2/blob/main/notebook/agentchat_agenteval_cq_math.ipynb).
6060

6161
* The goal of `QuantifierAgent` is to quantify each of the suggested criteria (Fig. 1), providing us with an idea of the utility of this system for the given task. Here is an example of how it can be defined:
6262

0 commit comments

Comments
 (0)