From 176c860099994b43ef21a8041a8a9e35039e45a7 Mon Sep 17 00:00:00 2001 From: "S.P. Mohanty" <spmohanty91@gmail.com> Date: Mon, 18 Mar 2024 05:04:29 +0000 Subject: [PATCH] Update README.md --- README.md | 7 +++---- 1 file changed, 3 insertions(+), 4 deletions(-) diff --git a/README.md b/README.md index f3b36f6..2d0b3ff 100644 --- a/README.md +++ b/README.md @@ -80,6 +80,9 @@ As all tasks are converted into text generation tasks, rule-based parsers will p Since all these metrics range from [0, 1], we calculate the average metric for all tasks within each track (macro-averaged) to determine the overall score for a track and identify track winners. The overall score of Track 5 will be calculated by averaging scores in Tracks 1-4. +Please refer to [local_evaluation.py](local_evaluation.py) for more details on how we will evaluate your submissions. + + # ðŸ—ƒï¸ Submission The challenge would be evaluated as a code competition. Participants must submit their code and essential resources, such as fine-tuned model weights and indices for Retrieval-Augmented Generation (RAG), which will be run on our servers to generate results and then for evaluation. @@ -225,10 +228,6 @@ Please follow the instructions in [docs/submission.md](ocs/submission.md) to mak **Note**: **Remember to accept the Challenge Rules** on the challenge page, and task page before making your first submission. -## Evaluation Metrics & Local Evaluation -Please refer to [local_evaluation.py](local_evaluation.py) for more details on how we will evaluate your submissions. - - ## Hardware and System Configuration We apply a limit on the hardware available to each participant to run their solutions. Specifically, -- GitLab