From 896710a755c1cc7f84c794c9f4b3c38ebf13acf1 Mon Sep 17 00:00:00 2001
From: mohanty <mohanty@aicrowd.com>
Date: Fri, 24 May 2024 17:01:47 +0000
Subject: [PATCH] remove redundant comment

---
 models/vanilla_llama3_baseline.py | 4 ----
 1 file changed, 4 deletions(-)

diff --git a/models/vanilla_llama3_baseline.py b/models/vanilla_llama3_baseline.py
index 294f1aa..b3c4bc3 100644
--- a/models/vanilla_llama3_baseline.py
+++ b/models/vanilla_llama3_baseline.py
@@ -116,10 +116,6 @@ class Llama3_8B_ZeroShotModel(ShopBenchBaseModel):
                 seed=AICROWD_RUN_SEED, # Seed for reprodicibility
                 skip_special_tokens=True,  # Whether to skip special tokens in the output.
                 max_tokens=max_new_tokens,  # Maximum number of tokens to generate per output sequence.
-                # Note: We are using 50 max new tokens instead of 75,
-                # because the 75 max token limit is checked using the Llama2 tokenizer.
-                # The Llama3 model instead uses a differet tokenizer with a larger vocabulary
-                # This allows it to represent the same content more efficiently, using fewer tokens.
             ),
             use_tqdm = False
         )
-- 
GitLab