From 896710a755c1cc7f84c794c9f4b3c38ebf13acf1 Mon Sep 17 00:00:00 2001 From: mohanty <mohanty@aicrowd.com> Date: Fri, 24 May 2024 17:01:47 +0000 Subject: [PATCH] remove redundant comment --- models/vanilla_llama3_baseline.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/models/vanilla_llama3_baseline.py b/models/vanilla_llama3_baseline.py index 294f1aa..b3c4bc3 100644 --- a/models/vanilla_llama3_baseline.py +++ b/models/vanilla_llama3_baseline.py @@ -116,10 +116,6 @@ class Llama3_8B_ZeroShotModel(ShopBenchBaseModel): seed=AICROWD_RUN_SEED, # Seed for reprodicibility skip_special_tokens=True, # Whether to skip special tokens in the output. max_tokens=max_new_tokens, # Maximum number of tokens to generate per output sequence. - # Note: We are using 50 max new tokens instead of 75, - # because the 75 max token limit is checked using the Llama2 tokenizer. - # The Llama3 model instead uses a differet tokenizer with a larger vocabulary - # This allows it to represent the same content more efficiently, using fewer tokens. ), use_tqdm = False ) -- GitLab