diff --git a/models/vanilla_llama3_baseline.py b/models/vanilla_llama3_baseline.py
index 294f1aa20cc6d9b808e4a0d41899b6c66eaae153..b3c4bc388ec1e8fe35b7c49c74d209d4fef5f380 100644
--- a/models/vanilla_llama3_baseline.py
+++ b/models/vanilla_llama3_baseline.py
@@ -116,10 +116,6 @@ class Llama3_8B_ZeroShotModel(ShopBenchBaseModel):
                 seed=AICROWD_RUN_SEED, # Seed for reprodicibility
                 skip_special_tokens=True,  # Whether to skip special tokens in the output.
                 max_tokens=max_new_tokens,  # Maximum number of tokens to generate per output sequence.
-                # Note: We are using 50 max new tokens instead of 75,
-                # because the 75 max token limit is checked using the Llama2 tokenizer.
-                # The Llama3 model instead uses a differet tokenizer with a larger vocabulary
-                # This allows it to represent the same content more efficiently, using fewer tokens.
             ),
             use_tqdm = False
         )