To access these LLMs, we use the Mistral AI and the open-source weights of the remaining models, available at the HuggingFace Hub repositories.
The pretrained LLM weights are used through the transformers python library, with models loaded using standard configurations and their respective default tokenizers.
For the instruction models, inputs were passed through each model's chat template, ensuring responses were formatted appropriately and cleaned for evaluation.
All models were optimized for GPU resources through specific loading options, which focus on maximizing performance based on available hardware.
Collection
[
|
...
]