Skip to content

armbues/SiLLM-examples

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Examples for using SiLLM

Make sure to install the requirements for these examples:

pip install -r requirements-examples.txt

HelpSteer LoRA

LoRA training Mistral-7B-Instruct-v0.2 with the Nvidia HelpSteer dataset.

Run train.sh in the helpsteer directory to download the dataset & model from HuggingFace and start the LoRA training. You can customize the training configuration by editing config.yml.

DPO-Mix-7K

DPO training Qwen1.5-7B-Chat with the DPO Mix 7K dataset. The training consists of a supervised fine tuning (SFT) followed by direct preference optimization (DPO).

Run train.sh in the dpo-mix-7k directory to download the dataset & model from HuggingFace and start the training. You can customize the training configuration by editing the config files sft.yml and dpo.yml.

MMLU Benchmark

Implementation of the "Massive Multitask Language Understanding" benchmark using the MMLU dataset.

Run mmlu.py with the model you would like to evaluate.

MMLU-Pro Benchmark

Implementation of the MMLU-Pro benchmark using the MMLU-Pro dataset.

Run mmlu-pro.py with the model you would like to evaluate.

Perplexity

Calculating perplexity scores for a sample dataset of entry paragraphs from Wikipedia articles.

Run perplexity.py with the model you would like to evaluate. Add quantization options to evaluate perplexity with quantized models.

About

Examples for using the SiLLM framework for training and running Large Language Models (LLMs) on Apple Silicon

Topics

Resources

License

Stars

Watchers

Forks