Skip to content

aayushxrj/Llama-2-7B-Chat-OpenOrca

Repository files navigation

Fine-tuning Llama2-Chat Model with QLoRA on OpenOrca dataset

This repository contains the code and data used to fine-tune the Llama2-Chat model using the 4-bit quantisation QLoRA (Quantization with Low Rank Approximation) PEFT technique on the OpenOrca dataset.

Dataset Preprocessing

OpenOrca-Clean

The OpenOrca-Clean dataset is a refined version derived from the original OpenOrca dataset.

Llama2-OpenOrca-Clean

The Llama2-OpenOrca-Clean dataset is tailored specifically for fine-tuning the Llama2-Chat model. It is derived from the OpenOrca-Clean dataset, further adapted to fit the llama prompt template. The dataset comprises a single column labeled "text," structured in the given format-

Screenshot 2024-04-06 115410

Model Fine-tuning

Model Details

The fine-tuning process involves training the Llama2-Chat model with 4-bit quantization using the QLoRA technique. This technique allows for efficient representation of model parameters while minimizing computational overhead.

Official Models

Llama-2-7B-Chat-OpenOrca

Our latest model, fine-tuned with 1000 examples using 4-bit quantization QLoRA from Llama2-OpenOrca-Clean dataset, is now available.

About

Fine tuned Llama-2-7b-chat-hf using 4-bit quantization QLoRA PEFT on the OpenOrca dataset.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published