Skip to content

Parameter Efficient Fine Tuning (PEFT) to create a chatbot from Facebook's LLaMA Large Language Model (LLM) on a public corpus (subreddit submissions and comments rearranged as chats).

License

calhounpaul/LLaMA-PEFT-LoRa-subreddit-chatbot-colab

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 

Repository files navigation

LLaMA PEFT LoRa subreddit chatbot

ChatGPT(v4) summary of this notebook: This Jupyter notebook demonstrates how to create a chatbot using Parameter Efficient Fine Tuning (PEFT) on Facebook's LLaMA Large Language Model (LLM) and a public Reddit corpus. The chatbot is fine-tuned using an RTX 3090 GPU. The code extracts subreddit data using the ps_reddit_tool, cleans the text with redditcleaner, and organizes the data into linear conversations for training. It uses a range of months from 2022 as an example and focuses on the "wholesome" subreddit. Finally, the chats are formatted and saved as a dataset for further training. Various packages are installed and important hyperparameters are set. The LLaMA model is trained using a dataset, which is shuffled and tokenized. The PEFT weights are then integrated into the original model for uploading to HuggingFace. Afterward, a HuggingFace-compatible model is created and a rudimentary Gradio chatbot interface is demonstrated.

/r/NewHampshire after ~4 hours of training on one rtx 3090:

llama

About

Parameter Efficient Fine Tuning (PEFT) to create a chatbot from Facebook's LLaMA Large Language Model (LLM) on a public corpus (subreddit submissions and comments rearranged as chats).

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published