Skip to content

gOsuzu/Efficient-LLM-Few-Examples-Supervised-Fine-Tuning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

95 Commits
 
 
 
 
 
 

Repository files navigation

Efficient-LLM-Few-Examples-Supervised-Fine-Tuning

This is a final project repository for Georgia Tech CS7643 (Spring 2024).

Because we need GPU for this project, we used Google Colab to run our code.

Project Abstract

In-context learning (ICL) in Large Language Models (LLMs) has been shown to be one of their key capabilities. ICL allows to provide a reasoned answer to a question based on a series of examples passed in the context to an LLM, this way providing a more accurate response. However, this comes at the expense of large-context utilization each time a question is asked, which has an associated higher computational cost and time penalty at inference. The project aims to evaluate different recently developed fine-tuning alternatives to ICL:

  • Context Distillation (CD),
  • Selective Context Distillation (SCD),
  • Virtual Token Embedding Fine-tuning applied with Context Distillation (VTEFT-CD)
  • Context Distillation with Low-Ranked Adaptation (LoRA) (CD-LoRA)
  • Vanilla Fine-Tuning,
  • Pattern-Based Fine-Tuning (PBFT)

We compare them to ICL in terms of accuracy for both in-domain and out-of-domain questions.

Project Member

Project Structure

notebooks folder contains ipynb used for this project. results folder contains the summary excel file of our results, and plot of in-domain accuracies of ICL and CDs with code and data.

About

This is a final porject repository for Goergia Tech CS7643.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •