Repository for My HuggingFace Natural Language Processing Projects
-
Updated
Aug 31, 2023 - Jupyter Notebook
Repository for My HuggingFace Natural Language Processing Projects
Auto-regressive causal language model for molecule (SMILES) and reaction template (SMARTS) generation based on the Hugging Face implementation of OpenAI's GPT-2 transformer decoder model
Transformers Intuition
Causal language modeling and intent classification using GPT-2.
A quick and easy way to interact with open-source LLMs.
Course materials for the Machine Learning for NLP course taught by Sameer Singh for the Cognitive Science summer school 2022.
An AI generated picturebook.
Fine-tuning (or training from scratch) the library models for language modeling on a text dataset for GPT, GPT-2, ALBERT, BERT, DitilBERT, RoBERTa, XLNet... GPT and GPT-2 are trained or fine-tuned using a causal language modeling (CLM) loss while ALBERT, BERT, DistilBERT and RoBERTa are trained or fine-tuned using a masked language modeling (MLM…
Rescoring Automatic Speech Recognition using Large Language Models
Dataset and model fine-tuning for function calling
This is the implementation of low rank adaptation (LoRA) which is a subset of parameter efficient fine tuning (PEFT).
Add a description, image, and links to the causal-language-modeling topic page so that developers can more easily learn about it.
To associate your repository with the causal-language-modeling topic, visit your repo's landing page and select "manage topics."