a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
-
Updated
May 9, 2024 - Jupyter Notebook
a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
Experimental project for AI and NLP based on Transformer Architecture
Generate caption on images using CNN Encoder- LSTM Decoder structure
Transformers Intuition
An LLM based tool for generation of cheese advirtisements
A miniGPT inspired from the original NanoGPT released by OpenAI. This is a notebook to walk through the decoder part of the transformer architecture with details outlined.
Implementation of the GPT-3 paper: Language Models are Few-Shot Learners
a dna sequence generation/classification using transformers
A Decoder Based Semantic Parser that can be tested on four benchmark datasets (ATIS, GeoQuery, Jobs640 and Django)
Decoder model for language modelling
Add a description, image, and links to the decoder-model topic page so that developers can more easily learn about it.
To associate your repository with the decoder-model topic, visit your repo's landing page and select "manage topics."