Skip to content

avramdj/transformers-in-pytorch

Repository files navigation

Transformers

This repository contains implementations of various transformer models for different natural language processing and computer vision tasks.

BERT

Paper: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

  • Fill-mask language model pretraining for downstream tasks ✅
  • Sequence classification ✅
  • Token classification 💠
  • Next sentence prediction 💠

GPT-2

Paper: Language Models are Unsupervised Multitask Learners

  • Semi-supervised training for sequence generation 💠

ViT

Paper: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

  • Image inpainting 💠
  • Image classification 💠

About

My implementation of various popular transformer architectures

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages