Skip to content

qbxlvnf11/GPT-series

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 

Repository files navigation

Description

- GPT-2

  • Auto-regression model
    • Like traditional language models such as RNN
    • Outputs one token at a time
      • Step 1. after each token is produced, that token is added to the sequence of inputs
      • Step 2. And that new sequence becomes the input to the model in its next step.

- KOGPT2

  • GPT series of Korean version
  • Pretrained with Korean Wiki, news, corpus etc.

- Upload code as a Jupiter Notebook file (.ipynb) for immediate understanding

Contents

- KoGPT2 zero-shot learning

Datasets

- KR3: Korean Restaurant Reviews with Ratings

https://www.kaggle.com/datasets/ninetyninenewton/kr3-korean-restaurant-reviews-with-ratings

References

- Papers

@article{GPT-2,
  title={Language Models are Unsupervised Multitask Learners},
  author={Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever},
  year={2018}
}

- GPT-2

https://github.com/openai/gpt-2

- KoGPT2

https://github.com/SKT-AI/KoGPT2

https://www.kaggle.com/code/ninetyninenewton/zero-shot-sentiment-classification-using-gpt-2/notebook

Author

Releases

No releases published

Packages

No packages published