Skip to content

LaurentVeyssier/TRAX_transformer_abstractive_summarization_model

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 

Repository files navigation

TRAX_transformer_abstractive_summarization_model

Use Trax and Transformer architecture for generative summarizer deep learning model

This colab notebook is based on Deeplearning.ai / coursera course "Natural Language Processing with Attention Models" week 2 assignment. This module is part of Deeplearning.ai 4-course Natural Language Processing specialization.

Requierements

  • Trax==1.3.4
  • T5==0.6.2

The model uses large 'cnn-dailymail' tensorflow dataset to train a transformer model on news article and associated summary. The model is then able to generate summary for a text provided as an input (abstractive summarization). You can refer to a couple of my previous projects covering the different types of summary generation (extractive vs abstractive). This model generates a summary from the input text by elaborating its own sentences. Greedy approach used for next token. The vocab used is based on subwords and not full words. This provides efficiency through the combination of subwords to form final words. Generated tokens are therefore subwords to form full words, sentence and full summary.

Output

image

image

more difficult context and language domain:

image

About

Use Trax and Transformer architecture for generative summarizer deep learning model

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published