ARTICLEAttentions Mechanism In Neural Machine Translation

Covers: theory of Attention Mechanism
Estimated time needed: 20 minutes
Questions this item adddesses:
  • What is attention and type of attention mechanisms?
How to use this item?

Read the shortlist Article 2 which referes to Attention Attention blog post. Easy to understand and covers types of Attention mechanism.

Author(s) / creator(s) / reference(s)
Lilian Wang

Understand the paper : Learning Longformer:The Long-Document Transformer

Sharvari DhoteTotal time needed: ~2 hours
Learning Objectives
With this list you can learn about Longformer and how to implement it.
Potential Use Cases
Long text summarization, Long text question answering.
Target Audience
ADVANCEDNLP Data scientist from all audience levels
Go through the following annotated items in order:
ARTICLE 1. Introduction to Transformer Encoder Decoder Model
  • What is transformer and how it works?
10 minutes
ARTICLE 2. Attentions Mechanism In Neural Machine Translation
  • What is attention and type of attention mechanisms?
20 minutes
PAPER 3. Longformer : The Long-Document Transformer (Original Paper)
  • What is long former and attention mechanism to process long sequence?
20 minutes
ARTICLE 4. Understanding Transformer-Based Self-Supervised Architectures - LongFormer
  • How can we reduce the computational cost of the attention calculations, which grow quadratically with sequence length?
7 minutes
REPO 5. Longformer : The Long-Document Transformer Github Repo
  • How longformer is implemented?
10 minutes
ARTICLE 6. Train a Longformer for Detecting Hyper-partisan News
  • How to train a longformer?
7 minutes
ARTICLE 7. Train a Longformer for the Question Answering
  • How to implement longformer for question answering task?
7 minutes

Concepts Convered