[Information] Papers Must Read

Sejin Jeong·2022년 12월 22일
2

Information

목록 보기
4/4

**The above picture is a laptop power saving screen, and has nothing to do with this post.

Neural Machine Translation by Jointly Learning to Align and Translate (a.k.a. Bahdanau Attention)

Go to Read

Effective Approaches to Attention-based Neural Machine Translation (a.k.a. Luong Attention)

Go to Read

Attention Is All You Need (a.k.a. Transformer)

Go to Read

Improving Language Understanding by Generative Pre-Training (a.k.a. GPT-1)

Go to Read

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (a.k.a. BERT)

Go to Read

Language Models are Unsupervised Multitask Learners (a.k.a. GPT-2)

Go to Read

Language Models are Few-Shot Learner (a.k.a. GPT-3)

Go to Read

Multi-Task Deep Neural Networks for Natural Language Understanding (a.k.a. MT-DNN)

Go to Read

MASS: Masked Sequence to Sequence Pre-training for Language Generation (a.k.a. MASS)

Go to Read

XLNet: Generalized Autoregressive Pretraining for Language Understanding (a.k.a. XLNet)

Go to Read

RoBERTa: A Robustly Optimized BERT Pretraining Approach (a.k.a. RoBERTa)

Go to Read

BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension (a.k.a. BART)

Go to Read

Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (a.k.a. T5)

Go to Read

Robust Speech Recognition via Large-Scale Weak Supervision (a.k.a. Whisper)

Go to Read

If you see this post, please recommend papers.
profile
Soli Deo Gloria. / Sapere Aude.

3개의 댓글

comment-user-thumbnail
2022년 12월 22일

인터스피치 컨퍼런스
Data2vec(메타AI)
NVIDIA - Language Understanding Model
ERNIE
ULMFit
ELMo
HyperClova(Naver)
KoGPT(kakao)
KorBERT(ETRI)
Learning to Identify Ambiguous and Misleading News Headlines
Why Does Unsupervised Pre-training Help Deep Learning?
PaLM: Scaling Language Modeling with Pathways
LaMDA
Gopher
GLaM
DALL-E (to attain for language)
YOLO (computer vision, but to attain for language)
Chinchilla
BIG-bench
DialogBERT
A Neural Conversational Model(Google, Paper) - Generative Conversational model(생성 대화 모델)
Meena
LUKE(Deep Contextualized Entity Representations with Entity-aware Self-attention)
RNN

답글 달기
comment-user-thumbnail
2023년 1월 8일
  1. Translation Background
  1. Translation Datasets
  1. Translation Efficiency
  1. Translation Training
  1. Translation Architecture
답글 달기
comment-user-thumbnail
2023년 1월 9일

Word2Vec (2013, 1, 16)
Glove (2014, 1, 2)
FastText (2016, 7, 15)
Transformer (2017, 6, 12)
ELMo (2018, 2, 15)
BERT (2018, 10, 11)

답글 달기