Paper Reading List¶
Step1: NMT, Attention and Transformer¶
TODO: 2014-Google-Sequence to Sequence¶
Sequence to Sequence Learning with Neural Networks
TODO: 2014-Bengio-Attention¶
Neural Machine Translation by Jointly Learning to Align and Translate. 首次提出了Attention的概念
PyTorch Practice
See: NLP From Scratch: Translation with a Sequence to Sequence Network and Attention
TODO: 2017-Transformers¶
Attention Is All You Need
Step2: GPT Series¶
2018-OpenAI-GPT1.0¶
Improving Language Understanding by Generative Pre-Training
2018-Google-BERT¶
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
2019-OpenAI-GPT2.0¶
Language Models are Unsupervised Multitask Learners
2020-OpenAI-GPT3.0¶
Language models are few-shot learners
Step3: Engineering in LLM¶
TBD
Reference¶
最后更新: November 23, 2023