Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/FranxYao/Language-Model-Pretraining-for-Text-Generation
LM pretraining for generation, reading list, resources, conference mappings.
https://github.com/FranxYao/Language-Model-Pretraining-for-Text-Generation
bert bert-model gpt language-generation language-model pretrained-models text-generation
Last synced: about 1 month ago
JSON representation
LM pretraining for generation, reading list, resources, conference mappings.
- Host: GitHub
- URL: https://github.com/FranxYao/Language-Model-Pretraining-for-Text-Generation
- Owner: FranxYao
- Created: 2019-04-23T04:13:58.000Z (over 5 years ago)
- Default Branch: master
- Last Pushed: 2020-02-25T06:54:30.000Z (almost 5 years ago)
- Last Synced: 2024-08-03T02:03:38.308Z (4 months ago)
- Topics: bert, bert-model, gpt, language-generation, language-model, pretrained-models, text-generation
- Size: 40 KB
- Stars: 20
- Watchers: 4
- Forks: 1
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- Awesome-Paper-List - Pre-trained LM for Text Generation - franxyao-be8abf) ![](https://img.shields.io/github/stars/FranxYao/Language-Model-Pretraining-for-Text-Generation) (Natural Language Processing)
README
# Language-Model-Pretraining-for-Text-Generation
LM pretraining for generation, reading list, resources, conference mappings.
![](cover.jpeg)
----
* Deep contextualized word representations (ELMo). NAACL 18
* I just like ELMo and AI2.* BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, NAACL 19
* BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension.
* Current SOTA on summarization* Language Models are Unsupervised Multitask Learners (GPT2)
* Improving Language Understanding by Generative Pre-Training (GPT)
* Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (T5)
* Unified Language Model Pre-training for Natural Language Understanding and Generation (UniLM), NeurIPS 19
* Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, Hsiao-Wuen Hon. MSR
* Text Summarization with Pretrained Encoders. ACL 19
* Unsupervised Pretraining for Sequence to Sequence Learning, EMNLP 17
* Prajit Ramachandran and Peter J. Liu and Quoc V. Le, Google Brain
* The LM pretraining idea, before BERT. Effective on MT and Summarization.* MASS: Masked Sequence to Sequence Pre-training for Language Generation, ICML 19
* Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, Tie-Yan Liu, NJU and MSRA* Pre-trained Language Model Representations for Language Generation, NAACL 19
* Sergey Edunov, Alexei Baevski, Michael Auli, FAIR
* LM pretraining fine tuned on different size of MT data. The smaller the more performance gain.* Semi-Supervised Sequence Modeling with Cross-View Training. (CVT) EMNLP 18
* Kevin Clark, Minh-Thang Luong, Christopher D. Manning, Quoc V. Le
* Semi-supervised learning with multi-task learning. MT is one application. Also improvements on other tasks* Few-shot NLG with Pre-trained Language Model
* Zhiyu Chen, Harini Eavani, Yinyin Liu, and William Yang Wang
* Extreme few-shot supervision: **50 training instances get 26 BLEU*** Cross-lingual Language Model Pretraining, Arxiv 19
* Guillaume Lample, Alexis Conneau, FAIR
* Application to Cross-lingual classification, Unsupervised machine translation, Supervised machine translation (low-resource), Low-resource language model* BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language Model, Arxiv 19
* Alex Wang, Kyunghyun Cho, NYU
* Directly sample from BERT, not as effective.* Constant-Time Machine Translation with Conditional Masked Language Models ([link](https://arxiv.org/pdf/1904.09324.pdf))
* Marjan Ghazvininejad, Omer Levy, Yinhan Liu and Luke Zettlemoyer
* Iterative decoding: generate with BERT, then polish with BERT, then continue polish