Fact-checked by Grok 2 weeks ago
References
-
[1]
Improving language understanding with unsupervised learningJun 11, 2018 · These results provide a convincing example that pairing supervised learning methods with unsupervised pre-training works very well.<|control11|><|separator|>
-
[2]
[PDF] Improving Language Understanding by Generative Pre-TrainingNatural language understanding comprises a wide range of diverse tasks such as textual entailment, question answering, semantic similarity assessment, and.
-
[3]
[PDF] Improving Language Understanding by Generative Pre-TrainingImproving Language Understanding by Generative Pre-Training ... +4 authors. S. Fidler. Computer Science. NIPS. 2015. We describe an approach for ...
-
[4]
[PDF] arXiv:2105.05241v1 [cs.CL] 11 May 2021May 11, 2021 · This paper aims to help address documentation debt for BookCorpus, a popular text dataset for training large language models. Notably, ...
-
[5]
[PDF] A Retrospective Datasheet for BookCorpusThis paper contributes a formal case study in retrospective dataset documentation and pinpoints several problems with the influential BookCorpus dataset.
-
[6]
Overview of data used to train language models - Our-HometownJul 17, 2023 · This post provides a brief summary of several corpora used for training Large Language Models (LLMs), categorized into six groups.
-
[7]
[PDF] Language Models are Unsupervised Multitask Learners | OpenAIThe smallest model is equivalent to the original GPT, and the second smallest equivalent to the largest model from BERT (Devlin et al., 2018). Our largest.
-
[8]
openai-community/openai-gpt - Hugging FaceThe model is a causal (unidirectional) transformer pre-trained using language modeling on a large corpus with long range dependencies.<|separator|>
- [9]
- [10]
-
[11]
Alec Radford - Google ScholarCo-authors ; Improving language understanding by generative pre-training. A Radford, K Narasimhan, T Salimans, I Sutskever. 17330, 2018 ; Improved techniques for ...