Fact-checked by Grok 2 weeks ago
References
-
[1]
[PDF] Deep Contextualized Word Representations - ACL AnthologyWe show that similar signals are also induced by the modified language model objective of our ELMo representations, and it can be very beneficial to learn ...
-
[2]
Deep Contextualized Word Representations - ACL AnthologyCite (ACL):: Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner ... Deep Contextualized Word Representations. In Proceedings of the 2018 ...
-
[3]
Deep contextualized word representations - Google Scholar- **Paper Identified**: "Deep contextualized word representations" by Peters et al. (2018) is not directly listed in the provided content with citation count.
- [4]
-
[5]
[1802.05365] Deep contextualized word representations - arXivFeb 15, 2018 · This paper introduces deep contextualized word representations modeling word use and context, using a deep bidirectional language model, and ...
-
[6]
Efficient Estimation of Word Representations in Vector Space - arXivJan 16, 2013 · We propose two novel model architectures for computing continuous vector representations of words from very large data sets.
-
[7]
GloVe: Global Vectors for Word Representation - ACL AnthologyJeffrey Pennington, Richard Socher, and Christopher Manning. 2014. GloVe: Global Vectors for Word Representation. In Proceedings of the 2014 Conference on ...
-
[8]
Learned in Translation: Contextualized Word Vectors - arXivAug 1, 2017 · In this paper, we use a deep LSTM encoder from an attentional sequence-to-sequence model trained for machine translation (MT) to contextualize word vectors.
-
[9]
elmo - AllenNLP v2.10.1Compute ELMo representations using a pre-trained bidirectional language model. See "Deep contextualized word representations", Peters et al. for details ...Missing: Benchmark corpus
-
[10]
allenai/allennlp: An open-source NLP research library, built ... - GitHubDec 16, 2022 · An Apache 2.0 NLP research library, built on PyTorch, for developing state-of-the-art deep learning models on a wide variety of linguistic tasks.AllenNLP models · Issues 80 · Pull requests 11 · Discussions
-
[11]
One Billion Word Benchmark for Measuring Progress in Statistical ...Dec 11, 2013 · We propose a new benchmark corpus to be used for measuring progress in statistical language modeling. With almost one billion words of training data.
-
[12]
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding### Summary of ELMo and Its Influence on BERT from arXiv:1810.04805
-
[13]
Universal Language Model Fine-tuning for Text Classification - arXivJan 18, 2018 · We propose Universal Language Model Fine-tuning (ULMFiT), an effective transfer learning method that can be applied to any task in NLP.
-
[14]
Pre-trained models: Past, present and future - ScienceDirect.comIn this paper, we take a deep look into the history of pre-training, especially its special relation with transfer learning and self-supervised learning.Missing: became | Show results with:became
-
[15]
Pre-trained ELMo Representations for Many Languages - GitHub1 for the first LSTM hidden layer; 2 for the second LSTM hidden layer; -1 for an average of 3 layers. (default); -2 for all 3 layers. Training Your Own ELMo.
- [16]