Fact-checked by Grok 2 weeks ago
References
-
[1]
What Is NLP (Natural Language Processing)? - IBMNatural language processing (NLP) is a subfield of artificial intelligence (AI) that uses machine learning to help computers communicate with human ...
-
[2]
What is Natural Language Processing? - NLP Explained - AWSNatural language processing (NLP) is technology that allows computers to interpret, manipulate, and comprehend human language.What Is Natural Language... · What are the approaches to... · What are NLP tasks?
-
[3]
Natural Language Processing (NLP): What it is and why it mattersNatural language processing (NLP) is a branch of artificial intelligence that helps computers understand, interpret and manipulate human language.
-
[4]
What Is Natural Language Processing and How Does It Relate to AI?NLP has a wide range of real-world applications, including: Virtual assistants; Chatbots; Autocomplete tools; Language translation; Sentiment analysis; Text ...What Is Natural Language... · How Does Natural Language... · NLP Examples
-
[5]
Key Milestones in Natural Language Processing (NLP) 1950 - 2024May 23, 2024 · Key NLP milestones include foundational concepts, symbolic approaches, statistical methods, deep learning, and the rise of large language ...
-
[6]
The Transformative Journey of Natural Language Processing | AI ...Aug 9, 2024 · A key innovation in this era was the development of word embeddings, which allow words to be represented as dense vectors in a continuous space.
-
[7]
Emerging Technology – Advancements in Natural Language ...Some anticipated uses of NLP will include better real-time translation of voice and text, smarter search engines, and advancements in business intelligence.Missing: key | Show results with:key
-
[8]
12 Applications of Natural Language ProcessingAug 23, 2021 · Examples of Natural Language Processing · 1. Autocorrect and Spell-check · 2. Text Classification · 3. Sentiment Analysis · 4. Question ...
-
[9]
Op-ed: Tackling biases in natural language processingNLP biases include gender bias, where models favor males for high-level jobs, and racial bias, where systems negatively score non-standard African-American ...Missing: controversies | Show results with:controversies
-
[10]
Detecting and mitigating bias in natural language processingMay 10, 2021 · Biased NLP algorithms cause instant negative effect on society by discriminating against certain social groups and shaping the biased ...Missing: controversies | Show results with:controversies
-
[11]
What are some controversies surrounding natural language ...May 25, 2023 · Natural language processing can have implicit biases, create a significant carbon footprint, and stoke concerns about AI sentience.<|separator|>
-
[12]
The First Public Demonstration of Machine Translation OccursThe first public demonstration of Russian-English machine translation occurred on January 7, 1954, in New York, using an IBM 701 computer. It was a ...
-
[13]
The Georgetown-IBM experiment demonstrated in January 1954The Georgetown-IBM experiment was a public demonstration of a Russian-English machine translation system, a small-scale experiment of 250 words and six grammar ...
-
[14]
NLP - overview - Stanford Computer ScienceThe field of natural language processing began in the 1940s, after World War II. At this time, people recognized the importance of translation from one ...
-
[15]
A Brief History of Natural Language Processing - DataversityJul 6, 2023 · NLP Begins and Stops Noam Chomsky published Syntactic Structures in 1957. In this book, he revolutionized linguistic concepts and concluded ...Missing: influence | Show results with:influence
-
[16]
ELIZA—a computer program for the study of natural language ...ELIZA—a computer program for the study of natural language communication between man and machine. Author: Joseph Weizenbaum.
-
[17]
[PDF] ALPAC-1966.pdf - The John W. Hutchins Machine Translation ArchiveIn this report, the Automatic Language. Processing Advisory Committee of the National Research Council describes the state of development of these applications.
-
[18]
[PDF] ALPAC -- the (in)famous report - ACL AnthologyThe best known event in the history of machine translation is without doubt the publication thirty years ago in November 1966 of the report by the Automatic ...
-
[19]
[PDF] SHRDLU - Computer Science00056001 SHRDLU, created by Terry Winograd, was a com- puter program that could understand instructions and carry on conversations about a world consist ...
-
[20]
A Brief Timeline of NLP - MediumSep 20, 2022 · The 1950s, 1960s, and 1970s: Hype and the First AI Winter. The first application that sparked interest in NLP was machine translation. The first ...
- [21]
-
[22]
History and Evolution of NLP - GeeksforGeeksJul 23, 2025 · In the 1950s, the dream of effortless communication across languages fueled the birth of NLP. Machine translation (MT) was the driving force, ...
-
[23]
[PDF] A STATISTICAL APPROACH TO MACHINE TRANSLATIONComputational Linguistics Volume 16, Number 2, June 1990. Page 7. Peter F. Brown et al. A Statistical Approach to Machine Translation. REFERENCES. Bahl, L. R. ...
-
[24]
[PDF] A Neural Probabilistic Language ModelAbstract. A goal of statistical language modeling is to learn the joint probability function of sequences of words in a language.
-
[25]
[PDF] A Unified Architecture for Natural Language Processing: Deep ...In this work we attempt to define a unified architecture for Natural Language Processing that learns features that are relevant to the tasks at hand given very ...
-
[26]
Deep learning: Historical overview from inception to actualization ...This study aims to provide a historical narrative of deep learning, tracing its origins from the cybernetic era to its current state-of-the-art status.
-
[27]
[PDF] LONG SHORT-TERM MEMORY 1 INTRODUCTIONHochreiter, S. and Schmidhuber, J. (1997). LSTM can solve hard long time lag problems. In. Advances in Neural Information Processing Systems 9. MIT ...
-
[28]
Efficient Estimation of Word Representations in Vector Space - arXivJan 16, 2013 · We propose two novel model architectures for computing continuous vector representations of words from very large data sets.
-
[29]
[1706.03762] Attention Is All You Need - arXivJun 12, 2017 · Abstract page for arXiv paper 1706.03762: Attention Is All You Need. ... We propose a new simple network architecture, the Transformer ...
-
[30]
[1810.04805] BERT: Pre-training of Deep Bidirectional Transformers ...Oct 11, 2018 · BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers.
-
[31]
[2005.14165] Language Models are Few-Shot Learners - arXivMay 28, 2020 · GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks ...
-
[32]
[2001.08361] Scaling Laws for Neural Language Models - arXivJan 23, 2020 · We study empirical scaling laws for language model performance on the cross-entropy loss. The loss scales as a power-law with model size, dataset size, and the ...
-
[33]
Training language models to follow instructions with human feedbackMar 4, 2022 · In this paper, we show an avenue for aligning language models with user intent on a wide range of tasks by fine-tuning with human feedback.
-
[34]
Natural Language Processing RELIES on LinguisticsWe argue our case around the acronym RELIES, which encapsulates six major facets where linguistics contributes to NLP: Resources, Evaluation, Low-resource ...
-
[35]
[PDF] Natural Language ProcessingJan 14, 2015 · • Phonology. • Morphology. • Syntax. • Semantics. • Pragmatics. • Discourse. Each kind of knowledge has associated with it an encapsulated set ...
-
[36]
[PDF] CS 545: Natural Language Processingphonetics phonology morphology syntax semantics pragmatics discourse orthography. What are the utterances? Page 8. syntax. What are the utterances? Noah gave ...
-
[37]
Chapter 2. A Crash Course in Linguistics - CUNY Pressbooks NetworkPhonology: The patterns of sounds in language. Morphology: Word formation. Syntax: The arrangement of words into larger structural units such as phrases and ...
-
[38]
Linguistic Fundamentals for Natural Language ProcessingMay 31, 2022 · The purpose of this book is to present in a succinct and accessible fashion information about the morphological and syntactic structure of human languages.
-
[39]
Linguistic Fundamentals for Natural Language Processing IIJun 1, 2020 · The book covers most of the key issues in semantics and pragmatics, ranging from “meaning of words” to “meaning of utterances in dialogue,” ...
-
[40]
The History of Natural Language Processing - LeximancerDec 4, 2024 · NLP started with rule-based systems in the 1950s, shifted to statistical models in the 1980s, machine learning in the 2000s, deep learning in ...
-
[41]
[PDF] N-gram Language Models - Stanford UniversityAn n-gram is a sequence of n words: a 2-gram (which we'll call bigram) is a two-word sequence of words like The water, or water of, and a 3- gram (a trigram) is ...
-
[42]
Modeling Natural Language with N-Gram Models | Kevin SookocheffJul 25, 2015 · This article explains what an n-gram model is, how it is computed, and what the probabilities of an n-gram model tell us.
- [43]
-
[44]
[PDF] Tagging Problems, and Hidden Markov Models - Columbia CSWe first discuss two important examples of tagging problems in NLP, part-of- speech (POS) tagging, and named-entity recognition. Figure 2.1 gives an example ...
-
[45]
[PDF] Probabilistic Context-Free Grammars (PCFGs) - Columbia CS3.4.2 Parsing using the CKY Algorithm. We now describe an algorithm for parsing with a PCFG in CNF. The input to the algorithm is a PCFG G = (N,Σ, S, R, q) ...
-
[46]
[PDF] Lecture 10: Statistical Parsing with PCFGsProbabilistic Context-Free Grammars. For every nonterminal X, define a ... A lexicalized PCFG assigns zero probability to any word that does not appear ...
-
[47]
Naive Bayes and Text Classification - Sebastian RaschkaOct 4, 2014 · In this first part of a series, we will take a look at the theory of naive Bayes classifiers and introduce the basic concepts of text classification.
-
[48]
[PDF] 13 Text classification and Naive - Bayes - Stanford NLP GroupThis list shows the general importance of classification in IR. Most retrieval systems today contain multiple components that use some form of classifier. The ...<|separator|>
-
[49]
Named Entity Recognition(NER) using Conditional Random Fields ...May 17, 2020 · CRF is amongst the most prominent approach used for NER. A linear chain CRF confers to a labeler in which tag assignment(for present word, ...
-
[50]
Building a named entity recognition model using a BiLSTM-CRF ...Jul 13, 2023 · Conditional random field (CRF) is a statistical model well suited for handling NER problems, because it takes context into account. In other ...
-
[51]
Analysis Methods in Neural Language Processing: A SurveyApr 1, 2019 · In this survey paper, we review analysis methods in neural language processing, categorize them according to prominent research trends, ...
-
[52]
Long Short-Term Memory | Neural Computation - MIT Press DirectNov 15, 1997 · We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient based method called ...
-
[53]
Empirical Evaluation of Gated Recurrent Neural Networks on ... - arXivDec 11, 2014 · In this paper we compare different types of recurrent units in recurrent neural networks (RNNs). Especially, we focus on more sophisticated units that ...
-
[54]
Convolutional Neural Networks for Sentence Classification - arXivAug 25, 2014 · We report on a series of experiments with convolutional neural networks (CNN) trained on top of pre-trained word vectors for sentence-level classification ...
-
[55]
Automatic Metrics in Natural Language Generation: A Survey ... - arXivAug 17, 2024 · Across both INLG and ACL papers, BLEU and ROUGE are the predominant metrics used for NLG automatic evaluations, as seen in Table 2. This is in ...
-
[56]
[PDF] We Need to Talk About Classification Evaluation Metrics in NLPSome of the most widely used classification metrics for measuring classifier performance in NLP tasks are Accuracy, F1-Measure and the Area Under the Curve - ...
-
[57]
We Need to Talk About Classification Evaluation Metrics in NLP - arXivJan 8, 2024 · We compare several standard classification metrics with more 'exotic' metrics and demonstrate that a random-guess normalised Informedness metric is a ...
-
[58]
[PDF] A Closer Look at Classification Evaluation Metrics and a Critical ...This paper aims to serve as a handy reference for anyone who wishes to better understand classi- fication evaluation, how evaluation metrics align with ...
-
[59]
Two minutes NLP — Learn the BLEU metric by examples - MediumJan 11, 2022 · BLEU, or the Bilingual Evaluation Understudy, is a metric for comparing a candidate translation to one or more reference translations.
-
[60]
Two minutes NLP — Learn the ROUGE metric by examples - MediumJan 19, 2022 · ROUGE focuses on recall: how much the words (and/or n-grams) in the human references appear in the candidate model outputs. These results are ...
-
[61]
Evaluation Metrics for Language Modeling - The GradientOct 18, 2019 · Intuitively, perplexity can be understood as a measure of uncertainty. The perplexity of a language model can be seen as the level of perplexity ...Understanding Perplexity... · Reasoning About Entropy As A... · Empirical Entropy
-
[62]
Perplexity - a Hugging Face Space by evaluate-metricPerplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a ...
-
[63]
A Survey of Evaluation Metrics Used for NLG SystemsIn particular, we highlight that the existing NLG metrics have poor correlations with human judgements, are uninterpretable, have certain biases and fail to ...
-
[64]
The Illusion of a Perfect Metric: Why Evaluating AI's Words Is Harder ...Aug 19, 2025 · Based on their underlying scoring methodologies, evaluation metrics can be categorized into three groups: lexical similarity, which measures the ...<|separator|>
-
[65]
[PDF] Tokenization Is More Than Compression - ACL AnthologyTokenization is an essential step in NLP that trans- lates human-readable text into a sequence of dis- tinct tokens that can be subsequently ...
- [66]
-
[67]
(PDF) Tokenization as the initial phase in NLP - ResearchGateIn this paper, the authors address the significance and complexity of tokenization, the beginning step of NLP. Notions of word and token are discussed and ...
-
[68]
Text Preprocessing in NLP - GeeksforGeeksJul 23, 2025 · Example - Text Preprocessing in NLP · 1. Text Cleaning · 2. Tokenization · 3. Stop Words Removal · 4. Stemming and Lemmatization · 5. Handling ...
-
[69]
Tokenization in NLP: Types, Challenges, Examples, ToolsLet's discuss the challenges and limitations of the tokenization task. In general, this task is used for text corpus written in English or French where these ...Missing: multilingual | Show results with:multilingual
-
[70]
Tokenization and Representation Biases in Multilingual Models on ...Sep 24, 2025 · Abstract page for arXiv paper 2509.20045: Tokenization and Representation Biases in Multilingual Models on Dialectal NLP Tasks.
-
[71]
What Are Word Embeddings? | IBMA brief history of word embeddings ... In the 2000s, researchers began exploring neural language models (NLMs), which use neural networks to model the ...
-
[72]
On word embeddings - Part 1 - ruder.ioApr 11, 2016 · A brief history of word embeddings. Since the 1990s, vector space models have been used in distributional semantics. During this time, many ...
-
[73]
[PDF] GloVe: Global Vectors for Word Representation - Stanford NLP GroupRecent methods for learning vector space representations of words have succeeded in capturing fine-grained semantic and syntactic regularities using vector ...
-
[74]
[1607.04606] Enriching Word Vectors with Subword Information - arXivIn this paper, we propose a new approach based on the skipgram model, where each word is represented as a bag of character n-grams.
-
[75]
[1802.05365] Deep contextualized word representations - arXivFeb 15, 2018 · We introduce a new type of deep contextualized word representation that models both (1) complex characteristics of word use (eg, syntax and semantics),
-
[76]
(PDF) A Critical Review of Recurrent Neural Networks for Sequence ...Recurrent neural networks (RNNs) are a powerful family of connectionist models that capture time dynamics via cycles in the graph.<|separator|>
-
[77]
RNN-LSTM: From applications to modeling techniques and beyond ...Since their introduction in 1997 by Hochreiter and Schmidhuber (1997), LSTMs have become widely used and highly effective in various sequential data tasks, from ...
-
[78]
Neural Machine Translation by Jointly Learning to Align and ... - arXivSep 1, 2014 · Access Paper: View a PDF of the paper titled Neural Machine Translation by Jointly Learning to Align and Translate, by Dzmitry Bahdanau and ...Missing: date | Show results with:date
-
[79]
Multi-Head Attention Mechanism - GeeksforGeeksOct 7, 2025 · Multi-head attention extends self-attention by splitting the input into multiple heads, enabling the model to capture diverse relationships and patterns.
-
[80]
What is the difference between a generative and a discriminative ...May 18, 2009 · A generative model learns the joint probability distribution p(x,y) and a discriminative model learns the conditional probability distribution p(y|x).
-
[81]
[PDF] On Discriminative vs. Generative classifiers: A comparison of logistic ...Abstract. We compare discriminative and generative learning as typified by logistic regression and naive Bayes. We show, contrary to a widely-.
-
[82]
[PPT] Generative and Discriminative Models in NLP: A Survey (ppt)generative (HMM); discriminative (maxent, memory-based, decision tree, neural network, linear models(boosting,perceptron) ). NN.<|separator|>
-
[83]
Generative vs. discriminative - Cross Validated - Stack ExchangeJun 27, 2011 · Discriminative models learn the (hard or soft) boundary between classes. Generative models model the distribution of individual classes.
-
[84]
Generative vs Discriminative Models: Differences & Use CasesSep 2, 2024 · This article explains the core differences between generative and discriminative models, covering their principles, use cases, and practical examplesWhat Are Discriminative Models? · Neural networks · Generative vs Discriminative...
-
[85]
[PDF] arXiv:1905.11912v2 [cs.CL] 9 Jul 2019Jul 9, 2019 · We propose a simple yet effective lo- cal discriminative neural model which retains the advantages of generative models while address- ing the ...
-
[86]
Decoding Generative and Discriminative Models | Analytics VidhyaDec 13, 2024 · A generative model explains how the data generates, while a discriminative model focuses on predicting the data labels.
-
[87]
Are We Really Making Much Progress in Text Classification? A ...Jan 19, 2025 · We emphasize the superiority of discriminative language models like BERT over generative models for supervised tasks. Additionally, we highlight ...
-
[88]
Generative models vs Discriminative models for Deep Learning.Apr 22, 2022 · Discriminative models separate data into classes, while generative models can generate new data points and model data distribution. ...
-
[89]
Best Tools for Natural Language Processing in 2025 - GeeksforGeeksJul 23, 2025 · Best Tools for Natural Language Processing in 2025 · spaCy · NLTK (Natural Language Toolkit) · Hugging Face Transformers · Stanford CoreNLP.
-
[90]
NLTK :: Natural Language ToolkitNLTK is a leading platform for building Python programs to work with human language data. It provides easy-to-use interfaces to over 50 corpora and lexical ...Installing NLTKBookExample UsageNltk packageInstalling NLTK Data
-
[91]
NLTK: The Natural Language Toolkit - ACL AnthologyCite (ACL):: Steven Bird and Edward Loper. 2004. NLTK: The Natural Language Toolkit. In Proceedings of the ACL Interactive Poster and Demonstration Sessions ...
-
[92]
nltk - PyPIThe Natural Language Toolkit (NLTK) is a Python package for natural language processing. NLTK requires Python 3.9, 3.10, 3.11, 3.12, or 3.13.<|control11|><|separator|>
-
[93]
Natural Language Processing With Python's NLTK PackageNLTK is a Python package for NLP, used for text preprocessing, analysis, and creating visualizations. It helps make natural language usable by programs.
-
[94]
spaCy · Industrial-strength Natural Language Processing in PythonspaCy is a free open-source library for Natural Language Processing in Python. It features NER, POS tagging, dependency parsing, word vectors and more.spaCy 101 · Usage · Models · Projects
-
[95]
explosion/spaCy: Industrial-strength Natural Language Processing ...spaCy is a library for advanced Natural Language Processing in Python and Cython. It's built on the very latest research, and was designed from day one to be ...
-
[96]
spaCy 101: Everything you need to knowspaCy is a free, open-source library for advanced Natural Language Processing (NLP) in Python. If you're working with a lot of text, you'll eventually want ...
-
[97]
Natural Language Processing With spaCy in Python - Real PythonFeb 1, 2025 · spaCy is a free, open-source library for NLP in Python written in Cython. It's a modern, production-focused NLP library that emphasizes speed, streamlined ...
-
[98]
Transformers - Hugging FaceTransformers acts as the model-definition framework for state-of-the-art machine learning models in text, computer vision, audio, video, and multimodal model.
-
[99]
How do Transformers work? - Hugging Face LLM CourseThe Transformer architecture was introduced in June 2017. The focus of the original research was on translation tasks. This was followed by the introduction of ...How 🤗 Transformers solve tasks · CO2 Emissions and the 🤗 Hub · SmolLM2 paper
-
[100]
What are Hugging Face Transformers? - Azure DatabricksNov 7, 2024 · Hugging Face Transformers is an open-source framework for deep learning created by Hugging Face. It provides APIs and tools to download state-of-the-art pre- ...
-
[101]
An Introduction To HuggingFace Transformers for NLP - WandbJan 17, 2024 · A Brief History of HuggingFace. Founded in 2016, HuggingFace (named after the popular emoji ) started as a chatbot company and later ...
-
[102]
7 Top NLP Libraries For NLP Development [Updated] - LabellerrOct 26, 2024 · Explore the fascinating world of Natural Language Processing (NLP) and its libraries, including NLTK, Gensim, spaCy, and more.
-
[103]
Top Open-Source AI/ML Frameworks in 2025 - Atlantic.NetSep 22, 2025 · Top Open-Source AI Frameworks for Machine Learning · #1: TensorFlow · #2: PyTorch · #3: Scikit-learn · #4: Keras · #5: Hugging Face Transformers · #6: ...
-
[104]
Google launches new API to help you parse natural languageJul 20, 2016 · Google today announced the public beta launch of its Cloud Natural Language API, a new service that gives developers access to Google ...
-
[105]
Natural Language AI - Google CloudIn this lab, you'll learn how to create an API key Use the Cloud Natural Language API and extract "entities" (e.g. people, places, and events) from a snippet ...Pricing · Cloud Natural Language · REST API Reference · Language Support
-
[106]
Introducing Amazon Comprehend – Discover Insights from Text - AWSNov 29, 2017 · Amazon Comprehend is a natural language processing (NLP) service that uses machine learning to analyze your text.
- [107]
-
[108]
Introducing the Realtime API - OpenAIOct 1, 2024 · Update on August 28, 2025: We announced the general availability of the Realtime API. · Update on February 3, 2025: We no longer limit the number ...
- [109]
-
[110]
What is Azure AI Language - Microsoft LearnSep 26, 2025 · Azure AI Language is a cloud-based service that provides Natural Language Processing (NLP) features for understanding and analyzing text.Named Entity Recognition · Conversational Language... · Language Detection
-
[111]
IBM Watson Natural Language UnderstandingWatson Natural Language Understanding is an API uses machine learning to extract meaning and metadata from unstructured text data.
-
[112]
Release notes for Natural Language Understanding - IBM Cloud DocsBase version. 17 October 2024. Retired Summarization (Experimental) Feature: The Summarization (Experimental) feature of Watson Natural Language Understanding ...
-
[113]
Best Natural Language Processing (NLP) APIs in 2025 - Eden AITop NLP APIs in 2025: AWS · Google Cloud · IBM · Meaning Cloud · Neural Space · Open AI and many more!
-
[114]
History of Machine TranslationMar 24, 2024 · Machine translation began during WWII, with the Georgetown-IBM experiment in 1954. Rule-based systems evolved to statistical models, and neural ...Missing: key | Show results with:key
-
[115]
Progress in Machine Translation - ScienceDirectIn this article, we first review the history of machine translation from rule-based machine translation to example-based machine translation and statistical ...Missing: milestones | Show results with:milestones
-
[116]
Neural Machine Translation: Evolution and Impact - BlogApr 17, 2025 · Improved translation quality. Technical milestones. Key developments in NMT include: 2014: Introduction of sequence-to-sequence models; 2015 ...Missing: advancements | Show results with:advancements
-
[117]
[PDF] Performance Comparison of Statistical vs. Neural-Based Translation ...Feb 22, 2023 · Figure 9: BLEU score generated by NMT and SMT for Eng.–Hindi language pairs. MT, machine translation; SGD, stochastic gradient descent; SMT, ...
-
[118]
Machine Translation Performance for Low-Resource LanguagesApr 21, 2025 · This review provides a detailed evaluation of the current state of MT for low-resource languages and emphasizes the need for further research into ...
-
[119]
[PDF] Understanding In-Context Machine Translation for Low-Resource ...Jul 27, 2025 · Recent advancements in multilingual NMT also show that models trained on multiple language pairs can better deal with low-resource languages. ( ...
-
[120]
GLUE BenchmarkGLUE is a benchmark for training and evaluating natural language understanding systems, including nine tasks, a diagnostic dataset, and a leaderboard.
-
[121]
SuperGLUE BenchmarkA new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard.SuperGLUE Diagnostic Dataset · Leaderboard · Tasks · FAQ
-
[122]
SuperGLUE: A Stickier Benchmark for General-Purpose Language ...May 2, 2019 · In this paper we present SuperGLUE, a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, a software toolkit, and a ...
-
[123]
Optimizing translation for low-resource languages: Efficient fine ...In their research, Andersland (2024) focused on improving translation and language understanding for low-resource languages, specifically Amharic. They ...
-
[124]
A Survey on Large Language Model Benchmarks - arXivAug 21, 2025 · These benchmarks primarily focused on natural language understanding (NLU) through relatively small-scale, single-task evaluations. However, as ...
-
[125]
A Comparative Study of Sentiment Analysis on Customer Reviews ...Sentiment analysis is a growing research area in natural language processing that enables computers to interpret and classify human emotions expressed in text.
-
[126]
Sentiment analysis: A survey on design framework, applications and ...Mar 20, 2023 · This survey presents a systematic and in-depth knowledge of different techniques, algorithms, and other factors associated with designing an effective ...Missing: peer- | Show results with:peer-<|separator|>
-
[127]
A comprehensive survey of text classification techniques and their ...This process involves natural language processing (NLP) methods to transform raw text into structured data, which can then be analyzed and classified using ...
-
[128]
[PDF] Text Classification: A Review of Deep learning Methods - arXivSep 24, 2023 · Also, text classification systems can classify text by its size, such as document level, paragraph level, sentence level, and clause level [1].
-
[129]
Natural language processing for analyzing online customer reviewsThis section covers the taxonomy of NLP applications, including sentiment analysis and opinion mining, review analysis and management, customer experience and ...
-
[130]
Sentiment analysis in public health: a systematic review of ... - FrontiersThis systematic review provides a comprehensive overview of sentiment analysis in public health, examining methodologies, applications, data sources, ...
-
[131]
Sentiment analysis methods, applications, and challengesSarcasm and Ridicule: The problem of identifying sarcasm and ridicule ... Aspect-level sentiment analysis with aspect-specific context position information.
-
[132]
Top 7 Sentiment Analysis Challenges - Research AIMultipleJul 9, 2025 · 1. Context-dependent errors Sarcasm People tend to use sarcasm as a way of expressing their negative sentiment, but the words used can be positive.
-
[133]
A systematic review of aspect-based sentiment analysis - SpringerLinkSep 17, 2024 · This paper presents a systematic literature review (SLR) of ABSA studies with a focus on trends and high-level relationships among these fundamental components.
-
[134]
Deep Learning is Transforming ASR and TTS AlgorithmsDec 16, 2022 · This post provides an overview of how automatic speech recognition (ASR) and text-to-speech (TTS) technologies have evolved due to deep learning.
-
[135]
Navigating the Evolution of Automatic Speech Recognition (ASR)Apr 10, 2024 · Learn more about the evolution and challenges of Automatic Speech Recognition (ASR) technology, from statistical models to advanced neural ...
-
[136]
Introducing Whisper - OpenAISep 21, 2022 · However, when we measure Whisper's zero-shot performance across many diverse datasets we find it is much more robust and makes 50% fewer errors ...
-
[137]
openai/whisper: Robust Speech Recognition via Large ... - GitHubWhisper is a general-purpose speech recognition model. It is trained on a large dataset of diverse audio and is also a multitasking model.
-
[138]
openai/whisper-large-v3 - Hugging FaceThe large-v3 model shows improved performance over a wide variety of languages, showing 10% to 20% reduction of errors compared to Whisper large-v2 . For more ...
-
[139]
[PDF] Dong Yu Li Deng A Deep Learning Approach - David Hason RuddThis is the first book on automatic speech recognition (ASR) that is focused on the deep learning approach, and in particular, deep neural network (DNN) ...
-
[140]
WaveNet: A generative model for raw audio - Google DeepMindSep 8, 2016 · This post presents WaveNet, a deep generative model of raw audio waveforms. We show that WaveNets are able to generate speech which mimics any human voice.
-
[141]
[1609.03499] WaveNet: A Generative Model for Raw Audio - arXivSep 12, 2016 · This paper introduces WaveNet, a deep neural network for generating raw audio waveforms. The model is fully probabilistic and autoregressive.
-
[142]
[PDF] Text to Speech Synthesis: A Systematic Review, Deep Learning ...In this literature, a taxonomy is introduced which represents some of the deep learning-based architectures and models popularly used in speech synthesis.
-
[143]
The Rise of Multimodal AI: Combining Text, Image, and Audio ...Oct 7, 2024 · Models like LipNet and AVSpeech integrate visual lip movements with audio signals to improve speech-to-text systems. Multimodal Pretrained ...
-
[144]
A Comprehensive Survey and Guide to Multimodal Large Language ...Nov 11, 2024 · This survey and application guide to multimodal large language models (MLLMs) explores the rapidly developing field of MLLMs, examining their architectures, ...<|separator|>
-
[145]
Audio Language Models and Multimodal Architecture - MediumMar 31, 2024 · Multimodal models are creating a synergy between previously separate research areas such as language, vision, and speech.
-
[146]
An overview of high-resource automatic speech recognition ...This paper evaluates state-of-the-art ASR models trained on high-resource data for LREs. We demonstrate that deeper model structures are not efficient for low- ...
-
[147]
A Comprehensive Survey on Document-Level Information ExtractionDocument-level information extraction (doc-IE) plays a pivotal role in the realm of natural language processing (NLP). This paper embarks on a comprehensive ...
-
[148]
[PDF] A Gold-Standard Multilingual Named Entity Recognition BenchmarkJun 16, 2024 · We introduce Universal NER (UNER), an open, community-driven project to develop gold- standard NER benchmarks in many languages.
-
[149]
A Comprehensive Survey on Deep Learning for Relation ExtractionJun 3, 2023 · Relation extraction (RE) involves identifying the relations between entities from unstructured texts. RE serves as the foundation for many ...
-
[150]
A Survey on Open Information Extraction from Rule-based Model to ...Open Information Extraction (OpenIE) represents a crucial NLP task aimed at deriving structured information from unstructured text, unrestricted by relation ...
-
[151]
[PDF] A Survey of Generative Information Extraction - ACL AnthologyJan 19, 2025 · Adaptability and generalization have consistently been key focus areas in information extraction tasks. (Details in Appendix A, B, C, I).
-
[152]
Relation extraction: advancements through deep learning and entity ...Jun 10, 2023 · This paper introduces an approach that fuses entity-related features under convolutional neural networks and graph convolution neural networks.
-
[153]
[PDF] Get the Best out of 1B LLMs: Insights from Information Extraction on ...Aug 16, 2024 · 2.2 LLMs and Information Extraction. Named Entity Recognition (NER) is a key NLP task involving the identification and classification of ...<|separator|>
-
[154]
Survey on Abstractive Text Summarization: Dataset, Models, and ...Dec 22, 2024 · This survey examines the state of the art in text summarization models, with a specific focus on the abstractive summarization approach.
-
[155]
A survey of text summarization: Techniques, evaluation and ...This paper explores the complex field of text summarization in Natural Language Processing (NLP), with particular attention to the development and importance ...
-
[156]
Abstractive Text Summarization: State of the Art, Challenges ... - arXivSep 4, 2024 · The results showed that the attention-based model outperformed a number of baselines, including the fundamental Seq2Seq model and a state-of-the ...
-
[157]
Deep learning for text summarization using NLP for automated news ...Oct 17, 2025 · Abstractive summarization This involves generating a summary that may contain new words or sentences that are absent from the source material.
-
[158]
A Survey on Bias and Fairness in Natural Language ProcessingMar 6, 2022 · This survey analyzes the origins of biases, definitions of fairness, how NLP subfields mitigate bias, and how to eradicate pernicious biases.
-
[159]
Five sources of bias in natural language processing - PMCWe outline five sources where bias can occur in NLP systems: (1) the data, (2) the annotation process, (3) the input representations, (4) the models, and ...
-
[160]
Bias and Fairness in Natural Language Processing - ACL AnthologyIn this tutorial, we will review the history of bias and fairness studies in machine learning and language processing and present recent community effort.
-
[161]
On Measurements of Bias and Fairness in NLP - Google ResearchThis work presents a comprehensive survey of existing bias measures in NLP---both intrinsic measures of representations and extrinsic measures of downstream ...
-
[162]
Bias and Fairness in Large Language Models: A Survey - arXivSep 2, 2023 · This survey covers bias evaluation and mitigation techniques for LLMs, including taxonomies for metrics, datasets, and mitigation methods.
-
[163]
Measuring gender and racial biases in large language modelsExperimental studies provide empirical support for these concerns. Glazko et al. (49), for example, find that GPT-4 exhibits biases against resumes that ...
-
[164]
Generative language models exhibit social identity biases - NatureDec 12, 2024 · Here we show that large language models (LLMs) exhibit patterns of social identity bias, similarly to humans.
-
[165]
[PDF] An Empirical Survey of the Effectiveness of Debiasing Techniques ...May 22, 2022 · To investigate which technique is most effective in mitigating bias (Q1), we evaluate debiased BERT, ALBERT, RoBERTa, and GPT-2 models against ...
-
[166]
Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus ...Dec 17, 2021 · The main findings are that self-debiasing effectively reduces bias across the six attributes; that it is particularly effective for high λ, at ...
-
[167]
Debiasing Methods in Natural Language Understanding Make Bias ...Sep 9, 2021 · Recent debiasing methods in natural language understanding (NLU) improve performance on such datasets by pressuring models into making unbiased ...
-
[168]
Explainability in Neural Networks for Natural Language Processing ...Dec 23, 2024 · Neural networks are widely regarded as black-box models, creating significant challenges in understanding their inner workings, especially in ...
-
[169]
Challenges and Opportunities in Text Generation Explainability - arXivMay 14, 2024 · This paper outlines 17 challenges categorized into three groups that arise during the development and assessment of attribution-based explainability methods.
-
[170]
Explainability in Neural Networks for Natural Language Processing ...Dec 23, 2024 · Techniques such as Local Interpretable Model-Agnostic Explanations (LIME), SHapley Additive exPlanations (SHAP), and perturbation analysis ...
-
[171]
Local Interpretations for Explainable Natural Language ProcessingThis work investigates various methods to improve the interpretability of deep neural networks for Natural Language Processing (NLP) tasks.
-
[172]
[PDF] Trends in NLP Model Interpretability in the Era of LLMsApr 29, 2025 · This surge in usage has led to an explosion in NLP model interpretability and analysis research, ac- companied by numerous technical surveys.
-
[173]
Survey and analysis of hallucinations in large language modelsSep 29, 2025 · Hallucination in Large Language Models (LLMs) refers to outputs that appear fluent and coherent but are factually incorrect, ...
-
[174]
[PDF] Why Language Models Hallucinate - OpenAISep 4, 2025 · Hallucinations are inevitable only for base models. Indeed, empirical studies (Fig. 2) show that base models are often found to be calibrated, ...
-
[175]
Profiling Legal Hallucinations in Large Language Models | Journal ...Jun 26, 2024 · We present the first systematic evidence of these hallucinations in public-facing LLMs, documenting trends across jurisdictions, courts, time periods, and ...
-
[176]
An Empirical Study on Factuality Hallucination in Large Language ...This work presents a systematic empirical study on LLM hallucination, focused on the the three aspects of hallucination detection, source and mitigation.
-
[177]
Detecting hallucinations in large language models using semantic ...Jun 19, 2024 · Here we develop new methods grounded in statistics, proposing entropy-based uncertainty estimators for LLMs to detect a subset of hallucinations.
-
[178]
[PDF] Mitigating Hallucinations via Dual Process of Fast and Slow ThinkingJul 27, 2025 · In this paper, we in- vestigate whether tree search-based slow thinking can effectively leverage accurate internal knowl- edge from LLMs to ...
-
[179]
Knowledge-Driven Hallucination in Large Language Models - arXivThis paper investigates the knowledge-driven hallucination of LLMs through a systematic, empirical study within the domain of Business Process Management (BPM).
-
[180]
FLOPS used for GPT-4 if released - MetaculusGPT-3 took 3.14E+23 FLOPS to train; Deepmind's GOPHER took 6.31E+23 FLOPS to train; The largest disclosed ML experiment to date (Megatron-Turing NLG 530B) ...
-
[181]
OpenAI's GPT-3 Language Model: A Technical Overview - LambdaJun 3, 2020 · Even at theoretical 28 TFLOPS for V100 and lowest 3 year reserved cloud pricing we could find, this will take 355 GPU-years and cost $4.6M for a ...
-
[182]
GPT-4 Architecture, Infrastructure, Training Dataset, Costs, Vision ...Jul 10, 2023 · OpenAI's training FLOPS for GPT-4 is ~2.15e25, on ~25,000 A100s ... 3rd party hardware support much more easily. A wave of huge models ...<|control11|><|separator|>
-
[183]
What is the cost of training large language models? - CUDO ComputeMay 12, 2025 · It is estimated that GPT-4's training consumed 2.1 × 1025 FLOPs (21 billion petaFLOPs), and models like Gemini Ultra might be around 5.0 × 1025 ...
-
[184]
[PDF] Training Compute-Optimal Large Language Models - arXivMar 29, 2022 · We investigate the optimal model size and number of tokens for training a transformer language model under a given compute budget.
-
[185]
Over 30 AI models have been trained at the scale of GPT-4Jan 30, 2025 · The largest AI models today are trained with over 1025 floating-point operations (FLOP) of compute. The first model trained at this scale was ...
-
[186]
How much AI inference can we do? - LessWrongMay 14, 2024 · ... GPT-4 requires 5.6e11 FLOP per forward pass. So that would ... GPT-4 is more than 10x the size of GPT-3. We believe it has a total ...
-
[187]
A systematic review of electricity demand for large language modelsThe primary challenge stems from their immense power consumption [7]. Currently, the power of LLM-serving data centers has risen to hundreds of megawatts and ...
-
[188]
The Energy Hunger of AI: Large Language Models as Challenges ...The exponential growth of AI workloads, especially from LLMs, is shifting data center electricity demand from a marginal to a system-relevant load. Meeting this ...
-
[189]
How Much Energy Do LLMs Consume? Unveiling the Power Behind ...Jul 3, 2024 · This article would help to unfold the hidden energy costs of training and inference these sophisticated AI models, exploring their environmental impact.
-
[190]
Scaling Large Language Models: Navigating the Challenges of Cost ...Running multiple instances to serve concurrent users escalates the demand for powerful processors, leading to increased energy consumption and operational costs ...
-
[191]
Investigating Energy Efficiency and Performance Trade-offs in LLM ...Jan 14, 2025 · In this work, we investigate the effect of important parameters on the performance and energy efficiency of LLMs during inference and examine their trade-offs.<|separator|>
-
[192]
Quantifying Social Biases in NLP: A Generalization and Empirical ...Jun 28, 2021 · This paper quantifies social biases in NLP by unifying and comparing fairness metrics, which measure differences in model behavior across ...
-
[193]
[PDF] Ethical and social risks of harm from Language Models - arXivDec 8, 2021 · Language models pose risks including discrimination, information hazards, misinformation, malicious uses, human-computer interaction harms, and ...
-
[194]
[PDF] Ethical Challenges and Solutions in Neural Machine TranslationApr 1, 2024 · Ethical challenges in NMT include data handling, privacy, ownership, consent, and the need for human oversight, as the system mirrors the ...
-
[195]
Ethical Concern Identification in NLP: A Corpus of ACL Anthology ...Nov 12, 2024 · The most frequent ethical concerns are privacy, unemployment, bias, human replacement, misuse, and impersonation. The complete survey, including ...4 Automatic Ethical Concern... · 7 Taxonomies · Appendix A Ethicon Dataset
-
[196]
Aligning Large Language Models with Human: A Survey - arXivJul 24, 2023 · This survey presents a comprehensive overview of these alignment technologies, including the following aspects.
-
[197]
Fundamental Limitations of Alignment in Large Language ModelsApr 19, 2023 · Our results expose fundamental limitations in alignment of LLMs and bring to the forefront the need to devise reliable mechanisms for ensuring AI safety.
-
[198]
Strong and weak alignment of large language models with human ...Aug 21, 2024 · The Alignment Problem that we deal with in this paper refers to the specific issue of AI systems alignment with human moral values. Moreover, we ...
-
[199]
How should the advancement of large language models affect the ...Conclusion. In conclusion, LLMs are often mischaracterized, misused, and overhyped, yet they will certainly impact the way we do science, from search to ...
-
[200]
The Working Limitations of Large Language ModelsNov 30, 2023 · The Working Limitations of Large Language Models. Overestimating the capabilities of AI models like ChatGPT can lead to unreliable applications.Missing: overhype NLP empirical
-
[201]
On the Dangers of Stochastic Parrots - ACM Digital LibraryMar 1, 2021 · In this paper, we take a step back and ask: How big is too big? What are the possible risks associated with this technology and what paths are available for ...
-
[202]
A Survey on Hallucination in Large Language ModelsJan 24, 2025 · Moreover, recent research has exposed that LLMs can occasionally exhibit unpredictable reasoning hallucinations spanning both long-range and ...
-
[203]
Lost in translation: AI's impact on translators and foreign language ...Mar 22, 2025 · In fact, for each 1 percentage point increase in MT usage, translator employment growth dropped by approximately 0.7 percentage points.
-
[204]
Advancements in natural language processing: Implications ...The 1980s and 1990s saw a shift towards statistical methods, leveraging large corpora of text data and probabilistic models to improve language processing ...
-
[205]
NLP in Customer Service: Benefits, Use Cases, and Future TrendsSep 4, 2025 · How Natural Language Processing (NLP) is transforming customer service with AI-driven chatbots, sentiment analysis, and predictive support.
-
[206]
AI-induced job impact: Complementary or substitution? Empirical ...This study utilizes 3,682 full-time workers to examine perceptions of AI-induced job displacement risk and evaluate AI's potential complementary effects on ...
-
[207]
[PDF] Comprehensive Research Report AI Job Displacement Analysis ...By 2025, 85 million jobs will be displaced by AI, but 97 million new roles will emerge, resulting in a net positive job creation of 12 million.Missing: statistics | Show results with:statistics
-
[208]
The impact of artificial intelligence on employment: the role of virtual ...Jan 18, 2024 · The positive effect of artificial intelligence on employment exhibits an inevitable heterogeneity, and it serves to relatively improves the job share of women ...
-
[209]
With language models on the rise, how can Natural Language ...Jun 2, 2023 · For instance, they found that one of the most researched areas among social good-related NLP papers has been health and well-being. Another area ...
-
[210]
Natural Language Processing Influence on Digital Socialization and ...The Metaverse and Natural Language Processing (NLP) technologies have combined to fundamentally change the nature of digital sociability.
-
[211]
[PDF] The Social Impact of Natural Language Processing - ACL AnthologyIn particular, we want to explore the impact of NLP on social justice,. i.e., equal opportunities for individuals and groups. (such as minorities) within ...
-
[212]
The Fearless Future: 2025 Global AI Jobs Barometer - PwCJun 3, 2025 · Skills for AI-exposed jobs are changing 66% faster than for other jobs: more than 2.5x faster than last year. The AI-driven skills earthquake is ...Missing: NLP | Show results with:NLP
-
[213]
LoRA: Low-Rank Adaptation of Large Language Models - arXivJun 17, 2021 · We propose Low-Rank Adaptation, or LoRA, which freezes the pre-trained model weights and injects trainable rank decomposition matrices into each layer of the ...
-
[214]
FlashAttention: Fast and Memory-Efficient Exact Attention with IO ...May 27, 2022 · We propose FlashAttention, an IO-aware exact attention algorithm that uses tiling to reduce the number of memory reads/writes between GPU high bandwidth memory ...
-
[215]
GPTQ: Accurate Post-Training Quantization for Generative Pre ...Oct 31, 2022 · GPTQ can quantize GPT models with 175 billion parameters in approximately four GPU hours, reducing the bitwidth down to 3 or 4 bits per weight.
-
[216]
Learning Transferable Visual Models From Natural Language ...Feb 26, 2021 · The paper proposes learning visual models by predicting image-caption pairs, then using natural language for zero-shot transfer to downstream ...
-
[217]
Flamingo: a Visual Language Model for Few-Shot Learning - arXivApr 29, 2022 · We introduce Flamingo, a family of Visual Language Models (VLM) with this ability. We propose key architectural innovations to: (i) bridge powerful pretrained ...
-
[218]
[2304.08485] Visual Instruction Tuning - arXivApr 17, 2023 · In this paper, we present the first attempt to use language-only GPT-4 to generate multimodal language-image instruction-following data.
-
[219]
[PDF] Neuro-Symbolic Methods in Natural Language Processing: A ReviewThis review paper explores recent advancements in neurosymbolic NLP methods. We carefully highlight the benefits and drawbacks of differ- ent approaches in ...
-
[220]
[PDF] Towards Neuro-Symbolic Approaches for Referring Expression ...Sep 8, 2025 · Sequential Sequential architectures are current the dominant approach in Deep Learning when the input and output of neural networks are symbolic.
-
[221]
Enhancing Large Language Models through Neuro-Symbolic ... - arXivApr 10, 2025 · We propose a neuro-symbolic approach integrating symbolic ontological reasoning and machine learning methods to enhance the consistency and reliability of LLM ...
-
[222]
Neurosymbolic AI Could Be the Answer to Hallucination in Large ...Jun 2, 2025 · Neurosymbolic AI combines the predictive learning of neural networks with teaching the AI a series of formal rules that humans learn to be able ...
-
[223]
Neurosymbolic AI Approach to Attribution in Large Language Models... interpretability ... Integrating neural adaptability with symbolic reasoning significantly reduces hallucinations and improves the transparency of outputs.
-
[224]
(PDF) Natural Language Processing and Neurosymbolic AIFeb 24, 2024 · Neurosymbolic AI (NeSy AI) represents a groundbreaking approach in the realm of Natural Language Processing (NLP), merging the pattern ...
-
[225]
Large language models empowered agent-based modeling and ...Sep 27, 2024 · This paper surveys the landscape of utilizing large language models in agent-based modeling and simulation, discussing their challenges and promising future ...
-
[226]
Multi-agent systems powered by large language models - FrontiersThis work examines the integration of large language models (LLMs) into multi-agent simulations by replacing the hard-coded programs of agents with LLM-driven ...
-
[227]
LLMs and generative agent-based models for complex systems ...This paper briefly reviews the disruptive role LLMs are playing in fields such as network science, evolutionary game theory, social dynamics, and epidemic ...
-
[228]
LLM-Based Agents for Tool Learning: A Survey | Data Science and ...Jun 26, 2025 · In the multi-agent framework, a more powerful agent is used as a supervised external reflection to evaluate the appropriateness of tool ...
-
[229]
[PDF] Designing LLM based agents to interact with the embodied worldMay 14, 2025 · In this work, we study methods to bridge the gap between LLMs and physical robotic systems through structured observation and action interfaces.
-
[230]
LLM-Powered AI Agent Systems and Their Applications in IndustryMay 22, 2025 · This paper comprehensively examines the evolution of agent systems from the pre-LLM era to current LLM-powered architectures.
-
[231]
What is Multimodal AI? | IBMMultimodal AI refers to machine learning models capable of processing and integrating information from multiple modalities or types of data.The Latest Ai News +... · How Multimodal Ai Works · Trends In Multimodal Ai
-
[232]
Multimodal NLP: The Next Powerful Shift In AI - Spot IntelligenceDec 19, 2023 · Multimodal NLP refers to the intersection of natural language processing (NLP) with other data or modalities, such as images, videos, audio, and sensor data.
-
[233]
NLP in AI Agents: A Comprehensive Guide to FunctionalityRating 4.0 (5) This introduction will explore the fundamentals of NLP and its significance in enhancing the capabilities of AI agents.
-
[234]
Transformation of industrial robotics with natural language modelsNLP is a branch of artificial intelligence that focuses on the interaction between humans and computers through natural language, as first introduced in [32].
-
[235]
Multimodal NLP for Robotics - Naver Labs EuropeApplications in robotics. Multimodal NLP research is applicable to robotics in numerous ways and whenever human-robot interaction or collaboration is necessary.
-
[236]
Integrating AI Planning with Natural Language ProcessingAug 18, 2025 · ... NLP techniques helps planning systems analyze multimodal data and allows humans to interact with intelligent systems. Moreover, the ...
-
[237]
Collaborative Agentic AI Needs Interoperability Across EcosystemsMay 25, 2025 · Notable examples include the \AcfA2A protocol [26] , released by Google in April 2025, the \AcfMCP [45] , released by Anthropic in November 2024 ...