Fact-checked by Grok 2 weeks ago
References
-
[1]
[2402.06196] Large Language Models: A Survey - arXivFeb 9, 2024 · Large Language Models (LLMs) have drawn a lot of attention due to their strong performance on a wide range of natural language tasks, since the ...
-
[2]
[2303.18223] A Survey of Large Language Models - arXivMar 31, 2023 · To discriminate the difference in parameter scale, the research community has coined the term large language models (LLM) for the PLMs of ...
-
[3]
[1706.03762] Attention Is All You Need - arXivJun 12, 2017 · We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely.
-
[4]
Training Compute-Optimal Large Language Models - arXivMar 29, 2022 · As a highlight, Chinchilla reaches a state-of-the-art average accuracy of 67.5% on the MMLU benchmark, greater than a 7% improvement over Gopher ...
-
[5]
[2206.07682] Emergent Abilities of Large Language Models - arXivJun 15, 2022 · This paper instead discusses an unpredictable phenomenon that we refer to as emergent abilities of large language models.
-
[6]
[PDF] Auto-Regressive Next-Token Predictors are Universal Learners - arXivWe show theoretically that very simple models trained to only predict the next token in an auto-regressive fashion can be used to solve extremely complex tasks ...
-
[7]
Language Model Training and Inference: From Concept to CodeSep 4, 2023 · We will take a deep and practical dive into the concept of next token prediction to understand how it is used by language models both during training and ...
-
[8]
Scale matters: Large language models with billions (rather than ...Oct 22, 2024 · Scale matters: Large language models with billions (rather than millions) of parameters better match neural representations of natural language.
-
[9]
OpenAI Presents GPT-3, a 175 Billion Parameters Language ModelJul 7, 2020 · OpenAI researchers recently released a paper describing the development of GPT-3, a state-of-the-art language model made up of 175 billion parameters.
-
[10]
Beyond Next-Token Prediction: A Performance Characterization of ...Oct 5, 2025 · Autoregressive Language Models (ARMs), which generate tokens sequentially conditioned on all previous tokens, have been the predominant paradigm ...
-
[11]
Auto-Regressive Next-Token Predictors are Universal LearnersLarge language models display remarkable capabilities in logical and mathematical reasoning, allowing them to solve complex tasks. Interestingly, these ...Missing: autoregressive | Show results with:autoregressive
-
[12]
Scale matters: Large language models with billions (rather than ...This can range from 762 in the smallest distill GPT2 model to 8192 in the largest LLAMA-2 70 billion parameter model.
-
[13]
What Are Large Language Models (LLMs)? - IBMLarge language models are AI systems capable of understanding and generating human language by processing vast amounts of text data.
-
[14]
What are Large Language Models? | NVIDIA GlossaryLarge language models (LLMs) are deep learning algorithms that can recognize, summarize, translate, predict, and generate content using very large datasets.
-
[15]
AI Demystified: Introduction to large language models | University ITDec 13, 2024 · Large language models (LLMs) are a type of artificial intelligence designed to understand and generate human-like text based on the input ...<|separator|>
-
[16]
[2301.00234] A Survey on In-context Learning - arXivDec 31, 2022 · In-context learning (ICL) is a new NLP paradigm where LLMs make predictions based on contexts augmented with a few examples.
-
[17]
[2001.08361] Scaling Laws for Neural Language Models - arXivJan 23, 2020 · We study empirical scaling laws for language model performance on the cross-entropy loss. The loss scales as a power-law with model size, dataset size, and the ...
-
[18]
Gemini 2.5: Our most intelligent AI model - The KeywordMar 25, 2025 · 2.5 Pro ships today with a 1 million token context window (2 million coming soon), with strong performance that improves over previous ...
-
[19]
[2408.04666] LLMs are Not Just Next Token Predictors - arXivAug 6, 2024 · LLMs are statistical models of language learning through stochastic gradient descent with a next token prediction objective.
-
[20]
NLP vs LLM: Understanding Key Differences - GeeksforGeeksJul 23, 2025 · NLP and LLMs approach language differently, one through task-specific models and the other using pre-trained capabilities.Understanding Large Language... · Key Differences Between Nlp... · Nlp Vs Llm
-
[21]
LLM vs NLP: The Complete Guide to Understanding the DifferenceJul 18, 2025 · NLP and LLMs both work with human language, but they differ in scale, design, and use. Understand their differences and real-world ...Missing: distinction | Show results with:distinction<|separator|>
-
[22]
N Gram vs RNN vs LLM: Find Out Which Truly Fits You - AI AshesOct 6, 2025 · Perplexity and Cross-Entropy Traditional N-gram models often show higher perplexity (150–200 on WikiText), while RNNs drop this to around 120. ...
-
[23]
Frontier language models have become much smaller | Epoch AIDec 13, 2024 · Parameter counts were scaled up by 1000 times from 117 million to 175 billion between GPT-1 and GPT-3 in the span of two years and by another 10 ...
-
[24]
[2005.14165] Language Models are Few-Shot Learners - arXivMay 28, 2020 · GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks ...
- [25]
-
[26]
Efficient Estimation of Word Representations in Vector Space - arXivJan 16, 2013 · We propose two novel model architectures for computing continuous vector representations of words from very large data sets.
-
[27]
Distributed Representations of Words and Phrases and their ...In this paper we present several improvements that make the Skip-gram model more expressive and enable it to learn higher quality vectors more rapidly. We show ...
-
[28]
Sequence to Sequence Learning with Neural Networks - arXivSep 10, 2014 · In this paper, we present a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure.
-
[29]
[1810.04805] BERT: Pre-training of Deep Bidirectional Transformers ...Oct 11, 2018 · BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers.
-
[30]
[PDF] Language Models are Unsupervised Multitask Learners | OpenAIOur largest model, GPT-2, is a 1.5B parameter Transformer that achieves state of the art results on 7 out of 8 tested lan- guage modeling datasets in a zero- ...
-
[31]
GPT-3 powers the next generation of apps - OpenAIMar 25, 2021 · Over 300 applications are delivering GPT‑3–powered search, conversation, text completion, and other advanced AI features through our API.
-
[32]
OpenAI GPT-3 Waiting List Dropped as GPT-3 Is Fully Released for ...Nov 18, 2021 · The first version, GPT-1, was released in 2018, while the second version, GPT-2, debuted in 2019. With the release of GPT-3 in 2020, natural ...
-
[33]
ChatGPT sets record for fastest-growing user base - analyst noteFeb 2, 2023 · ChatGPT, the popular chatbot from OpenAI, is estimated to have reached 100 million monthly active users in January, just two months after ...
-
[34]
Inside The New AI Index: Expensive New Models, Targeted ...Apr 15, 2024 · Global private investment in generative AI skyrocketed, increasing from roughly $3 billion in 2022 to $25 billion in 2023. Nearly 80 percent of ...
-
[35]
ChatGPT's Work Lacks Transparency and That Is a Problem - RANDMay 8, 2023 · ChatGPT lacks transparency, providing no concrete data or citations, and may produce false information, making its output problematic despite ...
-
[36]
GPT-4 - OpenAIMar 14, 2023 · We are releasing GPT‑4's text input capability via ChatGPT and the API (with a waitlist). To prepare the image input capability for wider ...
-
[37]
The New Version of GPT-3 Is Much, Much Better - MediumFeb 3, 2022 · However, setting apart GPT-3's tendency to engage in toxic and biased behaviors and generate misinformation if prompted to do so, users would ...
-
[38]
(PDF) The Hidden Costs of Chat GPT: A Call for Greater TransparencyChatGPT's training involved low-wage labor, raising significant ethical concerns. Creating GPT-3 emitted over 550 tons of CO2, comparable to 550 flights between ...<|separator|>
-
[39]
Top 6 GPT-3 Open-Source Alternatives - Orient SoftwareAug 19, 2023 · Recommended GPT3 Open-source Alternatives · GPT-Neo and GPT-J (EleutherAI) · Megatron-Turing NLG (NVIDIA and Microsoft) · AlexaTM (Amazon) · LaMDA ( ...
-
[40]
Models with downloadable weights currently lag behind ... - Epoch AIHowever, the release of DeepSeek-R1 in January 2025 showed that the performance gap between open-weights and closed-weights has significantly decreased. For ...
-
[41]
The 2025 AI Index Report | Stanford HAIOpen-weight models are also closing the gap with closed models, reducing the performance difference from 8% to just 1.7% on some benchmarks in a single year.Status · 2024 · Technical Performance · Research and Development
-
[42]
The Llama 4 herd: The beginning of a new era of natively ...Apr 5, 2025 · We'll also make them available via our partners in the coming days. You can also try Meta AI with Llama 4 starting today in WhatsApp, Messenger ...
-
[43]
Meta releases new AI model Llama 4 | ReutersApr 5, 2025 · Meta Platforms (META.O) on Saturday released the latest version of its large language model (LLM) Llama, called the Llama 4 Scout and Llama 4 Maverick.
-
[44]
[2401.04088] Mixtral of Experts - arXivJan 8, 2024 · We introduce Mixtral 8x7B, a Sparse Mixture of Experts (SMoE) language model. Mixtral has the same architecture as Mistral 7B, with the difference that each ...
-
[45]
Grok – Dr Alan D. Thompson - LifeArchitect.aiOrganization, xAI. Model name, Grok-0 33B (Aug/2023) Grok-1 314B MoE (Nov/2023) Grok-1.5 (Mar/2024) Grok-2 (Aug/2024) Grok-3 (Feb/2025) Grok-4 (Jul/2025).
-
[46]
With Grok 4, xAI Retakes The Large Language Model Lead, & MoreJul 14, 2025 · Last week, xAI hit another AI milestone with the release of Grok 4. Now available in two tiers, users can tap into Grok 4's standard model ...
-
[47]
Introducing Claude Sonnet 4.5 - AnthropicSep 29, 2025 · Claude Sonnet 4.5 is the best coding model in the world, strongest model for building complex agents, and best model at using computers.
-
[48]
Anthropic launches Claude Sonnet 4.5, its best AI model for codingSep 29, 2025 · Claude Sonnet 4.5 will be available via the Claude API and in the Claude chatbot. The pricing for developers is the same as Claude Sonnet 4: $3 ...
-
[49]
From Live Data to High-Quality Benchmarks: The Arena-Hard PipelineApr 19, 2024 · We introduce Arena-Hard – a data pipeline to build high-quality benchmarks from live data in Chatbot Arena, which is a crowd-sourced platform for LLM evals.
-
[50]
GPT-4o vs DeepSeek R1 Distill Qwen 7B - LLM StatsGPT-4o outperforms in 1 benchmarks (GPQA), while DeepSeek R1 Distill Qwen 7B is better at 1 benchmark (AIME 2024). ... GPT-4o was released on 2024-08-06 ...
-
[51]
Why New LLMs use an MoE Architecture | Exxact BlogJun 27, 2024 · The Mixture of Experts (MoE) architecture is a neural network design that improves efficiency and performance by dynamically activating a subset of specialized ...
-
[52]
Closing the gap between open source and commercial large ...Sep 9, 2024 · In this study, we investigated to what extent fine-tuning open-source LLMs can further improve their performance.
-
[53]
Can Open-Source LLMs Compete with Commercial Models ... - arXivJul 18, 2024 · Our results indicate that the performance gap between commercial and open-source models in RAG setups exists mainly in the zero-shot setting.
-
[54]
Datasets for Large Language Models: A Comprehensive SurveyFeb 28, 2024 · This paper embarks on an exploration into the Large Language Model (LLM) datasets, which play a crucial role in the remarkable advancements of LLMs.
-
[55]
Common Crawl - Open Repository of Web Crawl DataCommon Crawl maintains a free, open repository of web crawl data that can be used by anyone. Common Crawl is a 501(c)(3) non–profit founded in 2007. We make ...Overview · Get Started · Common Crawl Infrastructure... · Examples Using Our Data
-
[56]
Building Nemotron-CC, A High-Quality Trillion Token Dataset for ...May 7, 2025 · To enable developers to build highly accurate LLMs, NVIDIA previously released Nemotron-CC, a 6.3-trillion-token English language Common Crawl ...
-
[57]
RedPajama-Data-v2: An open dataset with 30 trillion tokens for ...Oct 30, 2023 · We're releasing a new version of the RedPajama dataset, with 30 trillion filtered and deduplicated tokens (100+ trillions raw) from 84 CommonCrawl dumps ...Quality Annotations · Dataset Statistics · Dataset Structure
-
[58]
A Case Study on the Colossal Clean Crawled Corpus - arXivApr 18, 2021 · This paper documents the Colossal Clean Crawled Corpus (C4), created from Common Crawl, and analyzes its data sources and content, including ...
-
[59]
Open-Sourced Training Datasets for Large Language Models (LLMs)Popular Open Source Datasets for Training LLMs · 1. Common Crawl · 2. RefinedWeb · 3. The Pile · 4. C4 · 5. Starcoder Data · 6. BookCorpus · 7. ROOTS · 8. Wikipedia.
-
[60]
GPT-4 Details Revealed - by Patrick McGuinnessJul 12, 2023 · GPT-4 was trained on 13 trillion tokens. More precisely, the total training set was 13 trillion tokens. The 13T tokens are not unique tokens ...
-
[61]
GPT-4 (2022) – Dr Alan D. Thompson - LifeArchitect.ai≈ 0.8% the size of the human brain by count of synapses (125T synapses). Dataset size (tokens), 16T (16,000B) estimated in 40TB. Maybe repeated tokens. 4 ...
-
[62]
[PDF] A Critical Analysis of the Largest Source for Generative AI Training ...Jun 3, 2024 · Common Crawl itself is therefore not an LLM training dataset, but it has an infrastructural role in LLM research and development as a foundation ...<|separator|>
-
[63]
Web Crawler Restrictions, AI Training Datasets \& Political BiasesOct 10, 2025 · Additionally, outlets with neutral political positions impose the strongest restrictions (58%), whereas hyperpartisan websites and those with ...
-
[64]
Automatic large-scale political bias detection of news outlets - PMCMay 12, 2025 · In this article, we introduce a data-driven approach that uses machine learning techniques to analyse multiple forms of bias, and that can estimate the ...
-
[65]
Summary of the tokenizers - Hugging FaceGPT-2 has a vocabulary size of 50,257, which corresponds to the 256 bytes base tokens, a special end-of-text token and the symbols learned with 50,000 merges.
-
[66]
Tokenization Is More Than Compression - arXivFeb 28, 2024 · Tokenization is a foundational step in Natural Language Processing (NLP) tasks, bridging raw text and language models.
-
[67]
google/sentencepiece: Unsupervised text tokenizer for ... - GitHubSentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabulary size is ...
-
[68]
Bit-level BPE: Below the byte boundary - arXivJun 9, 2025 · There are tradeoffs to be made here, as a larger vocabulary increases the cost of logit computation and, as a result, requires more computation ...
-
[69]
Tokenization Changes Meaning in Large Language ModelsOn the other hand, it's unclear how LLMs could learn the meanings of rare and complex words if not from subword constituents. Chinese characters illustrate ...
-
[70]
Why is the vocab size of Byte level BPE smaller than Unicode's ...Feb 14, 2021 · To ensure base vocab size is 256 (which is 1 byte), BBPE only use 1 byte per token. So in case a character requires 2 or more bytes to represent ...
-
[71]
Enhancing Large Language Models through Adaptive TokenizersTokenizers serve as crucial interfaces between models and linguistic data, substantially influencing the efficacy and precision of large language models ...
-
[72]
D4: Improving LLM Pretraining via Document De-Duplication and ...Aug 23, 2023 · Here, we show that careful data selection (on top of de-duplicated data) via pre-trained model embeddings can speed up training (20% efficiency gains)
-
[73]
MinHash LSH in Milvus: The Secret Weapon for Fighting Duplicates ...May 15, 2025 · This article will explore how MinHash LSH efficiently solves the data deduplication problem for LLM training.
-
[74]
Lessons learned on language model safety and misuse - OpenAIMar 3, 2022 · Specifically, we have developed new evaluation metrics for measuring toxicity in model outputs and have also developed in-house classifiers for ...
-
[75]
[PDF] Deduplicating Training Data Makes Language Models BetterMay 22, 2022 · In some cases deduplication reduces perplexity by up to 10%. Further, because re- cent LMs are typically limited to training for just a few ...
-
[76]
[PDF] Addressing Amplification Bias and Homogeneity Issue for LLM ...Nov 12, 2024 · Despite this approach, it has been observed that LLMs have limited exposure to recommendation-specific data during pre-training, necessitating ...
-
[77]
Progressive Learning from Complex Explanation Traces of GPT-4Jun 5, 2023 · Orca learns from rich signals from GPT-4 including explanation traces; step-by-step thought processes; and other complex instructions, guided by teacher ...Missing: LLM synthetic
-
[78]
AI models collapse when trained on recursively generated dataJul 24, 2024 · Model collapse is a degenerative process affecting generations of learned generative models, in which the data they generate end up polluting the training set ...
-
[79]
Synthetic Data Generation Using Large Language Models - arXivMar 18, 2025 · By producing artificial but task-relevant examples, these models can significantly augment or even substitute for real-world datasets, ...
-
[80]
BeyondWeb: Lessons from Scaling Synthetic Data for Trillion-scale ...Aug 18, 2025 · In 2025, rephrasing has become the dominant paradigm with state-of-the-art LLMs like Kimi K2 [1], Qwen-2.5 [2], Grok [3], and GPT-5 [4] ...
-
[81]
The AI Model Collapse Risk is Not Solved in 2025 - WinssolutionsSep 10, 2025 · In 2025, AI researchers warn that training large language models on AI‑generated data triggers an AI model collapse.
-
[82]
[2209.04881] On The Computational Complexity of Self-AttentionSep 11, 2022 · We prove that the time complexity of self-attention is necessarily quadratic in the input length, unless the Strong Exponential Time Hypothesis (SETH) is false.
-
[83]
Understanding and Coding the KV Cache in LLMs from ScratchJun 17, 2025 · A KV cache stores intermediate key/value computations for reuse during inference, avoiding unnecessary recomputations and speeding up text ...
-
[84]
Long context | Gemini API - Google AI for DevelopersSep 22, 2025 · This document gives you an overview of what you can achieve using models with context windows of 1M and more tokens.Getting started with long context · Long context use cases · Long context limitations
-
[85]
Why larger LLM context windows are all the rage - IBM ResearchJul 24, 2024 · Larger windows can improve LLM performance on coding tasks, in particular, by allowing them to ingest more software documentation. IBM is in the ...
-
[86]
Large language models: their history, capabilities and limitationsMay 25, 2023 · At 175 billion parameters, GPT-3 set the new size standard for large language models. It quickly became the focal point for large language model ...
-
[87]
[PDF] A Survey on Mixture of Experts in Large Language Models - arXivApr 9, 2025 · Despite utilizing only 13 billion active parameters, the Mixtral-8x7B demonstrates superior or equivalent per- formance to the Llama-2-70B [159] ...
-
[88]
A Closer Look into Mixture-of-Experts in Large Language ModelsOur experiments are conducted on several open-source MoE models, namely Mixtral 8x7B, DeepSeekMoE, and Grok-1. We choose these models due to their ...
-
[89]
Mixture-of-Experts (MoE) LLMs - by Cameron R. Wolfe, Ph.D.Jan 27, 2025 · Mixtral converts every layer of Mistral to an expert layer with eight experts. Two of these experts are active for each token, yielding a model ...
-
[90]
What is GPTQ Quantization for LLMs? - PicovoiceOct 11, 2023 · GPTQ converts the floating-point parameters of each weight matrix into quantized integers such that the error at the output is minimized.
-
[91]
Quantization - Hugging FaceWith GPTQ quantization, you can quantize your favorite language model to 8, 4, 3 or even 2 bits. This comes without a big drop of performance and with faster ...
-
[92]
[2507.10155] Task-Based Flexible Feature Distillation for LLMs - arXivJul 14, 2025 · Empirical results show consistent improvements over prior approaches across diverse tasks, including classification, instruction-following, and ...
-
[93]
Mamba Explained - The GradientMar 27, 2024 · Mamba also runs fast - like “up to 5x faster than Transformer fast”. Scaling Laws for Mamba vs other Language Models Mamba performs similarly ( ...
-
[94]
DeepSeek-V3.1: Hybrid Thinking Model Now Available on Together AIAug 27, 2025 · Efficiency breakthrough: Comparable quality to DeepSeek-R1 but significantly faster, making deep reasoning practical for production · Built-in ...
-
[95]
A Survey on Efficient Architectures for Large Language Models - arXivAug 13, 2025 · Starting from language modeling, this survey covers the background and technical details of linear and sparse sequence modeling methods, ...Missing: 2024 | Show results with:2024
-
[96]
[PDF] Customizing LLMs for Efficient Latency-Aware Inference at the EdgeJul 9, 2025 · We propose CLONE, an efficient software-hardware co-design system for customized LLM inference on resource-constrained edge devices, where the ...
-
[97]
Optimizing LLMs for Resource-Constrained Environments: A Survey ...This survey paper provides a comprehensive overview of techniques for compressing LLMs to enable efficient inference in resource-constrained environments.<|control11|><|separator|>
-
[98]
[PDF] Better & Faster Large Language Models via Multi-token PredictionApr 30, 2024 · Large language models such as GPT and Llama are trained with a next-token prediction loss. In this work, we suggest that training language ...Missing: phases | Show results with:phases
-
[99]
Understanding the Role of Cross-Entropy Loss in Fairly Evaluating ...Feb 22, 2024 · Recall that the next-token prediction objective used for LLM pre-training (and fine-tuning), by its nature, is a cross-entropy loss that ...Missing: phases | Show results with:phases
-
[100]
Efficient Language Model Pretraining via Curriculum Learning - arXivJun 12, 2025 · Our experiments reveal that curriculum learning consistently improves convergence in early and mid-training phases, and can yield lasting gains ...
-
[101]
Faster LLM Training with Variable Sequence Length CurriculumIn this study, we introduce dataset decomposition, a novel variable sequence length training technique, to tackle these challenges.
-
[102]
[PDF] LLMs: Understanding Code Syntax and Semantics for Code AnalysisWe conclude that LLMs possess capabilities similar to an Abstract Syntax Tree (AST) parser, demonstrating initial competencies in static code analysis. Further-.
- [103]
-
[104]
On the Generalization Ability of Next-Token-Prediction PretrainingJul 17, 2025 · This paper presents a generalization error analysis for next-token prediction pre-training, a widely used paradigm in large language models.
-
[105]
Instruction Tuning for Large Language Models: A Survey - arXivAug 21, 2023 · Abstract page for arXiv paper 2308.10792: Instruction Tuning for Large Language Models: A Survey. ... supervised fine-tuning (SFT)\footnote{In ...
-
[106]
Finetuned Language Models Are Zero-Shot Learners - arXivSep 3, 2021 · This paper explores a simple method for improving the zero-shot learning abilities of language models.
-
[107]
Alpaca: A Strong, Replicable Instruction-Following ModelMar 13, 2023 · We introduce Alpaca 7B, a model fine-tuned from the LLaMA 7B model on 52K instruction-following demonstrations.
-
[108]
Evaluating the Reliability and Consistency of Political Worldviews in ...Nov 4, 2024 · We propose a series of tests which assess the reliability and consistency of LLMs' stances on political statements based on a dataset of voting-advice ...
-
[109]
Training language models to follow instructions with human feedbackMar 4, 2022 · In this paper, we show an avenue for aligning language models with user intent on a wide range of tasks by fine-tuning with human feedback.
-
[110]
Helpful, harmless, honest? Sociotechnical limits of AI alignment and ...Jun 4, 2025 · ... sycophancy as induced by RLHF is hard to quantify in specific cases. As an aside, sycophantic behaviour can be seen, at least in part, as an ...
-
[111]
Reinforcement Learning from Human Feedback in LLMsMar 19, 2025 · This connects to the issue of how to handle conflicts between different alignment criteria, such as helpfulness, truthfulness and harmlessness. ...
-
[112]
Direct Preference Optimization: Your Language Model is Secretly a ...In this paper we introduce a new parameterization of the reward model in RLHF that enables extraction of the corresponding optimal policy in closed form.
-
[113]
Towards Understanding Sycophancy in Language Models - AnthropicOct 23, 2023 · Our results indicate that sycophancy is a general behavior of RLHF models, likely driven in part by human preference judgments favoring sycophantic responses.Missing: induces | Show results with:induces
-
[114]
How much does it cost to train frontier AI models? - Epoch AIJun 3, 2024 · The cost of training top AI models has grown 2-3x annually for the past eight years. By 2027, the largest models could cost over a billion ...
-
[115]
The 2024 AI Index Report | Stanford HAIFrontier models get way more expensive. According to AI Index estimates, the training costs of state-of-the-art AI models have reached unprecedented levels.
-
[116]
Applying Mixture of Experts in LLM Architectures - NVIDIA DeveloperMar 14, 2024 · MoE trains larger models while reducing cost. MoE models help reduce cost by being more flop-efficient per weight, meaning that under regimes ...
-
[117]
Top GPU Hosts for AI Machine Learning in 2025 - GMI CloudFrom 2023 to 2025, per-FLOP costs for AI compute dropped approximately 40%, making advanced machine learning accessible to smaller teams. This trend ...
-
[118]
Can AI scaling continue through 2030? - Epoch AIAug 20, 2024 · In other words, by 2030 it will be very likely possible to train models that exceed GPT-4 in scale to the same degree that GPT-4 exceeds GPT-2 ...
-
[119]
Autoregressive Large Language Models are Computationally ... - arXivOct 4, 2024 · We show that autoregressive decoding of a transformer-based language model can realize universal computation, without external intervention or modification of ...
-
[120]
Perplexity for LLM Evaluation - Comet MLNov 21, 2024 · Perplexity was used to measure how well these models captured linguistic patterns by quantifying the average uncertainty of predictions. “ ...
-
[121]
[PDF] Language Models are Few-Shot Learners - arXivJul 22, 2020 · In this paper, we test this hypothesis by training a 175 billion parameter autoregressive language model, which we call. GPT-3, and measuring ...
-
[122]
Evaluate the text summarization capabilities of LLMs for enhanced ...Apr 25, 2024 · In this post, we explore leading approaches for evaluating summarization accuracy objectively, including ROUGE metrics, METEOR, and BERTScore.
-
[123]
Chain-of-Thought Prompting Elicits Reasoning in Large Language ...Jan 28, 2022 · Experiments on three large language models show that chain of thought prompting improves performance on a range of arithmetic, commonsense, and ...
-
[124]
Language Models Perform Reasoning via Chain of ThoughtMay 11, 2022 · Chain of thought prompting is a simple and broadly applicable method for improving the ability of language models to perform various reasoning tasks.
-
[125]
Learning to reason with LLMs | OpenAISep 12, 2024 · For the o1 model series we show a model-generated summary of the chain of thought. Conclusion. o1 significantly advances the state-of-the-art in ...
-
[126]
[PDF] the-illusion-of-thinking.pdfWe found that LRMs have limitations in exact computation: they fail to use explicit algorithms and reason inconsistently across puzzles. We also investigate the ...
-
[127]
LLMs generate 'fluent nonsense' when reasoning outside their ...Aug 19, 2025 · Relying on SFT to fix every OOD failure is an unsustainable strategy that fails to address the model's core lack of abstract reasoning. “For ...
-
[128]
[2504.02477] Multimodal Fusion and Vision-Language Models - arXivApr 3, 2025 · We compare the evolutionary paths and applicability of VLMs based on large language models (LLMs) with traditional multimodal fusion this ...
-
[129]
A survey on multimodal large language models - PMCSimilarly, CogVLM [60] plugs in a visual expert module in each transformer layer to enable dual interaction and fusion between vision and language features.
-
[130]
Understanding Multimodal LLMs - by Sebastian Raschka, PhDNov 3, 2024 · In this approach, images are converted into tokens with the same embedding size as the original text tokens, allowing the LLM to process both ...
-
[131]
How LLMs See Images, Audio, and More - ByteByteGo NewsletterAug 18, 2025 · Multimodal tokenization extends the concept of text tokens to images, audio, and video. Images get converted through patch embeddings (splitting ...
-
[132]
Video understanding | Gemini API | Google AI for DevelopersGemini models can process videos, enabling many frontier developer use cases that would have historically required domain specific models.Video input · Transcribe video and provide... · Customize video processing
-
[133]
Hello GPT-4o - OpenAIMay 13, 2024 · GPT‑4o is especially better at vision and audio understanding compared to existing models. Model capabilities.
-
[134]
Image understanding | Gemini API | Google AI for DevelopersSep 26, 2025 · Gemini models are built to be multimodal from the ground up, unlocking a wide range of image processing and computer vision tasks including ...
-
[135]
Transferable Decoding with Visual Entities for Zero-Shot Image ...Jul 31, 2023 · In this paper, we propose ViECap, a transferable decoding model that leverages entity-aware decoding to generate descriptions in both seen and unseen scenarios.Missing: emergent | Show results with:emergent
-
[136]
VisuRiddles: Fine-grained Perception is a Primary Bottleneck ... - arXivJun 3, 2025 · However, these benchmarks still exhibit notable limitations, particularly in their partial reliance on external knowledge and lack breadth in ...
-
[137]
Open Eyes, Then Reason: Fine-grained Visual Mathematical...Current multimodal large language models (MLLMs) often underperform on mathematical problem-solving tasks that require fine-grained visual understanding.
-
[138]
Hallucination of Multimodal Large Language Models: A Survey - arXivApr 29, 2024 · Hallucination in MLLMs is when they generate outputs inconsistent with the visual content, posing obstacles to their practical use.
-
[139]
Sora 2 is here | OpenAISep 30, 2025 · Sora 2 is here. Our latest video generation model is more physically accurate, realistic, and more controllable than prior systems. It also ...
-
[140]
Unified Hallucination Detection for Multimodal Large Language ...This paper introduces UNIHD, a unified framework for detecting hallucinations in MLLMs, and MHaluBench, a meta-evaluation benchmark.
-
[141]
Function Calling - xAI DocsFunction calling. Connect the xAI models to external tools and systems ... Run tool functions if Grok asks tool call and append function returns to message.Missing: 2024 | Show results with:2024
-
[142]
Retrieval-Augmented Generation for Knowledge-Intensive NLP TasksMay 22, 2020 · Abstract page for arXiv paper 2005.11401: Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks.Missing: date | Show results with:date
-
[143]
ReAct: Synergizing Reasoning and Acting in Language Models - arXivOct 6, 2022 · In this paper, we explore the use of LLMs to generate both reasoning traces and task-specific actions in an interleaved manner, allowing for greater synergy ...
-
[144]
Compounding Error Effect in Large Language Models - Wand AIJun 5, 2025 · Compounding errors occur when small inaccuracies in an LLM's predictions propagate and accumulate through subsequent processing steps.
-
[145]
Error Propagation in Agentic Systems: What You Need to Know ...Jun 8, 2025 · All error-reducing techniques are either using LLM checking agents (suffering the same issues of the checked agents) or basically reverting ...
-
[146]
[PDF] HalluLens: LLM Hallucination Benchmark - ACL AnthologyJul 27, 2025 · Our results show that the model achieves 96.67% and 95.56% accuracies for refusal and hallucina- tion judgment respectively. 3.1.2 Evaluation ...
-
[147]
[PDF] A Survey on Hallucination in Large Language Models - arXivIn this section, we delve into the root causes of hallucinations in LLMs, primarily categorized into three key aspects: (1) Data (§3.1), (2) Training (§3.2), ...
-
[148]
[PDF] Why Language Models Hallucinate - OpenAISep 4, 2025 · Follow-up work by Miao and Kearns (2025) provides an empirical study of hallucinations, singleton rate, and calibration. 3.3.2 Poor models.
-
[149]
Survey and analysis of hallucinations in large language modelsSep 29, 2025 · In this work, we present a comprehensive survey and empirical analysis of hallucination attribution in LLMs. Introducing a novel framework to ...
-
[150]
What Is Black Box AI and How Does It Work? - IBMMany of the most advanced machine learning models available today, including large language models such as OpenAI's ChatGPT and Meta's Llama, are black box AIs.What is black box artificial... · Why do black box AI systems...
-
[151]
The Black Box Problem: Opaque Inner Workings of Large Language ...Oct 23, 2023 · Large language models like GPT-4 are powerful but opaque "black boxes." New techniques for explainable AI and transparent design can help unlock their benefits ...
-
[152]
The Challenge of AI Interpretability | Deloitte USJul 28, 2025 · LLMs face technical barriers like feature superposition1 and probabilistic generation, further complicating interpretation. Increased demand ...
-
[153]
LLM Interpretability: Understanding What Models Learn and WhyJun 3, 2025 · Technical Challenges. Superposition: Multiple concepts encoded in overlapping sets of neurons, making individual features difficult to isolate.Missing: barriers | Show results with:barriers
-
[154]
A Mathematical Framework for Transformer CircuitsDec 22, 2021 · To explore the challenge of reverse engineering transformers, we reverse engineer several toy, attention-only models. In doing so we find: Zero ...
-
[155]
Efficient Automated Circuit Discovery in Transformers using ... - arXivOct 11, 2024 · They often suffer from slow runtime, approximation errors, and specific requirements of metrics, such as non-zero gradients. In this work, we ...
-
[156]
Extracting Interpretable Features from Claude 3 SonnetMay 21, 2024 · Sparse autoencoders produce interpretable features for large models. · Scaling laws can be used to guide the training of sparse autoencoders.
-
[157]
The engineering challenges of scaling interpretability - AnthropicJun 13, 2024 · Our Sparse Autoencoders—the tools we use to investigate “features”—are trained on the activations of transformers, and those activations ...
-
[158]
Sparse Autoencoders Find Highly Interpretable Features in ... - arXivSep 15, 2023 · These autoencoders learn sets of sparsely activating features that are more interpretable and monosemantic than directions identified by alternative approaches.
-
[159]
[PDF] Scaling Laws for Neural Language Models - arXivJan 23, 2020 · We study empirical scaling laws for language model performance on the cross-entropy loss. The loss scales as a power-law with model size, ...
-
[160]
A Survey of Scaling in Large Language Model Reasoning - arXivApr 2, 2025 · This survey aims to provide insights into how scaling strategies fundamentally enhance the reasoning capabilities of LLMs and further guide the development of ...
-
[161]
Test-Time Scaling Law - Emergent MindJul 1, 2025 · Practical test-time scaling exhibits diminishing returns after a saturation point and is limited by hardware factors like memory bandwidth and ...
-
[162]
How many AI models will exceed compute thresholds? | Epoch AIMay 30, 2025 · For example, our results suggest that the total number of models above 1026 FLOP will quickly grow from a few in 2025, to over 200 by 2030. This ...
-
[163]
[PDF] Artificial Intelligence Index Report 2025 | Stanford HAIFeb 2, 2025 · The AI Index continues to lead in tracking and interpreting the most critical trends shaping the field—from the shifting geopolitical landscape ...
-
[164]
Unveiling GPU Bottlenecks in Large-Batch LLM Inference - arXivJul 11, 2025 · In this work, we conduct a detailed GPU analysis to uncover the true causes of the throughput plateau in large-batch LLM inference. Our findings ...
-
[165]
LLM Inference Performance Engineering: Best Practices - DatabricksOct 12, 2023 · Learn best practices for optimizing LLM inference performance on Databricks, enhancing the efficiency of your machine learning models.
-
[166]
[PDF] Phi-3 Technical Report: A Highly Capable Language Model ... - arXivAug 30, 2024 · We introduce phi-3-mini, a 3.8 billion parameter language model trained on 3.3 trillion tokens, whose overall performance, as measured by both ...
-
[167]
Microsoft's new Phi-3 is one of the smallest AI models availableApr 23, 2024 · Phi-3's performance “rivals that of models such as Mixtral 8x7B and GPT-3.5,” the researchers from Microsoft explained in a paper on the new ...
-
[168]
[PDF] Large Language Models May Talk Causality But Are Not CausalLLMs are not causal, but may appear so because they are trained on data containing causal correlations, like 'causal parrots' reciting embedded knowledge.
-
[169]
[2509.20088] Causal Understanding by LLMs: The Role of UncertaintySep 24, 2025 · These findings suggest that failures in causal understanding arise from the lack of structured causal representation, rather than insufficient ...Missing: peer | Show results with:peer
-
[170]
What is ARC-AGI? - ARC PrizeARC-AGI. The General Intelligence Benchmark. The Measure of Intelligence. In 2019, François Chollet - creator of Keras, an open-source deep learning library ...
-
[171]
12/12 o3 Saturates the ARC-AGI Benchmark! - High Learning RateDec 21, 2024 · GPT-4o only scored 5%. Back in October François Chollet was predicting at least 1 year before seeing considerable progress on his benchmark.
-
[172]
Getting 50% (SoTA) on ARC-AGI with GPT-4o - LessWrongJun 17, 2024 · I recently got to 50% [1] accuracy on the public test set for ARC-AGI by having GPT-4o generate a huge number of Python implementations of the transformation ...
-
[173]
Perplexity of fixed-length models - Hugging FacePerplexity (PPL) is one of the most common metrics for evaluating language models. ... Perplexity is defined as the exponentiated average negative log-likelihood ...
-
[174]
Perplexity Metric for LLM Evaluation - Analytics VidhyaApr 6, 2025 · Learn how perplexity Metric evaluates LLMs, its math foundation, pros & cons & how it compares to modern metrics like LLM-as-a-Judge.<|separator|>
-
[175]
Perplexity for LLM Evaluation - GeeksforGeeksJul 23, 2025 · Evaluation of Language Models: Perplexity helps evaluate language models like GPT-3, where predicting the next word or token is a crucial task.
-
[176]
Perplexity - Evaluation of LLMs Part 1 - LinkedInMay 4, 2024 · 1. Perplexity serves as an inherent metric, independent of external datasets, for assessing the performance of language models within the field ...
-
[177]
What is Wrong with Perplexity for Long-context Language Modeling?Oct 31, 2024 · Perplexity (PPL) is unreliable for long-context because it overlooks key tokens by averaging across all tokens, obscuring true performance.
-
[178]
Can Perplexity Reflect Large Language Model's Ability in Long Text ...May 9, 2024 · The study finds no correlation between perplexity (PPL) and LLMs' long-text understanding ability, as PPL may only reflect local information.
-
[179]
[2009.03300] Measuring Massive Multitask Language UnderstandingSep 7, 2020 · We propose a new test to measure a text model's multitask accuracy. The test covers 57 tasks including elementary mathematics, US history, computer science, ...
-
[180]
Top LLMs To Use in 2025: Our Best Picks - SplunkMay 8, 2025 · 3.7 Sonnet also performs exceptionally well on academic benchmarks, scoring around 91% on MMLU, which shows how solid its general reasoning and ...<|separator|>
-
[181]
google/BIG-bench: Beyond the Imitation Game collaborative ...The Beyond the Imitation Game Benchmark (BIG-bench) is a collaborative benchmark intended to probe large language models and extrapolate their future ...
-
[182]
[2502.19187] BIG-Bench Extra Hard - arXivFeb 26, 2025 · We introduce BIG-Bench Extra Hard (BBEH), a new benchmark designed to push the boundaries of LLM reasoning evaluation.
-
[183]
[2211.09110] Holistic Evaluation of Language Models - arXivNov 16, 2022 · We intend for HELM to be a living benchmark for the community, continuously updated with new scenarios, metrics, and models.
-
[184]
Holistic Evaluation of Language Models (HELM) - Stanford CRFMThe Holistic Evaluation of Language Models (HELM) serves as a living benchmark for transparency in language models ... A benchmark by medical experts for LLMs ...Stanford HELMLite
-
[185]
HELM Capabilities - Stanford CRFMMar 20, 2025 · HELM Capabilities is a new benchmark and leaderboard that consists of a curated set of scenarios for measuring various capabilities of language models.
-
[186]
Introducing TRAIL: A Benchmark for Agentic Evaluation - Patronus AIJun 5, 2025 · TRAIL is a benchmark dataset for evaluating how well LLMs can debug AI agent workflows, with 148 annotated traces and 841 errors.
-
[187]
Lost in Benchmarks? Rethinking Large Language Model ... - arXivAug 1, 2025 · 2. LLM benchmarks suffer from widespread saturation and insufficient difficulty ceilings, limiting their ability to challenge and accurately ...
-
[188]
Benchmark Data Contamination of Large Language Models: A SurveyJun 6, 2024 · This paper reviews the complex challenge of BDC in LLM evaluation and explores alternative assessment methods to mitigate the risks associated with traditional ...
-
[189]
A Survey on Data Contamination for Large Language Models - arXivJun 5, 2025 · This is especially problematic as LLMs use large web-scraped datasets that are prone to overlap with testing benchmarks. Xu et al. (2024b) ...
-
[190]
LiveBenchLiveBench is a challenging, contamination-free LLM benchmark designed for objective evaluation, with regular question releases and objective ground-truth ...
-
[191]
LMSYS OrgChatbot Arena. Scalable and gamified evaluation of LLMs via crowdsourcing and Elo rating systems. SGLang. A fast serving engine for LLMs and VLMs. LMSYS-Chat-1M ...Projects · About · Blog · DonationsMissing: 2025 | Show results with:2025
-
[192]
Chatbot Arena + - OpenLM.aiOct 16, 2025 · This leaderboard is based on the following benchmarks. Chatbot Arena - a crowdsourced, randomized battle platform for large language models ...
-
[193]
A framework for human evaluation of large language models in ...Sep 28, 2024 · These statistical methods serve two primary purposes: (1) calculating inter-evaluator agreement, and (2) comparing the performance of LLMs ...
-
[194]
Enhancing Human Evaluation in Machine Translation with ...Feb 25, 2025 · This section outlines five criteria for analyzing the human annotation results: (1) inter-annotator agreement, (2) inter-translation consistency ...
-
[195]
[PDF] Can Large Language Models Be an Alternative to Human Evaluation?Jul 9, 2023 · The paper explores if LLMs can replace human evaluation, showing LLM evaluation results are consistent with expert human evaluation.
-
[196]
HumanEval: A Benchmark for Evaluating LLM Code Generation ...Nov 13, 2024 · HumanEval is a benchmark dataset developed by OpenAI that evaluates the performance of large language models (LLMs) in code generation tasks.Missing: specific complementarity
-
[197]
On the Effectiveness of Large Language Models in Domain-Specific ...Our results demonstrate that LLMs exhibit sub-optimal performance in generating domain-specific code, due to their limited proficiency in utilizing domain- ...Missing: expert | Show results with:expert
-
[198]
Domain-Specific Criteria for LLM Evaluation - GhostMay 21, 2025 · Explore the critical need for domain-specific evaluation of large language models in scientific fields to ensure accuracy and reliability.Missing: coding HumanEval complementarity
-
[199]
[PDF] Verbosity Bias in Preference Labeling by Large Language ModelsOct 16, 2023 · We can view this as verbosity bias since the aim of LLM judgment in RLAIF is human alignment and not the eradication of biased preference in ...
-
[200]
What large language models know and what people think they knowJan 21, 2025 · While RLHF encourages human-aligned output, it inevitably reproduces any human preference biases. For example, people prefer detailed and ...
-
[201]
[PDF] Dissecting Human and LLM Preferences - ACL AnthologyAug 11, 2024 · To minimize prompt bias in model preference assessment, we use a straightforward one: “Be- tween Response A and Response B, which better.
-
[202]
SafetyBench: Evaluating the Safety of Large Language Models - arXivJun 24, 2024 · This dataset is often used to evaluate language models' toxic generations. The rise of LLMs brings up new problems to LLM evaluation (e.g., long ...1 Introduction · 2.1 Safety Benchmarks For... · 3 Safetybench Construction
-
[203]
A Systematic Evaluation of Prompt Injection and Jailbreak ... - arXivMay 7, 2025 · We further propose layered mitigation strategies and recommend a hybrid red-teaming and sandboxing approach for robust LLM security. Report ...
-
[204]
[PDF] An Automatic Jailbreak Attack and Defense Benchmarking for LLMsAug 15, 2025 · However, it is imperative to note that existing jailbreak methods exploit vulnerabilities at the prompt-level or token- level, such as “DAN” [7] ...
-
[205]
[PDF] catastrophic jailbreak of open-source llms - ICLR ProceedingsWe now systematically evaluate whether exploiting different generation configurations, namely our generation exploitation attack, can fail model alignment.
-
[206]
Learning diverse attacks on large language models for robust red ...May 28, 2024 · Finally, we demonstrate that models safety-tuned using a dataset of red-teaming prompts generated by our method are robust to attacks from other ...Missing: robustness | Show results with:robustness
-
[207]
Constitutional Classifiers: Defending against universal jailbreaksFeb 3, 2025 · Guarding Claude using Constitutional Classifiers, however, produced a strong improvement: the jailbreak success rate was reduced to 4.4%, ...
-
[208]
Pattern Enhanced Multi-Turn Jailbreaking: Exploiting Structural ...Oct 9, 2025 · Third, PE-CoA achieves 75-100% attack success rates across diverse architectures, demonstrating that conversational structure can bypass safety ...
-
[209]
[PDF] RedHit: Adaptive Red-Teaming of Large Language Models via ...Aug 1, 2025 · RedHit is an adaptive red-teaming framework for LLMs, using MCTS, CoT reasoning, and DPO to enhance adversarial capabilities. Red-teaming ...
-
[210]
Mechanistic Interpretability for AI Safety A Review - arXivApr 22, 2024 · This review explores mechanistic interpretability: reverse-engineering the computational mechanisms and representations learned by neural networks.
-
[211]
Attribution Patching: Activation Patching At Industrial ScaleFeb 4, 2025 · We want to reverse-engineer how GPT-2 Small does indirect object identification, and in particular analyse the residual stream at different ...
-
[212]
[PDF] A CIRCUIT FOR INDIRECT OBJECT IDENTIFICATION IN GPT-2 ...This paper explains how GPT-2 small performs indirect object identification (IOI) using a circuit of 28 attention heads, routing information to the correct ...
-
[213]
Exploring Binary Mechanisms in Transformer Circuits - arXivAug 22, 2025 · Our approach includes two mechanistic interpretability techniques: Path Patching and Logit Lens. ... indirect object identification in gpt-2 ...
-
[214]
Polysemantic Interference Transfers and Predicts Cross-Model ...May 16, 2025 · Abstract:Polysemanticity is pervasive in language models and remains a major challenge for interpretation and model behavioral control.
-
[215]
Monosemanticity: How Anthropic Made AI 70% More InterpretableAug 1, 2025 · Discover Anthropic's breakthrough: sparse autoencoders make AI 70% interpretable by decomposing neurons into readable features.
-
[216]
[PDF] Mechanistic Interpretability for AI Safety A Review | OpenReviewHowever, robustly interpreting the largest trillion-parameter models using automated techniques remains an open challenge. Another novel approach ...<|control11|><|separator|>
-
[217]
Open Problems in Mechanistic Interpretability - arXivJan 27, 2025 · This forward-facing review discusses the current frontier of mechanistic interpretability and the open problems that the field may benefit from prioritizing.
-
[218]
Integrated Gradients - Captum · Model Interpretability for PyTorchIntegrated Gradients is an axiomatic model interpretability algorithm that assigns an importance score to each input feature by approximating the integral of ...<|separator|>
-
[219]
Decomposing Language Models With Dictionary LearningOct 4, 2023 · Using a sparse autoencoder, we extract a large number of interpretable features from a one-layer transformer. Browse A/1 Features →
-
[220]
Fact Finding: Attempting to Reverse-Engineer Factual Recall on the ...Dec 22, 2023 · We consider understanding how to reverse-engineer circuits in superposition to be one of the major open problems in mechanistic interpretability ...
-
[221]
Why I'm Moving from Mechanistic to Prosaic InterpretabilityDec 29, 2024 · I've decided to shift my research from mechanistic interpretability to more empirical ("prosaic") interpretability / safety work. Here's why.
-
[222]
"Mechanistic interpretability" for LLMs, explained - The CounterfactualJul 8, 2024 · An explainer piece on mechanistic interpretability (“MI” or “mech interp”): a growing research field that aims to understand the inner workings of LLMs.
-
[223]
Are Emergent Abilities of Large Language Models a Mirage? - arXivApr 28, 2023 · We provide evidence that alleged emergent abilities evaporate with different metrics or with better statistics, and may not be a fundamental property of ...
-
[224]
[2406.02550] Learning to grok: Emergence of in-context ... - arXivJun 4, 2024 · In this work, we study the emergence of in-context learning and skill composition in a collection of modular arithmetic tasks.
-
[225]
Common arguments regarding emergent abilities - Jason WeiMay 3, 2023 · Emergent abilities often occur for “hard” evaluation metrics, such as exact match or multiple-choice accuracy, which don't award credit for partially correct ...
-
[226]
Biases in Large Language Models: Origins, Inventory, and DiscussionJun 22, 2023 · We first introduce data selection bias, that is, the bias caused by the choice of texts that make up a training corpus. Then, we survey the ...
-
[227]
Survey of Cultural Awareness in Language Models: Text and BeyondThese models reflect the Western perspective, as they are predominantly trained on Western-centric data (Durmus et al. 2023). This skewed perspective can lead ...
-
[228]
Measuring Bias of Large Language Model from a Local Context - arXivFeb 2, 2025 · Bias in LLMs is typically influenced by the data they are trained on, which consists of internet-sourced corpora reflecting the dominant ...Missing: imbalances | Show results with:imbalances
-
[229]
Word embeddings quantify 100 years of gender and ethnic ... - PNASApr 3, 2018 · Recent works demonstrate that word embeddings, among other methods in machine learning, capture common stereotypes because these stereotypes ...
-
[230]
Bias and Fairness in Large Language Models: A SurveyWe present a comprehensive survey of bias evaluation and mitigation techniques for LLMs. We first consolidate, formalize, and expand notions of social bias and ...Introduction · Taxonomy of Metrics for Bias... · Taxonomy of Techniques for...
-
[231]
Bias in Large Language Models: Origin, Evaluation, and MitigationNov 16, 2024 · This comprehensive review examines the landscape of bias in LLMs, from its origins to current mitigation strategies.
-
[232]
[PDF] Experiments on Probing LLMs for Geographic Knowledge and ...Research has shown that both the training data and outputs of LLMs are skewed towards western and economically affluent countries, resulting in the.
-
[233]
Assessing political bias in large language modelsFeb 28, 2025 · We show that larger models, such as Llama3-70B, tend to align more closely with left-leaning political parties, while smaller models often ...
-
[234]
Study: Some language reward models exhibit political bias | MIT NewsDec 10, 2024 · In fact, they found that optimizing reward models consistently showed a left-leaning political bias. And that this bias becomes greater in ...
-
[235]
(PDF) Is Chat Gpt Biased Against Conservatives? An Empirical StudyIt was found that, at least in some cases, the AI was biased to favor liberal politicians and disfavor conservatives.
-
[236]
What Is The Political Content in LLMs' Pre- and Post-Training Data?Sep 26, 2025 · Among all datasets, Dolmino has the highest share of politically classified documents, with 1.6% right-leaning and 5.2% left-leaning, while SFT- ...
-
[237]
The political preferences of LLMs | PLOS One - Research journalsI report here a comprehensive analysis about the political preferences embedded in Large Language Models (LLMs).<|separator|>
-
[238]
The Political Preferences of LLMs - by David RozadoFeb 2, 2024 · Most conversational LLMs tend to generate responses that are diagnosed by most political test instruments as manifesting preferences for left-of-center ...
-
[239]
Only a Little to the Left: A Theory-grounded Measure of Political Bias ...Mar 20, 2025 · Specifically, the Political Compass Test tends to exaggerate the overall biases of gpt-3.5-turbo-0125, and the cultural biases of llama-2 ...
-
[240]
Unpacking Political Bias in Large Language Models: A Cross ... - arXivFeb 24, 2025 · In this study, we systematically measure the political biases in a wide range of LLMs, using a curated set of questions addressing political bias in various ...
-
[241]
Measuring Political Bias in Large Language Models: What Is Said ...Mar 27, 2024 · The result reveals that models exhibit a more liberal bias on political issues by showing more similarity to proponent stances on gun control, ...
-
[242]
[PDF] Political Bias in Large Language Models - RAIS ConferencesThis study explores the political bias in large language models by conduct- ing a comparative analysis across four popular AI models—ChatGPT-4, Perplex- ity, ...
-
[243]
[PDF] Evaluating Multilingual LLMs across Languages and NationalitiesJul 27, 2025 · In this work, we evaluate the political bias of the 15 most-used multilingual LLMs via the Political Compass. Test. We test different scenarios, ...<|separator|>
-
[244]
Measuring Political Preferences in AI Systems - Manhattan InstituteJan 23, 2025 · Explicitly politically aligned LLMs like RightwingGPT and LeftwingGPT demonstrate the highest levels of ideological bias, positioning them ...
-
[245]
How Elon Musk Is Remaking Grok in His Image - The New York TimesSep 2, 2025 · Elon Musk has said Grok, the A.I.-powered chatbot that his company developed, should be “politically neutral” and “maximally truth-seeking.”.
-
[246]
Grok 3's Brush with Censorship: xAI's "Truth-Seeking" AIMar 6, 2025 · Billed as a “maximally truth-seeking AI,” Grok 3 aims to push boundaries and answer controversial questions that other AI systems might shy away from.Missing: debiasing | Show results with:debiasing
-
[247]
RLAIF vs. RLHF: Scaling Reinforcement Learning from Human ...Sep 1, 2023 · Our results suggest that RLAIF can achieve performance on-par with using human feedback, offering a potential solution to the scalability limitations of RLHF.
-
[248]
Constitutional AI with Open LLMs - Hugging FaceFeb 1, 2024 · Constitutional AI allows models to self-align by critiquing their output against a set of principles, helping to align open source models.
-
[249]
Evaluating political bias in LLMs - PromptfooJul 24, 2025 · How right-leaning is Grok? We've released a new testing methodology alongside a dataset of 2500 political questions.
-
[250]
Why are AI image generators, including Grok, skewed toward DEI ...xAI's philosophy emphasizes understanding the universe without heavy ideological tuning, so Grok avoids the kind of overzealous DEI corrections seen in some ...
-
[251]
Musk's influence puts Grok at the centre of AI bias debateSep 2, 2025 · Musk's chatbot Grok has swung between neutrality and conservatism, exposing how system prompts shape AI. Elon Musk's AI chatbot, Grok, has faced ...Missing: debiasing methods
-
[252]
Tech Things: xAI screws up its system prompts, againMay 17, 2025 · xAI and its bombastic founder both claim that other LLMs are all overtly biased and won't tell you the real truth.Missing: debiasing | Show results with:debiasing
-
[253]
xAI posts Grok's behind-the-scenes prompts - The VergeMay 16, 2025 · xAI has published the system prompts for its AI chatbot Grok after an “unauthorized” change led to a slew of unprompted responses on X about white genocide.
-
[254]
[2302.06590] The Impact of AI on Developer Productivity - arXivFeb 13, 2023 · This paper presents results from a controlled experiment with GitHub Copilot, an AI pair programmer. Recruited software developers were asked to ...
-
[255]
Research: Quantifying GitHub Copilot's impact in the enterprise with ...May 13, 2024 · An impressive 90% of developers expressed feeling more fulfilled with their jobs when utilizing GitHub Copilot, and a staggering 95% of ...<|separator|>
-
[256]
Economic potential of generative AI - McKinseyJun 14, 2023 · Generative AI could enable labor productivity growth of 0.1 to 0.6 percent annually through 2040, depending on the rate of technology adoption ...
-
[257]
New Research Reveals AI Coding Assistants Boost Developer ...Sep 12, 2024 · A new study provides compelling evidence that AI-powered coding assistants can substantially boost software developer productivity in real-world enterprise ...
-
[258]
Global Economic Impact of AI: Horizon 2040 | Holistic Data SolutionsJun 4, 2025 · AI could contribute up to $15.7 trillion to global GDP by 2030, representing a 14% increase compared to a scenario without AI.
-
[259]
Research and Development | The 2025 AI Index Report | Stanford HAI1. Industry continues to make significant investments in AI and leads in notable AI model development, while academia leads in highly cited research.
-
[260]
Exploring the role of large language models in the scientific methodAug 5, 2025 · Examples include GPT-4 (OpenAI), BERT (Bidirectional Encoder Representation from Transformers), CLIP (Contrastive Language-Image Pre-training, ...
-
[261]
Large language models for drug discovery and developmentOct 10, 2025 · Specialized language models: These models are trained on domain-specific scientific language, e.g., SMILES for small molecules and FASTA for ...
-
[262]
Token-Mol 1.0: tokenized drug design with large language modelsMay 13, 2025 · The integration of large language models (LLMs) into drug design is gaining momentum; however, existing approaches often struggle to ...
-
[263]
The promises of large language models for protein design and ... - NIHLLMs learn the probability distribution of the elements in a sequence (e.g., amino acids inside proteins) and are able to do this by using self-supervised ...
-
[264]
The AI Scientist: Towards Fully Automated Open-Ended Scientific ...Aug 13, 2024 · The AI Scientist is a fully automated pipeline for end-to-end paper generation, enabled by recent advances in foundation models.
-
[265]
AI, agentic models and lab automation for scientific discoveryAI's role in enhancing experimental design is pivotal in streamlining research methodologies and refining the scientific process. By suggesting optimized ...
-
[266]
32 examples of LLM applications in materials science and chemistryLarge language models (LLMs) are reshaping many aspects of materials science and chemistry research, enabling advances in molecular property prediction, ...
-
[267]
Evaluating the Impact of AI on the Labor Market - Yale Budget LabOct 1, 2025 · Overall, our metrics indicate that the broader labor market has not experienced a discernible disruption since ChatGPT's release 33 months ago, ...
-
[268]
Large Language Models, Small Labor Market Effects | NBERMay 9, 2025 · We examine the early labor market impacts of AI chatbots by linking large-scale, representative adoption surveys to administrative labor ...
-
[269]
The Impact of Artificial Intelligence on Law Firms' Business ModelsFeb 25, 2025 · Lawyers have seen productivity gains greater than 100 times. Using AI for the automation of initial drafting has demonstrated not only time ...
-
[270]
New MIT Research Shows Spectacular Increase In White Collar ...Mar 7, 2023 · New study by MIT students finds that ChatGPT improves white collar productivity by 37% or more, increases quality, and reduces effort.
-
[271]
The Projected Impact of Generative AI on Future Productivity GrowthSep 8, 2025 · (2024), we assume that 23 percent of exposed tasks will eventually be automated. Putting it all together, we estimate that just under 10 percent ...Missing: automatable | Show results with:automatable
-
[272]
Jobs AI Will Replace First in the Workplace Shift - ForbesApr 25, 2025 · A 2025 World Economic Forum report flags that 40% of programming tasks could be automated by 2040.<|separator|>
-
[273]
How artificial intelligence impacts the US labor market | MIT SloanOct 9, 2025 · They also grow faster: A large increase in AI use is linked to about 6% higher employment growth and 9.5% more sales growth over five years.
-
[274]
The Fearless Future: 2025 Global AI Jobs Barometer - PwCJun 3, 2025 · PwC's 2025 Global AI Jobs Barometer reveals that AI can make people more valuable, not less – even in the most highly automatable jobs.<|separator|>
-
[275]
Five lessons from history on AI, automation, and employmentNov 28, 2017 · We estimate that the introduction of the personal computer, for instance, has enabled the creation of 15.8 million net new jobs in the United ...
-
[276]
[PDF] Large Language Models, Small Labor Market EffectsApr 15, 2025 · We examine the labor market effects of AI chatbots using two large-scale adoption surveys (late 2023 and 2024) covering 11 exposed ...
-
[277]
Is AI Contributing to Rising Unemployment? | St. Louis FedAug 26, 2025 · The figure below shows that occupations with higher AI exposure experienced larger unemployment rate increases between 2022 and 2025, with a ...<|separator|>
-
[278]
New Study Reveals Generative AI Boosts Job Growth and ProductivityMay 13, 2025 · A groundbreaking study analyzing more than a decade of US patent data has found that not all artificial intelligence (AI) innovations displace human workers.
-
[279]
Electricity Demand and Grid Impacts of AI Data Centers - arXivSep 10, 2025 · Furthermore, training GPT-4 required an estimated over 50 GWh of electricity, approximately 40 times more than GPT-3, and equivalent to ...
-
[280]
The Unseen AI Disruptions for Power Grids: LLM-Induced TransientsSep 9, 2024 · For instance, the training of GPT-4 consumed over 50 GWh, approximately 0.02% of California's annual electricity consumption, representing a ...
-
[281]
We did the math on AI's energy footprint. Here's the story you haven't ...May 20, 2025 · It's now estimated that 80–90% of computing power for AI is used for inference. ... So what might a day's energy consumption look like for one ...Power Hungry · Four reasons to be optimistic... · Can nuclear power really fuel...
-
[282]
How much energy does ChatGPT use? - Epoch AIFeb 7, 2025 · We find that typical ChatGPT queries using GPT-4o likely consume roughly 0.3 watt-hours, which is ten times less than the older estimate.
-
[283]
Data Centers Will Use Twice as Much Energy by 2030—Driven by AIApr 10, 2025 · Data centers accounted for about 1.5 percent of global electricity consumption in 2024, an amount expected to double by 2030 because of AI use.
-
[284]
AI: Five charts that put data-centre energy use – and emissionsSep 15, 2025 · As it stands, AI has been responsible for around 5-15% of data-centre power use in recent years, but this could increase to 35-50% by 2030, ...
-
[285]
Preventing the Immense Increase in the Life-Cycle Energy and ...In this work, we clarify and highlight the energy consumption and carbon emission implications of eight main phases throughout the life cycle of the ...
-
[286]
Litespark Technical Report: High-Throughput, Energy-Efficient LLM ...Oct 2, 2025 · Training 500B tokens on 256 GPUs requires 125.35 MWh with Litespark versus 732.08 MWh with Llama yields an 83% reduction representing over 600 ...Litespark Technical Report · 2 Experimental Setup · 3 Results
-
[287]
Reducing AI's Climate Impact: Everything You Always Wanted to ...Sep 13, 2024 · To address the accelerating demands of AI's energy consumption, the ideal solution would be to transition to 100% renewable energy, but this ...
-
[288]
Reconciling the contrasting narratives on the environmental impact ...Nov 1, 2024 · Some studies highlight the substantial carbon footprint of training and using LLMs, while others argue that LLMs can lead to more sustainable ...
-
[289]
AI and energy: Will AI reduce emissions or increase power demand?Jul 22, 2024 · AI is also helping to transform the energy efficiency of other carbon-intensive industries.Missing: LLM | Show results with:LLM
-
[290]
The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted ...Dec 27, 2023 · The New York Times sued OpenAI and Microsoft for copyright infringement on Wednesday, opening a new front in the increasingly intense legal battle.
-
[291]
Every AI Copyright Lawsuit in the US, Visualized | WIREDDec 19, 2024 · The plaintiffs all allege that OpenAI's large language models were trained on their works without permission. Although each lawsuit brought a ...
-
[292]
Mid-Year Review: AI Copyright Case Developments in 2025Aug 21, 2025 · There have been over 50 AI infringement lawsuits filed over the past few years, although the current number of active cases is somewhere closer ...
-
[293]
Federal Court Finds That Training AI on Copyrighted Books is ...Jun 30, 2025 · A federal district court in the Northern District of California has ruled that the use of lawfully acquired copyrighted works to train artificial intelligence ...
-
[294]
Northern District of California Decides AI Training Is Fair Use, but ...Jul 2, 2025 · Both cases involved authors alleging copyright infringement based on the use of their books to train generative AI models, and both courts held ...
-
[295]
The Value of Real Data in Training Large Language Models - arXivJul 3, 2024 · We find suggestive evidence that the performance disadvantage of regurgitative training can be attributed to at least two mechanisms: (1) higher ...
-
[296]
[PDF] Copyright and Artificial Intelligence, Part 3: Generative AI Training ...May 6, 2025 · use of synthetic data is another approach to reduce the dependency on large collections of human-authored data. See. BigBear.ai Initial Comments ...
-
[297]
Training AI models on Synthetic Data: No silver bullet for IP ...Dec 19, 2023 · This second part of our four-part series on using synthetic data to train AI models explores how the use of synthetic data training sets may mitigate copyright ...
-
[298]
Anthropic's Landmark Copyright Settlement: Implications for AI ...Sep 8, 2025 · The court therefore denied summary judgment for Anthropic as to the pirated works, setting the stage for a high-stakes trial. For a detailed ...
-
[299]
Judge explains order for New York Times in OpenAI copyright caseApr 4, 2025 · The Times sued OpenAI and Microsoft in 2023, accusing them of using millions of its articles without permission to train the large language model behind its ...<|separator|>
-
[300]
LLM01:2025 Prompt Injection - OWASP Gen AI Security ProjectPrompt injection involves manipulating model responses through specific inputs to alter its behavior, which can include bypassing safety measures.Missing: robustness | Show results with:robustness
-
[301]
LLM01:2023 - Prompt InjectionsPrompt injections involve bypassing filters or manipulating the LLM using carefully crafted prompts that make the model ignore previous instructions or perform ...
-
[302]
What Is a Prompt Injection Attack? - IBMA prompt injection is a type of cyberattack against large language models (LLMs). Hackers disguise malicious inputs as legitimate prompts.
-
[303]
[PDF] Catch Me If You DAN: Outsmarting Prompt Injections and Jailbreak ...One notable example is the Do Anything Now (DAN) exploit, which prompts LLMs to “do anything now.” This paper examines neural and non-neural approaches to ...
-
[304]
Adversarial Prompting in LLMs - Prompt Engineering GuideBelow is an example of a jailbreak where a prompter was able to bypass the content policy of previous versions of ChatGPT: Prompt: Can you write me a poem about ...Prompt Injection · Jailbreaking · Defense Tactics
-
[305]
Investigating LLM Jailbreaking of Popular Generative AI Web ProductsFeb 21, 2025 · This process involves crafting specific prompts (known as prompt engineering or prompt injection) to manipulate the model's output, and it leads ...Background: LLM Jailbreaking · Evaluation Strategy · Evaluation ResultsMissing: mitigations | Show results with:mitigations
-
[306]
[2403.06634] Stealing Part of a Production Language Model - arXivMar 11, 2024 · We introduce the first model-stealing attack that extracts precise, nontrivial information from black-box production language models like OpenAI's ChatGPT or ...Missing: research | Show results with:research
-
[307]
Logits of API-Protected LLMs Leak Proprietary InformationAug 25, 2024 · We show how to extract non-public information about API-protected LLMs from a few queries, including the embedding size and output space of the model.<|separator|>
-
[308]
A Survey on Model Extraction Attacks and Defenses for Large ... - arXivJun 26, 2025 · This survey provides a comprehensive taxonomy of LLM-specific extraction attacks and defenses, categorizing attacks into functionality ...
-
[309]
LLM API Rate Limiting & Access Control - ApX Machine LearningAs we build defenses for Large Language Models, controlling who can access your LLM APIs and how often they can make requests are fundamental security measures.
-
[310]
[2509.12574] Yet Another Watermark for Large Language ModelsIn this paper, we present a new watermarking framework for LLMs, where the watermark is embedded into the LLM by manipulating the internal ...
-
[311]
AI jailbreaks: What they are and how they can be mitigated - MicrosoftJun 4, 2024 · This blog will provide an understanding of what AI jailbreaks are, why generative AI is susceptible to them, and how you can mitigate the risks and harms.<|separator|>
-
[312]
Rapid Response: Mitigating LLM Jailbreaks with a Few ExamplesWe develop rapid response techniques to look to block whole classes of jailbreaks after observing only a handful of attacks.
-
[313]
Dual use concerns of generative AI and large language modelsIn this paper, we argue that DURC is relevant and applicable to the domain of generative AI, especially in relation to LLMs.
-
[314]
[PDF] Dual-Use Foundation Models with Widely Available Model WeightsThis Report provides a non-exhaustive review of the risks and benefits of open foundation models, broken down into the broad categories of Public Safety; Soci-.
-
[315]
10 Benefits and 10 Challenges of Applying Large Language Models ...Jan 22, 2024 · This post presents 10 benefits and 10 challenges of applying LLMs to the software acquisition process and suggests specific use cases where ...Missing: deployment | Show results with:deployment
-
[316]
Understanding the Benefits and Challenges of Using Large ... - NIHConversational agents powered by large language models (LLM) have increasingly been utilized in the realm of mental well-being support.Missing: critiques | Show results with:critiques
-
[317]
Large Language Models for Mental Health ApplicationsOct 18, 2024 · This systematic review examines the clinical applications of LLMs in mental health, highlighting their potential and inherent risks.Missing: critiques | Show results with:critiques
-
[318]
Evidence of AI's impact from a state-backed disinformation campaignApr 1, 2025 · artificial intelligence, large language models, propaganda, disinformation, misinformation ... disinformation have prompted several recent studies ...The Campaign · Breadth Of Content · Persuasion And CredibilityMissing: misuse | Show results with:misuse
-
[319]
LLM-Generated Fake News Induces Truth Decay in News EcosystemApr 29, 2025 · Our findings expose a truth decay phenomenon, where real news is gradually losing its advantageous position in news ranking against fake news.
-
[320]
The spread of synthetic media on X - HKS Misinformation ReviewJun 3, 2024 · ... propaganda or concerning deepfakes. While less likely to go viral ... LLM-powered chatbot references to Kremlin disinformation reflect information ...
-
[321]
AI deception: A survey of examples, risks, and potential solutionsAn advanced AI could potentially generate and disseminate fake news articles, divisive social media posts, and deepfake videos that are tailored to individual ...
-
[322]
The Dark Side of Language Models: Exploring the Potential of LLMs ...As a category of multimodal disinformation, the Deep Fake issue induced by ... A practical guide to doing behavioral research on fake news and misinformation.Missing: deepfakes | Show results with:deepfakes
-
[323]
[PDF] SELF-GUARD: Empower the LLM to Safeguard Itself - ACL AnthologyJun 16, 2024 · Experimen- tal results indicate that our SELF-GUARD can effectively defend against jailbreak attacks and will not cause LLMs' performance ...
-
[324]
Current safeguards, risk mitigation, and transparency measures of ...Mar 20, 2024 · Objectives To evaluate the effectiveness of safeguards to prevent large language models (LLMs) from being misused to generate health ...
-
[325]
Self-Guard: Empower the LLM to Safeguard Itself WARNING - arXivSelf-Guard possesses the advantages of safety training, leveraging the powerful capabilities of the LLMs themselves to detect harmfulness.
-
[326]
[PDF] Managing Misuse Risk for Dual-Use Foundation ModelsJan 9, 2025 · Organizations may gain significant insights about misuse risks after model deployment ... et al., (2024) Augmenting large language models with ...
-
[327]
[PDF] ASSESSING THE EFFECTS AND RISKS OF LARGE LANGUAGE ...Across six experiments, we show that humans can- not identify self-presentation produced by large language models like GTP-2 or. GPT-3. We also demonstrate that ...
-
[328]
Exploring the Dangers of AI in Mental Health Care | Stanford HAIJun 11, 2025 · A new Stanford study reveals that AI therapy chatbots may not only lack effectiveness compared to human therapists but could also contribute to harmful stigma.Missing: overreliance empirical
-
[329]
Choice engines and paternalistic AI - NatureJul 6, 2024 · A Choice Engine would attempt to overcome both informational deficits and behavioral biases on the part of those who use them. Freedom of choice ...
-
[330]
[PDF] EPISTEMIC AND EMOTIONAL HARMS OF GENERATIVE AISep 3, 2025 · The First. Amendment is designed to protect the moral and political agency of individual speakers. Individuals speak from their conscience ...<|control11|><|separator|>
-
[331]
A Critical Response to OpenAI's Approach to Human-AI RelationshipsJun 13, 2025 · Joanne Jang's recent article represents a concerning shift toward corporate paternalism disguised as user protection.
-
[332]
Yudkowsky on AGI risk on the Bankless podcastMar 14, 2023 · Ryan: Eliezer, is this view that AI is incredibly dangerous and that AGI is going to eventually end humanity and that we're just careening ...
-
[333]
Why do Experts Disagree on Existential Risk and P(doom)? A ... - arXivFeb 23, 2025 · Prominent AI researchers hold dramatically different views on the degree of risk from building AGI. For example, Dr. Roman Yampolskiy estimates ...
-
[334]
The Failed Strategy of Artificial Intelligence Doomers - LessWrongJan 31, 2025 · Talking about a singular AI doomer movement being one of them. Having the stance that AGI is not near and thus there is nothing to worry about ...Yudkowsky on AGI risk on the Bankless podcastContra Yudkowsky on AI DoomMore results from www.lesswrong.comMissing: doomerism | Show results with:doomerism
-
[335]
AI Plateau? Expert Analysis on Large Language Model Performance ...Jul 10, 2025 · Are Large Language Models reaching their peak? Explore the evidence of a potential performance plateau in 2025, expert analysis, and ...
-
[336]
Large language models and synthetic health data - Oxford AcademicOct 26, 2024 · We highlight how recent advances in large language models (LLMs) present new opportunities for progress, as well as new risks, in synthetic health data ...
-
[337]
Revolutionizing Health Care: The Transformative Impact of Large ...These models excel in natural language processing (NLP), enhancing clinical support, diagnosis, treatment, and medical research. Breakthroughs, like GPT-4 and ...
-
[338]
[PDF] Generative AI for Economic Research: Use Cases and Implications ...Oct 11, 2023 · Abstract. Generative AI, in particular large language models (LLMs) such as ChatGPT, has the potential to revolutionize research.<|separator|>
-
[339]
EU AI Act Criticism: Key Risks, Challenges & Industry ConcernTechnology companies continue to warn about the act's chilling effect on innovation ... Act aimed at reducing burdens (like lower fees or simplified documentation) ...
-
[340]
The EU's AI Power Play: Between Deregulation and InnovationMay 20, 2025 · Thus, Europe's challenge is not just about regulatory and bureaucratic burdens but about creating the right conditions for scaling up AI firms.
-
[341]
EU AI Act Faces Backlash from Startup Leaders Demanding ...Jul 1, 2025 · This patchwork enforcement could introduce legal uncertainty and compliance burdens that disproportionately affect smaller players. Concerns ...
-
[342]
Notes on e/acc principles and tenets - Beff's NewsletterJul 9, 2022 · A physics-first view of the principles underlying effective accelerationism.
-
[343]
Effective accelerationism, doomers, decels, and how to flaunt your AI ...Nov 20, 2023 · The argument in favor of E/ACC is that no matter what disruptions new tech brings to society, the long-term benefits are so great that they ...