Fact-checked by Grok 2 weeks ago
References
-
[1]
On the Opportunities and Risks of Foundation Models - arXivAug 16, 2021 · This report provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities (eg, language, vision, robotics, ...
-
[2]
On the Opportunities and Risks of Foundation ModelsFundamentally, foundation models are a high-leverage single point of failure, making them a prime target for attack: existing work demonstrates a variety of ...Missing: achievements controversies
-
[3]
[PDF] On the Opportunities and Risks of Foundation Models2.6.1 What is a foundation model? There is not a precise technical definition of foundation model. Rather, this is an informal label for a large family of ...
-
[4]
What are Foundation Models? - AiseraRating 8.9/10 (147) Scale: Foundation models are parameter-intensive, often containing billions or trillions of parameters. Parameters include weights and biases in the feedforward ...
-
[5]
What Are Foundation Models? | IBM1 On the Opportunities and Risks of Foundation Models, Stanford Center for Research on Foundation Models and Stanford Institute for Human-Centered ...
-
[6]
What is a foundation model? - Ada Lovelace InstituteJul 17, 2023 · A defining characteristic of foundation models is the scale of data and computational resources involved in building them. They require datasets ...AI technologies and... · Foundation models: applicationsMissing: key | Show results with:key
-
[7]
What are Foundation Models? - DataCampAug 15, 2023 · Narrow AI refers to AI systems designed for specific tasks but which are unable to perform tasks outside their planned scope.What are Foundation Models... · How Do Foundation Models... · Modality
-
[8]
[2005.14165] Language Models are Few-Shot Learners - arXivMay 28, 2020 · GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks ...
-
[9]
[PDF] Can Foundation Models Talk Causality? - OpenReviewIn this work, we argue that foundation models might be exploiting a “loop hole” in the CHT2. Namely, what happens if the causal assumptions (which are required,.
-
[10]
[1706.03762] Attention Is All You Need - arXivJun 12, 2017 · Access Paper: View a PDF of the paper titled Attention Is All You Need, by Ashish Vaswani and 7 other authors. View PDF · HTML (experimental) ...
-
[11]
What are foundation models? | Google CloudFoundation models are AI models trained on massive datasets to perform a wide range of tasks with minimal fine-tuning. Learn more from Google Cloud.<|separator|>
-
[12]
Foundation models: 2022's AI paradigm shift - VentureBeatSep 13, 2022 · 2022 has seen incredible growth in foundation models, or large-scale AI models trained on a massive scale. What does the future hold?
-
[13]
Pathways Language Model (PaLM): Scaling to 540 Billion ...Apr 4, 2022 · We introduce the Pathways Language Model (PaLM), a 540-billion parameter, dense decoder-only Transformer model trained with the Pathways system.
-
[14]
AI Index: State of AI in 13 Charts | Stanford HAIApr 15, 2024 · This past year, organizations released 149 foundation models, more than double the number released in 2022. Of these newly released models, 65.7 ...Biggest Players · Prices Skyrocket · What Ai Race?
-
[15]
The Llama 4 herd: The beginning of a new era of natively ...Apr 5, 2025 · We're introducing Llama 4 Scout and Llama 4 Maverick, the first open-weight natively multimodal models with unprecedented context length support.
-
[16]
Frontier AI regulation: Managing emerging risks to public safetyJul 6, 2023 · In this paper, we focus on what we term “frontier AI” models: highly capable foundation models that could possess dangerous capabilities ...
-
[17]
Frontier AI: capabilities and risks – discussion paper - GOV.UKApr 28, 2025 · Increasingly, frontier AI models are multi-modal. In addition to text, they can generate and process other data types such as images, video, and ...What is the current state of... · What risks do frontier AI present? · Glossary
-
[18]
GPT-4 - OpenAIMar 14, 2023 · GPT-4 is a large multimodal model (accepting image and text inputs, emitting text outputs) that, while less capable than humans in many real-world scenarios,
-
[19]
OpenAI announces GPT-4, says beats 90% of humans on SAT - CNBCMar 14, 2023 · GPT-4 performed at the 90th percentile on a simulated bar exam, the 93rd percentile on an SAT reading exam, and the 89th percentile on the SAT ...
-
[20]
Introducing the next generation of Claude - AnthropicClaude 3 model family · A new standard for intelligence · Near-instant results · Strong vision capabilities · Fewer refusals · Improved accuracy.
-
[21]
[PDF] The United States Artificial Intelligence Safety Institute: Vision ...May 21, 2024 · AISI's research, testing, and guidance will enable more rigorous assessment of AI risk; more effective internal and external safeguards for. AI ...
-
[22]
Artificial Intelligence Safety Institute Consortium (AISIC) | NISTThe Consortium brings together more than 280 organizations to develop science-based and empirically backed guidelines and standards for AI measurement.
-
[23]
General-Purpose AI Models in the AI Act – Questions & AnswersJul 10, 2025 · General-purpose AI models are trained with large data, display significant generality, perform many tasks, and can be integrated into various ...
-
[24]
[PDF] General Purpose AI and the AI ActThese systems are sometimes referred to as 'foundation models' and are characterised by their widespread use as pre-trained models for other, more specialised ...
-
[25]
What Are Generative AI, Large Language Models, and Foundation ...May 12, 2023 · This post aims to clarify what each of these three terms mean, how they overlap, and how they differ.Missing: narrow | Show results with:narrow
-
[26]
ARC-AGI-2 A New Challenge for Frontier AI Reasoning SystemsMay 20, 2025 · The design goals of ARC-AGI-2 are intended to improve upon the limitations of ARC-AGI-1 and expand the depth and quantity of its datasets. Here ...
-
[27]
OpenAI o3 Breakthrough High Score on ARC-AGI-PubDec 20, 2024 · ARC-AGI serves as a critical benchmark for detecting such breakthroughs, highlighting generalization power in a way that saturated or less ...
-
[28]
AI Models Struggle with New ARC-AGI-2 Benchmark ... - MediumMar 25, 2025 · Unlike previous benchmarks, ARC-AGI-2 prevents models from brute-forcing their way through problems using vast computational resources; it ...Missing: limitations | Show results with:limitations
-
[29]
EU AI Act News: Rules on General-Purpose AI Start ... - Mayer BrownAug 1, 2025 · Obligations relating to general-purpose artificial intelligence (“GPAI”) models under the EU AI Act enter into force on 2 August 2025.
-
[30]
General-Purpose Artificial Intelligence (GPAI) Models and ... - RANDAug 8, 2024 · Under the EU AI Act, most foundation models are categorized as general-purpose AI (GPAI) and have special requirements imposed on them in ...
-
[31]
General-purpose AI regulation and the European Union AI ActAug 1, 2024 · The future-proofing of the AI Act needs to focus specifically on general-purpose AI and foundation models, as these types of AI are the most ...
-
[32]
The Failed Strategy of Artificial Intelligence DoomersJan 31, 2025 · The AI Doomers' plans are based on an urgency which is widely assumed but never justified. For many of them, the urgency leads to a rush to do ...
-
[33]
Genie 3: A new frontier for world models - Google DeepMindAug 5, 2025 · Genie 3 is our first world model to allow interaction in real-time, while also improving consistency and realism compared to Genie 2.
-
[34]
RT-2: New model translates vision and language into actionJul 28, 2023 · RT-2 is a novel vision-language-action (VLA) model that learns from both web and robotics data, and translates this knowledge into generalised instructions for ...
-
[35]
[2307.15818] RT-2: Vision-Language-Action Models Transfer Web ...We study how vision-language models trained on Internet-scale data can be incorporated directly into end-to-end robotic control to boost generalization.
-
[36]
Harvard and MIT Study: AI Models Are Not Ready to Make Scientific ...Jul 15, 2025 · They concluded that AI models make accurate predictions but fail to encode the world model of Newton's laws and instead resort to case-specific ...
-
[37]
Foundation models are going multimodal - Twelve LabsMar 31, 2023 · In 2021, OpenAI introduced CLIP (Contrastive Language–Image Pre-training). The input to CLIP is 400 million image-text pairs that were crawled ...
-
[38]
[2507.12496] FOUNDER: Grounding Foundation Models in World ...Jul 15, 2025 · FOUNDER integrates Foundation Models (FMs) and World Models (WMs) for open-ended task solving, using a mapping function to ground FM ...
-
[39]
[2401.04088] Mixtral of Experts - arXivJan 8, 2024 · We introduce Mixtral 8x7B, a Sparse Mixture of Experts (SMoE) language model. Mixtral has the same architecture as Mistral 7B, with the difference that each ...
-
[40]
Mixtral of experts - Mistral AIDec 11, 2023 · Mixtral is a sparse mixture-of-experts network. It is a decoder-only model where the feedforward block picks from a set of 8 distinct groups of parameters.
-
[41]
Large language model data pipelines and Common Crawl (WARC ...Jun 3, 2023 · This article provides a short introduction to the pipeline used to create the data to train LLaMA, but it allows for many variations.
-
[42]
Training Data for the Price of a Sandwich - Mozilla FoundationFeb 6, 2024 · While Common Crawl was never primarily about providing AI training data, it now positions itself as an important building block for LLM ...
-
[43]
How to Ensure Sufficient Data for AI Foundation ModelsJan 8, 2024 · As large language models (LLMs), Meta's LLaMA has 65 billion parameters and 4.5 TB of training data, while OpenAI's GPT-3.5 has 175 billion ...
-
[44]
Open-Sourced Training Datasets for Large Language Models (LLMs)The Common Crawl dataset comprises terabytes of raw web data extracted from billions of web pages. It releases new data files that the crawler obtains each ...
-
[45]
Datasets used for training LLM's: All types of data used to create ...Aug 21, 2025 · Major Sources of LLM Training Data · 1. Books · 2. Websites · 3. Articles and Journals · 4. Conversations and Dialogue · 5. Common Crawl · 6.
-
[46]
The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted ...Dec 27, 2023 · The New York Times sued OpenAI and Microsoft for copyright infringement on Wednesday, opening a new front in the increasingly intense legal battle.
-
[47]
Mastering LLM Techniques: Text Data Processing - NVIDIA DeveloperNov 13, 2024 · Text processing for LLMs includes: download, cleaning, heuristic filtering, deduplication, model-based filtering, and blending/shuffling.
-
[48]
In search of the next generation of training sets for language modelsJun 17, 2024 · Participants in the DCLM benchmark can experiment with data curation strategies such as deduplication, filtering, and data mixing at model ...Missing: large | Show results with:large
-
[49]
Curating Non-English Datasets for LLM Training with NVIDIA NeMo ...Jul 10, 2024 · Heuristic filtering helps remove low-quality content from the dataset, using simple, efficient-to-compute rules. By applying well-designed ...
-
[50]
[2411.15821] Is Training Data Quality or Quantity More Impactful to ...Nov 24, 2024 · This study investigates the relative impact of training data quality versus quantity on the performance of small language models (SLMs)<|control11|><|separator|>
-
[51]
AI models collapse when trained on recursively generated dataJul 24, 2024 · Model collapse is a degenerative process affecting generations of learned generative models, in which the data they generate end up polluting the training set ...
-
[52]
[2404.01413] Is Model Collapse Inevitable? Breaking the Curse of ...Apr 1, 2024 · We confirm that replacing the original real data by each generation's synthetic data does indeed tend towards model collapse.
-
[53]
Distributed Training: Guide for Data ScientistsDistributed training divides training workload across multiple processors, running subtasks in parallel to speed up deep learning model training.
-
[54]
Training and Serving System of Foundation Models - arXivThis survey explores methods for training and serving foundation models, which are built on data, models, computing, and algorithms, and face challenges in ...2.2 Transformer For... · 3 Model Training · 4 Model Serving<|separator|>
-
[55]
GPT-4 Details Revealed - by Patrick McGuinnessJul 12, 2023 · OpenAI's pre-training for GPT-4 required about 2.15 x 10^25 FLOPS. This meant running on 25,000 A100s for 90 to 100 days, with a total pre- ...
-
[56]
How Do GPUs and TPUs Differ in Training Large Transformer ...Aug 25, 2025 · TPUs outperform GPUs for massive batch processing and models directly compatible with their architecture, including most TensorFlow-based LLMs ...
-
[57]
How much power will frontier AI training demand in 2030? - Epoch AIAug 11, 2025 · The power required to train the largest frontier models is growing by more than 2x per year, and is on trend to reaching multiple gigawatts ...Missing: shortages Neptune.
-
[58]
[PDF] AI's Power Requirements Under Exponential Growth - RANDWe find that globally, AI data centers could need ten gigawatts (GW) of additional power capacity in 2025 alone, which is more than the total power capacity of ...Missing: Neptune. | Show results with:Neptune.
-
[59]
Deep Learning Model Optimization Methods - Neptune.aiPruning reduces model size by removing less important neurons, involving identification, elimination, and optional fine-tuning. · Quantization decreases memory ...
-
[60]
[2307.02973] Pruning vs Quantization: Which is Better? - arXivJul 6, 2023 · Our results show that in most cases quantization outperforms pruning. Only in some scenarios with very high compression ratio, pruning might be beneficial.
-
[61]
[2001.08361] Scaling Laws for Neural Language Models - arXivJan 23, 2020 · View a PDF of the paper titled Scaling Laws for Neural Language Models, by Jared Kaplan and 9 other authors. View PDF. Abstract:We study ...
-
[62]
Training Compute-Optimal Large Language Models - arXivMar 29, 2022 · As a highlight, Chinchilla reaches a state-of-the-art average accuracy of 67.5% on the MMLU benchmark, greater than a 7% improvement over Gopher ...
-
[63]
The Race to Efficiency: A New Perspective on AI Scaling Laws - arXivJan 4, 2025 · Empirical trends suggest that sustained efficiency gains can push AI scaling well into the coming decade, providing a new perspective on the ...
-
[64]
Current AI scaling laws are showing diminishing returns, forcing AI ...Nov 20, 2024 · AI. Current AI scaling laws are showing diminishing returns, forcing AI labs to change course. Maxwell Zeff. 6:00 AM PST · November 20, 2024.Missing: empirical | Show results with:empirical
-
[65]
Algorithmic Improvement Is Probably Faster Than Scaling NowJun 5, 2023 · Back in 2020, a group at OpenAI ran a conceptually simple test to quantify how much AI progress was attributable to algorithmic improvements.Anthropic: Core Views on AI Safety: When, Why, What, and How ...Revisiting algorithmic progress - LessWrongMore results from www.lesswrong.comMissing: ignoring | Show results with:ignoring
-
[66]
What if A.I. Doesn't Get Much Better Than This? | The New YorkerAug 12, 2025 · Orion's failure helped cement the creeping fear within the industry that the A.I. scaling law wasn't a law after all. If building ever-bigger ...
-
[67]
[2501.13787] Parameter-Efficient Fine-Tuning for Foundation ModelsJan 23, 2025 · PEFT is a cost-effective fine-tuning technique that minimizes parameters and computational complexity while striving for optimal downstream ...
-
[68]
What is parameter-efficient fine-tuning (PEFT)? - IBMPEFT is a method of improving the performance of pretrained large language models (LLMs) and neural networks for specific tasks or data sets.Overview · How does parameter-efficient...
-
[69]
Zero-Shot Prompting - Prompt Engineering GuideThe zero-shot prompt directly instructs the model to perform a task without any additional examples to steer it. We tried a few zero-shot examples in the ...
-
[70]
What is zero-shot prompting? - IBMIn zero-shot prompting, the model is prompted to generate a response without receiving an example of the desired output for the use case. Zero-shot prompting is ...What is zero-shot prompting? · Zero-shot prompting vs few...
-
[71]
PEFT - Hugging FacePEFT (Parameter-Efficient Fine-Tuning) is a library for efficiently adapting large pretrained models to various downstream applications.
-
[72]
LoRA: Low-Rank Adaptation of Large Language Models - arXivJun 17, 2021 · We propose Low-Rank Adaptation, or LoRA, which freezes the pre-trained model weights and injects trainable rank decomposition matrices into each layer of the ...
-
[73]
A Survey on Parameter-Efficient Fine-Tuning for Foundation Models ...Apr 29, 2025 · This survey provides a comprehensive review of the integration of PEFT techniques within federated learning environments.
-
[74]
Fine Tuning Large Language Model for Secure Code Generation... fine-tuned models. Our experiments on GPT-J show that the fine-tuned GPT-J achieved 70.4% and 64.5% ratios of non-vulnerable code generation for C and C++ ...
-
[75]
[2308.08747] An Empirical Study of Catastrophic Forgetting in Large ...Aug 17, 2023 · This study empirically evaluates the forgetting phenomenon in LLMs' knowledge during continual instruction tuning from the perspectives of domain knowledge, ...
-
[76]
What is Catastrophic Forgetting? - IBMCatastrophic forgetting happens when the training process for the new tasks interferes with the model's understanding of old tasks.
-
[77]
GLUE BenchmarkThe General Language Understanding Evaluation (GLUE) benchmark is a collection of resources for training, evaluating, and analyzing natural language ...SuperGLUE Benchmark · GLUE Diagnostic Dataset · Leaderboard · TasksMissing: foundation | Show results with:foundation
-
[78]
SuperGLUE: A Stickier Benchmark for General-Purpose Language ...May 2, 2019 · In this paper we present SuperGLUE, a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, a software toolkit, and a ...
-
[79]
SuperGLUE BenchmarkA new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard.SuperGLUE Diagnostic Dataset · Leaderboard · Tasks · FAQMissing: foundation | Show results with:foundation
-
[80]
[PDF] arXiv:2206.07682v2 [cs.CL] 26 Oct 2022Oct 26, 2022 · Figure 2 shows eight such emergent abilities spanning five language model families from various work. BIG-Bench. Figure 2A–D depicts four ...
-
[81]
Are Emergent Abilities of Large Language Models a Mirage? - arXivApr 28, 2023 · Recent work claims that large language models display emergent abilities, abilities not present in smaller-scale models that are present in ...
-
[82]
Benchmarking is Broken - Don't Let AI be its Own Judge - arXivOct 15, 2025 · Many scholars have raised concerns about the limitations of AI benchmarking, with some describing current evaluation practices as a “minefield” ...
-
[83]
NLP Evaluation in trouble: On the Need to Measure LLM Data ...This position paper defines different levels of data contamination and argues for a community effort, including the development of automatic and semi-automatic ...
-
[84]
A Survey on Data Contamination for Large Language Models - arXivJun 5, 2025 · In this paper, we first examine the definition and impacts of data contamination. Secondly, we review methods for contamination-free evaluation.
-
[85]
Benchmark Data Contamination of Large Language Models: A SurveyJun 6, 2024 · This paper reviews the complex challenge of BDC in LLM evaluation and explores alternative assessment methods to mitigate the risks associated with traditional ...
-
[86]
The ARC Benchmark: Evaluating LLMs' Reasoning AbilitiesThe ARC Benchmark challenges AI with puzzles that measure reasoning, not recall. In 2025, GPT, Claude, and Gemini still lag behind human performance ...
-
[87]
System 2 Reasoning for Human-AI Alignment: Generality and ... - arXivAug 13, 2025 · We examine weaknesses on ARC-AGI tasks, revealing gaps in compositional generalization and novel-rule adaptation, and argue that closing these ...
-
[88]
Nvidia dominates GPU shipments with 94% share - Tom's HardwareSep 3, 2025 · The total number of GPUs sold for the second quarter of 2025 hit 11.6 million units, while desktop PC CPUs went up to 21.7 million units ...
-
[89]
Nvidia and TSMC produce the first Blackwell wafer made in the U.S.Oct 18, 2025 · Nvidia and TSMC have produced the first Blackwell wafer at TSMC's Arizona fab, marking a historic step in bringing advanced AI chip ...
-
[90]
Exclusive: Nvidia and TSMC unveil first Blackwell chip wafer ... - AxiosOct 17, 2025 · Nvidia and TSMC announced Friday their first completed U.S.-made wafer that will eventually become Blackwell chips for AI purposes, Nvidia first ...
-
[91]
Overly Stringent Export Controls Chip Away at American AI LeadershipMay 5, 2025 · The Biden administration issued its first AI chip export controls in October 2022, restricting the export of AI chips to China, as well as the technology to ...
-
[92]
Jensen says Nvidia's China AI GPU market share has plummeted ...the Chinese market previously amounted to 20% to 25% of the ...
-
[93]
U.S. Export Controls and China: Advanced SemiconductorsSep 19, 2025 · Tables. Table A-1. Select Details of U.S. Advanced Chip Controls on China (2022-2025); Table A-2. Nvidia's Modified Chips for China. Appendixes.
-
[94]
What is Hugging Face? - IBMThe Hugging Face Hub is a central web-based platform where users can share, discover and collaborate on models, datasets and applications. It acts like a " ...
-
[95]
AI's Power Requirements Under Exponential Growth - RANDJan 28, 2025 · AI data centers could need ten gigawatts (GW) of additional power capacity in 2025, which is more than the total power capacity of the state of Utah.
- [96]
-
[97]
Securing America's Critical Minerals SupplyOct 8, 2025 · ... rare earths poses the main bottleneck in securing this supply chain. ... AI Action Plan from July 2025. 10 U.S. factories for lithium-ion ...<|control11|><|separator|>
-
[98]
Surveying the Future of U.S. Open Foundation Model Policy - CSISMar 21, 2024 · How might foreign adversaries use open weight models to exacerbate security risks? And how do those kinds of marginal risks compare to closed ...
-
[99]
Open-Source AI is a National Security Imperative - Third WayJan 30, 2025 · In this paper, we explore the benefits and drawbacks of open-source AI and conclude that open-source can help balance the safety and security we want from AI.What Is ``open'' Ai? · Androids 'r' Us · Ai: Made In America
-
[100]
Open Release of Grok-1 - xAIMar 17, 2024 · This is the raw base model checkpoint from the Grok-1 pre-training phase, which concluded in October 2023.
-
[101]
Defense Priorities in the Open-Source AI Debate - CSISAug 19, 2024 · Because open models can be retrained, fine-tuned, and broadly customized, they can serve as a basis for national-security-specific applications.
-
[102]
The Murky State of Frontier AI TransparencyJan 16, 2025 · Our analysis reveals four critical problems: closed models remain largely opaque about technical details, documentation is failing to keep pace with rapid ...
-
[103]
With Open Source Artificial Intelligence, Don't Forget the Lessons of ...Jul 29, 2024 · At CISA, we see significant value in open foundation models to help strengthen cybersecurity, increase competition, and promote innovation.Missing: iteration | Show results with:iteration
-
[104]
Can machine translation match human expertise? Quantifying the ...Jul 25, 2025 · Our findings suggest that large language models provide high-quality PROM translations to support human translations to reduce costs. However, ...
-
[105]
Sora: Creating video from text - OpenAIFeb 15, 2025 · Introducing Sora, our text-to-video model. Sora can generate videos up to a minute long while maintaining visual quality and adherence to the user's prompt.
-
[106]
Sora 2 is here | OpenAISep 30, 2025 · Sora 2 is here. Our latest video generation model is more physically accurate, realistic, and more controllable than prior systems. It also ...Settings · Launching Responsibly · Sora 2 Availability And...
-
[107]
Foundation Models in Robotics: Applications, Challenges, and the ...We explore how foundation models contribute to improving robot capabilities in the domains of perception, decision-making, and control.
-
[108]
Foundation Model Driven Robotics: A Comprehensive Review - arXivJul 14, 2025 · In summary, foundation models have made robot planning more general and flexible.Missing: achievements | Show results with:achievements
-
[109]
Large language models surpass human experts in predicting ...Nov 27, 2024 · Pre-trained LLMs can provide a foundation for further training in neuroscience with the aim of improving performance, as assessed by BrainBench.
-
[110]
[PDF] State of Foundation Models - 2025 (Innovation Endeavors)Scaling continues across all dimensions – All technical metrics for models continue to improve >10x year-over-year, including cost, intelligence, context ...Missing: milestones | Show results with:milestones
-
[111]
[2302.06590] The Impact of AI on Developer Productivity - arXivFeb 13, 2023 · Generative AI tools hold promise to increase human productivity. This paper presents results from a controlled experiment with GitHub Copilot, an AI pair ...<|separator|>
-
[112]
Research: Quantifying GitHub Copilot's impact in the enterprise with ...May 13, 2024 · We conducted research with developers at Accenture to understand GitHub Copilot's real-world impact in enterprise organizations.
-
[113]
Method of the Year 2021: Protein structure prediction - NatureJan 11, 2022 · In the past year, the deep-learning-based methods AlphaFold2 and RoseTTAfold have managed to achieve this feat over a range of targets, forever ...
-
[114]
AlphaFold Protein Structure DatabaseIn CASP14, AlphaFold was the top-ranked protein structure prediction method by a large margin, producing predictions with high accuracy. While the system still ...AlphaFold · View protein · Downloads · P70490Missing: achievements | Show results with:achievements
-
[115]
quantifying GitHub Copilot's impact on developer productivity and ...Sep 7, 2022 · In our research, we saw that GitHub Copilot supports faster completion times, conserves developers' mental energy, helps them focus on more satisfying work.
-
[116]
Why language models hallucinate | OpenAISep 5, 2025 · While evaluations themselves do not directly cause hallucinations, most evaluations measure model performance in a way that encourages guessing ...
-
[117]
Detecting hallucinations in large language models using semantic ...Jun 19, 2024 · From each biography generated by GPT-4, we automatically extract propositional factual claims about the individual (150 factual claims in total ...<|separator|>
-
[118]
[PDF] Understanding Inductive Bias in the Era of Large-Scale Pretraining ...This thesis challenges the conventional wisdom that strict architectural constraints are necessary for modeling numerical data, par- ticularly in physics and ...
-
[119]
Edge-First Language Model Inference: Models, Metrics, and TradeoffsMay 22, 2025 · The challenge with deploying SLMs on mobile and edge devices lies in their limited computing capability, which directly restricts the loading of ...
-
[120]
A survey of edge efficient LLMs and techniques - ScienceDirectThis survey provides a comprehensive overview of the state-of-the-art techniques and strategies for enabling efficient inference of LLMs on edge devices.<|control11|><|separator|>
-
[121]
The Sequence Opinion #485: What's Wrong With AI BenchmarksFeb 6, 2025 · As AI models advance, many benchmarks have become obsolete. In 2024, several key benchmarks saw near-perfect scores from leading models:.
-
[122]
MMLU-Pro: A More Robust and Challenging Multi-Task Language ...Jun 3, 2024 · This paper introduces MMLU-Pro, an enhanced dataset designed to extend the mostly knowledge-driven MMLU benchmark by integrating more challenging, reasoning- ...Missing: saturation | Show results with:saturation
-
[123]
MMLU-Pro: A More Robust and Challenging Multi-Task Language ...By incorporating more complex, reasoning-intensive tasks, MMLU-Pro addresses the performance saturation observed in previous benchmarks, effectively ...
- [124]
-
[125]
The big AI story right now: Pure scaling has failed to produce AGIFeb 19, 2025 · The most underreported and important story in AI right now is that pure scaling has failed to produce AGI.Missing: AIGuys | Show results with:AIGuys
- [126]
-
[127]
Martin Henz - ARCprize: How LLMs fail on ARC benchmark - LinkedInJul 3, 2024 · ARCprize: How LLMs fail on ARC benchmark ... LLMs perform very poorly on the ARC benchmark: https://arcprize.org. The inventor of ARC, Francois ...Missing: failures emergent
-
[128]
Emergent Abilities in Large Language Models: A Survey - arXivFeb 28, 2025 · Emergent abilities appear abruptly when a critical scale is reached rather than via smooth extrapolation. LLMs exhibit sudden performance jumps ...
-
[129]
ARC-AGI-1: Abstract Reasoning Benchmark - Emergent MindSep 16, 2025 · ... failures of prior search attempts. SOAR exemplifies this, combining LLM guided evolutionary search with automatic hindsight learning ...
-
[130]
Open Problems and Fundamental Limitations of Reinforcement ...Jul 27, 2023 · Our work emphasizes the limitations of RLHF and highlights the importance of a multi-faceted approach to the development of safer AI systems.Missing: jailbreaks 2023-2025
-
[131]
[PDF] Open Problems and Fundamental Limitations of Reinforcement ...RLHF has also not made models robust to adversarial attacks from jailbreaking (i.e., subverting the constraints the system is normally meant to operate under) ...Missing: 2023-2025 | Show results with:2023-2025
-
[132]
[PDF] Weak-to-Strong Jailbreaking on Large Language Models - arXivIn this paper, we propose the weak-to- strong jailbreaking attack, an efficient inference time attack for aligned LLMs to produce harmful text. Our key ...
-
[133]
Semantic Jailbreaks and RLHF Limitations in LLMsAug 2, 2025 · In this paper, various production scale model responses have been evaluated against encoded and cleverly paraphrased, obfuscated, ...
-
[134]
Human performance in detecting deepfakes: A systematic review ...Deepfake technology has been misused to create pornographic content containing a fake version of a real, often famous, person. Around 4000 celebrities have ...
-
[135]
[PDF] On the Societal Impact of Open Foundation Models - arXivFeb 27, 2024 · Open foundation models have benefits like innovation and distributed power, but also risks such as misuse, biosecurity, and cybersecurity ...
-
[136]
[PDF] Open-Sourcing Highly Capable Foundation Models - arXivSep 29, 2023 · Open-sourcing AI models offers benefits like oversight and progress, but also risks such as misuse and potential for dangerous AI diffusion.
-
[137]
AI deception: A survey of examples, risks, and potential solutionsThis paper argues that a range of current AI systems have learned how to deceive humans. We define deception as the systematic inducement of false beliefs.
-
[138]
A Closer Look at the Existing Risks of Generative AI - arXivMay 28, 2025 · Through a systematic analysis of 499 publicly reported incidents, we describe what harms are reported, how they arose, and who they impact. We ...Missing: rates | Show results with:rates
-
[139]
Ten Ways the Precautionary Principle Undermines Progress in ...Feb 4, 2019 · If policymakers apply the “precautionary principle” to AI, which says it's better to be safe than sorry, they will limit innovation and discourage adoption.
-
[140]
The Precautionary Principle, Safety Regulation, and AI: This Time, It ...Sep 4, 2024 · The PP has long been important in managing risks associated with technological innovations that have no explicit scientific knowledge.
-
[141]
What drives the divide in transatlantic AI strategy? - Atlantic CouncilSep 29, 2025 · The US and EU share AI ambitions but diverge on regulation, risking a fractured Western front. Nowhere is this tension sharper than in ...<|separator|>
-
[142]
How Europe's AI Act could affect innovation and competitivenessJul 4, 2024 · We caught up with ESCP's Philip Meissner to assess the impact of the EU AI Act on the broader political and economic landscape.
-
[143]
High-level summary of the AI Act | EU Artificial Intelligence ActOn 18 July 2025, the European Commission published draft Guidelines clarifying key provisions of the EU AI Act applicable to General Purpose AI (GPAI) models.
-
[144]
European Commission publishes guidelines on obligations for ...Jul 24, 2025 · The Guidelines clarify that providers placing GPAI models on the market before August 2, 2025 have until August 2, 2027 to comply with their ...
-
[145]
EU AI Act Criticized for Granting US Tech Firms Excessive InfluenceJul 6, 2025 · Strategic uncertainty from the AI Act's vague implementation affects EU research, potentially diverting talent and investment overseas.
-
[146]
Safe, Secure, and Trustworthy Development and Use of Artificial ...Nov 1, 2023 · It is the policy of my Administration to advance and govern the development and use of AI in accordance with eight guiding principles and priorities.
-
[147]
Executive Order on the Safe, Secure, and Trustworthy Development ...Oct 30, 2023 · It is the policy of my Administration to advance and govern the development and use of AI in accordance with eight guiding principles and priorities.
-
[148]
Dual-Use Foundation Models with Widely Available Model Weights ...Jul 30, 2024 · In October 2023 President Biden signed the Executive Order (EO) on “Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence ...
-
[149]
[PDF] America's AI Action Plan - The White HouseJul 10, 2025 · President Trump has already taken multiple steps toward this goal, including rescinding Biden Executive Order 14110 on AI that foreshadowed an ...
-
[150]
AI Doomers Versus AI Accelerationists Locked In Battle For Future ...Feb 18, 2025 · AI is advancing rapidly. AI doomers say we must stop and think. AI accelerationists say full speed ahead. Here is a head-to-head comparison.
-
[151]
Effective accelerationism, doomers, decels, and how to flaunt your AI ...Nov 20, 2023 · For the AI doomers, the decels are akin to attempts to build a third American political party; ineffective at best and insidious at worst. AI ...
-
[152]
Two warring visions of AI - Prospect MagazineJan 16, 2024 · The power struggle between “doomers” and “accelerationists” will define the way this world-changing technology evolves.
-
[153]
The 2025 AI Index Report | Stanford HAIIn 2024, U.S.-based institutions produced 40 notable AI models, significantly outpacing China's 15 and Europe's three. While the U.S. maintains its lead in ...
-
[154]
The US Is Winning the AI Race - But for How Long? - Project SyndicateSep 24, 2025 · On January 20, DeepSeek unveiled its hyper-efficient R1 model, making it clear that US sanctions would not hold back China's AI ambitions.
-
[155]
How will AI influence US-China relations in the next 5 years?Jun 18, 2025 · Already, the performance gap between the best Chinese and U.S. AI models had shrunk from 9.3% in 2024 to 1.7% in February. This will be the new ...
-
[156]
China now leads the U.S. in this key part of the AI raceOct 13, 2025 · China's open AI models are now more powerful and popular than those released by American rivals, a shift with implications for the future of ...Missing: foundation | Show results with:foundation
-
[157]
Unleash developer productivity with generative AI - McKinseyJun 27, 2023 · A McKinsey study shows that software developers can complete coding tasks up to twice as fast with generative AI. Four actions can maximize productivity and ...
-
[158]
Measuring the Impact of Early-2025 AI on Experienced ... - METRJul 10, 2025 · We conduct a randomized controlled trial (RCT) to understand how early-2025 AI tools affect the productivity of experienced open-source developers.Missing: benefits | Show results with:benefits
-
[159]
The reality of AI-Assisted software engineering productivityAug 16, 2025 · Voege concedes that “AI helps with boilerplate” and routine coding, estimating perhaps a 20–50% speed-up on certain sub-tasks for many engineers ...
-
[160]
AI Adoption Statistics in 2025 - NetguruSep 4, 2025 · Generative AI usage specifically has jumped from 33% in 2023 to 71% in 2024, showing how quickly businesses have gained confidence in these ...<|separator|>
-
[161]
Enterprise transformation and extreme productivity with AI | IBMAug 26, 2025 · How AI and automation enabled USD 4.5 billion in productivity gains across IBM.
-
[162]
AI in Action 2024 Report - IBMThe AI in Action report brings real-word data to the AI conversation to explore a simple question—during this critical period, what lessons can be learned ...Missing: transformation | Show results with:transformation
-
[163]
Foundation Models Explained: Why Every Startup and Enterprise ...Jun 13, 2025 · This scalability makes foundation models ideal for platformization, one model, many services. 4. Emergent Abilities and Compositional Reasoning.
-
[164]
(PDF) Foundation Models and AI Innovation: Evidence from the ...Oct 17, 2025 · Less experienced developers engage in broader exploration across various downstream applications, benefiting from lower entry barriers. These ...
-
[165]
5 Top Benefits of Foundation Models - DeepchecksJul 22, 2025 · This article will explore the top five benefits of foundation models that enable developers to harness their powerful features across a variety of applications.Superior Generalization... · Breakthroughs in Multimodal... · Democratization of AI...
-
[166]
AI and Labor Markets: What We Know and Don't KnowOct 14, 2025 · Hosseini and Lichtinger (2025) find employment declines for young workers in US firms that adopt AI technologies, with adoption measured by the ...Missing: foundation | Show results with:foundation
-
[167]
[PDF] CHAPTER 4: Economy - Stanford HAIOct 17, 2024 · In 2024,. 15.7% of all AI job postings in the United. States were for jobs based in California, followed by Texas (8.8%) and New York. (5.8%).
-
[168]
How artificial intelligence impacts the US labor market | MIT SloanOct 9, 2025 · What you'll learn: AI adoption leads to increased company growth in revenue, profits, employment, and profitability.Missing: foundation disruptions 2024
-
[169]
Artificial intelligence, job quality and inclusiveness - OECDJul 11, 2023 · The available evidence for workers with AI skills in the United States suggests that they receive a significant wage premium. Alekseeva et al. ( ...
-
[170]
[PDF] AI Adoption and Wage Growth in U.S. Industries: A Sectoral AnalysisAug 27, 2025 · AI raises wages in tech/finance sectors where it complements humans, but keeps wages down in labor-intensive sectors where it replaces them. ...Missing: foundation | Show results with:foundation
-
[171]
Artificial intelligence and the skill premium: A numerical analysis of ...The results show that AI widens the skill premium by substituting low-skilled labor with industrial robots and performing high-skilled labor tasks.
-
[172]
History Repeats: The Longstanding Fear of Technology Replacing ...Jul 1, 2024 · With the unemployment rate around 20%, many feared that advancements in technology were eliminating jobs faster than they could be created.
-
[173]
The impact of AI on the labour market: is this time different? - OECD.AIDespite fears, the technological progress experienced in recent decades has not led to mass unemployment. In fact, employment in OECD countries has risen.
-
[174]
Evaluating the Impact of AI on the Labor Market - Yale Budget LabOct 1, 2025 · Overall, our metrics indicate that the broader labor market has not experienced a discernible disruption since ChatGPT's release 33 months ago, ...Missing: foundation empirical
-
[175]
The fear of technology-driven unemployment and its empirical baseJun 10, 2022 · This column suggests that the empirical support for the labour-creating effects of technological change dominates that for labour-replacement.Missing: comparison | Show results with:comparison
-
[176]
The Fed - The State of AI Competition in Advanced EconomiesOct 6, 2025 · 2025. "China's AI Models Are Closing the Gap – but America's Real Advantage Lies Elsewhere." RAND Corporation. May 2. Hoffman, Mia and Laura ...
-
[177]
[PDF] Winning the Defining Contest: The US-China Artificial Intelligence ...Jul 7, 2025 · Narratives of US artificial intelligence (AI) leadership collapsed in January 2025 after the announcement of. DeepSeek's innovative large ...
-
[178]
Meta changes course on open-source AI as China pushes ahead ...Aug 3, 2025 · In fact, China has probably found the path to “surpass the US in AI” thanks to the momentum in the country's vibrant open-source AI ecosystem, ...
-
[179]
China's drive toward self-reliance in artificial intelligence: from chips ...Jul 22, 2025 · Ample private funding and access to global open-source models have allowed for rapid Chinese progress. Although China lagged in the first years ...
-
[180]
Export Controls on Open-Source Models Will Not Win the AI RaceFeb 25, 2025 · One emphasizes geopolitical risk and global power dynamics, with a focus on Chinese misuse of U.S. open-source AI. The other is rooted in ...
-
[181]
AI principles - OECDThe OECD AI Principles promote use of AI that is innovative and trustworthy and that respects human rights and democratic values. Adopted in May 2019, they set ...Key Links · Latest Insights · Related PublicationsMissing: bilateral | Show results with:bilateral
-
[182]
AI governance through global red lines can help prevent ...Sep 22, 2025 · Global AI red lines are one way to set enforceable limits on dangerous AI uses and behaviours to prevent unacceptable risks and build trust.
-
[183]
AI Act | Shaping Europe's digital future - European UnionThe AI Act is the first-ever legal framework on AI, which addresses the risks of AI and positions Europe to play a leading role globally.Regulation - EU - 2024/1689 · AI Pact · AI Factories · European AI OfficeMissing: geopolitical competition
-
[184]
The Annual AI Governance Report 2025: Steering the Future of AIBilateral Agreements for AI Safety and Innovation: Bilateral agreements are quickly becoming the responsive layer of AI safety governance. In April 2024 ...
-
[185]
With AI, we need both competition and safety - Brookings InstitutionJul 8, 2024 · AI regulation must promote safety and protect competition through industry-government cooperation and enforceable standards.