Fact-checked by Grok 2 weeks ago

Knowledge acquisition

Knowledge acquisition is the process of eliciting, structuring, and formalizing from human experts, documents, and other sources to construct and maintain knowledge bases for , particularly in the field of and . This activity encompasses stages such as knowledge identification, conceptualization, integration, and validation, forming a foundational step in developing expert systems that mimic human expertise in specific domains. Historically, knowledge acquisition emerged as a critical component during the rise of first-generation expert systems in the and , where it was identified as the primary "" limiting the scalability and efficiency of system development. Pioneering work by highlighted the challenges of transferring tacit expert knowledge into computable forms, emphasizing that acquiring sufficient high-quality knowledge often required extensive interaction between knowledge engineers and domain specialists. This persisted because experts frequently struggled to articulate their reasoning explicitly, leading to incomplete or inconsistent representations in early systems like and . Key methods for knowledge acquisition include manual techniques such as structured and unstructured interviews, protocol analysis (where experts verbalize their thought processes during task performance), and repertory grid analysis to uncover conceptual structures. Automated approaches, including algorithms for pattern extraction from data and for , have evolved to complement these, reducing reliance on human intervention in modern applications. Frameworks like KADS (Knowledge Acquisition and Design Support) provide structured modeling at the knowledge level to guide elicitation and ensure reusability across domains. In contemporary contexts, knowledge acquisition extends beyond traditional expert systems to support technologies, ontologies, and hybrid systems, where integrating diverse data sources—such as and sensor inputs—addresses ongoing challenges in knowledge validation and maintenance. Despite advancements, the process remains resource-intensive, underscoring the need for interdisciplinary collaboration between researchers, domain experts, and engineers to achieve robust, adaptable knowledge representations.

Overview

Definition and Scope

Knowledge acquisition refers to the transfer and transformation of problem-solving expertise from human experts or other sources into a computable form suitable for knowledge-based systems, such as rules, ontologies, and models. This process involves eliciting concepts, constraints, and heuristics to enable machine performance in specific domains. The scope of knowledge acquisition includes phases from elicitation—extracting tacit knowledge from experts—to structuring it into formal representations. It forms the initial and critical stage in the knowledge engineering lifecycle, which encompasses subsequent steps like translation, representation, validation, and maintenance to ensure the knowledge supports reasoning and decision-making. Unlike data collection in machine learning, which relies on statistical patterns from large datasets to induce general models, knowledge acquisition emphasizes domain-specific, symbolic knowledge that is explicit, interpretable, and often derived from human expertise rather than purely data-driven induction. Central to this process are key types of knowledge: declarative knowledge, consisting of static facts and descriptions about the world (e.g., "Paris is the capital of "); procedural knowledge, outlining steps or methods to achieve goals (e.g., algorithms for ); and heuristic knowledge, comprising practical rules of thumb or experiential shortcuts to guide efficient problem-solving (e.g., "try dividing the problem into smaller subproblems"). Knowledge acquisition emerged in the 1970s within research, particularly through early expert systems like (1976), which encoded medical diagnostic rules from infectious disease specialists.

Historical Development

Knowledge acquisition in originated during the foundational period of AI research in the and , when efforts at institutions like and emphasized symbolic reasoning and the explicit representation of knowledge through symbols and logical rules. The 1956 Dartmouth Summer Research Project, organized by John McCarthy, , , and , formally established as a field, focusing on machines that could simulate , including problem-solving and formation. Early projects, such as the (1956) by Allen Newell and Herbert Simon at and Carnegie Mellon, demonstrated symbolic manipulation for theorem proving, laying groundwork for knowledge-based approaches by highlighting the need to encode domain-specific rules and heuristics. A pivotal milestone came in 1965 with the project at Stanford, developed by , , and Bruce Buchanan, which became the first for deducing molecular structures from data in . encoded heuristic knowledge from chemists—such as fragmentation rules and structural constraints—requiring extensive manual elicitation, and introduced the principle that "knowledge is power," shifting from general-purpose search to domain-specific expertise. Over the 1960s and 1970s, extensions like Meta-DENDRAL automated rule discovery from data, underscoring knowledge acquisition as a core challenge in scaling systems. The 1970s marked the emergence of knowledge acquisition as a distinct process in expert systems, with , developed from 1972 to 1976 at Stanford by Bruce Buchanan and Edward Shortliffe, using backward-chaining rules to diagnose bacterial infections and recommend therapies, incorporating approximately 600 rules elicited from infectious disease experts. Similarly, , created in the late 1970s at by Richard Duda, Peter Hart, and colleagues, modeled geological reasoning for mineral exploration, successfully predicting a deposit in 1980 and demonstrating the practical value of structured via certainty factors. By the 1980s, the proliferation of expert systems—approximately 60 in practical or commercial use by the mid-decade—solidified as a discipline, but revealed the "knowledge acquisition bottleneck," where manual elicitation proved time-consuming and costly, often requiring thousands of hours per system. In the 1990s and 2000s, efforts shifted toward scalable ontologies and knowledge reuse to mitigate acquisition challenges, exemplified by the project, launched in 1984 by at Microelectronics and Computer Technology Corporation (MCC) and ongoing at Cycorp, which manually encoded over a million axioms of common-sense knowledge using a hierarchical ontology and inference engine to support assisted elicitation by non-experts. The initiative, envisioned by , James Hendler, and Ora Lassila in a 2001 article, advanced knowledge acquisition through web-scale standards like the (RDF, drafted 1997) and the (OWL, W3C recommendation 2004), enabling interoperable representation and automated discovery of structured data. These developments influenced knowledge reuse in domains like bioinformatics and , reducing reliance on ad hoc manual methods. Following 2010, knowledge acquisition increasingly integrated with ecosystems, leveraging distributed ontologies and principles from the to facilitate large-scale from unstructured sources. In the , hybrid neuro-symbolic AI systems have advanced this evolution by combining neural learning for with symbolic methods for reasoning, addressing gaps in explainability and generalization; notable examples include DeepMind's AlphaGeometry (2024), which integrates neural language models with symbolic deduction to solve International Mathematical Olympiad-level geometry problems, and its successor AlphaGeometry 2 (2025), which achieved gold-medal performance; these mark a resurgence in knowledge-infused AI.

Methods and Techniques

Manual Elicitation Methods

Manual elicitation methods involve direct interaction between knowledge engineers and domain experts to capture expertise through structured human-centered techniques. These approaches emphasize verbal and observational to uncover declarative, procedural, and reasoning knowledge essential for building . Unlike automated methods, manual elicitation relies on the expert's articulation and demonstration of cognitive processes, ensuring a deep understanding of domain-specific . Core methods include structured interviews and protocol analysis. Structured interviews employ predefined questions, such as questionnaires or laddering techniques, to systematically probe expert knowledge. In laddering, experts are asked hierarchical questions starting from attributes and progressing to underlying values or goals, revealing causal relationships in . Protocol analysis, often using think-aloud protocols, requires experts to verbalize their thoughts while performing tasks, capturing real-time cognitive processes and heuristics. This method, formalized by and Simon, distinguishes between verbalizable knowledge and subconscious elements by analyzing concurrent versus retrospective reports. Other techniques encompass the technique and elicitation. The , developed by , elicits personal constructs by having experts rate elements (e.g., domain concepts) on bipolar scales derived from triadic comparisons, mapping cognitive structures and similarities. elicitation involves analyzing past examples or scenarios, where experts explain outcomes and generalize rules from specific instances, facilitating the identification of exceptions and contextual nuances. Procedural steps in manual elicitation typically begin with preparation, including domain scoping to identify key concepts and select appropriate experts. Elicitation sessions follow, involving one-on-one interactions recorded via audio or video for accuracy. Subsequent transcription and analysis organize raw data into structured formats, such as decision trees, to model if-then rules or hierarchies from expert responses. Tools like software aid in visualizing and refining constructs during analysis. These methods offer advantages in high fidelity to expert intuition, capturing nuanced, that quantitative approaches might overlook. For instance, in the development of the XCON for computer configuration at in the , structured interviews and protocol analysis were used to encode over 10,000 rules from hardware experts, enabling automated order verification with significant error reduction.

Automated Acquisition Methods

Automated acquisition methods in knowledge acquisition leverage computational algorithms and tools to derive structured knowledge from data sources, minimizing human intervention and enabling for large datasets. These approaches contrast with manual by focusing on inductive and deductive processes that infer rules, patterns, or ontologies directly from raw data or text corpora. Key techniques include for rule induction and for pattern discovery, often integrated with (NLP) for text-based extraction. In , induction from examples allows systems to learn decision rules or classifiers by generalizing from instances. A seminal example is the , which constructs decision trees by recursively selecting attributes that maximize information gain to partition data. The information gain is computed using , defined as H(S) = -\sum_{i=1}^{c} p_i \log_2 p_i, where S is a set of examples, c is the number of classes, and p_i is the proportion of examples in class i. Developed by Ross Quinlan in 1986, ID3 has been foundational for knowledge acquisition in expert systems, enabling the automatic extraction of if-then rules from training examples in domains like . Subsequent extensions, such as C4.5, addressed limitations like continuous attributes. C4.5 maintains ID3's core inductive paradigm for scalable knowledge derivation. Data mining complements induction by uncovering associative patterns in unstructured or transactional data, facilitating knowledge acquisition through frequent itemset discovery. The , introduced by Agrawal and Srikant in 1994, exemplifies this by iteratively generating candidate itemsets and pruning those below a minimum , leveraging the that subsets of frequent itemsets must also be frequent. This method has been widely adopted for extracting association rules, such as analysis, where rules like {bread} → {butter} represent acquired from sales data. Apriori's efficiency scales to large databases, though it requires multiple passes, influencing modern variants for environments. Text-based methods employ to extract ontological structures from unstructured corpora, automating the identification of entities, relations, and taxonomies. (NER), a core NLP task, identifies and classifies entities like persons or organizations, serving as a building block for ontology population; for instance, systems use NER to tag entities in scientific texts before linking them to hierarchical concepts. Broader from text involves techniques like term extraction via statistical measures (e.g., TF-IDF) and relation discovery through or dependency parsing, as surveyed in foundational work on semi-automatic ontology construction from heterogeneous text sources. These methods have enabled the automated buildup of domain ontologies, such as in biomedical literature, where tools process corpora to infer subclass relations without manual annotation. Hybrid tools integrate and outputs into formal knowledge representations, such as s constructed using (RDF) and (). RDF models knowledge as directed graphs of subject-predicate-object triples, where resources (identified by or literals) form interconnected nodes, providing a standardized syntax for automated graph population from diverse sources. extends RDF with axioms for reasoning, supporting automated inference in ; for example, tools convert schemas to by mapping tables to classes, columns to properties, and foreign keys to relations, deriving ontologies like DBpedia from structured data. This approach automates construction for , as seen in projects extracting triples from dumps. As of 2025, large language models (LLMs) have advanced automated acquisition through , transferring structured insights from massive pre-trained models to efficient ones for domain-specific tasks. BERT, a bidirectional pre-trained on vast text, adapts it to downstream by adding task-specific layers, yielding state-of-the-art results in entity recognition and relation extraction with minimal architecture changes. For instance, domain-specific on legal or medical corpora enables BERT to distill rules like causal relations from text. Techniques like MiniLLM further optimize this by using reverse divergence in distillation, reducing exposure and enhancing for long-context knowledge generation across model sizes from 120M to 13B parameters. These LLM integrations scale automated methods, bridging raw data to verifiable knowledge bases.

Challenges and Limitations

The Knowledge Acquisition Bottleneck

The knowledge acquisition bottleneck represents a fundamental challenge in , characterized as the rate-limiting step in constructing due to the labor-intensive process of extracting and encoding expert knowledge into formal representations. Popularized by Edward A. Feigenbaum in the early , this bottleneck arises primarily during the phase, where domain experts' insights must be translated from informal human to computable structures, often dominating the entire development lifecycle. Studies from that era indicate it can consume 70-80% of project time, severely constraining the scalability and deployment of such systems. The impact is profound, as it not only inflates costs but also contributes to high project attrition, with empirical analyses revealing that approximately 60% of expert systems initiatives fail outright due to unresolved acquisition hurdles. Several interconnected causes exacerbate this bottleneck. A primary issue is the inherent difficulty in articulating —unwritten, intuitive expertise accumulated through experience that experts struggle to verbalize explicitly during interviews or sessions. Compounding this is the frequent unavailability of suitable experts, whose specialized roles in high-demand fields limit their participation in prolonged efforts. Additionally, cognitive biases in reporting, such as overconfidence or anchoring effects, distort the accuracy of elicited , leading to incomplete or inconsistent formalizations that require extensive revision. The bottleneck's significance was first systematically recognized in Feigenbaum's seminal work on , which highlighted its role as a barrier to advancing beyond prototype stages, and echoed in subsequent 1980s analyses of projects. These early recognitions underscored how acquisition failures propagated to broader system unreliability, prompting initial explorations into structured elicitation protocols. In contemporary as of 2025, while advancements in large models and automated tools have begun to alleviate some aspects of the by reducing manual needs, challenges persist in integrating and validating AI-generated knowledge with human expertise. To alleviate this challenge, one established mitigation strategy involves reusing pre-existing ontologies from shared repositories, such as those based on the (OWL), which provide reusable foundational knowledge structures to bypass much of the ground-up process and accelerate development.

Knowledge Representation and Validation Issues

Knowledge representation in knowledge acquisition involves selecting appropriate formal structures to encode elicited expertise, such as production rules, frames, or semantic networks, each with distinct strengths and limitations. Production rules, which express knowledge as conditional "if-then" statements, facilitate procedural reasoning but can lead to in large rule sets, making maintenance challenging. , introduced by Minsky in 1975, organize knowledge into structured slots representing stereotypical situations, enabling efficient inheritance and default reasoning, yet they struggle with dynamic or non-hierarchical domains due to rigid slot-filling assumptions. Semantic networks, pioneered by Quillian in 1968, depict knowledge as graph-based nodes and edges to model relationships, supporting associative inference, but they often suffer from ambiguous link semantics that hinder precise querying. Choosing among these formats requires balancing expressiveness with computational tractability, as no single method universally captures complex expertise without trade-offs. A primary challenge in representation is incompleteness, where acquired knowledge fails to cover all scenarios, leading to gaps in the knowledge base that propagate errors during inference. Inconsistency arises when conflicting rules or relations are encoded, such as contradictory facts from multiple experts, undermining the reliability of derived conclusions. Oversimplification occurs when nuanced human expertise is reduced to binary or deterministic forms, losing contextual subtleties essential for real-world application. These issues are exacerbated in hybrid systems combining multiple formats, where interoperability between rules and networks can introduce representational mismatches. Validation methods are crucial to ensure the represented knowledge is sound and usable. Consistency checks involve static analysis to detect redundancies, contradictions, or gaps, often using decision tables for rule-based systems. In logic-based representations, proving automates by attempting to derive contradictions from the , confirming logical coherence. Empirical testing through simulations applies the to synthetic or historical data, comparing outputs against judgments to assess behavioral fidelity. Common pitfalls include granularity mismatches, where the level of detail in representation—too abstract for broad applicability or overly detailed for efficiency—impedes integration with diverse data sources. Handling uncertainty is another hurdle, as deterministic formats like rules inadequately model probabilistic expertise; Bayesian networks address this by representing variables as nodes in a with conditional probabilities, enabling inference via : P(A|B) = \frac{P(B|A) P(A)}{P(B)}, where P(A|B) is the . This approach, formalized by Pearl in , quantifies dependencies but requires accurate elicitation, which can be error-prone in acquisition. Metrics for evaluating representation quality include coverage, measured as the completeness ratio of addressed scenarios to total possible ones, often computed as the proportion of test cases handled without defaulting to unknowns. Accuracy assesses prediction reliability through error rates, such as the percentage of incorrect inferences in validation simulations, providing a quantitative gauge of representational fidelity. These metrics guide iterative refinement, ensuring the knowledge base achieves sufficient and low error rates for practical deployment.

Applications

In Expert Systems and AI

Knowledge acquisition plays a central role in expert systems by eliciting domain-specific rules from human experts to populate the , enabling the to perform reasoning and decision-making. In these systems, the process involves structured interviews, protocol analysis, and iterative refinement to capture knowledge, such as conditional rules of the form "if-then" statements that guide problem-solving. For instance, the system, developed in the 1970s, acquired over 450 rules from infectious disease specialists to diagnose bacterial infections and recommend therapies, demonstrating how manual elicitation translates expert heuristics into executable logic for medical consultations. A prominent is the R1 (later XCON) , deployed by in the 1980s to configure VAX computer systems. Knowledge acquisition for R1 involved extracting configuration rules from hardware , resulting in a that automated order validation and component selection, achieving 95-98% accuracy in deployment and saving the company approximately $40 million annually by reducing configuration errors and manual labor. This success highlighted the practical impact of rule-based knowledge acquisition in industrial settings, where scalability allowed handling thousands of unique orders without expert intervention. In broader AI applications, knowledge acquisition integrates with cognitive architectures like SOAR, where chunking serves as a learning mechanism to automatically acquire new production rules from goal-based problem-solving experiences. During impasses in , SOAR compiles results from subgoals into "chunks"—generalized rules that summarize prior reasoning—enabling the system to refine its over time for improved efficiency in tasks like and . Similarly, hybrid approaches combine with expert systems by extracting interpretable rules from trained policies; for example, methods like policy extraction convert agents into decision trees, bridging black-box models with symbolic rule sets for verifiable and explainable AI behaviors. A modern example is , which leverages for acquisition in medical domains, ingesting vast corpora of clinical literature and guidelines to build probabilistic models of expert since its initial demonstrations around 2011. This enabled to analyze unstructured text for hypothesis generation in , supporting clinicians with evidence-based recommendations derived from acquired domain insights. The benefits of knowledge acquisition in these contexts include scalable that mimics performance without constant human involvement, with deployment metrics such as rule accuracy often exceeding 95% in validated systems like XCON, thereby enhancing reliability and cost-efficiency in AI-driven applications.

In Broader Domains

In , knowledge acquisition plays a pivotal role in developing intelligent systems (ITS) that adapt to learners' needs by incorporating pedagogical strategies and domain-specific expertise. For instance, AutoTutor, an ITS focused on natural language dialogue for subjects like and physics, relies on manual elicitation techniques such as interviews and analysis to build its , enabling the system to guide students toward deeper understanding through mixed-initiative conversations. This process ensures the tutor can assess and knowledge acquisition by drawing on elicited rules for prompting, , and content delivery. In healthcare, knowledge acquisition supports clinical decision support systems (CDSS) by formalizing medical expertise into structured for accurate diagnosis and treatment recommendations. , a comprehensive clinical terminology , is developed through collaborative efforts involving domain experts who contribute via structured input, validation committees, and iterative refinement to encode diagnostic knowledge, facilitating in CDSS applications like guideline-based alerts. This expert-driven acquisition process allows to represent over 370,000 clinical concepts (as of 2025), enabling systems to infer relationships for evidence-based decisions without ambiguity. Business applications leverage knowledge acquisition in knowledge management systems to capture tacit knowledge, which is often implicit and experience-based, enhancing customer relationship management (CRM). In CRM implementations, case studies demonstrate the use of interviews, observation, and narrative collection to elicit tacit insights from sales teams, which are then codified into reusable models for predictive analytics and personalized interactions. Similarly, in bioinformatics, gene ontology (GO) acquisition involves curating functional annotations from experimental data and literature through expert consortia, ensuring consistent representation of gene products for downstream analyses like pathway modeling. Domain-specific adaptations in knowledge acquisition often incorporate techniques to suit practical contexts, such as simulations where visual and interactive elements in capturing complex . In educational and business simulations, tools like polling—combining video scenarios, images, and real-time feedback—facilitate from experts by simulating real-world tasks, reducing and improving the of acquired models compared to text-only methods. These tweaks enable broader applicability, as seen in healthcare where simulations elicit clinical for updates.

Future Directions

Integration with Machine Learning

Knowledge acquisition has increasingly integrated with machine learning through hybrid neuro-symbolic approaches, which combine the pattern recognition capabilities of neural networks with the structured reasoning of symbolic rules. Neuro-symbolic AI frameworks enable the seamless learning and reasoning over both data and abstract knowledge by embedding logical constraints into differentiable tensor operations. A prominent example is Logic Tensor Networks (LTN), which formalizes first-order logic in a many-valued, end-to-end differentiable manner, allowing neural networks to perform tasks such as relational learning and query answering while respecting symbolic rules. This integration addresses limitations in pure neural methods by incorporating prior knowledge directly into the learning process, enhancing both accuracy and interpretability in knowledge-intensive applications. Recent advancements in the 2020s have further bridged knowledge acquisition and via techniques like symbolic from large models (LLMs). This transfers implicit encoded in LLMs—such as GPT-derived insights—into explicit symbolic representations, including bases, to create more transparent and efficient systems. Symbolic methods distill LLMs' reasoning into interpretable forms, such as logical or graphs, by leveraging techniques like prompting for extraction and optimization under symbolic constraints, thereby enabling the acquisition of structured without retraining from scratch. Additionally, facilitates privacy-preserving knowledge acquisition by allowing decentralized models to collaboratively learn from distributed data sources without sharing raw information, using mechanisms like differentially private to aggregate insights securely. These integrations yield significant benefits, particularly in improving explainability and scalability for complex domains. For instance, employs to acquire and predict knowledge from sequences, achieving near-atomic accuracy (median 0.96 Å r.m.s.d.95) by jointly embedding evolutionary relationships and spatial features through an Evoformer module and iterative structure refinement. By 2025, such synergies have led to widespread adoption in edge AI systems, where real-time knowledge updates occur via serverless edge-cloud architectures that process data locally to minimize and support dynamic applications like . This evolution enables continuous, on-device acquisition of domain-specific knowledge, fostering adaptive AI in resource-constrained environments.

Interdisciplinary and Ethical Considerations

Knowledge acquisition intersects with various disciplines, enriching its methodologies and theoretical foundations. In , provides models for elicitation that emphasize how is represented at multiple levels, from perceptual to conceptual, informing the design of elicitation techniques that align with human cognitive processes. These models highlight the limitations of simplistic in capturing high-level concepts, advocating for patterns that better reflect cognitive acquisition from text. Philosophical debates, particularly in , challenge practitioners to formalize common-sense , debating the feasibility of explicit ontologies for microworlds and the of in handling change and causation. Such discussions underscore the tension between 's practical implementations and philosophical rigor in knowledge structuring. In education, knowledge acquisition supports systems by fostering competencies through formal, nonformal, and informal pathways, as seen in frameworks that integrate theories to promote self-directed skill development across life stages. Ethical considerations in knowledge acquisition are paramount, particularly regarding and . Bias in systems can lead to skewed representations that perpetuate inequities; tractable probabilistic models can mitigate fairness issues by blending expert input with data-driven learning. Privacy concerns emerge in automated acquisition from user data, where processing must comply with regulations like the EU's (GDPR), Article 22, which prohibits solely automated decisions producing legal effects without human intervention, explicit consent, or legal authorization, alongside safeguards such as the right to contest outcomes. Sustainability challenges in knowledge acquisition, especially in machine learning hybrids, stem from high resource demands that strain environmental limits. Training large models consumes vast energy—equivalent to hundreds of households annually—and contributes to carbon emissions comparable to thousands of flights, with data centers projected to use 8% of electricity by 2030. Hardware production exacerbates this, requiring scarce materials like (with supplies under 15 years) and millions of liters of water daily for fabrication. To address fairness, metrics such as demographic parity ensure equal positive prediction rates across demographic groups in bases, promoting independence from sensitive attributes like or , though it may require adjusted thresholds to equity and accuracy. Future research as of 2025 emphasizes diverse panels and transparent auditing to advance ethical knowledge acquisition. Interdisciplinary panels, incorporating ethicists, scientists, and policymakers, are called for to embed fairness and in AI development, with surveys indicating 95% support for such collaborations to tackle biases and societal impacts. Transparent auditing frameworks are urged to monitor evolving systems, ensuring explainability and through robust methodologies that assess ethical without compromising .

References

  1. [1]
  2. [2]
  3. [3]
    [PDF] Expert Systems and Knowledge Acquisition
    The so-called Feigenbaum bottleneck problem, formulated in 1977, stated that “Expert knowledge acquisition is a problematic bottleneck in the construction of ...<|separator|>
  4. [4]
    [PDF] KNOWLEDGE ENGINEERING The Applied Side of Artificial ... - Stacks
    programs called "expert systems". The goal of an "expert system" project is to write a program that achieves a high level of performance on problems that ...
  5. [5]
    First generation expert systems: a review of knowledge acquisition ...
    >Journals; >The Knowledge Engineering Review; >Volume 3 Issue 2; >First ... knowledge acquisition” International Journal of Man-Machine Studies 26 231–243.
  6. [6]
  7. [7]
    [PDF] The Knowledge Reengineering Bottleneck - Semantic Web Journal
    Feigenbaum's knowledge acquisition bottleneck refers to the difficulty of correctly extracting expert knowledge into a knowledge base: 1In fact, this ...
  8. [8]
    The Knowledge Acquisition Bottleneck: Time for Reassessment?
    Abstract: Knowledge acquisition has long been considered to be the major constraint in the development of expert systems. Conventional wisdom also maintains ...
  9. [9]
    Knowledge acquisition, knowledge programming, and ... - RAND
    Knowledge acquisition, knowledge programming, and knowledge refinement. Frederick Hayes-Roth, Philip Klahr, David J. Mostow. ResearchPublished 1980. Share on ...
  10. [10]
    Knowledge Engineering Overview - Clinical Practice Guidelines v2.0.0
    Knowledge acquisition is the process of extracting, understanding, structuring and organizing knowledge from one source, often solely or largely from human/ ...
  11. [11]
    A Formal Framework for Knowledge Acquisition: Going beyond ...
    In our terminology, the purpose of machine learning is learning rather than knowledge acquisition. This can often be a disadvantage, since knowledge ...
  12. [12]
    [PDF] An Overview of Knowledge Representation Techniques - ijsrcseit.com
    May 28, 2018 · A. Types of knowledge. The types of knowledge include procedural knowledge, declarative knowledge and heuristic knowledge. Procedural knowledge: ...<|control11|><|separator|>
  13. [13]
    Mycin: A Knowledge-Based Computer Program Applied to Infectious ...
    1976 Jun;60(7):981–996. doi: 10.1016/0002-9343(76)90570-2. [DOI] ... acquisition capabilities of the MYCIN system. Comput Biomed Res. 1975 Aug;8 ...
  14. [14]
    [PDF] DENDRAL: a case study of the first expert system for scientific ... - MIT
    Whether DENDRAL was the first expert system is debatable; it was certainly the first application of AI to a problem of scientific reasoning. In his forward to ...
  15. [15]
    History Of AI In 33 Breakthroughs: The First Expert System - Forbes
    Oct 29, 2022 · Already in 1983, Feigenbaum identified the “key bottleneck” that led to their eventual demise, that of scaling the knowledge acquisition process ...
  16. [16]
    12 AI Milestones: 4. MYCIN, An Expert System For Infectious ...
    Apr 27, 2020 · MYCIN was an AI program developed at Stanford University in the early 1970s, designed to assist physicians by recommending treatments for certain infectious ...
  17. [17]
    PROSPECTOR computer-based expert system - SRI
    Dec 2, 1970 · The system, one of the world's first computer-based expert systems, attempted to represent the knowledge and reasoning process of geological experts.
  18. [18]
    Steps towards Assisted Knowledge Acquisition in Cyc - AAAI
    In this paper, we describe the Cyc knowledge base and inference system, enumerate the means that it provides for knowledge elicitation, including some means ...
  19. [19]
    A Review of the Semantic Web Field - Communications of the ACM
    Feb 1, 2021 · In 2004, the Web Ontology Language OWL became a W3C standard (the revision OWL 2 was established in 2012), providing further fuel for the field.
  20. [20]
    [PDF] Neuro-Symbolic AI in 2024: A Systematic Review - arXiv
    Apr 5, 2025 · Objective: This paper provides a systematic literature review of Neuro-Symbolic AI projects within the 2020-24 AI landscape, highlighting key ...
  21. [21]
    Knowledge Acquisition - an overview | ScienceDirect Topics
    Knowledge acquisition is defined as an activity in knowledge engineering crucial for building and updating a knowledge base, encompassing the initial gathering ...
  22. [22]
    Knowledge Acquisition - Kamel - Wiley Online Library
    Dec 14, 2007 · The two manual methods commonly used are interviews (structured, unstructured, and questionnaire) and task-based methods (protocol analysis, ...
  23. [23]
    (PDF) The Repertory Grid Technique - ResearchGate
    We introduce a basic procedure for construct elicitation followed by the laddering, pyramiding, and resistance to change techniques used to give greater ...
  24. [24]
    [PDF] Knowledge Acquisition Techniques for Expert Systems - DTIC
    May 1, 1988 · The Repertory Grid Method of Knowledge Elicitation. The Rep Grid Method (RGM) of knowledge acquisition uses SME judgments of similarity ...
  25. [25]
    Knowledge Acquisition for Expert Systems - A Practical Handbook
    SOFT SYSTEMS ANALYSIS. Personal construct psychology provides cognitive foundations for knowledge elicitation, and repertory grid techniques provide a corre- ...
  26. [26]
    Expert systems for configuration at Digital: XCON and beyond
    This paper discusses expert systems for configuration at Digital, including XCON, and lessons learned from a decade of designing configuration systems.
  27. [27]
    Ontology learning from text: A look back and into the future
    This together with the advanced state in related areas, such as natural language processing, have fueled research into ontology learning over the past decade.Abstract · Cited By · Information
  28. [28]
    Induction of decision trees | Machine Learning
    This paper summarizes an approach to synthesizing decision trees that has been used in a variety of systems, and it describes one such system, ID3, in detail.
  29. [29]
    [PDF] Fast Algorithms for Mining Association Rules - VLDB Endowment
    The AprioriTid algorithm has the additional prop- erty that the database is not used at all for count- ing the support of candidate itemsets after the first.
  30. [30]
    RDF 1.1 Concepts and Abstract Syntax
    ### Key Points on RDF for Knowledge Representation in Knowledge Graphs
  31. [31]
  32. [32]
    MiniLLM: Knowledge Distillation of Large Language Models
    Jan 16, 2024 · We propose a knowledge distillation method for Large Language Models with minimizing the reverse KL divergence as the objective.
  33. [33]
    Medical Expert System - an overview | ScienceDirect Topics
    ... knowledge acquisition and structuring can consume up to 70% of the total development time for an MES. 4. System validation is critical, as most MESs are ...
  34. [34]
    A STEP-Compliant Knowledge Based Systems for the process ...
    However, over half of those have struggled, with a 60% failure rate. ... Knowledge Acquisition in Expert Systems (ESs) development represents one of ...
  35. [35]
  36. [36]
    Using three minimally biasing elicitation techniques for knowledge ...
    Proceedings from 1st AAAI Knowledge Acquisition for Knowledge-Based Systems Workshop, ... Cleaves D.A.. Cognitive bias and corrective techniques: proposals for ...
  37. [37]
    [PDF] Ontologies and the Knowledge Acquisition Bottleneck - CEUR-WS.org
    In this paper we present an approach to rapid development of knowledge-based agents that illustrates several general methods and ideas related to ontology reuse.Missing: OWL | Show results with:OWL
  38. [38]
  39. [39]
    Expert system verification and validation: a survey and tutorial
    This paper surveys the issues, methods and techniques for verifying and validating expert systems.
  40. [40]
    Handling Granularity Differences in Knowledge Integration
    This analysis revealed four major types of granularity mismatches: filtering, generalization, abstraction, and co-reference across granularity difference (table ...
  41. [41]
    [PDF] BAYESIAN NETWORKS* Judea Pearl Cognitive Systems ...
    Bayesian networks are directed acyclic graphs modeling the environment, representing variables and their causal dependencies, and simulating environmental ...
  42. [42]
    (PDF) Metrics for Concept-Oriented Knowledge Bases - ResearchGate
    Aug 7, 2025 · We use the methodology to develop a series of metrics that measure the size and complexity of concept-oriented knowledge bases. Two of the ...
  43. [43]
    A rule-based computer program for advising physicians regarding ...
    MYCIN is an interactive computer program, relying to a large extent upon artificial intelligence (AI) techniques, which uses decision rules acquired from ...
  44. [44]
    [PDF] Rule-Based Expert Systems: The MYCIN Experiments of the ...
    Every rule in the MYCIN system belongs to one, and only one, of these categories. Furthermore, selecting the proper category for a newly ac- quired rule does ...
  45. [45]
    [PDF] Expert Systems Notes - Oxford statistics department
    By 1986 a $40 million per year saving was being claimed for XCON. A modern example (Heckerman and Wellman, 1995, Heckerman et al., 1995) is the. EPTS1 system ...
  46. [46]
    [PDF] R1: A Rule-Based Configurer of Computer Systems
    It uses Match as its principal problem solving method; it has sufficient knowledge of the configuration domain and of the peculiarities of the various.Missing: savings | Show results with:savings
  47. [47]
    [PDF] Chunking in SOAR: The Anatomy of a General Learning Mechanism.
    Abstract: In this article we describe an approach to the construction of a general learning mechanism based on chunking in Soar. Chunking is a learning ...
  48. [48]
    [PDF] Verifiable Reinforcement Learning via Policy Extraction
    We describe how to verify correctness (for the case of a toy game based on Pong), stability, and robustness of decision tree policies, and show that ...
  49. [49]
    IBM's Health Analytics and Clinical Decision Support - PMC - NIH
    Watson can read and analyze concepts in millions of pages of medical information in seconds, identify information that could be relevant to a decision ...
  50. [50]
    An Intelligent Tutoring System With Mixed-Initiative Dialogue
    Aug 6, 2025 · AutoTutor simulates a human tutor by holding a conversation with the learner in natural language. The dialogue is augmented by an animated conversational agent.
  51. [51]
    SNOMED clinical terms: overview of the development process ... - NIH
    This paper discusses the process and status of SNOMED CT development and how the resources and expertise of both organizations are being used to develop this ...Missing: knowledge | Show results with:knowledge
  52. [52]
    An encoding methodology for medical knowledge using SNOMED ...
    The encoding methodology uses SNOMED CT (SCT) to encode medical data, with four stages: collecting info, identifying concepts, mapping to SCT, and applying ...
  53. [53]
    Exploring the role of customer relationship management (CRM ...
    This study explores how customer relationship management (CRM) systems support customer knowledge creation processes.
  54. [54]
    [PDF] Tacit knowledge augmented customer relationship management ...
    Abstract. Tacit knowledge (TK) is a core value element important for obtaining a competitive edge for Customer Relationship Management (CRM).
  55. [55]
    Knowledge acquisition, consistency checking and concurrency ...
    The Gene Ontology (GO; http://www.geneontology.org) provides a taxonomy of concepts and their attributes for annotating gene products. As GO increases in size, ...
  56. [56]
    [PDF] Knowledge elicitation using multimedia poll techniques - rot13
    Abstract: This paper deals with usage of multimedia in the field of knowledge elicitation. It introduces multimedia poll as a technique suitable for ...
  57. [57]
    Using a simulation model for knowledge elicitation and knowledge ...
    The work reported in this paper is part of a project simulating maintenance operations in an automotive engine production facility.Missing: multimedia | Show results with:multimedia
  58. [58]
  59. [59]
    A Survey on Symbolic Knowledge Distillation of Large Language Models
    ### Summary of Symbolic Knowledge Distillation Survey
  60. [60]
    Differentially private knowledge transfer for federated learning - Nature
    Jun 24, 2023 · Federated learning that transfers knowledge from decentralized data into a shared intelligent model is widely used to reduce user privacy risks ...
  61. [61]
    Highly accurate protein structure prediction with AlphaFold - Nature
    Jul 15, 2021 · The AlphaFold network directly predicts the 3D coordinates of all heavy atoms for a given protein using the primary amino acid sequence and ...
  62. [62]
    AI edge cloud service provisioning for knowledge management ...
    Sep 1, 2025 · This capability minimizes latency and ensures real-time updates for critical knowledge databases, particularly in dynamic business environments.<|control11|><|separator|>
  63. [63]
    A cognitive science perspective on knowledge acquisition
    One of the outcomes of research in cognitive science shows that knowledge may be represented at different levels and in different formats: e.g. perceived ...
  64. [64]
    Logic and Artificial Intelligence - Stanford Encyclopedia of Philosophy
    Aug 27, 2003 · The relations between AI and philosophical logic are part of a larger story. It is hard to find a major philosophical theme that doesn't become ...
  65. [65]
    Lifelong Learning in the Educational Setting: A Systematic Literature ...
    May 13, 2023 · Meanwhile, the UNESCO definition of lifelong learning includes all intentional learning from birth to death that attempts to advance knowledge ...
  66. [66]
    Knowledge representation and acquisition for ethical AI: challenges ...
    Mar 11, 2023 · In this article, we advocate for a two-pronged approach ethical decision-making enabled using rich models of autonomous agency.
  67. [67]
    Art. 22 GDPR – Automated individual decision-making, including ...
    Rating 4.6 (10,111) The data subject shall have the right not to be subject to a decision based solely on automated processing, including profiling, which produces legal ...
  68. [68]
    Sustainable AI - ML Systems Textbook
    The computational demands of AI create sustainability challenges that extend beyond energy consumption, encompassing carbon emissions, resource extraction, ...
  69. [69]
    A clarification of the nuances in the fairness metrics landscape - PMC
    Fairness notions proposed in the literature are usually classified in broad areas, such as: definitions based on parity of statistical metrics across groups ...
  70. [70]
    [PDF] AAAI 2025 Presidential Panel on the Future of AI Research
    The panel covers AI reasoning, ethics, safety, social good, sustainability, AGI, and more, as AI research rapidly transforms.