[Representation]
Symbolic Knowledge Extraction vs Injection

Symbolic Knowledge Extraction vs Symbolic Knowledge Injection

This post is a study note of the paper “Symbolic knowledge extraction and injection with sub-symbolic predictors: A systematic literature review “ by Ciatto et al. (2024).

Introduction

Background

Conceptual Foundations

What is Symbolic Knowledge Extraction (SKE)?

Definition
Symbolic Knowledge Extraction (SKE) is the process of deriving intelligible and interpretable knowledge representations from sub-symbolic predictors, such as neural networks or other machine learning models.

The goal is to make the learned behavior of these models accessible and understandable to humans, enabling interpretability and transparency in decision-making processes. This process often involves translating the numerical or distributed representations of knowledge in sub-symbolic models into symbolic forms, such as logic rules, decision trees, or human-readable texts. SKE serves as a key enabler for Explainable AI (XAI), bridging the gap between complex machine learning models and human users.

What is Symbolic Knowledge Injection (SKI)?

Definition
Symbolic Knowledge Injection (SKI) refers to the process of integrating symbolic knowledge—often derived from human expertise, formal logic, or structured data—into sub-symbolic predictors.

This ensures that these models operate in alignment with predefined rules, constraints, or domain-specific knowledge. By injecting symbolic knowledge, SKI can guide or constrain the learning process, enhance model performance, or enforce consistency with ethical and practical considerations. This is achieved through techniques like embedding symbolic knowledge as input, structuring predictors to align with symbolic knowledge, or penalizing inconsistent behaviors during training.

Key Features and Dimensions of Comparison

Input/Output Nature

Data types and formats used in SKE and SKI.

SKE primarily processes sub-symbolic data such as numeric arrays (vectors, matrices, tensors) and outputs symbolic forms like logic rules or decision trees.
SKI, on the other hand, takes symbolic inputs, such as knowledge graphs or formal logic, and integrates them into sub-symbolic models.

Fidelity

Definition
Fidelity refers to the degree to which the extracted or injected symbolic knowledge accurately reflects the behavior or constraints of the sub-symbolic model.

Fidelity is about how much the extracted or injected symbolic knowledge accurately reflects the behavior or constraints of the sub-symbolic model. For SKE, high fidelity ensures that the extracted solymbic knowledge mirrors the decision-making logic of the predictor. For SKI, fidelity guarantees that the injected knowledge is faithfully incorporated into the model’s functioning.

Methods and Strategies

Extraction: Pedagogical (교육적인) vs Decompositional Approaches

Pedagogical methods prioritize generality and ease of application but may sacrifice some accuracy in the extracted knowledge. Decompositional methods provide higher fidelity and detailed insights but are more complex and less flexible, requiring access to model-specific internal details.

Injection: Knowledge Embedding, Guided Learning, and Predictor Structuring in SKI

Knowledge Embedding is versatile and enables models to utilize symbolic knowledge as part of their inputs. Predictor Structuring ensures the symbolic knowledge is inherently integrated into the model’s architecture, providing robust adherence to the rules. Guided Learning directly influences the model’s behavior by constraining its learning process.

Challenges and Opportunities

Opacity and interpretability issues addressed by SKE.

Symbolic Knowledge Extraction (SKE) addresses one of the most significant challenges in artificial intelligence: the opacity of sub-symbolic models like neural networks. These models often function as “black boxes,” making their decision-making processes difficult to interpret. SKE tackles this by converting these opaque models into interpretable forms, such as logic rules or decision trees, enabling transparency and trust in AI systems. However, achieving high fidelity between the extracted symbolic knowledge and the original model’s behavior remains a persistent challenge.

Consistency and control challenges addressed by SKI.

Symbolic Knowledge Injection (SKI) deals with consistency and control challenges in sub-symbolic models. By injecting symbolic knowledge, SKI ensures that models adhere to predefined rules, ethical considerations, or domain-specific logic. This enhances the reliability and robustness of AI systems, especially in critical applications. However, integrating symbolic knowledge effectively into sub-symbolic frameworks without sacrificing model performance or flexibility is a complex task.

Emerging trends highlight the potential synergies between SKE and SKI. Together, they can form a continuous loop where extracted knowledge is refined and reintegrated into models, improving both interpretability and performance. This interplay opens opportunities for hybrid AI systems that leverage the strengths of both symbolic and sub-symbolic approaches, paving the way for more transparent, adaptable, and controllable AI solutions.

Hybrid Systems and Interoperability

Combined Use of SKE and SKI

The combination of Symbolic Knowledge Extraction (SKE) and Symbolic Knowledge Injection (SKI) creates a powerful synergy, allowing hybrid systems to seamlessly integrate symbolic and sub-symbolic approaches. SKE makes sub-symbolic models interpretable by extracting knowledge in symbolic forms, while SKI ensures that models operate within human-defined constraints by injecting symbolic knowledge. Together, they form a theoretical foundation for hybrid symbolic-sub-symbolic systems, enhancing transparency, adaptability, and control.

TEFI (Train–Extract–Fix–Inject) Loop Process

  1. Train: Train the sub-symbolic model (e.g., neural network) on available data to achieve desired predictive performance.

  2. Extract: Use Symbolic Knowledge Extraction (SKE) to derive interpretable symbolic knowledge (e.g., logic rules or decision trees) from the trained model. This step provides insights into the model’s reasoning process.

  3. Fix: Analyze the extracted knowledge to identify inconsistencies, errors, or areas where alignment with domain-specific rules or objectives is needed. Modify and refine the symbolic knowledge accordingly.

  4. Inject: Reintegrate the refined symbolic knowledge back into the sub-symbolic model using Symbolic Knowledge Injection (SKI). This ensures that the model adheres to the corrected or enhanced symbolic rules.

  5. Iterate: Repeat the loop as necessary to continuously improve the model’s interpretability, consistency, and performance.

TEFI

The “Train–Extract–Fix–Inject (TEFI)” loop proposed in the literature exemplifies this synergy. In this loop, knowledge extracted via SKE is analyzed and refined to correct any inconsistencies or align with domain-specific requirements. This refined knowledge is then reintroduced into the model using SKI. The TEFI loop enables continuous improvement of models, ensuring that they remain interpretable, reliable, and aligned with human-defined objectives. It bridges the gap between interpretability and performance by iteratively refining both the symbolic and sub-symbolic components.

Practical Benefits

One of the most significant practical advantages of combining SKE and SKI is the ability to debug sub-symbolic models effectively. Through the TEFI loop, extracted symbolic knowledge can be inspected and corrected, enabling data scientists to identify and address errors in the model’s reasoning process. This approach ensures that sub-symbolic models operate within the boundaries of logical rules, ethical constraints, and domain-specific knowledge, creating a more robust and trustworthy AI system.

Additionally, SKE and SKI facilitate the use of symbolic knowledge as a lingua franca for intelligent agent systems. By representing knowledge in symbolic forms that are both human- and machine-interpretable, these techniques enable better interoperability among heterogeneous agents. Symbolic knowledge extracted from one model can be shared, modified, and reinjected into other models, promoting collaboration and adaptability in multi-agent systems. This capability is particularly valuable in complex systems where multiple agents with diverse learning paradigms must work together seamlessly.