KnowledgeSmith: Uncovering Knowledge Updating in LLMs with Model Editing and Unlearning
Overview
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors propose KnowledgeSmith, a unified framework that casts both knowledge editing and machine unlearning as instances of a single constrained optimization problem. This formulation enables systematic comparison and analysis of how LLMs update knowledge through these two complementary intervention strategies.
The authors develop an automatic pipeline that transforms existing knowledge graph datasets into dynamic benchmarks for evaluating knowledge interventions. The pipeline generates hierarchical probes across root, intermediate, and leaf levels, enabling controlled studies of how modifications propagate through model knowledge at multiple scales.
Through extensive experiments across multiple model families and domains, the authors uncover fundamental properties of knowledge updating in LLMs, including propagation asymmetry, plasticity scaling laws, consistency-capacity tradeoffs, subject-dependent update behavior, and unified failure modes. These findings reveal how editing and unlearning differ in their effects on model knowledge.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
[36] How new data permeates LLM knowledge and how to dilute it PDF
Contribution Analysis
Detailed comparisons for each claimed contribution
KnowledgeSmith unified framework for knowledge updating
The authors propose KnowledgeSmith, a unified framework that casts both knowledge editing and machine unlearning as instances of a single constrained optimization problem. This formulation enables systematic comparison and analysis of how LLMs update knowledge through these two complementary intervention strategies.
[22] Resolving editing-unlearning conflicts: A knowledge codebook framework for large language model updating PDF
[28] Towards Lifelong Learning of Large Language Models: A Survey PDF
[51] Rethinking machine unlearning for large language models PDF
[52] Co-occurrence is not factual association in language models PDF
[53] UniErase: Towards Balanced and Precise Unlearning in Language Models PDF
[54] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions PDF
[55] MMUnlearner: Reformulating Multimodal Machine Unlearning in the Era of Multimodal Large Language Models PDF
[56] Structured Knowledge Integration and Memory Modeling in Large Language Systems PDF
[57] UniErase: Unlearning Token as a Universal Erasure Primitive for Language Models PDF
[58] Lifelong learning of large language model based agents: A roadmap PDF
Automatic KG-based benchmark generation pipeline
The authors develop an automatic pipeline that transforms existing knowledge graph datasets into dynamic benchmarks for evaluating knowledge interventions. The pipeline generates hierarchical probes across root, intermediate, and leaf levels, enabling controlled studies of how modifications propagate through model knowledge at multiple scales.
[4] A comprehensive study of knowledge editing for large language models PDF
[60] Towards verifiable generation: A benchmark for knowledge-aware language model attribution PDF
[61] Biokgbench: A knowledge graph checking benchmark of ai agent for biomedical science PDF
[62] RARE: Retrieval-Aware Robustness Evaluation for Retrieval-Augmented Generation Systems PDF
[63] Assessing and Improving Factual Answers from Knowledge Graphs and Language Models PDF
[64] Benchmarking large language models in complex question answering attribution using knowledge graphs PDF
[65] A knowledge-graph-based intrinsic test for benchmarking medical concept embeddings and pretrained language models PDF
[66] GAPS: A Clinically Grounded, Automated Benchmark for Evaluating AI Clinicians PDF
[67] Towards Dynamically Generated KGQA Benchmark Datasets for Memorization-Resistant Evaluations PDF
[68] LENS: Layers of Evaluation of Hallucination in GenAI Systems PDF
Empirical insights on LLM knowledge updating mechanisms
Through extensive experiments across multiple model families and domains, the authors uncover fundamental properties of knowledge updating in LLMs, including propagation asymmetry, plasticity scaling laws, consistency-capacity tradeoffs, subject-dependent update behavior, and unified failure modes. These findings reveal how editing and unlearning differ in their effects on model knowledge.