UDIS: A User-query Driven Framework for Image Forgery Localization
Overview
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors propose a new conceptual framework where interpretability in image forgery localization should be driven by regional user queries instead of global outcome-based explanations. This paradigm shift establishes a foundation for addressing weak visual-text alignment in existing methods.
The authors develop a novel framework called UDIS that implements the user-query driven principle through two specialized modules: QGM aligns user queries with visual attention at the input level, while EAM aligns explanatory textual knowledge with forgery localization capability at the output level.
The authors curate a training dataset that includes both generic forensic questions and content-aware queries tailored to specific image regions, along with corresponding authenticity evidence annotations. This dataset enables training models under the user-query driven paradigm.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
Contribution Analysis
Detailed comparisons for each claimed contribution
User-query driven paradigm for IFL interpretability
The authors propose a new conceptual framework where interpretability in image forgery localization should be driven by regional user queries instead of global outcome-based explanations. This paradigm shift establishes a foundation for addressing weak visual-text alignment in existing methods.
[1] FakeShield: Explainable Image Forgery Detection and Localization via Multi-modal Large Language Models PDF
[22] Explainable Image-Centric Forgery Detection: A Survey PDF
[60] Towards Human Explainable Digital Forensics: Generating Human Interpretable Evidence for Semantic Understanding in Manipulated Images and Text PDF
[61] Multi-scale query-based transformer for image forgery localization PDF
[62] Rethinking Vision-Language Model in Face Forensics: Multi-Modal Interpretable Forged Face Detector PDF
[63] Learning patch-channel correspondence for interpretable face forgery detection PDF
[64] Explainable DualâStream Attention Network for Image Forgery Detection and Localisation Using Contrastive Learning PDF
[65] Literature Survey of Image Forgery Detection Using Machine Learning PDF
[66] IVY-FAKE: A Unified Explainable Framework and Benchmark for Image and Video AIGC Detection PDF
[67] So-Fake: Benchmarking and Explaining Social Media Image Forgery Detection PDF
UDIS framework with QGM and EAM modules
The authors develop a novel framework called UDIS that implements the user-query driven principle through two specialized modules: QGM aligns user queries with visual attention at the input level, while EAM aligns explanatory textual knowledge with forgery localization capability at the output level.
[22] Explainable Image-Centric Forgery Detection: A Survey PDF
[69] An information theoretic approach for attention-driven face forgery detection PDF
[70] Exploiting multi-domain visual information for fake news detection PDF
[71] Generalizing face forgery detection with high-frequency features PDF
[72] Lightweight End-to-End Patch-Based Self-Attention Network for Robust Image Forgery Detection PDF
[73] Zooming In on Fakes: A Novel Dataset for Localized AI-Generated Image Detection with Forgery Amplification Approach PDF
[74] A Similarity-Based Positional Attention-Aided Deep Learning Model for CopyâMove Forgery Detection PDF
[76] Forgery-aware adaptive learning with vision transformer for generalized face forgery detection PDF
Dataset with region-specific queries and authenticity evidence
The authors curate a training dataset that includes both generic forensic questions and content-aware queries tailored to specific image regions, along with corresponding authenticity evidence annotations. This dataset enables training models under the user-query driven paradigm.