DP-Fusion: Token-Level Differentially Private Inference for Large Language Models
Overview
Overall Novelty Assessment
The paper introduces DP-Fusion, a mechanism for differentially private LLM inference that bounds the influence of sensitive tokens on generated outputs. It resides in the 'Differential Privacy for Next-Token Prediction' leaf, which contains five papers total, including the original work. This leaf sits within the broader 'Privacy-Preserving Inference Mechanisms' branch, indicating a moderately populated research direction focused on runtime privacy protections. The taxonomy shows this is an active but not overcrowded area, with sibling papers exploring related noise-injection and decoding strategies for autoregressive generation.
The taxonomy reveals that DP-Fusion's leaf is one of four under Privacy-Preserving Inference Mechanisms, alongside 'Privacy-Preserving In-Context Learning and Prompting' (five papers), 'Cryptographic and Secure Computation for Inference' (four papers), and 'Instance Obfuscation and Masking for Inference Privacy' (two papers). These neighboring leaves address complementary challenges: protecting prompts and exemplars, leveraging cryptographic primitives, or perturbing inputs rather than outputs. The scope note for the parent branch explicitly excludes training-time privacy, clarifying that DP-Fusion's focus on inference-time token influence bounds distinguishes it from adaptation or fine-tuning methods found in other taxonomy branches.
Among twenty-three candidates examined across three contributions, no refutable prior work was identified. The core DP-Fusion mechanism examined ten candidates with zero refutations, the document privatization application examined ten candidates with zero refutations, and the per-group privacy budget framework examined three candidates with zero refutations. This limited search scope—top-K semantic matches plus citation expansion—suggests that within the examined set, the fusion-based approach and per-token influence bounding appear distinct from prior noise-injection or decoding strategies. However, the analysis does not claim exhaustive coverage of all differential privacy inference techniques in the broader literature.
Based on the limited search of twenty-three candidates, DP-Fusion appears to occupy a recognizable niche within differential privacy for next-token prediction, with no clear overlap detected in the examined set. The taxonomy context indicates a moderately active research direction with established sibling work on noise calibration and adaptive budgets, suggesting the paper builds on known challenges in balancing privacy and utility during autoregressive generation. The absence of refutations in this scope does not preclude related work outside the top-K matches or in adjacent taxonomy leaves.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors introduce DP-FUSION, a novel differentially private inference mechanism that provides provable token-level privacy guarantees for large language models. The method works by inferring the LLM with and without sensitive tokens, then blending the output distributions to bound the influence of sensitive tokens on generated outputs.
The authors apply DP-FUSION to document privatization, demonstrating that their method can paraphrase documents containing personally identifiable information while achieving substantially better privacy-utility trade-offs than existing methods, with 6× lower perplexity than related DPI approaches.
The authors develop a framework that allows assigning different privacy budgets to different groups of sensitive tokens and implements a parallelizable inference procedure that computes multiple distributions (one public and multiple private) per generation step, enabling efficient token-level privacy control.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
[5] Differentially Private Decoding in Large Language Models PDF
[6] Submix: Practical private prediction for large-scale language models PDF
[7] Differentially Private Next-Token Prediction of Large Language Models PDF
[13] Adaptively private next-token prediction of large language models PDF
Contribution Analysis
Detailed comparisons for each claimed contribution
DP-FUSION mechanism for token-level differentially private LLM inference
The authors introduce DP-FUSION, a novel differentially private inference mechanism that provides provable token-level privacy guarantees for large language models. The method works by inferring the LLM with and without sensitive tokens, then blending the output distributions to bound the influence of sensitive tokens on generated outputs.
[7] Differentially Private Next-Token Prediction of Large Language Models PDF
[10] InferDPT: Privacy-preserving Inference for Black-box Large Language Models PDF
[19] Multi-tier privacy protection for large language models using differential privacy PDF
[23] PrivInfer: Privacy-Preserving Inference for Black-box Large Language Model PDF
[51] Privacy-preserving retrieval-augmented generation with differential privacy PDF
[52] DP-MLM: Differentially private text rewriting using masked language models PDF
[53] Hidden no more: Attacking and defending private third-party LLM inference PDF
[54] Split-and-denoise: Protect large language model inference with local differential privacy PDF
[55] Rag with differential privacy PDF
[56] Differential privacy in the era of generative AI: promises and challenges PDF
Document privatization application with improved privacy-utility trade-off
The authors apply DP-FUSION to document privatization, demonstrating that their method can paraphrase documents containing personally identifiable information while achieving substantially better privacy-utility trade-offs than existing methods, with 6× lower perplexity than related DPI approaches.
[59] Mitigating the privacy issues in retrieval-augmented generation (rag) via pure synthetic data PDF
[60] Truthful text sanitization guided by inference attacks PDF
[61] Robust utility-preserving text anonymization based on large language models PDF
[62] Enhancing Privacy While Preserving Context in Text Transformations by Large Language Models PDF
[63] Idt: Dual-task adversarial rewriting for attribute anonymization PDF
[64] DP-VAE: Human-Readable Text Anonymization for Online Reviews with Differentially Private Variational Autoencoders PDF
[65] Tau-Eval: A Unified Evaluation Framework for Useful and Private Text Anonymization PDF
[66] Silencing the Risk, Not the Whistle: A Semi-automated Text Sanitization Tool for Mitigating the Risk of Whistleblower Re-Identification PDF
[67] Self-Refining Language Model Anonymizers via Adversarial Distillation PDF
[68] The text anonymization benchmark (tab): A dedicated corpus and evaluation framework for text anonymization PDF
Per-group privacy budget framework with parallelizable inference
The authors develop a framework that allows assigning different privacy budgets to different groups of sensitive tokens and implements a parallelizable inference procedure that computes multiple distributions (one public and multiple private) per generation step, enabling efficient token-level privacy control.