Tractability via Low Dimensionality: The Parameterized Complexity of Training Quantized Neural Networks
Overview
Overall Novelty Assessment
The paper establishes parameterized complexity results for training quantized ReLU networks, contributing both hardness and tractability findings. It resides in the Parameterized Complexity Analysis leaf under Complexity-Theoretic Foundations, where it is currently the sole paper in that specific leaf. This positioning reflects a relatively sparse research direction: while the broader taxonomy encompasses eight papers across multiple branches, the parameterized complexity perspective on quantized training remains underexplored. The taxonomy reveals that most related work either addresses general computational hardness without parameterization or focuses on algorithmic and hardware-oriented methods rather than fine-grained complexity analysis.
The taxonomy structure shows that neighboring work divides into three main branches. Complexity-Theoretic Foundations contains sibling leaves on General Computational Hardness and Discrete Neural Network Theory, which establish fundamental intractability results and theoretical properties of discrete networks but do not employ parameterized analysis. Algorithm Design and Training Methods develops practical low-precision training strategies and gradient estimation techniques, explicitly excluding complexity-theoretic analysis. Hardware-Oriented Inference Optimization targets deployment efficiency rather than training complexity. The paper's focus on parameterized tractability via structural parameters like treewidth and dimensionality distinguishes it from these neighboring directions, which either lack parameterization or emphasize empirical performance over theoretical boundaries.
Among the 22 candidates examined, the contribution on fixed-parameter tractability via input dimensionality combined with width and output/error parameters shows one refutable candidate out of ten examined, suggesting some overlap with prior parameterized approaches. The other two contributions—strong lower bounds for quantized training and tractability results using treewidth—each examined ten and two candidates respectively, with no clear refutations found. The limited search scope means these statistics reflect top semantic matches rather than exhaustive coverage. The lower bounds contribution appears more novel within this sample, while the dimensionality-based tractability result encounters at least one overlapping prior work among the candidates reviewed.
Given the limited search of 22 candidates and the sparse population of the Parameterized Complexity Analysis leaf, the paper appears to occupy a relatively underexplored niche. The taxonomy context suggests that while quantized network training has received attention from algorithmic and hardware perspectives, the systematic parameterized complexity treatment remains less developed. The analysis captures top semantic matches but does not claim exhaustive field coverage, leaving open the possibility of additional related work outside the examined candidate set.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors prove that training quantized ReLU neural networks remains NP-hard even in the binary (2-bit) case and under severe restrictions including constant depth, constant width, single output neurons, and zero error bounds. These results exclude fixed-parameter tractability for standard architectural parameters like depth and width.
The authors develop algorithms showing that quantized neural network training becomes fixed-parameter tractable when parameterized by the input dimensionality combined with network width and either the output dimensionality or the error bound. These are the first non-trivial tractable parameterizations for this problem.
The authors prove that their fixed-parameter tractability results can be strengthened by replacing the width parameter with treewidth, a more general graph-theoretic measure. This is achieved through a structural insight (Lemma 1) bounding the non-zero indegree of neurons in solutions, enabling dynamic programming over tree decompositions.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
Contribution Analysis
Detailed comparisons for each claimed contribution
Strong lower bounds for quantized ReLU neural network training
The authors prove that training quantized ReLU neural networks remains NP-hard even in the binary (2-bit) case and under severe restrictions including constant depth, constant width, single output neurons, and zero error bounds. These results exclude fixed-parameter tractability for standard architectural parameters like depth and width.
Fixed-parameter tractability via input dimensionality combined with width and output/error parameters
The authors develop algorithms showing that quantized neural network training becomes fixed-parameter tractable when parameterized by the input dimensionality combined with network width and either the output dimensionality or the error bound. These are the first non-trivial tractable parameterizations for this problem.
[26] Complexity of Training ReLU Neural Network PDF
[19] Training neural networks is np-hard in fixed dimension PDF
[20] The computational complexity of ReLU network training parameterized by data dimensionality PDF
[21] SHAP Meets Tensor Networks: Provably Tractable Explanations with Parallelism PDF
[22] Training Fully Connected Neural Networks is -Complete PDF
[23] Complexity of Injectivity and Verification of ReLU Neural Networks PDF
[24] On the complexity of learning neural networks PDF
[25] Nonlinear initialization methods for low-rank neural networks PDF
[27] Learning Deep ReLU Networks Is Fixed-Parameter Tractable PDF
[28] Beating the Perils of Non-Convexity: Guaranteed Training of Neural Networks using Tensor Methods PDF
Strengthened tractability results using treewidth instead of width
The authors prove that their fixed-parameter tractability results can be strengthened by replacing the width parameter with treewidth, a more general graph-theoretic measure. This is achieved through a structural insight (Lemma 1) bounding the non-zero indegree of neurons in solutions, enabling dynamic programming over tree decompositions.