<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Sparsification on Feng's Blog</title><link>http://fengwang.github.io/tags/sparsification/</link><description>Recent content in Sparsification on Feng's Blog</description><generator>Hugo -- gohugo.io</generator><language>en-us</language><lastBuildDate>Mon, 11 May 2026 00:00:00 +0000</lastBuildDate><atom:link href="http://fengwang.github.io/tags/sparsification/index.xml" rel="self" type="application/rss+xml"/><item><title>MoE Expert Pruning: What Works, What Doesn't, and What We Still Don't Know</title><link>http://fengwang.github.io/posts/moe-expert-pruning/</link><pubDate>Mon, 11 May 2026 00:00:00 +0000</pubDate><guid>http://fengwang.github.io/posts/moe-expert-pruning/</guid><description>&lt;p&gt;I spent the last week reading seven papers on expert compression for Mixture-of-Experts models. I went in assuming the landscape was settled: expert pruning was a useful technique for shrinking models, expert merging was a promising alternative, and the choice between them was mostly a matter of taste. I came out with a very different picture. The pruning-vs-merging debate flipped in later 2025, and the reason it flipped tells you something fundamental about how these models actually work.&lt;/p&gt;
&lt;p&gt;The core tension is simple. Sparse Mixture-of-Experts models activate only a fraction of their parameters per token, but they still occupy all of them in memory. Mixtral 8×7B activates 2 of 8 experts per layer — yet all 8 experts (45B of 47B total parameters) must sit on the GPU. NLLB-200 has 1,536 experts; you need four 32GB GPUs just to load the thing. Expert compression asks: can we drop or combine the experts that rarely get used, and how much does it cost?&lt;/p&gt;
&lt;p&gt;The answer, across all seven papers, is surprising in two ways. First, you can drop far more experts than I expected, and the costs are concentrated in specific capabilities rather than spread evenly. Second, and more unexpectedly, &lt;strong&gt;when it comes to actual generative tasks — code, math, creative writing — expert pruning is decisively better than expert merging.&lt;/strong&gt; The merging methods that looked good on multiple-choice benchmarks collapse on tasks that require the model to actually generate tokens. The reason is structural, not empirical: merging removes the router&amp;rsquo;s fine-grained control over experts, and on generative tasks, that control matters.&lt;/p&gt;
&lt;h2 id="why-expert-pruning-works-at-all"&gt;Why expert pruning works at all&lt;/h2&gt;
&lt;p&gt;Here&amp;rsquo;s the one-sentence model: &lt;strong&gt;expert utilization in MoE models is long-tailed — a handful of experts do most of the work, and the rest are along for the ride.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Think of it like a restaurant kitchen during a dinner rush. You have eight chefs at eight stations. On a given night, two chefs handle 80% of the orders. The other six are standing by, occasionally contributing a garnish, mostly drawing salary. If you fire four of them and redistribute their stations, dinner still gets served — maybe even faster, because the remaining chefs spend less time coordinating. That is expert pruning.&lt;/p&gt;
&lt;p&gt;The numbers back this up. Heatmap analysis of Mixtral 8×7B on MMLU shows stark unevenness: Expert #2 in Layers 26 and 30 is heavily activated while Expert #7 in Layers 22 and 23 is barely touched [1]. The same pre-trained MoE model produces substantially different expert contribution patterns when fine-tuned on different tasks [2]. This task-specificity cuts both ways — it means pruning must be calibrated to the deployment domain, but it also means aggressive pruning is possible for narrow use cases.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The distribution is not just long-tailed — it&amp;rsquo;s task-dependent.&lt;/strong&gt; An expert that dominates on MNLI might be silent on CoLA. This is the core insight behind every pruning strategy: you are not removing universal knowledge. You are removing specialists that your specific task does not call.&lt;/p&gt;
&lt;p&gt;Where the analogy breaks: experts are not independent chefs. The router learns to distribute tokens across experts during pre-training, and removing experts changes the routing distribution for the survivors. This is why naive pruning based on activation frequency alone performs &lt;em&gt;worse than random pruning&lt;/em&gt; [3]. The router expects a full kitchen.&lt;/p&gt;
&lt;h3 id="knowledge-redundancy-the-surprising-overcapacity-result"&gt;Knowledge redundancy: the surprising overcapacity result&lt;/h3&gt;
&lt;p&gt;Here is the finding that made me stop and re-read: pruning 4 of 8 experts in Mixtral 8×7B-Instruct &lt;em&gt;improves&lt;/em&gt; SQuAD accuracy from 53.4% to 75.4%, without updating any remaining expert parameters [4]. This is not a typo. Removing half the experts makes the model &lt;em&gt;better&lt;/em&gt; at question answering.&lt;/p&gt;
&lt;p&gt;The mechanism: pruning simplifies the routing problem. With 8 experts, the router must learn to partition the hidden space across many specialists — a hard optimization problem. With 4, the routing is easier, and the remaining experts each get a cleaner slice of the input distribution. The router stops sending ambiguous tokens to the wrong specialist.&lt;/p&gt;
&lt;p&gt;This overcapacity effect is not universal, but it recurs: on data-limited downstream tasks, a single-expert model can outperform the full multi-expert counterpart. After fine-tuning a pruned Mixtral 8×7B on MetaMathQA, the 7-expert model slightly &lt;em&gt;exceeds&lt;/em&gt; the original 8-expert model on GSM8K (81.50 vs. 81.43) [3]. A single expert in Mixtral 8×7B-Instruct operates without model collapse [4].&lt;/p&gt;
&lt;h2 id="pruning-vs-merging-the-debate-that-flipped-in-2025"&gt;Pruning vs. merging: the debate that flipped in 2025&lt;/h2&gt;
&lt;p&gt;Until mid-2025, the story seemed clear. Expert merging — clustering and averaging experts rather than discarding them — was winning. M-SMoE and HC-SMoE showed that merging outperformed pruning when measured by perplexity and multiple-choice question answering benchmarks. If you only looked at those numbers, merging was the smarter choice. Retain information from all experts. Avoid the binary brutality of pruning.&lt;/p&gt;
&lt;p&gt;Then REAP showed up and asked: what happens when you actually make these models &lt;em&gt;generate&lt;/em&gt; tokens?&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The answer is a head-on collision between the two approaches.&lt;/strong&gt; On code generation, REAP achieves a mean accuracy decrease of only 1.9% at 25% compression and 6.9% at 50% compression. Merging methods? HC-SMoE and M-SMoE degrade more than 5% at 25% and more than 20% at 50% [7]. On creative writing and mathematical reasoning, the same pattern holds. Merging is not slightly worse — it is qualitatively broken on generative tasks at 50% compression.&lt;/p&gt;
&lt;p&gt;Lasby et al. didn&amp;rsquo;t just report the numbers. They derived &lt;em&gt;why&lt;/em&gt; this has to be the case. When a router selects two experts f_i and f_j for a token, it produces a dynamic mixture r(x)·f_i(x) + (1−r(x))·f_j(x), where the mixing ratio r(x) depends on the input. After merging, the router must apply the summed gate to a constant convex combination — a static merged expert. The merged model must approximate a dynamic, input-dependent target with a static one. The resulting irreducible error is proportional to the router&amp;rsquo;s policy variability Var[r(x)] and the functional gap between the merged experts ∥Δ_ij∥ [7].&lt;/p&gt;
&lt;p&gt;Pruning doesn&amp;rsquo;t have this problem. When you prune expert j, the router still controls each surviving expert independently. Pruning only incurs error when the pruned expert was in the top-k set, and that error is proportional to its gate-value g_j — it does not penalize policy variability at all [7]. The mathematical difference is clean: pruning is a coordinate subspace operation that preserves the functional manifold&amp;rsquo;s topology. Merging introduces novel functions and collapses the manifold toward its center — by up to 100× reduction in spread in late layers of high-granularity models [7].&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Here&amp;rsquo;s what this means in practice.&lt;/strong&gt; I can now look at a compressed model and predict failure modes based on the operation, not just the sparsity level. Merged model outputs have significantly lower N-gram diversity and their logits diverge from the original model more rapidly during auto-regressive generation [7]. The tokens drift. The model stops sounding like itself. MC benchmarks missed this entirely because they never asked the model to string tokens together — they only asked it to rank answer choices in a single forward pass.&lt;/p&gt;
&lt;p&gt;One more uncomfortable finding: when merging &lt;em&gt;does&lt;/em&gt; work well, look closer. HC-SMoE produces a high prevalence of &lt;strong&gt;singleton clusters&lt;/strong&gt; — single-expert clusters that are functionally indistinguishable from keeping the expert unmerged [7]. The &amp;ldquo;merging&amp;rdquo; that succeeds is pruning plus a few mega-clusters of the truly redundant experts. And those mega-clusters are fragile: restricting the maximum cluster size to 32 experts causes large accuracy drops [7].&lt;/p&gt;
&lt;p&gt;A separate problem compounds this. The L2-distance between clustered expert weights, even after weight-matching permutation, greatly exceeds the distance between pretrained and instruction-fine-tuned checkpoints. Singular-vector alignment remains poor [7]. Merging experts is fundamentally harder than the widely successful technique of model merging, and we should stop assuming the two are similar problems.&lt;/p&gt;
&lt;h2 id="how-to-score-expert-importance"&gt;How to score expert importance&lt;/h2&gt;
&lt;p&gt;The choice of importance criterion is the single biggest lever in expert pruning. I organize the criteria by what information they use.&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Criterion&lt;/th&gt;
&lt;th&gt;Source&lt;/th&gt;
&lt;th&gt;What it measures&lt;/th&gt;
&lt;th&gt;Best result&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Alpha score (accumulated gating weight)&lt;/td&gt;
&lt;td&gt;Chen et al. 2022 [2]&lt;/td&gt;
&lt;td&gt;Weighted contribution to output&lt;/td&gt;
&lt;td&gt;Single expert preserves 99.3% of full MoE&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Soft counting (accumulated softmax)&lt;/td&gt;
&lt;td&gt;Muzio et al. 2024 [1]&lt;/td&gt;
&lt;td&gt;Confidence margin of selection&lt;/td&gt;
&lt;td&gt;25% sparsity: 3.85 pp MMLU drop&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Min-EAN (activation norm)&lt;/td&gt;
&lt;td&gt;Jaiswal et al. 2025 [5]&lt;/td&gt;
&lt;td&gt;Minimum activation magnitude&lt;/td&gt;
&lt;td&gt;14.02 PPL at 75% sparsity&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;REAP&lt;/strong&gt; (conditional g_j∥f_j∥)&lt;/td&gt;
&lt;td&gt;Lasby et al. 2025 [7]&lt;/td&gt;
&lt;td&gt;Gate × activation, conditional&lt;/td&gt;
&lt;td&gt;Near-lossless at 50%, up to 1T params&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Importance product (top1 × exp(conf))&lt;/td&gt;
&lt;td&gt;Koishekenov et al. 2023 [6]&lt;/td&gt;
&lt;td&gt;Combined activity and confidence&lt;/td&gt;
&lt;td&gt;80% pruning, chrF++ Δ = 0.29&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Activation frequency alone&lt;/td&gt;
&lt;td&gt;Lu et al. 2024 [3]&lt;/td&gt;
&lt;td&gt;Simple token count&lt;/td&gt;
&lt;td&gt;Worse than random&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;REAP deserves special attention because it&amp;rsquo;s the first criterion explicitly designed to minimize the reconstruction error bound. Its saliency score computes the conditional average of g_j(x)·∥f_j(x)∥ over only those tokens where expert j is active [7]. This decouples functional impact from usage frequency — a specialist expert that activates rarely but contributes heavily when it does won&amp;rsquo;t be pruned just because it&amp;rsquo;s infrequent. Min-EAN held the previous crown among 16 criteria benchmarked by MC-Suite [5]. REAP now looks like the new baseline for generative tasks, especially at scale.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The easy heuristic is still wrong.&lt;/strong&gt; Simple activation frequency — counting how many tokens each expert processes — does worse than random selection [3]. The router&amp;rsquo;s assignment frequency is not the same as contribution.&lt;/p&gt;
&lt;p&gt;Domain-specific calibration delivers the biggest gap I&amp;rsquo;ve seen in any compression result. When REAP calibrates on C4 (general pre-training data) instead of domain-specific data (evol-codealpaca for code), code generation accuracy collapses — some compressed models produce 0% accuracy, failing to output coherent code at all [7]. This is not a matter of degree. The calibrating dataset determines whether the pruned model works or is completely useless on the target task. And this was already visible in earlier work: using MATH instead of C4 for calibration shifts expert selections in 28 of 32 layers of Mixtral 8×7B [3].&lt;/p&gt;
&lt;h2 id="pruning-strategies-the-choices-that-matter"&gt;Pruning strategies: the choices that matter&lt;/h2&gt;
&lt;p&gt;Once you have an importance score, you need to decide &lt;em&gt;how&lt;/em&gt; to use it. Three choices define your strategy.&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Choice&lt;/th&gt;
&lt;th&gt;Options&lt;/th&gt;
&lt;th&gt;Trade-off&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Scope&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Global vs. layer-wise&lt;/td&gt;
&lt;td&gt;Global: better quality but variable per-layer counts. Layer-wise: fixed memory layout but lower ceiling&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Schedule&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;One-shot vs. iterative&lt;/td&gt;
&lt;td&gt;One-shot: fast but importance rankings are stale post-pruning. Iterative: ~2× better PPL but needs re-estimation&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Timing&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Eager vs. staged&lt;/td&gt;
&lt;td&gt;Eager: more optimization steps for survivors. Staged: better importance estimates from longer observation&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;&lt;strong&gt;Global vs. layer-wise.&lt;/strong&gt; Global pruning — sorting all experts across all layers by a single importance ranking — outperforms layer-wise on quality because it avoids the constraint of keeping a fixed number per layer [1]. But it creates deployment headaches: variable per-layer expert counts mean variable memory usage across tasks, requiring model recreation for each configuration [6]. Layer-wise pruning gives predictable memory layouts at the cost of some quality.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;One-shot vs. iterative.&lt;/strong&gt; One-shot pruning drops experts in a single pass. The problem: after you remove experts, the importance rankings of the survivors change. Iterative pruning re-estimates importance after each round, achieving ~2× better perplexity. Add task-agnostic finetuning between rounds and you get ~3× better [5]. One-shot and iterative pruning identify &lt;em&gt;substantially different subsets&lt;/em&gt; of experts at the same sparsity level — they produce effectively different subnetworks [5]. REAP demonstrates that with the right criterion, one-shot pruning can be remarkably effective even at 50% compression on models up to 1T parameters [7], but the iterative advantage likely still holds.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Eager vs. staged.&lt;/strong&gt; Eager (progressive) pruning drops experts early using a dynamic threshold T = β / Z where Z is the number of surviving experts [2]. The earlier you drop, the more training steps you can dedicate to the selected expert. Eager consistently wins [2].&lt;/p&gt;
&lt;h3 id="the-nllb-200-special-case-language-specific-pruning"&gt;The NLLB-200 special case: language-specific pruning&lt;/h3&gt;
&lt;p&gt;The NLLB-200 translation model surfaces a phenomenon that the Mixtral papers miss: &lt;strong&gt;language-specific expert emergence&lt;/strong&gt;. In the decoder, Jaccard similarity of selected experts is 68–87% for the same target language versus only 13–39% for different target languages [6]. Per-language pruning (source language for encoder, target language for decoder) performs as well as per-language-pair pruning while requiring only L configurations instead of L² [6]. An unbalanced 3:1 encoder-to-decoder ratio yields the best quality [6].&lt;/p&gt;
&lt;h2 id="beyond-pruning-complementary-techniques"&gt;Beyond pruning: complementary techniques&lt;/h2&gt;
&lt;p&gt;Static expert pruning rarely stands alone. Four complementary techniques compound its gains, and there&amp;rsquo;s now a clearer distinction between approaches that help and approaches that hurt.&lt;/p&gt;
&lt;h3 id="expert-merging-the-post-pruning-variant--and-why-its-different"&gt;Expert merging (the post-pruning variant — and why it&amp;rsquo;s different)&lt;/h3&gt;
&lt;p&gt;EEP&amp;rsquo;s expert merging is not the same thing as HC-SMoE or M-SMoE. EEP merges pruned expert knowledge into survivors &lt;em&gt;after&lt;/em&gt; pruning, using learned Router Mapping and Expert Merging matrices, adding 5–7% accuracy improvement [4]. This is a knowledge transfer operation — the pruned experts are already gone and their useful information is folded into the survivors. It&amp;rsquo;s fundamentally different from the HC-SMoE/M-SMoE approach of replacing entire expert groups with merged averages, which removes router independence and causes the collapse described above. The EEP variant is a net positive. The HC-SMoE variant is not, unless you&amp;rsquo;re only evaluating on multiple-choice.&lt;/p&gt;
&lt;h3 id="dynamic-expert-skipping"&gt;Dynamic expert skipping&lt;/h3&gt;
&lt;p&gt;Static pruning removes experts permanently. Dynamic skipping removes them &lt;em&gt;conditionally&lt;/em&gt; — dropping the second-ranked expert for a token when its routing weight is below a threshold β times the top expert&amp;rsquo;s weight, yielding ~50% skipping probability [3]. The key finding: skipping is &lt;em&gt;complementary&lt;/em&gt; to pruning. A model pruned to 6 experts with skipping achieves the same speedup as pruning alone to 4 experts, but with higher accuracy [3]. You get the speedup without the full accuracy cost.&lt;/p&gt;
&lt;h3 id="active-expert-reduction-and-finetuning"&gt;Active expert reduction and finetuning&lt;/h3&gt;
&lt;p&gt;Switching from top-2 to top-1 expert activation reduces forward-pass FLOPs by ~27% in Mixtral [1], but zero-shot top-1 routing drops SST5 accuracy from 50.8% to 42.6%. Recovery via entropy-based gating regularization plus annealing top-k reduction closes most of this gap (51.8% vs. 53.6% top-2) [1].&lt;/p&gt;
&lt;p&gt;Task-agnostic finetuning (~1M tokens; benefits saturate) corrects the skewed load distribution caused by removing router entries. It doesn&amp;rsquo;t change which experts are selected — it mitigates impact through load rebalancing. This finetuning is central enough that iterative prune-estimate-finetune cycles produce what Jaiswal et al. call &lt;strong&gt;MoE Lottery Subnetworks&lt;/strong&gt; [5].&lt;/p&gt;
&lt;h3 id="quantization-after-pruning"&gt;Quantization after pruning&lt;/h3&gt;
&lt;p&gt;Pruning combines naturally with quantization without additional steps, unlike merging which requires block-scale reconciliation for block quantization formats [7]. Combining REAP with 4-bit quantization on Kimi-K2 achieves 87.5% total size reduction — a compression rate neither technique can reach alone [7].&lt;/p&gt;
&lt;h2 id="what-the-numbers-actually-say"&gt;What the numbers actually say&lt;/h2&gt;
&lt;p&gt;Across all seven papers, the efficiency-performance trade-off is more favorable than I expected.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;At moderate sparsity (25–50% experts removed), the accuracy cost on generative tasks is remarkably low — provided you prune, not merge.&lt;/strong&gt; REAP achieves a 1.9% mean accuracy decrease at 25% compression and 6.9% at 50% on coding benchmarks [7]. On Qwen3-Coder-480B and Kimi-K2, pruning 50% of experts drops code generation accuracy by only 1.2% [7]. On SWE-Bench (agentic software engineering), REAP-pruned Kimi-K2 at 50% compression actually slightly &lt;em&gt;exceeds&lt;/em&gt; the baseline (0.576 vs. 0.554) [7].&lt;/p&gt;
&lt;p&gt;Compare with merging at the same compression: HC-SMoE and M-SMoE see &amp;gt;5% accuracy decrease at 25% and &amp;gt;20% at 50% on the same coding benchmarks [7]. Merging looks reasonable on MC benchmarks (~4% decrease at 25%) but the MC numbers don&amp;rsquo;t predict generative performance. This gap — between discriminative and generative evaluation — is what the pre-REAP literature missed.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;At high sparsity (75–80% experts removed), the numbers depend heavily on task type and recovery technique.&lt;/strong&gt; At 75% sparsity, Min-EAN achieves 14.02 PPL versus 34.47 random [5]. NLLB-200 at 80% pruning achieves chrF++ 36.61 versus 36.81 full — a delta of −0.2 [6]. Expert dropping predominantly degrades instruction-following, not pretraining knowledge or reasoning; these capabilities can be substantially restored through K-shot examples or fine-tuning [5].&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The fastest path to deployment, based on the evidence, is: Base model → expert pruning → finetuning → instruction tuning.&lt;/strong&gt; Expert dropping yields greater benefits before instruction tuning than after [5]. With SFT after pruning, high-sparsity models can outperform full counterparts on easier tasks like BoolQ and ARC-easy.&lt;/p&gt;
&lt;h2 id="where-the-standard-story-breaks"&gt;Where the standard story breaks&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;The standard story:&lt;/strong&gt; expert utilization is long-tailed. You prune the tail. Light finetuning recovers the loss. Any compression method that reduces the expert count should work about as well.&lt;/p&gt;
&lt;p&gt;This story is wrong in ways that matter, and REAP is the paper that forced the correction.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Pruning and merging are not interchangeable.&lt;/strong&gt; They produce qualitatively different models with different failure modes. Merging loses the router&amp;rsquo;s input-dependent control — an irreducible error proportional to the router&amp;rsquo;s policy variability. Pruning preserves it. On discriminative tasks, the difference is hidden because ranking answers in a single forward pass doesn&amp;rsquo;t require the model to maintain coherent generation. On generative tasks, the difference is dramatic [7].&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Discriminative metrics like perplexity and MC accuracy are poor proxies for generative quality.&lt;/strong&gt; This sounds obvious in retrospect, but the field relied on these metrics to claim merging &amp;gt; pruning. Jaiswal et al. had already warned that perplexity can be misleading for compressed LLMs [5]. REAP proved it with a clean experiment: merging methods that looked competitive on MC benchmarks collapsed on code generation to the point of producing 0% accuracy outputs [7]. If you evaluate a compressed model only on MC, you haven&amp;rsquo;t evaluated it at all for real-world use.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Expert-level sparsification still beats weight pruning, and the argument is now stronger.&lt;/strong&gt; Across equivalent sparsity levels, dropping whole experts outperforms Wanda by ~3.6% average accuracy and ~16.2% on ARC-c [5]. And expert pruning preserves manifold topology while weight pruning may not — the geometric analysis from REAP [7] provides a structural argument for why whole-expert removal is the more principled approach.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;High-vocabulary-coverage experts hurt when dropped&lt;/strong&gt; — the specialist-generalist tension is real. If an expert handles many distinct tokens, removing it does outsized damage [5]. This suggests that pre-training methods that push experts toward specialization may make pruning easier in one sense (more experts are &amp;ldquo;dispensable&amp;rdquo;) but harder in another (the remaining generalist experts carry structural load that can&amp;rsquo;t be removed).&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Dominant experts have lower stable-rank&lt;/strong&gt; — a clean signal for identification but not yet exploited for additional compression [5].&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The second-pass degradation puzzle.&lt;/strong&gt; Two-pass eager-drop pruning degrades performance compared to a single pass, with average GLUE dropping by 0.58 points [2]. More iteration is not always better. But the REAP paper shows that a strong criterion in a single pass can go remarkably far — one-shot REAP on a 1T-parameter model at 50% compression is near-lossless on code [7]. The lesson isn&amp;rsquo;t &amp;ldquo;one-shot &amp;gt; iterative.&amp;rdquo; It&amp;rsquo;s that criterion quality and scale-appropriate calibration dominate the one-shot vs. iterative trade-off.&lt;/p&gt;
&lt;h2 id="boundary-conditions"&gt;Boundary conditions&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;The enumeration-based pruning approach in Lu et al. [3] works for 4–8 experts per layer but becomes computationally intractable at 32+ experts. The combinatorial explosion is unresolved.&lt;/li&gt;
&lt;li&gt;Gradient-free methods like EEP&amp;rsquo;s evolutionary strategy [4] have been studied only on the Mixtral family. Whether they generalize to architectures with many more experts is unknown.&lt;/li&gt;
&lt;li&gt;HC-SMoE&amp;rsquo;s mega-clusters containing tens of experts are fragile — restricting maximum cluster size to 32 causes large accuracy drops [7]. Coherently merging many experts remains an open problem.&lt;/li&gt;
&lt;li&gt;Hallucination and over-generation have been observed in pruned translation models, with global threshold methods more sensitive than fixed-per-layer pruning [6].&lt;/li&gt;
&lt;li&gt;All seven papers study static expert counts per layer. None address dynamic architectures where expert count varies by input complexity.&lt;/li&gt;
&lt;li&gt;Qwen2-MoE experts are notably homogeneous — the &amp;ldquo;expert specialization&amp;rdquo; narrative is architecture-dependent [4].&lt;/li&gt;
&lt;li&gt;Merging methods require recording activations from every expert for every token during calibration, making them more expensive at scale than pruning methods [7].&lt;/li&gt;
&lt;li&gt;The pruning-vs-merging analysis from REAP [7] applies to one-shot, no-fine-tuning compression. Whether fine-tuning after merging can recover the policy variability loss is not addressed.&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id="open-questions"&gt;Open questions&lt;/h2&gt;
&lt;ol&gt;
&lt;li&gt;The REAP criterion&amp;rsquo;s derivation minimizes a reconstruction error bound assuming one-shot pruning. Can the same router-gate × activation-norm logic be extended to iterative pruning, and does it produce even better results?&lt;/li&gt;
&lt;li&gt;Merging fails on generative tasks because it removes router independence. But could you &lt;em&gt;train&lt;/em&gt; a model to be merge-friendly — by regularizing expert functional similarity or router policy smoothness — and get the memory savings of merging without the generative collapse?&lt;/li&gt;
&lt;li&gt;What is the interaction between expert pruning and quantization at scale? REAP showed the combination works [7], but only on one model family (Kimi-K2 at 4-bit). Do pruned experts tolerate lower-bit quantization better or worse than full experts?&lt;/li&gt;
&lt;li&gt;The &amp;ldquo;MoE Lottery Subnetworks&amp;rdquo; framing [5] has only been studied up to Mixtral 8×22B. Does it hold at the scale REAP demonstrated (480B–1T parameters)?&lt;/li&gt;
&lt;li&gt;The vocabulary coverage finding [5] — high-coverage experts hurt when dropped — implies a tension with specialization. If you make experts more specialized, you might make pruning easier, but you risk creating fragile specialists that cannot be removed. Which direction wins?&lt;/li&gt;
&lt;li&gt;No paper in this set studies pruning during pre-training rather than post-training. Could you train an MoE from scratch knowing it will be pruned and get a better result?&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;[[Q]] Six months from now: has the community converged on REAP as the default one-shot pruning criterion, or has the merging community produced a variant that recovers router independence and closes the generative-task gap?&lt;/p&gt;
&lt;h2 id="references"&gt;References&lt;/h2&gt;
&lt;ol&gt;
&lt;li&gt;Muzio et al., &amp;ldquo;SEER-MoE: Sparse Expert Efficiency through Regularization for Mixture-of-Experts&amp;rdquo;, arXiv:2404.05089, 2024.&lt;/li&gt;
&lt;li&gt;Chen et al., &amp;ldquo;Task-Specific Expert Pruning for Sparse Mixture-of-Experts&amp;rdquo;, arXiv:2206.00277, 2022.&lt;/li&gt;
&lt;li&gt;Lu et al., &amp;ldquo;Not All Experts are Equal: Efficient Expert Pruning and Skipping for Mixture-of-Experts Large Language Models&amp;rdquo;, ACL 2024.&lt;/li&gt;
&lt;li&gt;Liu et al., &amp;ldquo;Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference Costs&amp;rdquo;, arXiv:2407.00945, 2024.&lt;/li&gt;
&lt;li&gt;Jaiswal et al., &amp;ldquo;Finding Fantastic Experts in MoEs: A Unified Study for Expert Dropping Strategies and Observations&amp;rdquo;, arXiv:2504.05586, 2025.&lt;/li&gt;
&lt;li&gt;Koishekenov et al., &amp;ldquo;Memory-efficient NLLB-200: Language-specific Expert Pruning of a Massively Multilingual Machine Translation Model&amp;rdquo;, ACL 2023.&lt;/li&gt;
&lt;li&gt;Lasby et al., &amp;ldquo;REAP the Experts: Why Pruning Prevails for One-Shot MoE Compression&amp;rdquo;, arXiv:2510.13999, 2025.&lt;/li&gt;
&lt;/ol&gt;</description></item><item><title>LLM/VLM Compression Foundations</title><link>http://fengwang.github.io/posts/llm-vlm-compression-foundations-clean/</link><pubDate>Sun, 10 May 2026 00:00:00 +0000</pubDate><guid>http://fengwang.github.io/posts/llm-vlm-compression-foundations-clean/</guid><description>&lt;p&gt;I started looking at model compression because the numbers didn&amp;rsquo;t add up. My GPU has 24GB of VRAM and the models I want to run need 40GB. The gap is a factor of two, which quantization claims to solve. But then I found papers about pruning, and distillation, and token compression, and hardware-aware NAS, and suddenly the question wasn&amp;rsquo;t &amp;ldquo;which technique&amp;rdquo; but &amp;ldquo;which combination, in what order, for which hardware.&amp;rdquo;&lt;/p&gt;
&lt;p&gt;This article is my attempt to organize what I&amp;rsquo;ve learned into a coherent map. It is not a survey — there are good surveys for that. It is a working notebook: what I understand, what surprised me, and what I still can&amp;rsquo;t explain.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Thesis:&lt;/strong&gt; Compression works because neural networks are overparameterized for the expressivity they actually use. The hard part is knowing which bits are the ones that don&amp;rsquo;t matter — and that answer depends on what you&amp;rsquo;re compressing (text vs. vision-language), how you remove it (prune, quantize, or distill), what order you apply the steps (P-KD-Q), and what hardware runs the result.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Scope:&lt;/strong&gt; This covers the foundations of LLM and VLM compression — the three pillars (pruning, quantization, distillation), token compression, NAS for compression, the empirical ordering evidence, failure modes, and hardware decision rules. It does not cover training-from-scratch efficiency, inference serving systems (vLLM, TensorRT-LLM) beyond their connection to compression, or retrieval-augmented generation.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Prerequisites:&lt;/strong&gt; This assumes familiarity with transformer architectures, basic neural network training (backpropagation, gradient descent), floating-point representation, and cross-entropy loss.&lt;/p&gt;
&lt;h2 id="1-overparameterization-is-the-precondition"&gt;1. Overparameterization is the precondition&lt;/h2&gt;
&lt;p&gt;If models weren&amp;rsquo;t overparameterized, compression wouldn&amp;rsquo;t work. The Lottery Ticket Hypothesis established this formally in 2018: dense, randomly-initialized networks contain subnetworks that, trained in isolation, match the full network&amp;rsquo;s accuracy. For modern LLMs, the numbers are concrete — up to 30% of parameters can be pruned with negligible loss, and models hold 98-99% of original capabilities at just 15% pruning.&lt;/p&gt;
&lt;p&gt;This overparameterization isn&amp;rsquo;t a mistake. Sparse architectures are hard to train from scratch. We train dense and then compress because that&amp;rsquo;s what the optimization surface allows.&lt;/p&gt;
&lt;p&gt;The shape of the redundancy matters, and it differs by modality. This is something I initially underestimated — I thought all redundancy was weight-level, but the token-level and modality-dependent patterns are just as important for practical compression decisions.&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Modality&lt;/th&gt;
&lt;th&gt;Redundancy pattern&lt;/th&gt;
&lt;th&gt;Scale&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Images&lt;/td&gt;
&lt;td&gt;Spatial — neighboring patches share textures/colors&lt;/td&gt;
&lt;td&gt;—&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Video&lt;/td&gt;
&lt;td&gt;Spatiotemporal — consecutive frames share backgrounds; at 10fps, 1000 tokens/frame, a 90-min video yields ~54M tokens&lt;/td&gt;
&lt;td&gt;54M tokens/video&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Audio&lt;/td&gt;
&lt;td&gt;Salient info concentrates in sparse, brief segments and specific frequency bands&lt;/td&gt;
&lt;td&gt;—&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;MLLM sequences&lt;/td&gt;
&lt;td&gt;&amp;gt;50% of tokens get minimal attention; multimodal tokens are &amp;gt;80% of sequences in reasoning tasks&lt;/td&gt;
&lt;td&gt;&amp;gt;80% of sequence&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;All compression exploits some version of this: there are bits you can throw away because they don&amp;rsquo;t change the output. The question is which bits.&lt;/p&gt;
&lt;h2 id="2-the-three-pillars-and-two-newer-additions"&gt;2. The three pillars, and two newer additions&lt;/h2&gt;
&lt;p&gt;The literature converges on five categories. Three dominate practice:&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Method&lt;/th&gt;
&lt;th&gt;Mechanism&lt;/th&gt;
&lt;th&gt;What it reduces&lt;/th&gt;
&lt;th&gt;Tuning required&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Quantization&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Lower-bit weight/activation representation&lt;/td&gt;
&lt;td&gt;Memory, potentially speed&lt;/td&gt;
&lt;td&gt;Often tuning-free for LLMs&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Pruning&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Remove unimportant weights or structures&lt;/td&gt;
&lt;td&gt;Parameters, compute&lt;/td&gt;
&lt;td&gt;Recovery training at high ratios&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Distillation&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Transfer knowledge from large → small model&lt;/td&gt;
&lt;td&gt;Parameters, compute&lt;/td&gt;
&lt;td&gt;Training a student&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;Token compression and Neural Architecture Search sit alongside these — newer, less universal, but important for specific scenarios.&lt;/p&gt;
&lt;h3 id="21-quantization-the-hardware-sensitive-frontier"&gt;2.1 Quantization: the hardware-sensitive frontier&lt;/h3&gt;
&lt;p&gt;Quantization converts float32/float16 weights to fewer bits. The fundamental tension: non-uniform quantization achieves higher accuracy because weights aren&amp;rsquo;t uniformly distributed, but uniform quantization gets hardware support. You cannot have both accuracy and hardware efficiency simultaneously with existing methods.&lt;/p&gt;
&lt;p&gt;A critical asymmetry drives the research: &lt;strong&gt;weights are easy to quantize, activations are hard&lt;/strong&gt; because of outlier distributions. SmoothQuant addresses this by migrating quantization difficulty from activations to weights via per-channel scaling:&lt;/p&gt;
&lt;p&gt;$$Y = X \cdot \text{diag}(s)^{-1} \times \text{diag}(s) \cdot W$$&lt;/p&gt;
&lt;p&gt;This enables W8A8 quantization with minimal accuracy loss and a 2× throughput gain. The idea is simple — smooth the activation outliers into the weights where they do less damage — but the execution requires careful per-channel scaling factors.&lt;/p&gt;
&lt;h4 id="the-outlier-problem-quantified"&gt;The outlier problem, quantified&lt;/h4&gt;
&lt;p&gt;ICQuant reveals the structure of the problem in a way I find unusually clean: &lt;strong&gt;the top 5% of weight outliers consume about 50% of the total value range&lt;/strong&gt; — meaning one full quantization bit gets wasted on just 5% of the weights. About 97% of weight channels have uniformly-distributed outlier positions (verified across Llama2/3/4 and Qwen2.5 families), which enables a per-channel partitioning strategy: separate codebooks for outliers and inliers, combined with index coding that costs ≈0.3 bits/weight vs. ≈1 bit for prior approaches.&lt;/p&gt;
&lt;h4 id="the-production-baseline-and-the-frontier"&gt;The production baseline and the frontier&lt;/h4&gt;
&lt;p&gt;&lt;strong&gt;FP8 (E4M3) on NVIDIA H100/B200&lt;/strong&gt; is the modern production baseline — essentially lossless 50% memory reduction from FP16. &lt;strong&gt;4-bit PTQ (AWQ, GPTQ)&lt;/strong&gt; achieves virtually lossless quantization for models above 70B parameters. QuIP/QuIP# pushes to 2 bits by multiplying weight and Hessian matrices with randomized Hadamard transforms to make entries approximately i.i.d. Gaussian, enabling E8 lattice codebook quantization.&lt;/p&gt;
&lt;p&gt;At the extreme frontier: LittleBit reaches 0.1 bits/weight through latent factorization; iFairy uses complex numbers {±1, ±i} for 2-bit &amp;ldquo;multiplication-free&amp;rdquo; inference via sign flips.&lt;/p&gt;
&lt;h4 id="edge-and-vlm-specific-quantization"&gt;Edge and VLM-specific quantization&lt;/h4&gt;
&lt;p&gt;Edge deployment demands specialized methods. Q-VLM minimizes cross-layer dependency errors in LVLMs using activation entropy as a proxy. MBQ accounts for differential sensitivity between vision and language tokens, achieving up to 1.4× decoding speedup with a custom W3 kernel. P4Q introduces learnable prompts and a lightweight low-bit adapter to realign post-quantization feature distributions.&lt;/p&gt;
&lt;p&gt;KV-cache quantization deserves separate mention. In PaLM-540B with batch size 512 and context length 2048, the KV cache alone needs 3TB — three times the model parameters. KIVI-style KV-cache quantization is now table-stakes for long-context serving.&lt;/p&gt;
&lt;h3 id="22-pruning-three-strategies-three-hardware-outcomes"&gt;2.2 Pruning: three strategies, three hardware outcomes&lt;/h3&gt;
&lt;p&gt;Pruning&amp;rsquo;s real-world impact depends entirely on the pattern, because hardware can only exploit certain sparsity structures:&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Pruning type&lt;/th&gt;
&lt;th&gt;What&amp;rsquo;s removed&lt;/th&gt;
&lt;th&gt;Hardware speedup&lt;/th&gt;
&lt;th&gt;Examples&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Unstructured&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Individual weights&lt;/td&gt;
&lt;td&gt;None without sparse kernels&lt;/td&gt;
&lt;td&gt;SparseGPT, Wanda&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Semi-structured&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Fixed patterns (2:4, 4:8)&lt;/td&gt;
&lt;td&gt;Yes on NVIDIA Ampere+&lt;/td&gt;
&lt;td&gt;SparseGPT 2:4, Wanda N:M&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Structured&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Whole layers/heads/channels&lt;/td&gt;
&lt;td&gt;Yes on commodity hardware&lt;/td&gt;
&lt;td&gt;LLM-Pruner, NIRVANA, UKMP&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;The key insight — and the one I keep coming back to — is that unstructured sparsity achieves the best accuracy but delivers zero speedup without special hardware. Structured pruning physically reduces matrix dimensions — immediate gains on any hardware, at higher accuracy cost. Semi-structured 2:4 sparsity is NVIDIA&amp;rsquo;s compromise: hardware-supported on Ampere GPUs, but one-shot methods like SparseGPT and Wanda still suffer at 60-80% sparsity or with tight 2:4 constraints.&lt;/p&gt;
&lt;h4 id="beyond-uniform-sparsity-per-dimension-pruning"&gt;Beyond uniform sparsity: per-dimension pruning&lt;/h4&gt;
&lt;p&gt;A critical limitation of prior methods is uniform sparsity within layers — all output dimensions of a weight matrix get the same pruning ratio. TRIM demonstrates this is deeply suboptimal: individual output dimensions differ significantly in sensitivity. By assigning unique per-row sparsity ratios via iterative metric-driven adjustment, TRIM reduces OPT-13B perplexity at 80% sparsity from 6461 (Wanda-based OWL) to 324 — &lt;strong&gt;over 95% reduction in perplexity&lt;/strong&gt; at the same sparsity level.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;NIRVANA&lt;/strong&gt; redesigns structured pruning by combining magnitude-scaled gradient saliency ($|\partial f / \partial W \cdot W|$) with Adam-based NTK stability guarantees. The dual criterion balances output preservation with training stability — Proposition 4.1 proves $|\hat{\Theta} - \Theta| \leq O(\varepsilon)$ under the SignGD kernel.&lt;/p&gt;
&lt;p&gt;Key design choices:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Adaptive sparsity allocation&lt;/strong&gt;: parameter $\gamma$ controls MLP vs. attention pruning rates ($v_{\text{MLP}} = \gamma \cdot v_{\text{Attn}}$)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Hardware-aware dimension alignment&lt;/strong&gt;: all hidden dimensions forced to multiples of 8 for Tensor Core compatibility&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Global joint ranking&lt;/strong&gt; across all layers/modules with a safeguard retaining ≥1 unit per layer to prevent collapse&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;At 50% sparsity, NIRVANA achieves WikiText2 perplexity (PPL) of 48.94 vs. 215.94 for LLM-Pruner on Llama3.1-8B. Ablation reveals that magnitude-based scoring alone causes extreme collapse (PPL ≈ 10⁵–10⁶), and removing adaptive allocation $\gamma$ raises PPL from 48.94 to 102.00.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;FastForward Pruning&lt;/strong&gt; reformulates sparsity allocation as a single-step RL problem. The RL state is defined solely by the global target sparsity $\sigma_t$ (enabling transfer learning), with a ratio-based reward function ( $PPL_{\text{dense}} / PPL_{\text{pruned}}$ ) that is scale-invariant for portability across model sizes. Results: 3.4× faster than EAS (6.13 vs. 23.6 GPU-hr) on LLaMA-V1 7B at 20% sparsity, with better PPL (6.64 vs. 6.89).&lt;/p&gt;
&lt;h4 id="vlm-specific-structured-pruning-ukmp"&gt;VLM-specific structured pruning: UKMP&lt;/h4&gt;
&lt;p&gt;Text-only pruning methods fail on LVLMs because they treat the language backbone in isolation, ignoring the vision-language interface. &lt;strong&gt;UKMP (AAAI 2025)&lt;/strong&gt; introduces the first unified structured pruning framework purpose-built for LVLMs.&lt;/p&gt;
&lt;p&gt;The &lt;strong&gt;UKMI metric&lt;/strong&gt; combines three innovations:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Adaptive dual normalization&lt;/strong&gt;: block-wise normalization (by parameter volume) prevents large modules from dominating; modality-wise normalization balances vision and language components&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;First-order gradient saliency&lt;/strong&gt;: UKMP discards the second-order Fisher term because the convergence assumption of second-order derivatives is invalid when parameters are frozen — they retain first-order gradients&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Angle distribution entropy&lt;/strong&gt;: entropy over 100 cosine bins weights the Taylor importance, penalizing parameters whose removal would cause large angular shifts in feature space&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Recovery uses a &lt;strong&gt;weight recalling module&lt;/strong&gt;: low-rank $P_2 Q_2 W^p$ transformation parallel to LoRA, trained through three-phase progressive distillation (vision-only MSE → vision+language MSE → task loss + KL). This module is reparameterizable — it folds into base weights after training at no inference cost.&lt;/p&gt;
&lt;p&gt;At 50% pruning, UKMP achieves 47.81% VQAv2 accuracy (vs. 36.40% next-best) and 96.92 NoCaps CIDEr (vs. 85.51). Even at 20% pruning, the pruned BLIP-2 beats similarly-sized full BLIP-2 on OK-VQA and GQA.&lt;/p&gt;
&lt;h3 id="23-distillation-transfer-without-the-baggage"&gt;2.3 Distillation: transfer without the baggage&lt;/h3&gt;
&lt;p&gt;Knowledge distillation trains a smaller student model to mimic a larger teacher. The three challenges: what knowledge to transfer, which algorithm to use, and how to design the student-teacher pair.&lt;/p&gt;
&lt;p&gt;White-box distillation using KL-divergence at high temperature ($\tau = 4.0$) reveals the teacher&amp;rsquo;s confidence across the full vocabulary, enabling finer-grained transfer than black-box methods relying on text outputs alone.&lt;/p&gt;
&lt;h4 id="curriculum-distillation-with-selective-reflection"&gt;Curriculum distillation with selective reflection&lt;/h4&gt;
&lt;p&gt;&lt;strong&gt;SRD (Selective Reflection Distillation)&lt;/strong&gt; demonstrates that not all training samples contribute equally — and that curriculum ordering matters. Easy-to-hard curriculum significantly outperforms reverse hard-to-easy ordering. An increasing temperature schedule ($\tau_0 = 1 \to \tau_n = 2$) is a key effectiveness driver; reversing it severely degrades results.&lt;/p&gt;
&lt;p&gt;SRD achieves up to 39% training time reduction while using 75% of data, and consistently improves ROUGE-L by 3.92–15.53% across all 7 tested KD methods on 5 benchmarks. It is plug-and-play — no changes to model architectures, loss functions, or KD algorithms. It even enables distilled students to surpass teacher performance (26.07 vs. 25.15 ROUGE-L for OpenLLaMA2).&lt;/p&gt;
&lt;h4 id="vlm-specific-distillation"&gt;VLM-specific distillation&lt;/h4&gt;
&lt;p&gt;VLMs present unique challenges because cross-modal alignment must be preserved.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Switch-KD (CVPR 2026)&lt;/strong&gt; unifies vision-language knowledge transfer within a shared text-probability space. The Visual-Switch Distillation pathway switches student visual outputs into the teacher&amp;rsquo;s language pathway ($S\text{-ViT} \to T\text{-Projector} \to T\text{-LLM}$), producing visual-switch logits that represent the teacher&amp;rsquo;s output distribution conditioned on student-encoded visual representations. This is supervised by &lt;strong&gt;DBiLD loss&lt;/strong&gt;, which uses the Kneedle algorithm for adaptive top-k boundary detection and bidirectional reverse KL alignment on pairwise logit differences — outperforming forward KL by 0.5 points.&lt;/p&gt;
&lt;p&gt;Switch-KD-0.5B achieves +3.6 Avg10 over TinyLLaVA-0.5B across 10 multimodal benchmarks and matches the 3B teacher with half the parameters. However, it requires feature-space and vocabulary consistency between teacher and student.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Align-KD&lt;/strong&gt; rests on a critical architectural finding: cross-modal alignment in VLMs occurs primarily at the first attention layer&amp;rsquo;s text-query-vision component ($A_{1, t \leftarrow v}$). Distilling only this targeted attention map achieves the same performance as distilling all maps while saving up to 50% computation. Distilling the wrong component is harmful: vision-query-vision attention KD collapses performance to 43.7 (vs. 64.4 baseline).&lt;/p&gt;
&lt;h4 id="bridging-black-box-and-white-box-distillation"&gt;Bridging black-box and white-box distillation&lt;/h4&gt;
&lt;p&gt;The strongest teachers (GPT-4, proprietary models) are black-box — only text outputs available via API. White-box KD requires internal parameters. &lt;strong&gt;GrayKD (AAAI 2026)&lt;/strong&gt; bridges this with a single-stage framework using no proxy teacher. Black-box rationales are injected through a lightweight cross-attention module — student hidden states as queries, rationale embeddings as keys/values, with 15% random masking for augmentation.&lt;/p&gt;
&lt;p&gt;The efficiency gain is dramatic: GrayKD uses 610M parameters total vs. 2.06B for conventional KD pipelines. GrayKD Triple achieves 27.64 Avg Rouge-L, beating PromptKD + White Teacher (26.44) — despite using the same black-box GPT-4o-mini teacher as lower-scoring methods. Rationale diversity is the dominant factor: switching from multi-rationale to single-rationale reuse drops Rouge-L by 1.14 points.&lt;/p&gt;
&lt;h3 id="24-token-compression-compressing-the-input-not-the-model"&gt;2.4 Token compression: compressing the input, not the model&lt;/h3&gt;
&lt;p&gt;Token compression operates upstream of the three traditional pillars: instead of compressing model weights, it compresses the input. Approaches are categorized by modality (image/video/audio) and mechanism (transformation-based, similarity-based, attention-based, query-based). The key advantage: token compression is post-optimization, requiring no retraining.&lt;/p&gt;
&lt;p&gt;I find this category theoretically elegant but practically limited — it only helps when tokens dominate the compute budget, which is true for video and long-context multimodal tasks but less so for standard image+text inference.&lt;/p&gt;
&lt;h3 id="25-nas-for-compression"&gt;2.5 NAS for compression&lt;/h3&gt;
&lt;p&gt;&lt;strong&gt;CompressNAS&lt;/strong&gt; treats Tucker rank selection as a global search problem, using an MSE-based accuracy proxy comparing decomposed vs. reference layer feature vectors. Existing zero-cost proxies (NASWOT, GraSP, SNIP, ZiCo) fail monotonic trends at higher ranks. CompressNAS builds two lookup tables ($\Delta\text{acc}$, $\Delta\text{flash}$) and uses ILP-based NAS to select ranks globally given a hardware budget — 8× compression of ResNet-18 on ImageNet with &amp;lt;4% accuracy drop.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;LLM-NAS&lt;/strong&gt; solves a problem I hadn&amp;rsquo;t considered: LLM-driven architecture search exhibits exploration bias, repeatedly proposing designs within a narrow region of the search space. The fix is three innovations:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Complexity-driven partitioning&lt;/strong&gt; into 6 disjoint niches defined by architectural complexity (nor_conv_3×3 count)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;LLM-powered prompt co-evolution&lt;/strong&gt; — prompts and architectures co-evolve across rounds&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;XGBoost zero-cost predictor&lt;/strong&gt; aggregating 13 proxy metrics with Spearman correlation ~0.90 to ground truth&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;Search takes 3 minutes and 120 API calls vs. 2–17 GPU-days for supernet baselines. Removing partitioning drops hypervolume from 0.978 to 0.516. Removing the LLM entirely drops it to 0.843.&lt;/p&gt;
&lt;h2 id="3-the-p-kd-q-ordering-sequence-matters"&gt;3. The P-KD-Q ordering: sequence matters&lt;/h2&gt;
&lt;p&gt;A systematic study on Qwen2.5-3B shows that compression ordering is not a detail — it determines whether the pipeline works at all.&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Sequence&lt;/th&gt;
&lt;th&gt;Compression&lt;/th&gt;
&lt;th&gt;G-Eval&lt;/th&gt;
&lt;th&gt;PPL&lt;/th&gt;
&lt;th&gt;Verdict&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;P-KD-Q&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;3.68×&lt;/td&gt;
&lt;td&gt;0.733&lt;/td&gt;
&lt;td&gt;5.048&lt;/td&gt;
&lt;td&gt;Best&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;KD-P-Q&lt;/td&gt;
&lt;td&gt;3.68×&lt;/td&gt;
&lt;td&gt;0.644&lt;/td&gt;
&lt;td&gt;—&lt;/td&gt;
&lt;td&gt;Intermediate&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;P-Q-KD&lt;/td&gt;
&lt;td&gt;3.68×&lt;/td&gt;
&lt;td&gt;0.610&lt;/td&gt;
&lt;td&gt;—&lt;/td&gt;
&lt;td&gt;Intermediate&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;KD-Q-P&lt;/td&gt;
&lt;td&gt;3.68×&lt;/td&gt;
&lt;td&gt;—&lt;/td&gt;
&lt;td&gt;53.4&lt;/td&gt;
&lt;td&gt;Collapse&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Q-P-KD&lt;/td&gt;
&lt;td&gt;3.68×&lt;/td&gt;
&lt;td&gt;0.060&lt;/td&gt;
&lt;td&gt;34.5&lt;/td&gt;
&lt;td&gt;Near-zero&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Q-KD-P&lt;/td&gt;
&lt;td&gt;3.68×&lt;/td&gt;
&lt;td&gt;0.080&lt;/td&gt;
&lt;td&gt;24.1&lt;/td&gt;
&lt;td&gt;Near-zero&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;The mechanism is specific and instructive — and worth pausing on because it explains an entire class of pipeline failures: NF4 quantization produces inference-only models incompatible with gradient-based training. Any sequence with Q before training steps is dead on arrival. The P-KD-Q sequence lets each step compound: pruning reduces the search space, distillation transfers knowledge to the pruned architecture, quantization reduces precision with minimal added loss.&lt;/p&gt;
&lt;p&gt;A practical note: quantization alone achieves 3.00× compression (5886→1959 MB). Adding pruning and distillation adds only 0.68× more (to 3.68×) at significant complexity cost. For many use cases, quantization alone is the right answer.&lt;/p&gt;
&lt;h2 id="4-where-compression-fails"&gt;4. Where compression fails&lt;/h2&gt;
&lt;h3 id="41-the-alignment-cliff-in-vlms"&gt;4.1 The alignment cliff in VLMs&lt;/h3&gt;
&lt;p&gt;VLM compression has a failure mode absent in text-only LLMs. At low compression ratios, structural pruning damages multimodal alignment (vision ↔ language) more than the language backbone; at high ratios, both degrade. This means for mild compression, fine-tuning only the multimodal projector is sufficient — you are repairing the alignment bridge, not the entire model.&lt;/p&gt;
&lt;p&gt;UKMP addresses this directly through modality-wise adaptive normalization and its weight recalling module&amp;rsquo;s progressive three-phase distillation. Text-only importance metrics (magnitude, gradient) cannot detect which parameters mediate the vision-language interface. The convergence assumption of Fisher information is also invalid for VLMs: frozen parameters retain first-order gradients, making second-order importance estimates actively misleading.&lt;/p&gt;
&lt;h3 id="42-extreme-sparsity-collapse"&gt;4.2 Extreme sparsity collapse&lt;/h3&gt;
&lt;p&gt;One-shot pruning methods degrade severely at 60-80% sparsity with semi-structured patterns. NIRVANA&amp;rsquo;s ablation shows magnitude-based scoring alone causes PPL ≈ 10⁵–10⁶ at 50% sparsity. Attention-only pruning causes catastrophic collapse; joint pruning of attention and MLP yields the smoothest degradation.&lt;/p&gt;
&lt;h3 id="43-early-quantization-destroys-trainability"&gt;4.3 Early quantization destroys trainability&lt;/h3&gt;
&lt;p&gt;Applying NF4 quantization before any other technique destroys trainability. Q-KD-P and Q-P-KD sequences achieve near-zero G-Eval scores (0.080, 0.060). The gradient-free nature of NF4-quantized models means they cannot participate in subsequent distillation or pruning recovery.&lt;/p&gt;
&lt;h3 id="44-layer-sensitivity-isnt-uniform"&gt;4.4 Layer sensitivity isn&amp;rsquo;t uniform&lt;/h3&gt;
&lt;p&gt;In partial 2:4 sparsification, later layers are more sensitive than earlier ones — skipping the last third of the model yields the best accuracy. For LVLMs, widthwise pruning of attention heads and MLP neurons outperforms wholesale layer removal. And within a single layer, individual output dimensions differ dramatically in sensitivity.&lt;/p&gt;
&lt;h2 id="5-how-hardware-changes-everything"&gt;5. How hardware changes everything&lt;/h2&gt;
&lt;h3 id="51-the-hardware-taxonomy-mismatch"&gt;5.1 The hardware-taxonomy mismatch&lt;/h3&gt;
&lt;p&gt;The compression technique that looks best on paper often delivers zero real-world speedup. Models optimized for GPU do not run fast on CPU and mobile, and vice versa.&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;If your target is&amp;hellip;&lt;/th&gt;
&lt;th&gt;Prefer&amp;hellip;&lt;/th&gt;
&lt;th&gt;Avoid&amp;hellip;&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Datacenter GPU (A100/H100)&lt;/td&gt;
&lt;td&gt;Semi-structured 2:4 + quantization&lt;/td&gt;
&lt;td&gt;Pure unstructured&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Edge/CPU/Mobile&lt;/td&gt;
&lt;td&gt;Structured pruning (widthwise)&lt;/td&gt;
&lt;td&gt;Any unstructured or semi-structured&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Long-context serving&lt;/td&gt;
&lt;td&gt;KV-cache quantization&lt;/td&gt;
&lt;td&gt;—&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Extreme compression (≤45%)&lt;/td&gt;
&lt;td&gt;Structured + distillation recovery&lt;/td&gt;
&lt;td&gt;One-shot pruning alone&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;h3 id="52-memory-bandwidth-is-the-real-bottleneck"&gt;5.2 Memory bandwidth is the real bottleneck&lt;/h3&gt;
&lt;p&gt;During autoregressive decode, each token generation requires loading the entire model from memory — a classic memory-bandwidth-bound operation. This explains why quantization helps more than pruning for decode latency (smaller weights mean less data movement), why KV-cache quantization becomes critical at long contexts, and why joint algorithm-hardware optimization is the only path to order-of-magnitude gains.&lt;/p&gt;
&lt;p&gt;The Titanus accelerator takes this to the extreme: chiplet-based digital computing-in-memory stores all static weights on-chip, eliminating repeated weight reloading during decode — a 39.4× reduction in off-chip memory access.&lt;/p&gt;
&lt;h3 id="53-the-edge-reality"&gt;5.3 The edge reality&lt;/h3&gt;
&lt;p&gt;CLIP-B/16 at 149.6M parameters already exceeds Jetson Nano&amp;rsquo;s 4GB RAM (no dedicated GPU), causing frequent memory swaps that kill real-time performance. Edge deployment demands the full toolbox: pre-deployment compression, efficient fine-tuning, runtime optimization, and careful security/privacy handling.&lt;/p&gt;
&lt;h3 id="54-a-practical-decision-flow"&gt;5.4 A practical decision flow&lt;/h3&gt;
&lt;p&gt;For compressing an existing LVLM:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Extremely low resources, no recovery training&lt;/strong&gt;: Widthwise pruning only. Accept accuracy loss.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Moderate compression (≤30%)&lt;/strong&gt;: Layerwise pruning + multimodal projector fine-tuning (5% of original data suffices).&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;High compression (≤45%)&lt;/strong&gt;: Widthwise pruning + supervised fine-tuning + hidden-state distillation.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;For any combination&lt;/strong&gt;: 4-bit quantization adds ~+0.1 PPL on top of sparsity.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Always: P-KD-Q ordering.&lt;/strong&gt; Never quantize before training.&lt;/li&gt;
&lt;/ol&gt;
&lt;h2 id="6-where-we-are-in-2025"&gt;6. Where we are in 2025&lt;/h2&gt;
&lt;p&gt;The compression field has matured from experimental techniques into an engineering discipline with clear tiers:&lt;/p&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tier&lt;/th&gt;
&lt;th&gt;Category&lt;/th&gt;
&lt;th&gt;Examples&lt;/th&gt;
&lt;th&gt;Status&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Production&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Foundational pruning&lt;/td&gt;
&lt;td&gt;SparseGPT, Wanda, LLM-Pruner&lt;/td&gt;
&lt;td&gt;Deployed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Production&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;4-bit quantization&lt;/td&gt;
&lt;td&gt;AWQ, GPTQ, NF4 (QLoRA)&lt;/td&gt;
&lt;td&gt;Deployed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Production&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Inference engines&lt;/td&gt;
&lt;td&gt;vLLM (PagedAttention), TensorRT-LLM&lt;/td&gt;
&lt;td&gt;Deployed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Production&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;FP8 baseline&lt;/td&gt;
&lt;td&gt;H100/B200 hardware-native&lt;/td&gt;
&lt;td&gt;Deployed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Experimental&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Extreme pruning&lt;/td&gt;
&lt;td&gt;TRIM, NIRVANA, FastForward&lt;/td&gt;
&lt;td&gt;Active research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Experimental&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;VLM-specific pruning&lt;/td&gt;
&lt;td&gt;UKMP (UKMI + weight recalling)&lt;/td&gt;
&lt;td&gt;Active research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Experimental&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Ultra-low-bit quant&lt;/td&gt;
&lt;td&gt;LittleBit (0.1-bit), ICQuant, iFairy&lt;/td&gt;
&lt;td&gt;Active research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Experimental&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Curriculum distillation&lt;/td&gt;
&lt;td&gt;SRD&lt;/td&gt;
&lt;td&gt;Active research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Experimental&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;VLM distillation&lt;/td&gt;
&lt;td&gt;Switch-KD, Align-KD&lt;/td&gt;
&lt;td&gt;Active research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Experimental&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Black-box KD&lt;/td&gt;
&lt;td&gt;GrayKD (610M params, no proxy)&lt;/td&gt;
&lt;td&gt;Active research&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Experimental&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;NAS for compression&lt;/td&gt;
&lt;td&gt;CompressNAS, LLM-NAS, HAT&lt;/td&gt;
&lt;td&gt;Active research&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;I expect several of the experimental rows to move to production within 12-18 months. TRIM-style per-dimension pruning and ICQuant-style index coding are both conceptually simple enough to integrate into existing pipelines. UKMP&amp;rsquo;s modality-aware pruning is clearly the right approach for VLMs — the question is whether it generalizes beyond BLIP-2 to LLaVA-style architectures.&lt;/p&gt;
&lt;h2 id="7-a-unifying-mental-model"&gt;7. A unifying mental model&lt;/h2&gt;
&lt;p&gt;Compression works because networks use fewer bits of information than they allocate parameters. The art is knowing which parameters carry that information. The answer depends on four things:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;What redundancy exists&lt;/strong&gt;: token-level, weight-level, layer-level — and modality-dependent. Video has spatiotemporal redundancy. MLLM sequences are &amp;gt;80% multimodal tokens that receive minimal attention. Within a single weight matrix, individual rows differ in sensitivity by orders of magnitude.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;How you remove it&lt;/strong&gt;: quantize, prune, or distill. Quantization targets precision. Pruning targets structure. Distillation targets knowledge transfer. Token compression targets the input directly.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;What order you apply techniques&lt;/strong&gt;: P-KD-Q is empirically optimal. Any sequence with Q before training steps fails catastrophically. This is not a heuristic — it follows from NF4&amp;rsquo;s gradient-free nature.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;What hardware runs it&lt;/strong&gt;: This is what determines whether a 50% parameter reduction translates to a 50% latency reduction or no reduction at all. Unstructured sparsity wins on accuracy but loses on every hardware metric. Structured pruning is the opposite. Semi-structured 2:4 is NVIDIA&amp;rsquo;s compromise.&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;What I find striking about the failure modes is how cleanly they carve the parameter space. In VLMs, the vision-language alignment is more fragile than the language backbone. In deep transformers, later layers carry disproportionate importance, and individual output dimensions within the same layer differ dramatically. Hardware is not an implementation detail — it defines which removal patterns become faster.&lt;/p&gt;
&lt;p&gt;The frontier is hybrid, sequential, and precision-extreme: combining pruning, distillation, and quantization in the right order, with per-dimension granularity, pushing quantization to fractions of a bit — while ensuring the pipeline remains trainable throughout.&lt;/p&gt;
&lt;h2 id="boundary-conditions"&gt;Boundary conditions&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;This model assumes the pretrained model is available. If you are training from scratch with a compression target, the entire framework shifts — you would design the architecture sparse from the start rather than compressing post-hoc.&lt;/li&gt;
&lt;li&gt;The P-KD-Q ordering evidence comes from a single systematic study on Qwen2.5-3B. I have not seen replication on larger models or different architectures. The mechanism (NF4 gradient-free) is general, but the magnitude of the ordering effect at other scales is unknown.&lt;/li&gt;
&lt;li&gt;UKMP has been validated primarily on BLIP-2. Its generalization to LLaVA, InternVL, or other VLM architectures is an open question.&lt;/li&gt;
&lt;li&gt;Token compression is effective when tokens dominate compute (long video, long context). For single-image QA, the gains are modest.&lt;/li&gt;
&lt;li&gt;The practical decision flow (Section 5.4) assumes access to fine-tuning resources. For truly zero-shot deployment, only quantization and token compression apply.&lt;/li&gt;
&lt;li&gt;I have not covered training-time efficiency (mixed precision, gradient accumulation, ZeRO, FSDP), which interacts with compression in deployment pipelines but is a separate topic.&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id="open-questions"&gt;Open questions&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;Does the P-KD-Q ordering effect replicate on models above 70B parameters? The mechanism is architecture-agnostic, but the magnitude could scale differently.&lt;/li&gt;
&lt;li&gt;Why do some layers tolerate 4-bit quantization while structurally similar layers fall apart at 6-bit? I suspect effective rank or singular value distribution, but I have not tested this.&lt;/li&gt;
&lt;li&gt;Can UKMP&amp;rsquo;s modality-aware importance metric be adapted to video-language models, where the redundancy patterns are spatiotemporally structured rather than spatially structured?&lt;/li&gt;
&lt;li&gt;What is the minimum viable recovery data for structured pruning of LVLMs at 50%+ sparsity? The current evidence says 5% of original data for moderate compression and full SFT for high compression, but the boundary between these regimes is fuzzy.&lt;/li&gt;
&lt;li&gt;At what point does token compression become more practical than model compression for video understanding? The 54M-token number for 90-minute video suggests the crossover exists but I do not know where.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;[[Q]] Six months from now: has UKMP been extended to LLaVA-style architectures, and does the modality-aware importance metric generalize beyond BLIP-2?&lt;/p&gt;
&lt;h2 id="references"&gt;References&lt;/h2&gt;
&lt;ol&gt;
&lt;li&gt;Frankle &amp;amp; Carbin, &amp;ldquo;The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks&amp;rdquo;, ICLR 2019.&lt;/li&gt;
&lt;li&gt;Xiao et al., &amp;ldquo;SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models&amp;rdquo;, ICML 2023.&lt;/li&gt;
&lt;li&gt;Frantar &amp;amp; Alistarh, &amp;ldquo;SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot&amp;rdquo;, ICML 2023.&lt;/li&gt;
&lt;li&gt;Sun et al., &amp;ldquo;Wanda: A Simple and Effective Pruning Approach for Large Language Models&amp;rdquo;, ICLR 2024.&lt;/li&gt;
&lt;li&gt;Ashkboos et al., &amp;ldquo;SliceGPT: Compress Large Language Models by Deleting Rows and Columns&amp;rdquo;, ICLR 2024.&lt;/li&gt;
&lt;li&gt;Ma et al., &amp;ldquo;LLM-Pruner: On the Structural Pruning of Large Language Models&amp;rdquo;, NeurIPS 2023.&lt;/li&gt;
&lt;li&gt;NIRVANA: &amp;ldquo;NIRVANA: Neural Implicit Removal via Verifiable Adam-based NTK Alignment for Structured Pruning&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;TRIM: &amp;ldquo;TRIM: Per-Dimension Structured Pruning for Large Language Models&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;FastForward: &amp;ldquo;FastForward Pruning: Efficient LLM Pruning via Single-Step Reinforcement Learning&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;UKMP: &amp;ldquo;UKMP: Unified Knowledge Maintenance Pruning for Vision-Language Models&amp;rdquo;, AAAI 2025.&lt;/li&gt;
&lt;li&gt;ICQuant: &amp;ldquo;ICQuant: Index Coding Quantization for Large Language Models&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;Switch-KD: &amp;ldquo;Switch-KD: Knowledge Distillation with Visual Switch for Efficient Vision-Language Models&amp;rdquo;, CVPR 2026.&lt;/li&gt;
&lt;li&gt;Align-KD: &amp;ldquo;Align-KD: Shallow-Layer Attention Alignment for Mobile Vision-Language Models&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;GrayKD: &amp;ldquo;GrayKD: Gray-Box Knowledge Distillation for Large Language Models&amp;rdquo;, AAAI 2026.&lt;/li&gt;
&lt;li&gt;SRD: &amp;ldquo;Selective Reflection Distillation: Curriculum Knowledge Distillation for LLMs&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;CompressNAS: &amp;ldquo;CompressNAS: Neural Architecture Search for Model Compression&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;LLM-NAS: &amp;ldquo;LLM-NAS: Large Language Models for Neural Architecture Search&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;Compression Ordering: &amp;ldquo;A Systematic Study of Compression Ordering for Large Language Models&amp;rdquo;, 2025.&lt;/li&gt;
&lt;li&gt;Multimodal Token Compression Survey, arXiv 2507.20198, 2025.&lt;/li&gt;
&lt;li&gt;Efficient VLM Survey: &amp;ldquo;Efficient Vision-Language Models: A Survey&amp;rdquo;, 2025.&lt;/li&gt;
&lt;/ol&gt;</description></item><item><title>About</title><link>http://fengwang.github.io/about/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>http://fengwang.github.io/about/</guid><description>&lt;p&gt;A former physicist turned machine learning engineer, I have a passion for learning and sharing knowledge. With a background in physics, I bring a unique perspective to software development, combining analytical thinking with creativity. I enjoy exploring new technologies and applying them to solve real-world problems. In my free time, I like to read, travel, and experiment with new frontiers in deep learning and artificial intelligence. This blog is a platform for me to share my insights, experiences, and projects in the world of software engineering and beyond. I hope to inspire others to pursue their passions and contribute to the ever-evolving field of technology.&lt;/p&gt;</description></item></channel></rss>