EN KO
← All Publications

When to Speak, When to Abstain: Contrastive Decoding with Abstention

ACL 2025
Hyuhng Joon Kim, Youna Kim, Sang-goo Lee, Taeuk Kim

One-Line Summary

A training-free contrastive decoding method that dynamically blends parametric, contextual, and abstention distributions so that LLMs answer when they have relevant knowledge and gracefully abstain when they do not.

Four scenarios based on parametric and contextual knowledge availability
Figure 1. Four knowledge-access scenarios: depending on whether parametric knowledge (from pre-training) and contextual knowledge (from retrieval) are available, the model should answer or abstain.

Background & Motivation

LLMs acquire broad parametric knowledge during pre-training, yet inevitably lack information about underrepresented or rapidly-evolving topics. Retrieval-Augmented Generation (RAG) supplements this with external contextual knowledge, but sometimes neither source contains the answer. Forcing the model to respond in such cases leads to confident-sounding hallucinations, which is especially dangerous in high-stakes domains.

The Missing Scenario: Prior work on contrastive decoding (e.g., Context-Aware Decoding, Adaptive Contrastive Decoding) only handles cases where at least one knowledge source is relevant. None of them address the critical fourth scenario—where both parametric and contextual knowledge are absent—and therefore these methods never abstain, achieving near-zero F1abs scores.

The authors identify four distinct scenarios: (1) only parametric knowledge available, (2) only contextual knowledge available, (3) both available, and (4) neither available. They first construct a controlled testbed that explicitly labels each scenario, then propose Contrastive Decoding with Abstention (CDA)—a training-free method that robustly handles all four.

Proposed Method: Contrastive Decoding with Abstention (CDA)

Testbed construction pipeline
Figure 2. Testbed construction pipeline: from MRQA datasets, parametric and contextual knowledge availability is assessed to create balanced evaluation sets across all four scenarios.
1
Three-Way Distribution Blending
Extends standard two-way contrastive decoding (parametric + contextual) to a three-way formulation: do = wp·dp + wc·dc + (1 − wp − wc)·da, where da is an explicit abstention distribution obtained by prompting the model to decline answering.
2
Calibrated Uncertainty Estimation
Raw entropy values are not directly comparable across different prompts. CDA computes "content-free" null distributions using placeholder inputs, then calibrates confidence as the relative entropy reduction: rp = max(Hp − H̄p, 0) / H̄p. This ensures fair comparison between parametric and contextual knowledge signals.
3
Dynamic Weight Normalization
Calibrated confidence ratios are normalized so that wp + wc ≤ 1. When both confidences are low, the residual weight (1 − wp − wc) naturally flows to the abstention distribution, triggering the model to decline.
4
Momentum Stabilization (CDA-m)
Previously decoded tokens can unintentionally steer the model. CDA-m applies exponential moving average to the weights: wt ← α·wt−1 + (1−α)·wt, smoothing abrupt fluctuations across decoding steps.

Experimental Results

Evaluated on three QA benchmarks (Natural Questions, HotpotQA, TriviaQA) with four LLMs (Llama3-8B, Llama2-7B/13B, Mistral-7B). Metrics: F1ans (answerable accuracy), F1abs (abstention accuracy), and Reliability Score (RS).

DatasetMethodF1ansF1absRS
NQFSB (best baseline)69.2754.9459.64
NQCDA72.0655.4962.95
NQCDA-m73.1555.4763.72
HotpotQAFSB (best baseline)74.8958.5166.21
HotpotQACDA78.7162.5070.20
HotpotQACDA-m79.3262.5970.64
TriviaQAFSB (best baseline)77.0259.8468.55
TriviaQACDA80.3965.6772.35
TriviaQACDA-m80.9365.6672.74
Reliability scores in RAG settings
Figure 7. Reliability scores in a practical RAG setting using Contriever-msmarco retriever with Wikipedia. CDA-m achieves the highest RS (~68.7) across all baselines.

Why It Matters

For trustworthy AI deployment, knowing when not to answer is just as important as answering correctly. CDA is the first training-free decoding approach that integrates abstention directly into the contrastive decoding framework, handling all four knowledge-access scenarios without any parameter updates. Its calibrated uncertainty estimation ensures robust performance across diverse models and datasets, while momentum stabilization prevents error propagation during autoregressive generation. The method is immediately applicable to any instruction-tuned LLM with RAG, making it a practical step toward more reliable question-answering systems.

Links

RAG Knowledge Reasoning