返回论文列表
Paper Detail
Confidence-Based Decoding is Provably Efficient for Diffusion Language Models基于置信度的解码在扩散语言模型中的可证明高效性
cs.CL大语言模型端到端Transformer热门获取
Anonymous Authors
2026年03月24日
arXiv: 2603.22248v1

作者人数

1

标签数量

4

内容状态

含 PDF

原文 + 中文

同页查看标题和摘要的双语信息

PDF 预览

直接在详情页阅读或下载论文全文

深度分析

继续下钻到 AI 生成的结构化解读

摘要 / Abstract

This paper addresses the decoding efficiency challenge in diffusion language models (DLMs), which have emerged as promising alternatives to autoregressive models for language generation. The research focuses on confidence-based decoding strategies that adaptively select and unmask tokens based on prediction confidence during the generation process. The authors develop a theoretical analysis framework specifically for an entropy sum-based strategy that continues unmasking tokens until cumulative entropy exceeds a threshold. This work provides the first formal theoretical guarantees demonstrating the efficiency of confidence-based decoding in diffusion-based text generation, offering insights into how adaptive token selection can improve sampling efficiency compared to traditional decoding approaches.

本文探讨了扩散语言模型(DLMs)中的解码效率挑战,该模型是语言生成领域中自回归模型的有前景替代方案。研究聚焦于基于置信度的解码策略,该策略在生成过程中根据预测置信度自适应地选择和去掩标记。作者针对基于熵求和的策略建立了理论分析框架,该策略持续去掩标记直到累积熵超过阈值。这项工作首次提供了形式化理论保证,证明基于置信度的解码在扩散文本生成中的高效性。

PDF 预览
1
在 arXiv 查看下载 PDF

分类 / Categories

cs.CLcs.AI

深度分析

AI 深度理解论文内容,生成具有洞见性的总结