作者人数
标签数量
内容状态
原文 + 中文
同页查看标题和摘要的双语信息
PDF 预览
直接在详情页阅读或下载论文全文
深度分析
继续下钻到 AI 生成的结构化解读
摘要 / Abstract
This paper addresses the decoding efficiency challenge in diffusion language models (DLMs), which have emerged as promising alternatives to autoregressive models for language generation. The research focuses on confidence-based decoding strategies that adaptively select and unmask tokens based on prediction confidence during the generation process. The authors develop a theoretical analysis framework specifically for an entropy sum-based strategy that continues unmasking tokens until cumulative entropy exceeds a threshold. This work provides the first formal theoretical guarantees demonstrating the efficiency of confidence-based decoding in diffusion-based text generation, offering insights into how adaptive token selection can improve sampling efficiency compared to traditional decoding approaches.
本文探讨了扩散语言模型(DLMs)中的解码效率挑战,该模型是语言生成领域中自回归模型的有前景替代方案。研究聚焦于基于置信度的解码策略,该策略在生成过程中根据预测置信度自适应地选择和去掩标记。作者针对基于熵求和的策略建立了理论分析框架,该策略持续去掩标记直到累积熵超过阈值。这项工作首次提供了形式化理论保证,证明基于置信度的解码在扩散文本生成中的高效性。
分类 / Categories
深度分析
AI 深度理解论文内容,生成具有洞见性的总结