返回论文列表
Paper Detail
Partial Attention in Deep Reinforcement Learning for Safe Multi-Agent Control面向安全多智能体控制的深度强化学习部分注意力机制
cs.CV自动驾驶热门获取
Turki Bin Mohaya, Peter Seiler
2026年03月23日
arXiv: 2603.21810v1

作者人数

2

标签数量

2

内容状态

含 PDF

原文 + 中文

同页查看标题和摘要的双语信息

PDF 预览

直接在详情页阅读或下载论文全文

深度分析

继续下钻到 AI 生成的结构化解读

摘要 / Abstract

Attention mechanisms excel at learning sequential patterns by discriminating data based on relevance and importance. This provides state-of-the-art performance in advanced generative artificial intelligence models. This paper applies this concept of an attention mechanism for multi-agent safe control. We specifically consider the design of a neural network to control autonomous vehicles in a highway merging scenario. The environment is modeled as a Decentralized Partially Observable Markov Decision Process (Dec-POMDP). Within a QMIX framework, we include partial attention for each autonomous vehicle, thus allowing each ego vehicle to focus on the most relevant neighboring vehicles. Moreover, we propose a comprehensive reward signal that considers the global objectives of the environment (e.g., safety and vehicle flow) and the individual interests of each agent. Simulations are conducted in the Simulation of Urban Mobility (SUMO). The results show better performance compared to other driving algorithms in terms of safety, driving speed, and reward.

注意力机制通过区分数据的相关性和重要性来学习序列模式,在先进生成式人工智能模型中取得了最优性能。本文将注意力机制应用于多智能体安全控制,具体设计了一个神经网络用于控制高速公路汇入场景中的自动驾驶车辆。环境被建模为分散式部分可观测马尔可夫决策过程(Dec-POMDP),在QMIX框架中为每辆自动驾驶车引入部分注意力,使其能够关注最相关的邻近车辆。

PDF 预览
1
在 arXiv 查看下载 PDF

分类 / Categories

cs.CVcs.AI

深度分析

AI 深度理解论文内容,生成具有洞见性的总结