| تعداد نشریات | 9 |
| تعداد شمارهها | 446 |
| تعداد مقالات | 5,719 |
| تعداد مشاهده مقاله | 8,018,458 |
| تعداد دریافت فایل اصل مقاله | 6,527,290 |
Double Deep Q Network with Adaptive Prioritized Experience Replay | ||
| AUT Journal of Modeling and Simulation | ||
| دوره 57، شماره 1، شهریور 2025، صفحه 53-62 اصل مقاله (1.01 M) | ||
| نوع مقاله: Research Article | ||
| شناسه دیجیتال (DOI): 10.22060/miscj.2025.23426.5373 | ||
| نویسندگان | ||
| Majid Adibian؛ Mohammad Mahdi Ebadzadeh* | ||
| Department of Computer Engineering, Amirkabir University of Technology, Tehran, Iran | ||
| چکیده | ||
| In deep reinforcement learning, experience replay buffers are used to reduce the effects of sequential data and make better use of past experiences. Prioritized Experience Replay (PER) improves upon random sampling by selecting transitions based on their temporal difference (TD) error. However, PER does not consider how important each transition is or how many times it has been used during training. In this paper, we propose a new method for adaptive prioritization that takes into account three additional transition-level factors: reward, usage count (counter), and policy probability—collectively referred to as RCP values. These values are normalized and used alongside the TD error to calculate the probability of selecting each transition from the replay buffer. We evaluate our method on several Atari environments and show that using any of the RCP values individually can improve performance compared to standard PER. To combine all three RCP components, we explore three aggregation functions: minimum, maximum, and mean. Experimental results show that the best aggregation method depends on the environment. However, the mean function generally provides stable improvements across tasks, as it balances all RCP signals and avoids over-relying on any single factor. | ||
| کلیدواژهها | ||
| Deep Reinforcement Learning؛ Prioritized Experience Replay؛ Deep Q-Network | ||
| مراجع | ||
| ||
|
آمار تعداد مشاهده مقاله: 435 تعداد دریافت فایل اصل مقاله: 319 |
||