Bilgilendirme: Kurulum ve veri kapsamındaki çalışmalar devam etmektedir. Göstereceğiniz anlayış için teşekkür ederiz.
 

Task-Based Visual Attention for Continually Improving the Performance of Autonomous Game Agents

No Thumbnail Available

Date

2023

Journal Title

Journal ISSN

Volume Title

Publisher

Mdpi

Open Access Color

GOLD

Green Open Access

No

OpenAIRE Downloads

OpenAIRE Views

Publicly Funded

No
Impulse
Average
Influence
Average
Popularity
Average

Research Projects

Journal Issue

Abstract

Deep Reinforcement Learning (DRL) has been effectively performed in various complex environments, such as playing video games. In many game environments, DeepMind's baseline Deep Q-Network (DQN) game agents performed at a level comparable to that of humans. However, these DRL models require many experience samples to learn and lack the adaptability to changes in the environment and handling complexity. In this study, we propose Attention-Augmented Deep Q-Network (AADQN) by incorporating a combined top-down and bottom-up attention mechanism into the DQN game agent to highlight task-relevant features of input. Our AADQN model uses a particle-filter -based top-down attention that dynamically teaches an agent how to play a game by focusing on the most task-related information. In the evaluation of our agent's performance across eight games in the Atari 2600 domain, which vary in complexity, we demonstrate that our model surpasses the baseline DQN agent. Notably, our model can achieve greater flexibility and higher scores at a reduced number of time steps.Across eight game environments, AADQN achieved an average relative improvement of 134.93%. Pong and Breakout games both experienced improvements of 9.32% and 56.06%, respectively. Meanwhile, SpaceInvaders and Seaquest, which are more intricate games, demonstrated even higher percentage improvements, with 130.84% and 149.95%, respectively. This study reveals that AADQN is productive for complex environments and produces slightly better results in elementary contexts.

Description

Celikkale, Ismail Bora/0000-0002-2281-8773; Celikcan, Ufuk/0000-0001-6421-185X; Ulu, Eren/0009-0005-0993-2554

Keywords

Deep Reinforcement Learning, Deep Q-Learning, Layer-Wise Relevance Propagation, Particle Filter, Bottom-Up And Top-Down Visual Attention, Saliency Map, Convolutional Neural Network

Turkish CoHE Thesis Center URL

Fields of Science

0202 electrical engineering, electronic engineering, information engineering, 02 engineering and technology

Citation

WoS Q

Q2

Scopus Q

Q2
OpenCitations Logo
OpenCitations Citation Count
N/A

Source

Electronics

Volume

12

Issue

21

Start Page

4405

End Page

PlumX Metrics
Citations

Scopus : 2

Captures

Mendeley Readers : 4

SCOPUS™ Citations

2

checked on Feb 01, 2026

Web of Science™ Citations

1

checked on Feb 01, 2026

Page Views

3

checked on Feb 01, 2026

Google Scholar Logo
Google Scholar™
OpenAlex Logo
OpenAlex FWCI
0.51088578

Sustainable Development Goals

SDG data is not available