Reinforcement Learning for Adversarial Systems Using Relational Observations

Loading...
Thumbnail Image

Contributor

Advisor

Editor

Performer

Department

Instructor

Depositor

Speaker

Researcher

Consultant

Interviewer

Interviewee

Narrator

Transcriber

Annotator

Journal Title

Journal ISSN

Volume Title

Publisher

Journal Name

Volume

Number/Issue

Starting Page

1114

Ending Page

Alternative Title

Abstract

This paper investigates the integration of relational observations with the reinforcement learning (RL) framework for improved generalization capability. A hide-and-seek simulation environment is designed in Unity for proof-of-concept demonstration. Two observation representations—relational (analogical) and standard positional—are designed to evaluate agent learning and generalization capabilities. Agents are trained using the Proximal Policy Optimization (PPO) algorithm in a random-room environment and tested in both the random-room environment and a novel environment with greater spatial complexity and path obstructions. Comparative studies indicate that relational representation of objects in the adversarial environment could potentially improve the generalization capability of RL agents to novel and complex environments. Cross-testing results also suggest that relational observations may enhance agents’ effectiveness in pursuit and evasion tasks in adversarial environments.

Description

Citation

DOI

Extent

10 pages

Format

Type

Conference Paper

Geographic Location

Time Period

Related To

Proceedings of the 59th Hawaii International Conference on System Sciences

Related To (URI)

Table of Contents

Rights

Attribution-NonCommercial-NoDerivatives 4.0 International

Rights Holder

Catalog Record

Local Contexts

Email libraryada-l@lists.hawaii.edu if you need this content in ADA-compliant format.