Variational Auto-encoder Based Solutions to Interactive Dynamic Influence Diagrams
Addressing multiagent decision problems in AI, especially those involving collaborative or competitive agents acting concurrently in a partially observable and stochastic environment, remains a formidable challenge. While Interactive Dynamic Influence Diagrams~(I-DIDs) have offered a promising decis...
Saved in:
Main Authors | , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
30.09.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Addressing multiagent decision problems in AI, especially those involving
collaborative or competitive agents acting concurrently in a partially
observable and stochastic environment, remains a formidable challenge. While
Interactive Dynamic Influence Diagrams~(I-DIDs) have offered a promising
decision framework for such problems, they encounter limitations when the
subject agent encounters unknown behaviors exhibited by other agents that are
not explicitly modeled within the I-DID. This can lead to sub-optimal responses
from the subject agent. In this paper, we propose a novel data-driven approach
that utilizes an encoder-decoder architecture, particularly a variational
autoencoder, to enhance I-DID solutions. By integrating a perplexity-based tree
loss function into the optimization algorithm of the variational autoencoder,
coupled with the advantages of Zig-Zag One-Hot encoding and decoding, we
generate potential behaviors of other agents within the I-DID that are more
likely to contain their true behaviors, even from limited interactions. This
new approach enables the subject agent to respond more appropriately to unknown
behaviors, thus improving its decision quality. We empirically demonstrate the
effectiveness of the proposed approach in two well-established problem domains,
highlighting its potential for handling multi-agent decision problems with
unknown behaviors. This work is the first time of using neural networks based
approaches to deal with the I-DID challenge in agent planning and learning
problems. |
---|---|
DOI: | 10.48550/arxiv.2409.19965 |