Details
Original language | English |
---|---|
Title of host publication | Agents and Artificial Intelligence |
Subtitle of host publication | 13th International Conference, ICAART 2021, Virtual Event, February 4–6, 2021, Revised Selected Papers |
Editors | Ana Paula Rocha, Luc Steels, Jaap van den Herik |
Publisher | Springer Science and Business Media Deutschland GmbH |
Pages | 158-180 |
Number of pages | 23 |
ISBN (electronic) | 978-3-031-10161-8 |
ISBN (print) | 9783031101601 |
Publication status | Published - 19 Jul 2022 |
Event | 13th International Conference on Agents and Artificial Intelligence, ICAART 2021 - Virtual, Online, Austria Duration: 4 Feb 2021 → 6 Feb 2021 |
Publication series
Name | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
---|---|
Volume | 13251 LNAI |
ISSN (Print) | 0302-9743 |
ISSN (electronic) | 1611-3349 |
Abstract
Using multi-agent reinforcement learning to find solutions to complex decision-making problems in shared environments has become standard practice in many scenarios. However, this is not the case in safety-critical scenarios, where the reinforcement learning process, which uses stochastic mechanisms, could lead to highly unsafe outcomes. We proposed a novel, safe multi-agent reinforcement learning approach named Assured Multi-Agent Reinforcement Learning (AMARL) to address this issue. Distinct from other safe multi-agent reinforcement learning approaches, AMARL utilises quantitative verification, a model checking technique that guarantees agent compliance of safety, performance, and non-functional requirements, both during and after the learning process. We have previously evaluated AMARL in patrolling domains with various multi-agent reinforcement learning algorithms for both homogeneous and heterogeneous systems. In this work we extend AMARL through the use of deep multi-agent reinforcement learning. This approach is particularly appropriate for systems in which the rewards are sparse and hence extends the applicability of AMARL. We evaluate our approach within a new search and collection domain which demonstrates promising results in safety standards and performance compared to algorithms not using AMARL.
Keywords
- Assurance, Assured Multi-Agent Reinforcement Learning, Deep Reinforcement Learning, Multi-Agent Reinforcement Learning, Multi-Agent Systems, Quantitative verification, Reinforcement Learning, Safe Multi-Agent Reinforcement Learning, Safety-critical scenarios
ASJC Scopus subject areas
- Mathematics(all)
- Theoretical Computer Science
- Computer Science(all)
- General Computer Science
Cite this
- Standard
- Harvard
- Apa
- Vancouver
- BibTeX
- RIS
Agents and Artificial Intelligence : 13th International Conference, ICAART 2021, Virtual Event, February 4–6, 2021, Revised Selected Papers. ed. / Ana Paula Rocha; Luc Steels; Jaap van den Herik. Springer Science and Business Media Deutschland GmbH, 2022. p. 158-180 (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Vol. 13251 LNAI).
Research output: Chapter in book/report/conference proceeding › Conference contribution › Research › peer review
}
TY - GEN
T1 - Assured Deep Multi-Agent Reinforcement Learning for Safe Robotic Systems
AU - Riley, Joshua
AU - Calinescu, Radu
AU - Paterson, Colin
AU - Kudenko, Daniel
AU - Banks, Alec
N1 - Funding Information: Supported by the Defence Science and Technology Laboratory.
PY - 2022/7/19
Y1 - 2022/7/19
N2 - Using multi-agent reinforcement learning to find solutions to complex decision-making problems in shared environments has become standard practice in many scenarios. However, this is not the case in safety-critical scenarios, where the reinforcement learning process, which uses stochastic mechanisms, could lead to highly unsafe outcomes. We proposed a novel, safe multi-agent reinforcement learning approach named Assured Multi-Agent Reinforcement Learning (AMARL) to address this issue. Distinct from other safe multi-agent reinforcement learning approaches, AMARL utilises quantitative verification, a model checking technique that guarantees agent compliance of safety, performance, and non-functional requirements, both during and after the learning process. We have previously evaluated AMARL in patrolling domains with various multi-agent reinforcement learning algorithms for both homogeneous and heterogeneous systems. In this work we extend AMARL through the use of deep multi-agent reinforcement learning. This approach is particularly appropriate for systems in which the rewards are sparse and hence extends the applicability of AMARL. We evaluate our approach within a new search and collection domain which demonstrates promising results in safety standards and performance compared to algorithms not using AMARL.
AB - Using multi-agent reinforcement learning to find solutions to complex decision-making problems in shared environments has become standard practice in many scenarios. However, this is not the case in safety-critical scenarios, where the reinforcement learning process, which uses stochastic mechanisms, could lead to highly unsafe outcomes. We proposed a novel, safe multi-agent reinforcement learning approach named Assured Multi-Agent Reinforcement Learning (AMARL) to address this issue. Distinct from other safe multi-agent reinforcement learning approaches, AMARL utilises quantitative verification, a model checking technique that guarantees agent compliance of safety, performance, and non-functional requirements, both during and after the learning process. We have previously evaluated AMARL in patrolling domains with various multi-agent reinforcement learning algorithms for both homogeneous and heterogeneous systems. In this work we extend AMARL through the use of deep multi-agent reinforcement learning. This approach is particularly appropriate for systems in which the rewards are sparse and hence extends the applicability of AMARL. We evaluate our approach within a new search and collection domain which demonstrates promising results in safety standards and performance compared to algorithms not using AMARL.
KW - Assurance
KW - Assured Multi-Agent Reinforcement Learning
KW - Deep Reinforcement Learning
KW - Multi-Agent Reinforcement Learning
KW - Multi-Agent Systems
KW - Quantitative verification
KW - Reinforcement Learning
KW - Safe Multi-Agent Reinforcement Learning
KW - Safety-critical scenarios
UR - http://www.scopus.com/inward/record.url?scp=85135062258&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-10161-8_8
DO - 10.1007/978-3-031-10161-8_8
M3 - Conference contribution
AN - SCOPUS:85135062258
SN - 9783031101601
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 158
EP - 180
BT - Agents and Artificial Intelligence
A2 - Rocha, Ana Paula
A2 - Steels, Luc
A2 - van den Herik, Jaap
PB - Springer Science and Business Media Deutschland GmbH
T2 - 13th International Conference on Agents and Artificial Intelligence, ICAART 2021
Y2 - 4 February 2021 through 6 February 2021
ER -