Home  | Publications | SS20

Semi-Markov Reinforcement Learning for Stochastic Resource Collection

MCML Authors

Link to Profile Matthias Schubert PI Matchmaking

Matthias Schubert

Prof. Dr.

Principal Investigator

Abstract

We show that the task of collecting stochastic, spatially distributed resources (Stochastic Resource Collection, SRC) may be considered as a Semi-Markov-Decision-Process. Our Deep-Q-Network (DQN) based approach uses a novel scalable and transferable artificial neural network architecture. The concrete use-case of the SRC is an officer (single agent) trying to maximize the amount of fined parking violations in his area. We evaluate our approach on a environment based on the real-world parking data of the city of Melbourne. In small, hence simple, settings with short distances between resources and few simultaneous violations, our approach is comparable to previous work. When the size of the network grows (and hence the amount of resources) our solution significantly outperforms preceding methods. Moreover, applying a trained agent to a non-overlapping new area outperforms existing approaches.

inproceedings


IJCAI 2020

29th International Joint Conference on Artificial Intelligence. Yokohama, Japan (postponed due to the Corona pandemic), Jan 07-15, 2021.
Conference logo
A* Conference

Authors

S. Schmoll • M. Schubert

Links

DOI

Research Area

 A3 | Computational Models

BibTeXKey: SS20

Back to Top