2022
DOI: 10.1007/s10915-022-01876-x
|View full text |Cite
|
Sign up to set email alerts
|

A Sojourn-Based Approach to Semi-Markov Reinforcement Learning

Abstract: In this paper we introduce a new approach to discrete-time semi-Markov decision processes based on the sojourn time process. Different characterizations of discrete-time semi-Markov processes are exploited and decision processes are constructed by their means. With this new approach, the agent is allowed to consider different actions depending also on the sojourn time of the process in the current state. A numerical method based on Q-learning algorithms for finite horizon reinforcement learning and stochastic … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 29 publications
(71 reference statements)
0
1
0
Order By: Relevance
“…The idea of this method is to decompose a whole task into multi-level subtasks by introducing mechanisms such as State space decomposition ( Takahashi, 2001 ), State abstraction ( Abel, 2019 ), and Temporal abstraction ( Bacon and Precup, 2018 ) so that each subtask can be solved in a small-scale state space, thus speeding up the solution of the whole task. To model these abstract mechanisms, researchers introduced the semi-Markov Decision Process (SMDP) ( Ascione and Cuomo, 2022 ) model to handle actions that must be completed at multiple time steps. The state space decomposition approach decomposes the state space into different subsets.…”
Section: Related Workmentioning
confidence: 99%
“…The idea of this method is to decompose a whole task into multi-level subtasks by introducing mechanisms such as State space decomposition ( Takahashi, 2001 ), State abstraction ( Abel, 2019 ), and Temporal abstraction ( Bacon and Precup, 2018 ) so that each subtask can be solved in a small-scale state space, thus speeding up the solution of the whole task. To model these abstract mechanisms, researchers introduced the semi-Markov Decision Process (SMDP) ( Ascione and Cuomo, 2022 ) model to handle actions that must be completed at multiple time steps. The state space decomposition approach decomposes the state space into different subsets.…”
Section: Related Workmentioning
confidence: 99%