On Markov decision processes with the stochastic differential Bellman Equation
dc.affiliation.institute | Institut für Informationssysteme | |
dc.contributor.author | Cakir, Merve Nur | |
dc.contributor.referee | Möller, Ralf | |
dc.contributor.referee | Rössler, Andreas | |
dc.date.accepted | 2024-12-17 | |
dc.date.accessioned | 2025-02-18T09:32:49Z | |
dc.date.available | 2025-02-18T09:32:49Z | |
dc.date.issued | 2025 | |
dc.description.abstract | Stochastic differential equations play an important role in capturing the dynamics of complex systems, where uncertainty prevails in the form of noise. In complex systems noise is abundant, but its exact behaviour is unknown. However, noise can be simulated with stochastic processes. Stochastic calculi, such as the Itˆo formula, provide tools for navigating these systems. In this work, the adaptation of the Bellman equation, a cornerstone of dynamic programming, to the realm of stochastic differential equations is explored, facilitating the modeling of decision problems subject to noise. Value iteration and Q-learning, two well-known solution methods in machine learning, are extended to stochastic algorithms in order to approximate the solution for Markov decision processes with uncertainties modeled by the stochastic differential Bellman equation. These stochastic algorithms enable a realistic approach to modeling and solving decision problems in stochastic environments efficiently. The stochastic value iteration is applied when the environment is fully known, while the stochastic Q-learning extends its utility even in cases where transition probabilities remain unknown. Through theoretical analyses and case studies, these algorithms demonstrate their efficacy and applicability, delivering meaningful results. Additionally, the stochastic Q-learning achieves superior rewards compared to the deterministic algorithm, indicating its ability to optimize decision processes in stochastic environments more effectively by exploring more states. Finally, the stochastic differential Bellman equation is formulated as a system of ordinary equations, providing an alternative solution. For this, the concept of the random dynamical system is explored, of which a stochastic differential equation is an example. | |
dc.identifier.uri | https://epub.uni-luebeck.de/handle/zhb_hl/3377 | |
dc.identifier.urn | urn:nbn:de:gbv:841-202502181 | |
dc.language.iso | en | |
dc.subject | markov decision process | |
dc.subject | stochastic process | |
dc.subject | machine learning | |
dc.subject.ddc | 004 | |
dc.title | On Markov decision processes with the stochastic differential Bellman Equation | |
dc.type | thesis.doctoral |
Dateien
Originalbündel
1 - 1 von 1
Lizenzbündel
1 - 1 von 1
Vorschaubild nicht verfügbar
- Name:
- license.txt
- Größe:
- 5.07 KB
- Format:
- Item-specific license agreed to upon submission
- Beschreibung: