Please use this identifier to cite or link to this item:
https://scholarbank.nus.edu.sg/handle/10635/231419
Title: | VARIATIONAL DISTRIBUTION DESIGNS FOR APPROXIMATE THOMPSON SAMPLING IN DEEP REINFORCEMENT LEARNING | Authors: | SIDDHARTH ARAVINDAN | ORCID iD: | orcid.org/0000-0002-1782-5936 | Keywords: | Exploration, Thompson Sampling, Reinforcement Learning, Variational Learning, Deep Learning, Artifical Intelligence | Issue Date: | 23-Jan-2022 | Citation: | SIDDHARTH ARAVINDAN (2022-01-23). VARIATIONAL DISTRIBUTION DESIGNS FOR APPROXIMATE THOMPSON SAMPLING IN DEEP REINFORCEMENT LEARNING. ScholarBank@NUS Repository. | Abstract: | Exploration is a vital ingredient in reinforcement learning algorithms that has largely contributed to its success in various applications. Standard naive exploration strategies used in deep reinforcement learning are effective in simple tasks, but do not perform well in tasks with high dimensional state-action spaces as they are undirected. Thompson sampling is a directed, well-known and principled approach for balancing exploration and exploitation. But it requires the posterior distribution over the action-value functions or environment models to be maintained; this is generally computationally intractable for tasks that have a high dimensional state-action space. In this thesis, we argue that incorporating domain knowledge during the formulation of variational distributions for approximating these posterior distributions is useful in reinforcement learning. We explore this assertion by designing variational distributions, namely SANE and EVaDE for two different scenarios in the model-free and model-based reinforcement learning settings respectively. | URI: | https://scholarbank.nus.edu.sg/handle/10635/231419 |
Appears in Collections: | Ph.D Theses (Open) |
Show full item record
Files in This Item:
File | Description | Size | Format | Access Settings | Version | |
---|---|---|---|---|---|---|
AravindanS.pdf | 10.73 MB | Adobe PDF | OPEN | None | View/Download |
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.