Please use this identifier to cite or link to this item: https://scholarbank.nus.edu.sg/handle/10635/178992
DC FieldValue
dc.titleA STUDY AND ENHANCEMENTS OF REINFORCEMENT LEARNING METHODS
dc.contributor.authorLEE CHIN YEW
dc.date.accessioned2020-10-22T05:31:31Z
dc.date.available2020-10-22T05:31:31Z
dc.date.issued1998
dc.identifier.citationLEE CHIN YEW (1998). A STUDY AND ENHANCEMENTS OF REINFORCEMENT LEARNING METHODS. ScholarBank@NUS Repository.
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/178992
dc.description.abstractIn recent years, Reinforcement Learning (RL) has become one of the most actively studied area in machine learning with widespread interests in applying these methods to solving difficult control, optimization and scheduling problems. Many of these solutions make use of powerful generalizing function approximators where there are no guarantee of convergence or learning rates but they proved to be effective. Although the field of RL is well developed for the case of discrete tabular representations, many of these methods in its current form do not extend naturally to the use of function approximators. On the other hand, to solve large scale problems we need generalizing architectures in order to have a tractable solution. Effectiveness of this combination of RL with function approximators has been demonstrated in many applications. In this thesis we shall propose two simple ideas to be incorporated into RL methods using function approximators. The first approach involves the use of multi-steps backup to train the evaluation function. The next approach investigates the usefulness of initiating training cycles in states that have incurred large learning errors. We present arguments as to how such simple ideas may help to improve learning times and convergence rates. We also relate these ideas to several approaches that have been extensively studied in learning methods using discrete tabular representations. Experiments are conducted to evaluate the utility of these enhancements on three learning problems. They are the pole-cart task, mountain car task and the game of pursuit and evasion. We present better and stronger learning results on these problems using the proposed ideas. Lastly self training is used to train a pair of predator and prey and we present some of the interesting strategies learned by the agents.
dc.sourceCCK BATCHLOAD 20201023
dc.typeThesis
dc.contributor.departmentINFORMATION SYSTEMS & COMPUTER SCIENCE
dc.contributor.supervisorSUNG KAH KAY
dc.description.degreeMaster's
dc.description.degreeconferredMASTER OF SCIENCE
Appears in Collections:Master's Theses (Restricted)

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
B22107617.PDF2.1 MBAdobe PDF

RESTRICTED

NoneLog In

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.