Please use this identifier to cite or link to this item:
https://doi.org/10.1109/ADPRL.2013.6614997
DC Field | Value | |
---|---|---|
dc.title | A reinforcement learning algorithm developed to model GenCo strategic bidding behavior in multidimensional and continuous state and action spaces | |
dc.contributor.author | Lau, A.Y.F. | |
dc.contributor.author | Srinivasan, D. | |
dc.contributor.author | Reindl, T. | |
dc.date.accessioned | 2014-10-07T04:40:57Z | |
dc.date.available | 2014-10-07T04:40:57Z | |
dc.date.issued | 2013 | |
dc.identifier.citation | Lau, A.Y.F.,Srinivasan, D.,Reindl, T. (2013). A reinforcement learning algorithm developed to model GenCo strategic bidding behavior in multidimensional and continuous state and action spaces. IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, ADPRL : 116-123. ScholarBank@NUS Repository. <a href="https://doi.org/10.1109/ADPRL.2013.6614997" target="_blank">https://doi.org/10.1109/ADPRL.2013.6614997</a> | |
dc.identifier.isbn | 9781467359252 | |
dc.identifier.issn | 23251824 | |
dc.identifier.uri | http://scholarbank.nus.edu.sg/handle/10635/83413 | |
dc.description.abstract | The electricity market have provided a complex economic environment, and consequently have increased the requirement for advancement of learning methods. In the agent-based modeling and simulation framework of this economic system, the generation company's decision-making is modeled using reinforcement learning. Existing learning methods that models the generation company's strategic bidding behavior are not adapted to the non-stationary and non-Markovian environment involving multidimensional and continuous state and action spaces. This paper proposes a reinforcement learning method to overcome these limitations. The proposed method discovers the input space structure through the self-organizing map, exploits learned experience through Roth-Erev reinforcement learning and the explores through the actor critic map. Simulation results from experiments show that the proposed method outperforms Simulated Annealing Q-Learning and Variant Roth-Erev reinforcement learning. The proposed method is a step towards more realistic agent learning in Agent-based Computational Economics. © 2013 IEEE. | |
dc.description.uri | http://libproxy1.nus.edu.sg/login?url=http://dx.doi.org/10.1109/ADPRL.2013.6614997 | |
dc.source | Scopus | |
dc.subject | agent-based modeling | |
dc.subject | electricity market | |
dc.subject | reinforcement learning | |
dc.subject | strategic bidding behavior | |
dc.type | Conference Paper | |
dc.contributor.department | ELECTRICAL & COMPUTER ENGINEERING | |
dc.description.doi | 10.1109/ADPRL.2013.6614997 | |
dc.description.sourcetitle | IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, ADPRL | |
dc.description.page | 116-123 | |
dc.identifier.isiut | NOT_IN_WOS | |
Appears in Collections: | Staff Publications |
Show simple item record
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.