Heuristically accelerated reinforcement learning: Theoretical and experimental results
dc.contributor.author | Reinaldo Bianchi | |
dc.contributor.author | RIBEIRO, C. H. C. | |
dc.contributor.author | COSTA, A. H. R. | |
dc.date.accessioned | 2022-01-12T22:02:40Z | |
dc.date.available | 2022-01-12T22:02:40Z | |
dc.date.issued | 2012-08-05 | |
dc.description.abstract | Since finding control policies using Reinforcement Learning (RL) can be very time consuming, in recent years several authors have investigated how to speed up RL algorithms by making improved action selections based on heuristics. In this work we present new theoretical results - convergence and a superior limit for value estimation errors - for the class that encompasses all heuristics-based algorithms, called Heuristically Accelerated Reinforcement Learning. We also expand this new class by proposing three new algorithms, the Heuristically Accelerated Q(λ), SARSA(λ) and TD(λ), the first algorithms that uses both heuristics and eligibility traces. Empirical evaluations were conducted in traditional control problems and results show that using heuristics significantly enhances the performance of the learning process. © 2012 The Author(s). | |
dc.description.firstpage | 169 | |
dc.description.lastpage | 174 | |
dc.description.volume | 242 | |
dc.identifier.citation | BIANCHI, R.; RIBEIRO, C. H. C.; COSTA, A. H. R. Heuristically accelerated reinforcement learning: Theoretical and experimental results. Frontiers in Artificial Intelligence and Applications, v. 242, p. 169-174, Aug. 2012. | |
dc.identifier.doi | 10.3233/978-1-61499-098-7-169 | |
dc.identifier.issn | 0922-6389 | |
dc.identifier.uri | https://repositorio.fei.edu.br/handle/FEI/4151 | |
dc.relation.ispartof | Frontiers in Artificial Intelligence and Applications | |
dc.rights | Acesso Restrito | |
dc.title | Heuristically accelerated reinforcement learning: Theoretical and experimental results | |
dc.type | Artigo de evento | |
fei.scopus.citations | 21 | |
fei.scopus.eid | 2-s2.0-84878810338 | |
fei.scopus.subject | Action selection | |
fei.scopus.subject | Control policy | |
fei.scopus.subject | Control problems | |
fei.scopus.subject | Eligibility traces | |
fei.scopus.subject | Empirical evaluations | |
fei.scopus.subject | Learning process | |
fei.scopus.subject | Speed up | |
fei.scopus.subject | Value estimation | |
fei.scopus.updated | 2024-07-01 | |
fei.scopus.url | https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=84878810338&origin=inward |