URI | http://purl.tuc.gr/dl/dias/75E77769-957E-4070-8DCA-33D273034342 | - |
Identifier | http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.472.7494&rep=rep1&type=pdf | - |
Language | en | - |
Extent | 8 pages | en |
Title | Algorithm selection using reinforcement learning | en |
Creator | Lagoudakis Michael | en |
Creator | Λαγουδακης Μιχαηλ | el |
Creator | Littman, M. | en |
Content Summary | Many computational problems can be solved by
multiple algorithms, with different algorithms
fastest for different problem sizes, input distributions,
and hardware characteristics. We consider
the problem of algorithm selection: dynamically
choose an algorithm to attack an instance
of a problem with the goal of minimizing
the overall execution time. We formulate the
problem as a kind of Markov decision process
(MDP), and use ideas from reinforcement learning
to solve it. This paper introduces a kind of
MDP that models the algorithm selection problem
by allowing multiple state transitions. The well
known Q-learning algorithm is adapted for this
case in a way that combines both Monte-Carlo
and Temporal Difference methods. Also, this
work uses, and extends in a way to control problems,
the Least-Squares Temporal Difference algorithm
(LSTD(0)) of Boyan. The experimental
study focuses on the classic problems of order
statistic selection and sorting. The encouraging
results reveal the potential of applying learning
methods to traditional computational problems. | en |
Type of Item | Πλήρης Δημοσίευση σε Συνέδριο | el |
Type of Item | Conference Full Paper | en |
License | http://creativecommons.org/licenses/by/4.0/ | en |
Date of Item | 2015-11-14 | - |
Date of Publication | 2000 | - |
Subject | Intelligence, Computational | en |
Subject | computational intelligence | en |
Subject | intelligence computational | en |
Bibliographic Citation | Michail G. Lagoudakis and Michael L. Littman. (2000, June). Algorithm selection using reinforcement learning. [Online]. Available: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.472.7494&rep=rep1&type=pdf | en |