URI | http://purl.tuc.gr/dl/dias/75E77769-957E-4070-8DCA-33D273034342 | - |
Αναγνωριστικό | http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.472.7494&rep=rep1&type=pdf | - |
Γλώσσα | en | - |
Μέγεθος | 8 pages | en |
Τίτλος | Algorithm selection using reinforcement learning | en |
Δημιουργός | Lagoudakis Michael | en |
Δημιουργός | Λαγουδακης Μιχαηλ | el |
Δημιουργός | Littman, M. | en |
Περίληψη | Many computational problems can be solved by
multiple algorithms, with different algorithms
fastest for different problem sizes, input distributions,
and hardware characteristics. We consider
the problem of algorithm selection: dynamically
choose an algorithm to attack an instance
of a problem with the goal of minimizing
the overall execution time. We formulate the
problem as a kind of Markov decision process
(MDP), and use ideas from reinforcement learning
to solve it. This paper introduces a kind of
MDP that models the algorithm selection problem
by allowing multiple state transitions. The well
known Q-learning algorithm is adapted for this
case in a way that combines both Monte-Carlo
and Temporal Difference methods. Also, this
work uses, and extends in a way to control problems,
the Least-Squares Temporal Difference algorithm
(LSTD(0)) of Boyan. The experimental
study focuses on the classic problems of order
statistic selection and sorting. The encouraging
results reveal the potential of applying learning
methods to traditional computational problems. | en |
Τύπος | Πλήρης Δημοσίευση σε Συνέδριο | el |
Τύπος | Conference Full Paper | en |
Άδεια Χρήσης | http://creativecommons.org/licenses/by/4.0/ | en |
Ημερομηνία | 2015-11-14 | - |
Ημερομηνία Δημοσίευσης | 2000 | - |
Θεματική Κατηγορία | Intelligence, Computational | en |
Θεματική Κατηγορία | computational intelligence | en |
Θεματική Κατηγορία | intelligence computational | en |
Βιβλιογραφική Αναφορά | Michail G. Lagoudakis and Michael L. Littman. (2000, June). Algorithm selection using reinforcement learning. [Online]. Available: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.472.7494&rep=rep1&type=pdf | en |