Using the Q-learning algorithm in the constructive phase of the GRASP and reactive GRASP metaheuristics

Abstract

Currently many non-tractable considered problems have been solved satisfactorily through methods of approximate optimization called metaheuristic. These methods use non-deterministic approaches that find good solutions which, however, do not guarantee the determination of the global optimum. The success of a metaheuristic is conditioned by capacity to adequately alternate between exploration and exploitation of the solution space. A way to guide such algorithms while searching for better solutions is supplying them with more knowledge of the solution space (environment of the problem). This can to be made in terms of a mapping of such environment in states and actions using reinforcement learning. This paper proposes the use of a technique of reinforcement learning - Q-learning algorithm - for the constructive phase of GRASP and reactive GRASP metaheuristic. The proposed methods will be applied to the symmetrical traveling salesman problem.

16 Figures and Tables

Cite this paper

@article{Lima2008UsingTQ, title={Using the Q-learning algorithm in the constructive phase of the GRASP and reactive GRASP metaheuristics}, author={F de A de Lima and J. D. de Melo and A. D. Doria Neto}, journal={2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence)}, year={2008}, pages={4169-4176} }