Genetic algorithms with sharing for multimodal function optimization
Proceedings of the Second International Conference on Genetic Algorithms on Genetic algorithms and their application
Radial basis functions for multivariable interpolation: a review
Algorithms for approximation
Efficient reinforcement learning through symbiotic evolution
Machine Learning - Special issue on reinforcement learning
Reinforcement learning with replacing eligibility traces
Machine Learning - Special issue on reinforcement learning
Markov Decision Processes: Discrete Stochastic Dynamic Programming
Markov Decision Processes: Discrete Stochastic Dynamic Programming
Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
Evolving neural networks through augmenting topologies
Evolutionary Computation
Averaging Efficiently in the Presence of Noise
PPSN V Proceedings of the 5th International Conference on Parallel Problem Solving from Nature
Genetic Programming And Multi-agent Layered Learning By Reinforcements
GECCO '02 Proceedings of the Genetic and Evolutionary Computation Conference
Evolving Soccer Keepaway Players Through Task Decomposition
Machine Learning
Co-evolving recurrent neurons learn deep memory POMDPs
GECCO '05 Proceedings of the 7th annual conference on Genetic and evolutionary computation
Cooperative Coevolution: An Architecture for Evolving Coadapted Subcomponents
Evolutionary Computation
Evolutionary Function Approximation for Reinforcement Learning
The Journal of Machine Learning Research
A comparison between cellular encoding and direct encoding for genetic neural networks
GECCO '96 Proceedings of the 1st annual conference on Genetic and evolutionary computation
Competitive coevolution through evolutionary complexification
Journal of Artificial Intelligence Research
Proceedings of the 9th annual conference companion on Genetic and evolutionary computation
Empirical Studies in Action Selection with Reinforcement Learning
Adaptive Behavior - Animals, Animats, Software Agents, Robots, Adaptive Systems
A common genetic encoding for both direct and indirect encodings of networks
Proceedings of the 9th annual conference on Genetic and evolutionary computation
Transfer via inter-task mappings in policy search reinforcement learning
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Batch reinforcement learning in a complex domain
Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
Proceedings of the 10th annual conference companion on Genetic and evolutionary computation
Evolving neural networks for fractured domains
Proceedings of the 10th annual conference on Genetic and evolutionary computation
Analysis of an evolutionary reinforcement learning method in a multiagent domain
Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 1
Evolving Neural Networks for Online Reinforcement Learning
Proceedings of the 10th international conference on Parallel Problem Solving from Nature: PPSN X
Proceedings of the 17th international conference on Parallel architectures and compilation techniques
An empirical analysis of value function-based and policy search reinforcement learning
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 2
Evolving an autonomous agent for non-Markovian reinforcement learning
Proceedings of the 11th Annual conference on Genetic and evolutionary computation
Proceedings of the 11th Annual Conference Companion on Genetic and Evolutionary Computation Conference: Late Breaking Papers
Temporal difference and policy search methods for reinforcement learning: an empirical comparison
AAAI'07 Proceedings of the 22nd national conference on Artificial intelligence - Volume 2
Interactive evolution of particle systems for computer graphics and animation
IEEE Transactions on Evolutionary Computation
Evolving content in the galactic arms race video game
CIG'09 Proceedings of the 5th international conference on Computational Intelligence and Games
Kernel-based online NEAT for keepaway soccer
LSMS'07 Proceedings of the Life system modeling and simulation 2007 international conference on Bio-Inspired computational intelligence and applications
Autonomous Agents and Multi-Agent Systems
Transfer learning through indirect encoding
Proceedings of the 12th annual conference on Genetic and evolutionary computation
Proceedings of the 12th annual conference companion on Genetic and evolutionary computation
Evolving Static Representations for Task Transfer
The Journal of Machine Learning Research
Evolving plastic neural networks with novelty search
Adaptive Behavior - Animals, Animats, Software Agents, Robots, Adaptive Systems
Constraining connectivity to encourage modularity in HyperNEAT
Proceedings of the 13th annual conference on Genetic and evolutionary computation
Proceedings of the 13th annual conference companion on Genetic and evolutionary computation
Picbreeder: A case study in collaborative evolutionary exploration of design space
Evolutionary Computation
CMA-TWEANN: efficient optimization of neural networks via self-adaptation and seamless augmentation
Proceedings of the 14th annual conference on Genetic and evolutionary computation
Proceedings of the 14th annual conference companion on Genetic and evolutionary computation
Proceedings of the 15th annual conference companion on Genetic and evolutionary computation
Hi-index | 0.00 |
Both genetic algorithms (GAs) and temporal difference (TD) methods have proven effective at solving reinforcement learning (RL) problems. However, since few rigorous empirical comparisons have been conducted, there are no general guidelines describing the methods' relative strengths and weaknesses. This paper presents the results of a detailed empirical comparison between a GA and a TD method in Keepaway, a standard RL benchmark domain based on robot soccer. In particular, we compare the performance of NEAT [19], a GA that evolves neural networks, with Sarsa [16, 17], a popular TD method. The results demonstrate that NEAT can learn better policies in this task, though it requires more evaluations to do so. Additional experiments in two variations of Keepaway demonstrate that Sarsa learns better policies when the task is fully observable and NEAT learns faster when the task is deterministic. Together, these results help isolate the factors critical to the performance of each method and yield insights into their general strengths and weaknesses.