Iteratively improving models or algorithms by mimicking biological natural selection.
Directed evolution in machine learning applies principles borrowed from biological evolution—mutation, selection, and reproduction—to iteratively refine algorithms, neural architectures, or model parameters toward better performance. Rather than relying on gradient-based optimization or hand-crafted design, directed evolution maintains a population of candidate solutions, evaluates each according to a fitness function, and uses the best performers as parents for the next generation. This cycle of selection and variation continues until a sufficiently capable solution emerges, making the approach especially attractive for problems where gradients are unavailable, discontinuous, or misleading.
The mechanics typically involve three core operations. Mutation introduces random perturbations to individual candidates—flipping weights, altering hyperparameters, or modifying architectural choices. Crossover (or recombination) combines elements from two or more high-performing candidates to produce offspring that inherit traits from each parent. Selection pressure then determines which candidates survive into the next generation, often using tournament selection, fitness-proportionate sampling, or elitism to preserve top performers. Over many generations, this process can discover solutions that gradient descent might miss, particularly in rugged or multimodal fitness landscapes.
In modern ML, directed evolution overlaps substantially with neuroevolution—the evolution of neural network weights and topologies—and with neural architecture search (NAS) methods that treat network design as an optimization problem. Techniques like NEAT (NeuroEvolution of Augmenting Topologies) and evolutionary NAS have demonstrated competitive results on reinforcement learning benchmarks and image classification tasks, sometimes matching or exceeding manually designed architectures. The approach also connects to population-based training, where hyperparameters are evolved online during a training run rather than fixed in advance.
The practical appeal of directed evolution lies in its flexibility and parallelism: because candidates are evaluated independently, the method scales naturally across distributed compute. Its main drawback is sample inefficiency—evolutionary search typically requires far more fitness evaluations than gradient-based methods to converge. Nevertheless, for black-box optimization problems, hardware-aware architecture search, and settings where differentiability cannot be assumed, directed evolution remains a powerful and widely used tool in the ML practitioner's toolkit.