Automated method that discovers optimal neural network architectures without manual design.
Neural Architecture Search (NAS) is a subfield of automated machine learning (AutoML) that uses computational methods to discover high-performing neural network architectures automatically, replacing the traditionally manual and expertise-intensive process of network design. Rather than relying on human intuition to determine the number of layers, connection patterns, activation functions, and other structural choices, NAS treats architecture design as an optimization problem and searches systematically through a defined space of possible configurations.
NAS methods generally consist of three components: a search space defining which architectural choices are possible, a search strategy for exploring that space, and a performance estimation strategy for evaluating candidate architectures. Early approaches used reinforcement learning, training a controller network to generate architecture descriptions and rewarding it based on validation accuracy. Evolutionary algorithms offered an alternative by mutating and selecting architectures across generations. More recent gradient-based methods, such as DARTS (Differentiable Architecture Search), made the search dramatically more efficient by relaxing the discrete architecture choices into continuous parameters that can be optimized with standard backpropagation, reducing search time from thousands of GPU-hours to a handful.
NAS has produced architectures that rival or surpass hand-crafted designs on benchmarks in image classification, object detection, and natural language processing. Models like NASNet, EfficientNet, and MobileNetV3 emerged from or were informed by NAS pipelines and became widely adopted in production systems. The technique also enables hardware-aware search, where the optimization explicitly accounts for latency, memory footprint, or energy consumption on target devices such as mobile phones or edge hardware.
Despite its power, NAS carries significant computational costs and risks of overfitting to benchmark datasets, prompting ongoing research into more efficient and generalizable search methods. It represents a broader trend in machine learning toward automating the design decisions that once required deep specialist knowledge, lowering barriers to building state-of-the-art models across diverse application domains.