Convolutional neural networks that model sequential data using dilated, causal convolutions.
Temporal Convolutional Networks (TCNs) are a class of neural network architectures that apply convolutional operations across the time dimension to model sequential data. Rather than processing sequences step-by-step as recurrent models do, TCNs use causal convolutions — filters constrained so that the output at any time step depends only on current and past inputs, never future ones. This causality constraint makes TCNs appropriate for real-time and autoregressive tasks where future information is unavailable at inference time.
The key architectural innovation in TCNs is the use of dilated convolutions, where gaps are introduced between filter taps to exponentially expand the receptive field without increasing the number of parameters. By stacking layers with progressively larger dilation factors (e.g., 1, 2, 4, 8, ...), a TCN can capture dependencies spanning thousands of time steps while remaining computationally efficient. Residual connections are typically added between layers to stabilize training in deep networks and allow gradients to flow cleanly during backpropagation.
TCNs offer several practical advantages over recurrent architectures such as LSTMs and GRUs. Because convolutions across a sequence can be computed in parallel, TCNs train significantly faster on modern hardware. They also sidestep the vanishing gradient problem that plagues deep RNNs, since gradients in convolutional networks travel through fixed-depth paths rather than unrolled time steps. A landmark 2018 study by Bai, Kolter, and Koltun demonstrated that TCNs matched or outperformed canonical recurrent models across a wide range of sequence modeling benchmarks, prompting broader adoption of convolutional approaches for temporal data.
TCNs have found application in time series forecasting, audio generation, anomaly detection, and natural language processing. While transformer-based architectures have since become dominant for many sequence tasks, TCNs remain a strong baseline — particularly in resource-constrained settings where their efficiency and simplicity are valuable. Their design principles, especially dilated causal convolutions, have also influenced hybrid architectures that blend convolutional and attention-based components.