Google's open-source framework for building and deploying machine learning models.
TensorFlow is an open-source machine learning framework developed by Google Brain and released in 2015. At its core, TensorFlow represents computation as directed dataflow graphs, where nodes correspond to mathematical operations and edges carry multidimensional arrays called tensors between them. This abstraction allows the framework to efficiently distribute computation across CPUs, GPUs, and specialized hardware like Google's Tensor Processing Units (TPUs), making it well-suited for the demanding workloads of training large neural networks.
The framework operates by first constructing a computational graph that defines the structure of a model, then executing that graph within a session — a design known as static graph execution. Later versions introduced eager execution, which evaluates operations immediately as they are called, making debugging more intuitive and bringing TensorFlow closer to the dynamic execution style popularized by PyTorch. The high-level Keras API, integrated directly into TensorFlow 2.x, further simplified model building by providing modular, composable layers and training loops that reduce boilerplate code significantly.
TensorFlow's ecosystem extends well beyond model training. TensorFlow Serving enables production deployment of trained models, TensorFlow Lite targets mobile and embedded devices, and TensorFlow.js brings model inference and training to web browsers. This breadth makes TensorFlow one of the few frameworks that addresses the full machine learning lifecycle — from research experimentation to large-scale production deployment — within a single coherent toolset.
The framework has had an outsized influence on the field, accelerating research in computer vision, natural language processing, reinforcement learning, and generative modeling. While competition from PyTorch has intensified, particularly in academic research settings, TensorFlow remains widely deployed in industry due to its mature tooling, strong support for distributed training, and robust production infrastructure. Its development helped establish the expectation that serious deep learning work requires hardware-aware, graph-based computation frameworks capable of scaling across thousands of devices.