Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Siamese Network

Siamese Network

A twin neural network architecture that learns similarity by comparing two inputs.

Year: 1994Generality: 595
Back to Vocab

A Siamese network is a neural network architecture consisting of two or more identical subnetworks that share the same weights and parameters. Rather than classifying a single input, the network processes two inputs in parallel and produces embeddings that can be directly compared. Because both branches are structurally identical and weight-tied, any transformation applied to one input is applied in exactly the same way to the other, ensuring that the resulting representations live in a common feature space where meaningful distance comparisons are possible.

The comparison between outputs is typically performed using a distance metric such as Euclidean distance or cosine similarity, and the network is trained with a loss function designed to shape the embedding space appropriately. Contrastive loss is a classic choice: it penalizes large distances between embeddings of similar pairs and small distances between embeddings of dissimilar pairs. More recent variants use triplet loss, which anchors learning relative to a reference example, a positive match, and a negative mismatch simultaneously, often yielding better-structured embedding spaces.

Siamese networks are particularly well-suited to few-shot learning scenarios, where labeled data is scarce. Instead of learning to classify fixed categories, the network learns a general notion of similarity that can generalize to entirely new classes at inference time. This makes the architecture valuable in domains like face verification, signature authentication, medical image comparison, and one-shot object recognition, where collecting large labeled datasets for every possible class is impractical.

The architecture was introduced in 1993 for handwritten signature verification and gained renewed prominence in the deep learning era as researchers applied it to face recognition and image retrieval with convolutional backbones. Today, Siamese-style weight sharing and contrastive objectives underpin many self-supervised learning methods, including SimCLR and MoCo, making the core idea foundational to modern representation learning well beyond its original verification use case.

Related

Related

Similarity Learning
Similarity Learning

Training models to measure meaningful similarity between data points for comparison tasks.

Generality: 694
Non-Contrastive Learning
Non-Contrastive Learning

Self-supervised representation learning that requires no negative example pairs.

Generality: 575
Contrastive Learning
Contrastive Learning

A self-supervised technique that learns representations by comparing similar and dissimilar data pairs.

Generality: 694
Similarity Computation
Similarity Computation

Quantifying how alike two data objects are to support learning algorithms.

Generality: 709
Capsule Networks
Capsule Networks

Neural networks that encode spatial relationships between features using grouped neuron vectors called capsules.

Generality: 461
SNN (Spiking Neural Network)
SNN (Spiking Neural Network)

Neural networks that process information through discrete, time-dependent electrical spikes.

Generality: 583