Back to glossaryGLOSSARY · Concepts

Embedding

A numeric vector representation of text (or image, or audio) that captures semantic meaning. Two embeddings close in vector space mean two pieces of content close in meaning. Generated by embedding models like text-embedding-3-large, voyage-3, or multilingual-e5.

How it works

An embedding model takes a piece of content and outputs a fixed-size vector (typically 768-3072 dimensions). Similar content produces similar vectors. Cosine similarity or Euclidean distance measures how close two embeddings are in meaning.

Example

The phrases "How do I cancel my subscription" and "I need to end my account" produce similar embeddings even though they share few words, allowing a RAG system to retrieve the same help article for both.

Related terms

Need to actually use Embedding?

We build production AI systems that put these concepts to work. 30 minutes, we map your use case.