Back to Elasticsearch

Dense vector field type [dense-vector]

docs/reference/elasticsearch/mapping-reference/dense-vector.md

9.4.030.4 KB
Original Source

Dense vector field type [dense-vector]

The dense_vector field type stores dense vectors of numeric values. Dense vector fields are primarily used for k-nearest neighbor (kNN) search.

The dense_vector type does not support aggregations or sorting.

You can provide vectors in two different input formats:

  • Array of floats: A JSON array of numeric values representing each vector dimension.
  • Base64-encoded string: A Base64-encoded binary representation of the vector. More compact than float arrays, reducing payload size and improving efficiency for large vectors and bulk indexing.

The following example creates an index with a dense_vector field (using default element_type float) and indexes documents using each format:

::::{tab-set} :::{tab-item} Array of floats

console
PUT my-index
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 3,
        "index_options": {
          "type": "bbq_disk" <1>
        }
      },
      "my_text": {
        "type": "keyword"
      }
    }
  }
}

PUT my-index/_doc/1
{
  "my_text" : "text1",
  "my_vector" : [0.5, 10, 6]
}

PUT my-index/_doc/2
{
  "my_text" : "text2",
  "my_vector" : [-0.5, 10, 10]
}
  1. (Optional) Controls how vectors are indexed internally for kNN search. In this example, bbq_disk enables disk-based binary quantization, which can significantly reduce memory usage for large vector datasets. If you don’t specify index_options, {{es}} automatically selects a default indexing strategy based on the vector type and dimensions. To learn more about the available index options and how they affect vector quantization, refer to Automatically quantize vectors for kNN search.

::: :::{tab-item} Base64-encoded string

console
PUT my-index
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 3
      },
      "my_text" : {
        "type" : "keyword"
      }
    }
  }
}

PUT my-index/_doc/1
{
  "my_text" : "text1",
  "my_vector" : "PwAAAEEgAABAwAAA"
}

PUT my-index/_doc/2
{
  "my_text" : "text2",
  "my_vector" : "vwAAAEEgAABBIAAA"
}

::: ::::

::::{note} Unlike most other data types, dense vectors are always single-valued. It is not possible to store multiple values in one dense_vector field. ::::

Index vectors for kNN search [index-vectors-knn-search]

A k-nearest neighbor (kNN) search finds the k nearest vectors to a query vector, as measured by a similarity metric.

Dense vector fields can be used to rank documents in script_score queries. This lets you perform a brute-force kNN search by scanning all documents and ranking them by similarity.

In many cases, a brute-force kNN search is not efficient enough. For this reason, the dense_vector type supports indexing vectors into a specialized data structure to support fast kNN retrieval through the knn option in the search API

Unmapped array fields of float elements with size between 128 and 4096 are dynamically mapped as dense_vector with a default similarity of cosine. You can override the default similarity by explicitly mapping the field as dense_vector with the desired similarity.

Indexing is enabled by default for dense vector fields. When indexing float vectors, {{es}} uses a default index type.

:::{note} In {{stack}} 9.0, dense vector fields are always indexed as int8_hnsw. :::

When indexing is enabled, you can define the vector similarity to use in kNN search:

console
PUT my-index-2
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 3,
        "similarity": "dot_product"
      }
    }
  }
}

::::{note} Indexing vectors for approximate kNN search is an expensive process. It can take substantial time to ingest documents that contain vector fields with index enabled. See k-nearest neighbor (kNN) search to learn more about the memory requirements. ::::

You can disable indexing by setting the index parameter to false:

console
PUT my-index-2
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 3,
        "index": false
      }
    }
  }
}

{{es}} uses the HNSW algorithm to support efficient kNN search. Like most kNN algorithms, HNSW is an approximate method that sacrifices result accuracy for improved speed.

Accessing dense_vector fields in search responses

{applies_to}
stack: ga 9.2
serverless: ga

By default, dense_vector fields are not included in _source in responses from the _search, _msearch, _get, and _mget APIs. This helps reduce response size and improve performance, especially in scenarios where vectors are used solely for similarity scoring and not required in the output.

To retrieve vector values explicitly, you can use:

  • The fields option to request specific vector fields directly:

    console
    POST my-index-2/_search
    {
      "fields": ["my_vector"]
    }
    

    % TEST[continued]

  • The _source.exclude_vectors flag to re-enable vector inclusion in _source responses:

    console
    POST my-index-2/_search
    {
      "_source": {
        "exclude_vectors": false
      }
    }
    

    % TEST[continued]

:::{tip} For more context about the decision to exclude vectors from _source by default, read the blog post. :::

Docvalue output formats [dense-vector-docvalue-formats]

{applies_to}
stack: ga 9.4
serverless: ga

You can return dense vector doc values using the docvalue_fields search option. The response format can be controlled per field:

  • format: array (default) returns the decoded vector values as a JSON array.
  • format: binary returns the raw vector bytes encoded as base64. This works whether the vector was originally indexed from an array or from a binary string. Numeric element types (float, bfloat16) are emitted in big-endian order; byte and bit vectors are returned exactly as stored.

Example: retrieve dense vector doc values as arrays or base64-encoded bytes

console
PUT dv-format
{
  "mappings": {
    "properties": {
      "vec_float": {
        "type": "dense_vector",
        "element_type": "float",
        "dims": 3,
        "index": false
      }
    }
  }
}

POST dv-format/_bulk?refresh
{"index":{"_id":"1"}}
{"vec_float":[1.5, 2.0, -3.25]}
{"index":{"_id":"2"}}
{"vec_float":[1.25, -2.5, 4.0]}

POST dv-format/_search
{
  "_source": false,
  "query": { "match_all": {} },
  "docvalue_fields": ["vec_float"],
  "sort": "_id"
}

Sample response (array format):

console-result
{
  "hits": {
    "hits": [
      {
        "_id": "1",
        "fields": {
          "vec_float": [
            [1.5, 2.0, -3.25]
          ]
        }
      },
      {
        "_id": "2",
        "fields": {
          "vec_float": [
            [1.25, -2.5, 4.0]
          ]
        }
      }
    ]
  }
}

To retrieve the same vectors as base64-encoded bytes, request only the binary format:

console
POST dv-format/_search
{
  "_source": false,
  "query": { "match_all": {} },
  "docvalue_fields": [ { "field": "vec_float", "format": "binary" } ],
  "sort": "_id"
}

Sample response (binary format):

console-result
{
  "hits": {
    "hits": [
      {
        "_id": "1",
        "fields": {
          "vec_float": [
            "P8AAAEAAAADAUAAA"
          ]
        }
      },
      {
        "_id": "2",
        "fields": {
          "vec_float": [
            "P+AAAAAAAP+QEA=="
          ]
        }
      }
    ]
  }
}

Storage behavior and _source

By default, dense_vector fields are not stored in _source on disk. This is also controlled by the index setting index.mapping.exclude_source_vectors. This setting is enabled by default for newly created indices and can only be set at index creation time.

When enabled:

  • dense_vector fields are removed from _source and the rest of the _source is stored as usual.
  • If a request includes _source and vector values are needed (e.g., during recovery or reindex), the vectors are rehydrated from their internal format.

This setting is compatible with synthetic _source, where the entire _source document is reconstructed from columnar storage. In full synthetic mode, no _source is stored on disk, and all fields — including vectors — are rebuilt when needed.

Rehydration and precision

When vector values are rehydrated (e.g., for reindex, recovery, or explicit _source requests), they are restored from their internal format. By default, vectors are stored at float precision, so if they were originally indexed as higher-precision types (e.g., double or long), the rehydrated values will have reduced precision. This lossy representation is intended to save space while preserving search quality. Additionally, using an element_type of bfloat16 will cause a further loss in precision in restored vectors.

Storing original vectors in _source

If you want to preserve the original vector values exactly as they were provided, you can re-enable vector storage in _source:

console
PUT my-index-include-vectors
{
  "settings": {
    "index.mapping.exclude_source_vectors": false
  },
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector"
      }
    }
  }
}

When this setting is disabled:

  • dense_vector fields are stored as part of the _source, exactly as indexed.
  • The index will store both the original _source value and the internal representation used for vector search, resulting in increased storage usage.
  • Vectors are once again returned in _source by default in all relevant APIs, with no need to use exclude_vectors or fields.

This configuration is appropriate when full source fidelity is required, such as for auditing or round-tripping exact input values.

Automatically quantize vectors for kNN search [dense-vector-quantization]

The dense_vector field type supports quantization to reduce the memory footprint required when searching float vectors. The supported vector quantization strategies for dense_vector kNN indexing are:

Here is an example of configuring disk-based binary quantization using bbq_disk:

console
PUT my-bbq-disk-index
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 384,
        "index": true,
        "index_options": {
          "type": "bbq_disk"
        }
      }
    }
  }
}

Quantized vectors can use oversampling and rescoring to improve accuracy on approximate kNN search results.

::::{note} Quantization will continue to keep the raw float vector values on disk for reranking, reindexing, and quantization improvements over the lifetime of the data. This means disk usage will increase by ~25% for int8, ~12.5% for int4, and ~3.1% for bbq due to the overhead of storing the quantized and raw vectors. ::::

Default quantization types

::::{applies-switch}

:::{applies-item} stack: ga 9.0 When indexing float vectors, the default index type is int8_hnsw. :::

:::{applies-item} stack: ga 9.1+ When indexing float vectors, the default index type is:

  • bbq_hnsw for vectors with greater than or equal to 384 dimensions
  • int8_hnsw for vectors with less than 384 dimensions :::

:::{applies-item} { "stack": "ga 9.4", "serverless": "ga" } When indexing float vectors, the default index type is bbq_disk when available under the current license. ::: ::::

int8 [dense-vector-quantization-int8]

Quantizes each dimension of the vector to 1-byte integers. This reduces the memory footprint by 75% (or 4x) at the cost of some accuracy.

Here is an example of how to create a byte-quantized index:

console
PUT my-byte-quantized-index
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 3,
        "index": true,
        "index_options": {
          "type": "int8_hnsw"
        }
      }
    }
  }
}

int4 [dense-vector-quantization-int4]

Quantizes each dimension of the vector to half-byte integers. This reduces the memory footprint by 87% (or 8x) at the cost of accuracy.

Here is an example of how to create a half-byte-quantized index:

console
PUT my-byte-quantized-index
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 4,
        "index": true,
        "index_options": {
          "type": "int4_hnsw"
        }
      }
    }
  }
}

::::{note} int4 quantization requires an even number of vector dimensions. ::::

bbq [dense-vector-quantization-bbq]

bbq or Better binary quantization reduces each dimension to a single bit precision. This reduces the memory footprint by 96% (or 32x) at a larger cost of accuracy. Generally, oversampling during query time and reranking can help mitigate the accuracy loss. You can choose one of the following BBQ index types:

Here is an example of how to create a binary quantized index:

console
PUT my-byte-quantized-index
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 64,
        "index": true,
        "index_options": {
          "type": "bbq_hnsw"
        }
      }
    }
  }
}

::::{note} bbq quantization only supports vector dimensions that are greater than 64. ::::

Parameters for dense vector fields [dense-vector-params]

The following mapping parameters are accepted:

$$$dense-vector-element-type$$$

element_type : (Optional, string) The data type used to encode vectors.

::::{dropdown} Valid values for element_type float : indexes a 4-byte floating-point value per dimension. This is the default value.

bfloat16 {applies_to}stack: ga 9.3 : indexes a 2-byte floating-point value per dimension. This uses the bfloat16 encoding, not IEEE-754 float16, to maintain the same value range as 4-byte floats. Using bfloat16 is likely to cause a loss of precision in the stored values compared to float.

byte : indexes a 1-byte integer value per dimension.

bit : indexes a single bit per dimension. Useful for very high-dimensional vectors or models that specifically support bit vectors. NOTE: when using bit, the number of dimensions must be a multiple of 8 and must represent the number of bits.

::::

dims : (Optional, integer) Number of vector dimensions. Can’t exceed 4096. If dims is not specified, it will be set to the length of the first vector added to the field.

index : (Optional, Boolean) If true, you can search this field using the knn query or knn in _search . Defaults to true.

$$$dense-vector-similarity$$$

similarity : (Optional¹, string) The vector similarity metric to use in kNN search. Documents are ranked by their vector field’s similarity to the query vector. The _score of each document will be derived from the similarity, in a way that ensures scores are positive and that a larger score corresponds to a higher ranking. Defaults to l2_norm when element_type: bit otherwise defaults to cosine.

¹ $$$footnote-1$$$ This parameter can only be specified when `index` is `true`.

::::{note}
`bit` vectors only support `l2_norm` as their similarity metric.
::::

::::{dropdown} Valid values for similarity l2_norm : Computes similarity based on the L2 distance (also known as Euclidean distance) between the vectors. The document _score is computed as 1 / (1 + l2_norm(query, vector)^2).

For `bit` vectors, instead of using `l2_norm`, the `hamming` distance between the vectors is used. The `_score` transformation is `(numBits - hamming(a, b)) / numBits`

dot_product : Computes the dot product of two unit vectors. This option provides an optimized way to perform cosine similarity. The constraints and computed score are defined by element_type.

When `element_type` is `float`, all vectors must be unit length, including both document and query vectors. The document `_score` is computed as `(1 + dot_product(query, vector)) / 2`.

When `element_type` is `byte`, all vectors must have the same length including both document and query vectors or results will be inaccurate. The document `_score` is computed as `0.5 + (dot_product(query, vector) / (32768 * dims))` where `dims` is the number of dimensions per vector.

cosine : Computes the cosine similarity. During indexing {{es}} automatically normalizes vectors with cosine similarity to unit length. This allows to internally use dot_product for computing similarity, which is more efficient. Original un-normalized vectors can be still accessed through scripts. The document _score is computed as (1 + cosine(query, vector)) / 2. The cosine similarity does not allow vectors with zero magnitude, since cosine is not defined in this case.

max_inner_product : Computes the maximum inner product of two vectors. This is similar to dot_product, but doesn’t require vectors to be normalized. This means that each vector’s magnitude can significantly effect the score. The document _score is adjusted to prevent negative values. For max_inner_product values < 0, the _score is 1 / (1 + -1 * max_inner_product(query, vector)). For non-negative max_inner_product results the _score is calculated max_inner_product(query, vector) + 1.

::::

::::{note} Although they are conceptually related, the similarity parameter is different from text field similarity and accepts a distinct set of options. ::::

$$$dense-vector-index-options$$$

index_options : (Optional², object) An optional section that configures the kNN indexing algorithm. The HNSW algorithm has two internal parameters that influence how the data structure is built. These can be adjusted to improve the accuracy of results, at the expense of slower indexing speed.

² $$$footnote-2$$$ This parameter can only be specified when `index` is `true`.

:::::{dropdown} Properties of index_options type : (Required, string) The type of kNN algorithm to use. Can be either any of: * hnsw - This utilizes the HNSW algorithm for scalable approximate kNN search. This supports all element_type values. * int8_hnsw - The default index type for some float vectors: * {applies_to}stack: ga 9.1 Default for float vectors with less than 384 dimensions. * {applies_to}stack: ga 9.0 Default for float all vectors. This utilizes the HNSW algorithm in addition to automatically scalar quantization for scalable approximate kNN search with element_type of float or bfloat16. This can reduce the memory footprint by 4x at the cost of some accuracy. See Automatically quantize vectors for kNN search. * int4_hnsw - This utilizes the HNSW algorithm in addition to automatically scalar quantization for scalable approximate kNN search with element_type of float or bfloat16. This can reduce the memory footprint by 8x at the cost of some accuracy. See Automatically quantize vectors for kNN search. * bbq_hnsw - This utilizes the HNSW algorithm in addition to automatically binary quantization for scalable approximate kNN search with element_type of float or bfloat16. This can reduce the memory footprint by 32x at the cost of accuracy. See Automatically quantize vectors for kNN search.

  {applies_to}`stack: ga 9.1` `bbq_hnsw` is the default index type for float vectors with greater than or equal to 384 dimensions.
* `flat` - This utilizes a brute-force search algorithm for exact kNN search. This supports all `element_type` values.
* `int8_flat` - This utilizes a brute-force search algorithm in addition to automatic scalar quantization. Only supports `element_type` of `float` or `bfloat16`.
* `int4_flat` - This utilizes a brute-force search algorithm in addition to automatic half-byte scalar quantization. Only supports `element_type` of `float` or `bfloat16`.
* `bbq_flat` - This utilizes a brute-force search algorithm in addition to automatic binary quantization. Only supports `element_type` of `float` or `bfloat16`.
* {applies_to}`stack: ga 9.2` `bbq_disk` - This utilizes a variant of [k-means clustering algorithm](https://en.wikipedia.org/wiki/K-means_clustering) in addition to automatic binary quantization to partition vectors and search subspaces that may be more memory friendly in comparison to HNSW. Only supports `element_type` of `float` or `bfloat16`.  To learn more, refer to [bbq_disk](/reference/elasticsearch/mapping-reference/bbq.md#bbq-disk). This requires an [Enterprise subscription](https://www.elastic.co/subscriptions).

m : (Optional, integer) The number of neighbors each node will be connected to in the HNSW graph. Defaults to 16. Only applicable to hnsw, int8_hnsw, int4_hnsw and bbq_hnsw index types.

ef_construction : (Optional, integer) The number of candidates to track while assembling the list of nearest neighbors for each new node. Defaults to 100. Only applicable to hnsw, int8_hnsw, int4_hnsw and bbq_hnsw index types.

default_visit_percentage {applies_to}stack: ga 9.2 : (Optional, integer) Only applicable to bbq_disk. Must be between 0 and 100. 0 will default to using num_candidates for calculating the percent visited. Increasing default_visit_percentage tends to improve the accuracy of the final results. Defaults to ~1% per shard for every 1 million vectors.

precondition {applies_to}stack: ga 9.4 : (Optional, boolean) Only applicable to bbq_disk. When true transforms the indexed vectors using a random orthogonal projection. This can help improve accuracy when any of the vector components are not normally distributed. Defaults to false.

cluster_size {applies_to}stack: ga 9.2 : (Optional, integer) Only applicable to bbq_disk. The number of vectors per cluster. Smaller cluster sizes increases accuracy at the cost of performance. Defaults to 384. Must be a value between 64 and 65536.

rescore_vector {applies_to}stack: preview =9.0, ga 9.1+ : (Optional, object) An optional section that configures automatic vector rescoring on knn queries for the given field. Only applicable to quantized index types. ::::{dropdown} Properties of rescore_vector oversample : (required, float) The amount to oversample the search results by. This value should be one of the following: * Greater than 1.0 and less than 10.0 * Exactly 0 to indicate no oversampling and rescoring should occur {applies_to}stack: ga 9.1 : The higher the value, the more vectors will be gathered and rescored with the raw values per shard. : In case a knn query specifies a rescore_vector parameter, the query rescore_vector parameter will be used instead. : See oversampling and rescoring quantized vectors for details. ::::

on_disk_rescore {applies_to}stack: preview 9.3 {applies_to}serverless: unavailable : (Optional, boolean) Only applicable to quantized HNSW and bbq_disk index types. When true, vector rescoring will read the raw vector data directly from disk, and will not copy it in memory. This can improve performance when vector data is larger than the amount of available RAM. This setting only applies to newly-indexed vectors; after changing this setting, the vectors must be reindexed or force-merged to apply the new setting to the whole index. Defaults to false. :::::

Synthetic _source [dense-vector-synthetic-source]

dense_vector fields support synthetic _source .

Indexing and searching bit vectors [dense-vector-index-bit]

When using element_type: bit, this will treat all vectors as bit vectors. Bit vectors utilize only a single bit per dimension and are internally encoded as bytes. This can be useful for very high-dimensional vectors or models.

When using bit, the number of dimensions must be a multiple of 8 and must represent the number of bits. Additionally, with bit vectors, the typical vector similarity values are effectively all scored the same, e.g. with hamming distance.

Let’s compare two byte[] arrays, each representing 40 individual bits.

[-127, 0, 1, 42, 127] in bits 1000000100000000000000010010101001111111 [127, -127, 0, 1, 42] in bits 0111111110000001000000000000000100101010

When comparing these two bit, vectors, we first take the hamming distance.

xor result:

1000000100000000000000010010101001111111
^
0111111110000001000000000000000100101010
=
1111111010000001000000010010101101010101

Then, we gather the count of 1 bits in the xor result: 18. To scale for scoring, we subtract from the total number of bits and divide by the total number of bits: (40 - 18) / 40 = 0.55. This would be the _score between these two vectors.

Here is an example of indexing and searching bit vectors:

console
PUT my-bit-vectors
{
  "mappings": {
    "properties": {
      "my_vector": {
        "type": "dense_vector",
        "dims": 40, <1>
        "element_type": "bit"
      }
    }
  }
}
  1. The number of dimensions that represents the number of bits
console
POST /my-bit-vectors/_bulk?refresh
{"index": {"_id" : "1"}}
{"my_vector": [127, -127, 0, 1, 42]} <1>
{"index": {"_id" : "2"}}
{"my_vector": "8100012a7f"} <2>

% TEST[continued]

  1. 5 bytes representing the 40 bit dimensioned vector
  2. A hexadecimal string representing the 40 bit dimensioned vector

Then, when searching, you can use the knn query to search for similar bit vectors:

console
POST /my-bit-vectors/_search?filter_path=hits.hits
{
  "query": {
    "knn": {
      "query_vector": [127, -127, 0, 1, 42],
      "field": "my_vector"
    }
  }
}

% TEST[continued]

console-result
{
  "hits": {
    "hits": [
      {
        "_index": "my-bit-vectors",
        "_id": "1",
        "_score": 1,
        "_source": {
          "my_vector": [
            127,
            -127,
            0,
            1,
            42
          ]
        }
      },
      {
        "_index": "my-bit-vectors",
        "_id": "2",
        "_score": 0.55,
        "_source": {
          "my_vector": "8100012a7f"
        }
      }
    ]
  }
}

GPU vector indexing

{applies_to}
stack: preview 9.3

{{es}} can leverage GPU acceleration to speed up the indexing of dense vectors.

Updatable field type [_updatable_field_type]

To better accommodate scaling and performance needs, updating the type setting in index_options is possible with the Update Mapping API, according to the following graph (jumps allowed):

::::{applies-switch} :::{applies-item} stack: ga 9.1+

txt
flat --> int8_flat --> int4_flat --> bbq_flat --> hnsw --> int8_hnsw --> int4_hnsw --> bbq_hnsw

::: :::{applies-item} stack: ga =9.0

txt
flat --> int8_flat --> int4_flat --> hnsw --> int8_hnsw --> int4_hnsw

::: ::::

For updating all HNSW types (hnsw, int8_hnsw, int4_hnsw, bbq_hnsw) the number of connections m must either stay the same or increase.

Updating type in index_options will fail in all other scenarios.

Switching types won’t re-index vectors that have already been indexed (they will keep using their original type), vectors being indexed after the change will use the new type instead.

For example, it’s possible to define a dense vector field that utilizes the flat type (raw float32 arrays) for a first batch of data to be indexed.

console
PUT my-index-000001
{
    "mappings": {
        "properties": {
            "text_embedding": {
                "type": "dense_vector",
                "dims": 384,
                "index_options": {
                    "type": "flat"
                }
            }
        }
    }
}

Changing the type to int4_hnsw makes sure vectors indexed after the change will use an int4 scalar quantized representation and HNSW (e.g., for KNN queries). That includes new segments created by merging previously created segments.

console
PUT /my-index-000001/_mapping
{
    "properties": {
        "text_embedding": {
            "type": "dense_vector",
            "dims": 384,
            "index_options": {
                "type": "int4_hnsw"
            }
        }
    }
}

% TEST[setup:my_index]

Vectors indexed before this change will keep using the flat type (raw float32 representation and brute force search for KNN queries).

In order to have all the vectors updated to the new type, either reindexing or force merging should be used.

For debugging purposes, it’s possible to inspect how many segments (and docs) exist for each type with the Index Segments API.