Share via


ScalarQuantizationCompression Class

Definition

Contains configuration options specific to the scalar quantization compression method used during indexing and querying.

public class ScalarQuantizationCompression : Azure.Search.Documents.Indexes.Models.VectorSearchCompression
public class ScalarQuantizationCompression : Azure.Search.Documents.Indexes.Models.VectorSearchCompression, System.ClientModel.Primitives.IJsonModel<Azure.Search.Documents.Indexes.Models.ScalarQuantizationCompression>, System.ClientModel.Primitives.IPersistableModel<Azure.Search.Documents.Indexes.Models.ScalarQuantizationCompression>
type ScalarQuantizationCompression = class
    inherit VectorSearchCompression
type ScalarQuantizationCompression = class
    inherit VectorSearchCompression
    interface IJsonModel<ScalarQuantizationCompression>
    interface IPersistableModel<ScalarQuantizationCompression>
Public Class ScalarQuantizationCompression
Inherits VectorSearchCompression
Public Class ScalarQuantizationCompression
Inherits VectorSearchCompression
Implements IJsonModel(Of ScalarQuantizationCompression), IPersistableModel(Of ScalarQuantizationCompression)
Inheritance
ScalarQuantizationCompression
Implements

Constructors

ScalarQuantizationCompression(String)

Initializes a new instance of ScalarQuantizationCompression.

Properties

CompressionName

The name to associate with this particular configuration.

(Inherited from VectorSearchCompression)
DefaultOversampling

Default oversampling factor. Oversampling will internally request more documents (specified by this multiplier) in the initial search. This increases the set of results that will be reranked using recomputed similarity scores from full-precision vectors. Minimum value is 1, meaning no oversampling (1x). This parameter can only be set when rerankWithOriginalVectors is true. Higher values improve recall at the expense of latency.

(Inherited from VectorSearchCompression)
Parameters

Contains the parameters specific to Scalar Quantization.

RerankWithOriginalVectors

If set to true, once the ordered set of results calculated using compressed vectors are obtained, they will be reranked again by recalculating the full-precision similarity scores. This will improve recall at the expense of latency.

(Inherited from VectorSearchCompression)
RescoringOptions

Contains the options for rescoring.

(Inherited from VectorSearchCompression)
TruncationDimension

The number of dimensions to truncate the vectors to. Truncating the vectors reduces the size of the vectors and the amount of data that needs to be transferred during search. This can save storage cost and improve search performance at the expense of recall. It should be only used for embeddings trained with Matryoshka Representation Learning (MRL) such as OpenAI text-embedding-3-large (small). The default value is null, which means no truncation.

(Inherited from VectorSearchCompression)

Methods

JsonModelWriteCore(Utf8JsonWriter, ModelReaderWriterOptions)

Explicit Interface Implementations

IJsonModel<ScalarQuantizationCompression>.Create(Utf8JsonReader, ModelReaderWriterOptions)

Reads one JSON value (including objects or arrays) from the provided reader and converts it to a model.

IJsonModel<ScalarQuantizationCompression>.Write(Utf8JsonWriter, ModelReaderWriterOptions)

Writes the model to the provided Utf8JsonWriter.

IJsonModel<VectorSearchCompression>.Create(Utf8JsonReader, ModelReaderWriterOptions)

Reads one JSON value (including objects or arrays) from the provided reader and converts it to a model.

(Inherited from VectorSearchCompression)
IJsonModel<VectorSearchCompression>.Write(Utf8JsonWriter, ModelReaderWriterOptions)

Writes the model to the provided Utf8JsonWriter.

(Inherited from VectorSearchCompression)
IPersistableModel<ScalarQuantizationCompression>.Create(BinaryData, ModelReaderWriterOptions)

Converts the provided BinaryData into a model.

IPersistableModel<ScalarQuantizationCompression>.GetFormatFromOptions(ModelReaderWriterOptions)

Gets the data interchange format (JSON, Xml, etc) that the model uses when communicating with the service.

IPersistableModel<ScalarQuantizationCompression>.Write(ModelReaderWriterOptions)

Writes the model into a BinaryData.

IPersistableModel<VectorSearchCompression>.Create(BinaryData, ModelReaderWriterOptions)

Converts the provided BinaryData into a model.

(Inherited from VectorSearchCompression)
IPersistableModel<VectorSearchCompression>.GetFormatFromOptions(ModelReaderWriterOptions)

Gets the data interchange format (JSON, Xml, etc) that the model uses when communicating with the service.

(Inherited from VectorSearchCompression)
IPersistableModel<VectorSearchCompression>.Write(ModelReaderWriterOptions)

Writes the model into a BinaryData.

(Inherited from VectorSearchCompression)

Applies to