ONNX Pipeline Models : Text Embedding
ONNX pipeline models provides text embedding models that accepts text as input and produces embeddings as output. The pipeline models also provide the necessary pre-processing and post-processing needed for the text.
Note:
- This API is updated for 23.7. The version used for 23.6 and below used python packages EmbeddingModel and EmbeddingModelConfig. These packages are replaced with ONNXPipeline and ONNXPipelineConfig respectively. Oracle recommends that you use the latest version. You can find the details of the deprecated python classes in Python Classes to Convert Pretrained Models to ONNX Models (Deprecated).
-
This feature will only work on OML4Py 2.1 client. It is not supported on the OML4Py server.
-
In-database embedding models must include tokenization and post-processing. Providing only the core ONNX model to
DBMS_VECTOR.LOAD_ONNX_MODEL
is insufficient because you need to handle tokenization externally, pass tensors into the SQL operator, and convert output tensors into vectors. - Oracle is providing a Hugging Face
all-MiniLM-L12-v2 model in ONNX format, available to download
directly to the database using
DBMS_VECTOR.LOAD_ONNX_MODEL
. For more information, see the blog post Now Available! Pre-built Embedding Generation model for Oracle Database 23ai.
If you do not have a pretrained embedding model in ONNX-format to
generate embeddings for your data, Oracle offers a Python package that downloads
pretrained models from an external source, converts the model to ONNX format
augmented with pre-processing and post-processing steps, and imports the resulting
ONNX-format model into Oracle Database. Use the
DBMS_VECTOR.LOAD_ONNX_MODEL
procedure or
export2db()
function to import the file as a mining model.
(eport2db()
is a method on the ONNXPipeline
object).
Then leverage the in-database ONNX Runtime with the ONNX model to produce vector
embeddings.
- Downloads the pretrained model from external source (example: from Hugging Face repository) to your system
- Augments the model with pre-processing and post-processing steps and creates a new ONNX model
- Validates the augmented ONNX model
- Loads into the database as a data mining model or optionally exports to a file
The Python package can take any of the models in the preconfigured list as input. Alternatively, you can use the built-in template that contains common configurations for certain groups of models such as text or image models. To understand what a preconfigured list, what a built-in template is, and how to use them, read further.
Limitations
This table describes the limitations of the Python package.
Note:
This feature is available with the OML4Py 2.1 client only.Parameter | Description |
---|---|
Transformer Model Type |
Supports transformer models that supports text embedding. |
Model Size |
Model size should be less than 1GB. Quantization can help reduce the size. |
Tokenizers |
Must be either BERT ,
GPT2 , SENTENCEPIECE ,
CLIP, or ROBERTA .
|
Preconfigured List of Models
Preconfigured list of models are common models from external resource
repositories that are provided with the Python package. The preconfigured models
have an existing specification. Users can create their own specification using the
text template as a starting point. To get a list of all model names in the
preconfigured list, you can use the show_preconfigured
function.
Templates
The Python package provides built-in text template for you to configure the pretrained models with pre-processing and post-processing operations. The template has a default specification for the pretrained models. This specification can be changed or augmented to create custom configurations. The text template uses Mean Pooling and Normalization as post-processing operations by default.
End to End Example for Converting Text Models to ONNX Format and Storing them as a File or in a Database:
oml.utils
, ensure that you
have the following:
-
OML4Py 2.1 Client running on Linux X64 for On-Premises Databases
-
Python 3.12 (the earlier versions are not compatible)
DBMS_VECTOR.LOAD_ONNX_MODEL
is only needed if
export2file
was used to save the ONNX model file to the local
system instead of using export2db
to save the model in the database.
The DBMS_VECTOR.LOAD_ONNX_MODEL
imports the ONNX format model into the
Oracle Database to leverage the in-database ONNX Runtime to produce vector embeddings
using the VECTOR_EMBEDDING
SQL operator.
See Also:
- Oracle Database SQL
Language Reference for information about the
VECTOR_EMBEDDING
SQL function - Oracle Database PL/SQL
Packages and Types Reference for information about the
IMPORT_ONNX_MODEL
procedure - Oracle Database PL/SQL
Packages and Types Reference for information about the
LOAD_ONNX_MODEL
procedure - Oracle Machine Learning for SQL Concepts for more information about importing pretrained embedding models in ONNX format and generating vector embeddings
- https://onnx.ai/onnx/intro/ for ONNX documentation
Parent topic: Import Pretrained Models in ONNX Format