Install with HTTP with Models and Advanced Options
This tutorial is a superset of the HTTP with Configuration File tutorial that allows you to define vector embedding models that do not ship with the container. For a more advanced configuration, you can optionally choose to specify the HTTP port, container version, and or container name.
The API Key and SSL are neither configured nor used in this tutorial.
The default configuration is used with HTTP port 8080.
Choose Your ONNX Pipeline Models
There are many possible embedding models that can be used. Along with models that are shipped with the container, a more extensive list of embedding models that are known to work with the container can be found in Available Embedding Models.
Once you have chosen your desired ONNX Pipeline model, and built it with Oracle Machine Learning Client 2.1 (if you choose a model other than the pre-built options), you need to copy that ONNX file into a directory on the host machine in which the container will run.
You must also have a JSON configuration file that lists the desired ONNX Pipeline models to be used in the container.
In this tutorial, both the ONNX models and the config file are copied into their own directories and run as the container user with least privilege.
Parent topic: Configure the Private AI Services Container