Skip to main content
Version: 2.0.0

Generative AI

OCI Generative AI provides managed chat and embedding models through Spring-friendly abstractions.

Dependency Coordinates

Maven

<dependency>
<groupId>com.oracle.cloud.spring</groupId>
<artifactId>spring-cloud-oci-starter-gen-ai</artifactId>
</dependency>

Gradle

dependencies {
implementation("com.oracle.cloud.spring:spring-cloud-oci-starter-gen-ai")
}

Using Chat Models

The starter auto-configures a ChatModel bean:

@Autowired
private ChatModel chatModel;

public void chat() {
ChatResponse response = chatModel.chat("my chat prompt");
}

Using Embedding Models

The starter auto-configures an EmbeddingModel bean:

@Autowired
private EmbeddingModel embeddingModel;

public void embed() {
EmbedTextResponse response = embeddingModel.embed("my embedding text");
}

Configuration

NameDescriptionRequiredDefault
spring.cloud.oci.genai.enabledEnables the OCI Generative AI clientNotrue
spring.cloud.oci.genai.embedding.enabledEnables embedding APIsNofalse
spring.cloud.oci.genai.embedding.on-demand-model-idOn-demand embedding model IDNo
spring.cloud.oci.genai.embedding.dedicated-cluster-endpointDedicated embedding endpointNo
spring.cloud.oci.genai.embedding.compartmentCompartment for embeddingYes
spring.cloud.oci.genai.embedding.truncateTruncation mode: START, END, or NONENoNONE
spring.cloud.oci.genai.chat.enabledEnables chat APIsNofalse
spring.cloud.oci.genai.chat.on-demand-model-idOn-demand chat model IDNo
spring.cloud.oci.genai.chat.dedicated-cluster-endpointDedicated chat endpointNo
spring.cloud.oci.genai.chat.compartmentCompartment for chatYes
spring.cloud.oci.genai.chat.preample-overrideOverrides the model preambleNo
spring.cloud.oci.genai.chat.temperatureOutput temperatureNo1.0
spring.cloud.oci.genai.chat.top-pTop P sampling valueNo0.75
spring.cloud.oci.genai.chat.top-kTop K sampling valueNo0.0
spring.cloud.oci.genai.chat.frequency-penaltyPenalty for repeated tokensNo0.0
spring.cloud.oci.genai.chat.presence-penaltyPenalty when a token already appearsNo0.0
spring.cloud.oci.genai.chat.max-tokensMaximum output tokensNo600

For chat and embedding, specify either an on-demand model ID or a dedicated cluster endpoint for the respective feature.

Sample

See spring-cloud-oci-gen-ai-sample.