published on Tuesday, May 12, 2026 by kong
published on Tuesday, May 12, 2026 by kong
GatewayPluginAiProxy Resource
Example Usage
import * as pulumi from "@pulumi/pulumi";
import * as konnect from "@pulumi/konnect";
const myGatewaypluginaiproxy = new konnect.GatewayPluginAiProxy("my_gatewaypluginaiproxy", {
condition: "...my_condition...",
config: {
auth: {
allowOverride: false,
awsAccessKeyId: "...my_aws_access_key_id...",
awsSecretAccessKey: "...my_aws_secret_access_key...",
azureClientId: "...my_azure_client_id...",
azureClientSecret: "...my_azure_client_secret...",
azureTenantId: "...my_azure_tenant_id...",
azureUseManagedIdentity: false,
gcpMetadataUrl: "...my_gcp_metadata_url...",
gcpOauthTokenUrl: "...my_gcp_oauth_token_url...",
gcpServiceAccountJson: "...my_gcp_service_account_json...",
gcpUseServiceAccount: false,
headerName: "...my_header_name...",
headerValue: "...my_header_value...",
paramLocation: "query",
paramName: "...my_param_name...",
paramValue: "...my_param_value...",
},
genaiCategory: "text/generation",
llmFormat: "openai",
logging: {
logPayloads: false,
logStatistics: false,
},
maxRequestBodySize: 1048576,
model: {
modelAlias: "...my_model_alias...",
name: "...my_name...",
options: {
anthropicVersion: "...my_anthropic_version...",
azureApiVersion: "2023-05-15",
azureDeploymentId: "...my_azure_deployment_id...",
azureInstance: "...my_azure_instance...",
bedrock: {
awsAssumeRoleArn: "...my_aws_assume_role_arn...",
awsRegion: "...my_aws_region...",
awsRoleSessionName: "...my_aws_role_session_name...",
awsStsEndpointUrl: "...my_aws_sts_endpoint_url...",
batchBucketPrefix: "...my_batch_bucket_prefix...",
batchRoleArn: "...my_batch_role_arn...",
embeddingsNormalize: false,
performanceConfigLatency: "...my_performance_config_latency...",
videoOutputS3Uri: "...my_video_output_s3_uri...",
},
cohere: {
embeddingInputType: "classification",
waitForModel: false,
},
dashscope: {
international: true,
},
databricks: {
workspaceInstanceId: "...my_workspace_instance_id...",
},
embeddingsDimensions: 6,
gemini: {
apiEndpoint: "...my_api_endpoint...",
endpointId: "...my_endpoint_id...",
locationId: "...my_location_id...",
projectId: "...my_project_id...",
},
huggingface: {
useCache: true,
waitForModel: false,
},
inputCost: 7.42,
llama2Format: "openai",
maxTokens: 9,
mistralFormat: "ollama",
outputCost: 1.81,
temperature: 2.26,
topK: 359,
topP: 0.14,
upstreamPath: "...my_upstream_path...",
upstreamUrl: "...my_upstream_url...",
},
provider: "azure",
},
modelNameHeader: true,
responseStreaming: "allow",
routeType: "audio/v1/audio/speech",
},
consumer: {
id: "...my_id...",
},
consumerGroup: {
id: "...my_id...",
},
controlPlaneId: "9524ec7d-36d9-465d-a8c5-83a3c9390458",
createdAt: 9,
enabled: true,
gatewayPluginAiProxyId: "...my_id...",
instanceName: "...my_instance_name...",
ordering: {
after: {
accesses: ["..."],
},
before: {
accesses: ["..."],
},
},
partials: [{
id: "...my_id...",
name: "...my_name...",
path: "...my_path...",
}],
protocols: ["http"],
route: {
id: "...my_id...",
},
service: {
id: "...my_id...",
},
tags: ["..."],
updatedAt: 3,
});
import pulumi
import pulumi_konnect as konnect
my_gatewaypluginaiproxy = konnect.GatewayPluginAiProxy("my_gatewaypluginaiproxy",
condition="...my_condition...",
config={
"auth": {
"allow_override": False,
"aws_access_key_id": "...my_aws_access_key_id...",
"aws_secret_access_key": "...my_aws_secret_access_key...",
"azure_client_id": "...my_azure_client_id...",
"azure_client_secret": "...my_azure_client_secret...",
"azure_tenant_id": "...my_azure_tenant_id...",
"azure_use_managed_identity": False,
"gcp_metadata_url": "...my_gcp_metadata_url...",
"gcp_oauth_token_url": "...my_gcp_oauth_token_url...",
"gcp_service_account_json": "...my_gcp_service_account_json...",
"gcp_use_service_account": False,
"header_name": "...my_header_name...",
"header_value": "...my_header_value...",
"param_location": "query",
"param_name": "...my_param_name...",
"param_value": "...my_param_value...",
},
"genai_category": "text/generation",
"llm_format": "openai",
"logging": {
"log_payloads": False,
"log_statistics": False,
},
"max_request_body_size": 1048576,
"model": {
"model_alias": "...my_model_alias...",
"name": "...my_name...",
"options": {
"anthropic_version": "...my_anthropic_version...",
"azure_api_version": "2023-05-15",
"azure_deployment_id": "...my_azure_deployment_id...",
"azure_instance": "...my_azure_instance...",
"bedrock": {
"aws_assume_role_arn": "...my_aws_assume_role_arn...",
"aws_region": "...my_aws_region...",
"aws_role_session_name": "...my_aws_role_session_name...",
"aws_sts_endpoint_url": "...my_aws_sts_endpoint_url...",
"batch_bucket_prefix": "...my_batch_bucket_prefix...",
"batch_role_arn": "...my_batch_role_arn...",
"embeddings_normalize": False,
"performance_config_latency": "...my_performance_config_latency...",
"video_output_s3_uri": "...my_video_output_s3_uri...",
},
"cohere": {
"embedding_input_type": "classification",
"wait_for_model": False,
},
"dashscope": {
"international": True,
},
"databricks": {
"workspace_instance_id": "...my_workspace_instance_id...",
},
"embeddings_dimensions": 6,
"gemini": {
"api_endpoint": "...my_api_endpoint...",
"endpoint_id": "...my_endpoint_id...",
"location_id": "...my_location_id...",
"project_id": "...my_project_id...",
},
"huggingface": {
"use_cache": True,
"wait_for_model": False,
},
"input_cost": 7.42,
"llama2_format": "openai",
"max_tokens": 9,
"mistral_format": "ollama",
"output_cost": 1.81,
"temperature": 2.26,
"top_k": 359,
"top_p": 0.14,
"upstream_path": "...my_upstream_path...",
"upstream_url": "...my_upstream_url...",
},
"provider": "azure",
},
"model_name_header": True,
"response_streaming": "allow",
"route_type": "audio/v1/audio/speech",
},
consumer={
"id": "...my_id...",
},
consumer_group={
"id": "...my_id...",
},
control_plane_id="9524ec7d-36d9-465d-a8c5-83a3c9390458",
created_at=9,
enabled=True,
gateway_plugin_ai_proxy_id="...my_id...",
instance_name="...my_instance_name...",
ordering={
"after": {
"accesses": ["..."],
},
"before": {
"accesses": ["..."],
},
},
partials=[{
"id": "...my_id...",
"name": "...my_name...",
"path": "...my_path...",
}],
protocols=["http"],
route={
"id": "...my_id...",
},
service={
"id": "...my_id...",
},
tags=["..."],
updated_at=3)
package main
import (
"github.com/pulumi/pulumi-terraform-provider/sdks/go/konnect/v3/konnect"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
_, err := konnect.NewGatewayPluginAiProxy(ctx, "my_gatewaypluginaiproxy", &konnect.GatewayPluginAiProxyArgs{
Condition: pulumi.String("...my_condition..."),
Config: &konnect.GatewayPluginAiProxyConfigArgs{
Auth: &konnect.GatewayPluginAiProxyConfigAuthArgs{
AllowOverride: pulumi.Bool(false),
AwsAccessKeyId: pulumi.String("...my_aws_access_key_id..."),
AwsSecretAccessKey: pulumi.String("...my_aws_secret_access_key..."),
AzureClientId: pulumi.String("...my_azure_client_id..."),
AzureClientSecret: pulumi.String("...my_azure_client_secret..."),
AzureTenantId: pulumi.String("...my_azure_tenant_id..."),
AzureUseManagedIdentity: pulumi.Bool(false),
GcpMetadataUrl: pulumi.String("...my_gcp_metadata_url..."),
GcpOauthTokenUrl: pulumi.String("...my_gcp_oauth_token_url..."),
GcpServiceAccountJson: pulumi.String("...my_gcp_service_account_json..."),
GcpUseServiceAccount: pulumi.Bool(false),
HeaderName: pulumi.String("...my_header_name..."),
HeaderValue: pulumi.String("...my_header_value..."),
ParamLocation: pulumi.String("query"),
ParamName: pulumi.String("...my_param_name..."),
ParamValue: pulumi.String("...my_param_value..."),
},
GenaiCategory: pulumi.String("text/generation"),
LlmFormat: pulumi.String("openai"),
Logging: &konnect.GatewayPluginAiProxyConfigLoggingArgs{
LogPayloads: pulumi.Bool(false),
LogStatistics: pulumi.Bool(false),
},
MaxRequestBodySize: pulumi.Float64(1048576),
Model: &konnect.GatewayPluginAiProxyConfigModelArgs{
ModelAlias: pulumi.String("...my_model_alias..."),
Name: pulumi.String("...my_name..."),
Options: &konnect.GatewayPluginAiProxyConfigModelOptionsArgs{
AnthropicVersion: pulumi.String("...my_anthropic_version..."),
AzureApiVersion: pulumi.String("2023-05-15"),
AzureDeploymentId: pulumi.String("...my_azure_deployment_id..."),
AzureInstance: pulumi.String("...my_azure_instance..."),
Bedrock: &konnect.GatewayPluginAiProxyConfigModelOptionsBedrockArgs{
AwsAssumeRoleArn: pulumi.String("...my_aws_assume_role_arn..."),
AwsRegion: pulumi.String("...my_aws_region..."),
AwsRoleSessionName: pulumi.String("...my_aws_role_session_name..."),
AwsStsEndpointUrl: pulumi.String("...my_aws_sts_endpoint_url..."),
BatchBucketPrefix: pulumi.String("...my_batch_bucket_prefix..."),
BatchRoleArn: pulumi.String("...my_batch_role_arn..."),
EmbeddingsNormalize: pulumi.Bool(false),
PerformanceConfigLatency: pulumi.String("...my_performance_config_latency..."),
VideoOutputS3Uri: pulumi.String("...my_video_output_s3_uri..."),
},
Cohere: &konnect.GatewayPluginAiProxyConfigModelOptionsCohereArgs{
EmbeddingInputType: pulumi.String("classification"),
WaitForModel: pulumi.Bool(false),
},
Dashscope: &konnect.GatewayPluginAiProxyConfigModelOptionsDashscopeArgs{
International: pulumi.Bool(true),
},
Databricks: &konnect.GatewayPluginAiProxyConfigModelOptionsDatabricksArgs{
WorkspaceInstanceId: pulumi.String("...my_workspace_instance_id..."),
},
EmbeddingsDimensions: pulumi.Float64(6),
Gemini: &konnect.GatewayPluginAiProxyConfigModelOptionsGeminiArgs{
ApiEndpoint: pulumi.String("...my_api_endpoint..."),
EndpointId: pulumi.String("...my_endpoint_id..."),
LocationId: pulumi.String("...my_location_id..."),
ProjectId: pulumi.String("...my_project_id..."),
},
Huggingface: &konnect.GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs{
UseCache: pulumi.Bool(true),
WaitForModel: pulumi.Bool(false),
},
InputCost: pulumi.Float64(7.42),
Llama2Format: pulumi.String("openai"),
MaxTokens: pulumi.Float64(9),
MistralFormat: pulumi.String("ollama"),
OutputCost: pulumi.Float64(1.81),
Temperature: pulumi.Float64(2.26),
TopK: pulumi.Float64(359),
TopP: pulumi.Float64(0.14),
UpstreamPath: pulumi.String("...my_upstream_path..."),
UpstreamUrl: pulumi.String("...my_upstream_url..."),
},
Provider: pulumi.String("azure"),
},
ModelNameHeader: pulumi.Bool(true),
ResponseStreaming: pulumi.String("allow"),
RouteType: pulumi.String("audio/v1/audio/speech"),
},
Consumer: &konnect.GatewayPluginAiProxyConsumerArgs{
Id: pulumi.String("...my_id..."),
},
ConsumerGroup: &konnect.GatewayPluginAiProxyConsumerGroupArgs{
Id: pulumi.String("...my_id..."),
},
ControlPlaneId: pulumi.String("9524ec7d-36d9-465d-a8c5-83a3c9390458"),
CreatedAt: pulumi.Float64(9),
Enabled: pulumi.Bool(true),
GatewayPluginAiProxyId: pulumi.String("...my_id..."),
InstanceName: pulumi.String("...my_instance_name..."),
Ordering: &konnect.GatewayPluginAiProxyOrderingArgs{
After: &konnect.GatewayPluginAiProxyOrderingAfterArgs{
Accesses: pulumi.StringArray{
pulumi.String("..."),
},
},
Before: &konnect.GatewayPluginAiProxyOrderingBeforeArgs{
Accesses: pulumi.StringArray{
pulumi.String("..."),
},
},
},
Partials: konnect.GatewayPluginAiProxyPartialArray{
&konnect.GatewayPluginAiProxyPartialArgs{
Id: pulumi.String("...my_id..."),
Name: pulumi.String("...my_name..."),
Path: pulumi.String("...my_path..."),
},
},
Protocols: pulumi.StringArray{
pulumi.String("http"),
},
Route: &konnect.GatewayPluginAiProxyRouteArgs{
Id: pulumi.String("...my_id..."),
},
Service: &konnect.GatewayPluginAiProxyServiceArgs{
Id: pulumi.String("...my_id..."),
},
Tags: pulumi.StringArray{
pulumi.String("..."),
},
UpdatedAt: pulumi.Float64(3),
})
if err != nil {
return err
}
return nil
})
}
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Konnect = Pulumi.Konnect;
return await Deployment.RunAsync(() =>
{
var myGatewaypluginaiproxy = new Konnect.GatewayPluginAiProxy("my_gatewaypluginaiproxy", new()
{
Condition = "...my_condition...",
Config = new Konnect.Inputs.GatewayPluginAiProxyConfigArgs
{
Auth = new Konnect.Inputs.GatewayPluginAiProxyConfigAuthArgs
{
AllowOverride = false,
AwsAccessKeyId = "...my_aws_access_key_id...",
AwsSecretAccessKey = "...my_aws_secret_access_key...",
AzureClientId = "...my_azure_client_id...",
AzureClientSecret = "...my_azure_client_secret...",
AzureTenantId = "...my_azure_tenant_id...",
AzureUseManagedIdentity = false,
GcpMetadataUrl = "...my_gcp_metadata_url...",
GcpOauthTokenUrl = "...my_gcp_oauth_token_url...",
GcpServiceAccountJson = "...my_gcp_service_account_json...",
GcpUseServiceAccount = false,
HeaderName = "...my_header_name...",
HeaderValue = "...my_header_value...",
ParamLocation = "query",
ParamName = "...my_param_name...",
ParamValue = "...my_param_value...",
},
GenaiCategory = "text/generation",
LlmFormat = "openai",
Logging = new Konnect.Inputs.GatewayPluginAiProxyConfigLoggingArgs
{
LogPayloads = false,
LogStatistics = false,
},
MaxRequestBodySize = 1048576,
Model = new Konnect.Inputs.GatewayPluginAiProxyConfigModelArgs
{
ModelAlias = "...my_model_alias...",
Name = "...my_name...",
Options = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsArgs
{
AnthropicVersion = "...my_anthropic_version...",
AzureApiVersion = "2023-05-15",
AzureDeploymentId = "...my_azure_deployment_id...",
AzureInstance = "...my_azure_instance...",
Bedrock = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsBedrockArgs
{
AwsAssumeRoleArn = "...my_aws_assume_role_arn...",
AwsRegion = "...my_aws_region...",
AwsRoleSessionName = "...my_aws_role_session_name...",
AwsStsEndpointUrl = "...my_aws_sts_endpoint_url...",
BatchBucketPrefix = "...my_batch_bucket_prefix...",
BatchRoleArn = "...my_batch_role_arn...",
EmbeddingsNormalize = false,
PerformanceConfigLatency = "...my_performance_config_latency...",
VideoOutputS3Uri = "...my_video_output_s3_uri...",
},
Cohere = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsCohereArgs
{
EmbeddingInputType = "classification",
WaitForModel = false,
},
Dashscope = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsDashscopeArgs
{
International = true,
},
Databricks = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsDatabricksArgs
{
WorkspaceInstanceId = "...my_workspace_instance_id...",
},
EmbeddingsDimensions = 6,
Gemini = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsGeminiArgs
{
ApiEndpoint = "...my_api_endpoint...",
EndpointId = "...my_endpoint_id...",
LocationId = "...my_location_id...",
ProjectId = "...my_project_id...",
},
Huggingface = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs
{
UseCache = true,
WaitForModel = false,
},
InputCost = 7.42,
Llama2Format = "openai",
MaxTokens = 9,
MistralFormat = "ollama",
OutputCost = 1.81,
Temperature = 2.26,
TopK = 359,
TopP = 0.14,
UpstreamPath = "...my_upstream_path...",
UpstreamUrl = "...my_upstream_url...",
},
Provider = "azure",
},
ModelNameHeader = true,
ResponseStreaming = "allow",
RouteType = "audio/v1/audio/speech",
},
Consumer = new Konnect.Inputs.GatewayPluginAiProxyConsumerArgs
{
Id = "...my_id...",
},
ConsumerGroup = new Konnect.Inputs.GatewayPluginAiProxyConsumerGroupArgs
{
Id = "...my_id...",
},
ControlPlaneId = "9524ec7d-36d9-465d-a8c5-83a3c9390458",
CreatedAt = 9,
Enabled = true,
GatewayPluginAiProxyId = "...my_id...",
InstanceName = "...my_instance_name...",
Ordering = new Konnect.Inputs.GatewayPluginAiProxyOrderingArgs
{
After = new Konnect.Inputs.GatewayPluginAiProxyOrderingAfterArgs
{
Accesses = new[]
{
"...",
},
},
Before = new Konnect.Inputs.GatewayPluginAiProxyOrderingBeforeArgs
{
Accesses = new[]
{
"...",
},
},
},
Partials = new[]
{
new Konnect.Inputs.GatewayPluginAiProxyPartialArgs
{
Id = "...my_id...",
Name = "...my_name...",
Path = "...my_path...",
},
},
Protocols = new[]
{
"http",
},
Route = new Konnect.Inputs.GatewayPluginAiProxyRouteArgs
{
Id = "...my_id...",
},
Service = new Konnect.Inputs.GatewayPluginAiProxyServiceArgs
{
Id = "...my_id...",
},
Tags = new[]
{
"...",
},
UpdatedAt = 3,
});
});
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.konnect.GatewayPluginAiProxy;
import com.pulumi.konnect.GatewayPluginAiProxyArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigAuthArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigLoggingArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelOptionsArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelOptionsBedrockArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelOptionsCohereArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelOptionsDashscopeArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelOptionsDatabricksArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelOptionsGeminiArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConsumerArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyConsumerGroupArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyOrderingArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyOrderingAfterArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyOrderingBeforeArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyPartialArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyRouteArgs;
import com.pulumi.konnect.inputs.GatewayPluginAiProxyServiceArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var myGatewaypluginaiproxy = new GatewayPluginAiProxy("myGatewaypluginaiproxy", GatewayPluginAiProxyArgs.builder()
.condition("...my_condition...")
.config(GatewayPluginAiProxyConfigArgs.builder()
.auth(GatewayPluginAiProxyConfigAuthArgs.builder()
.allowOverride(false)
.awsAccessKeyId("...my_aws_access_key_id...")
.awsSecretAccessKey("...my_aws_secret_access_key...")
.azureClientId("...my_azure_client_id...")
.azureClientSecret("...my_azure_client_secret...")
.azureTenantId("...my_azure_tenant_id...")
.azureUseManagedIdentity(false)
.gcpMetadataUrl("...my_gcp_metadata_url...")
.gcpOauthTokenUrl("...my_gcp_oauth_token_url...")
.gcpServiceAccountJson("...my_gcp_service_account_json...")
.gcpUseServiceAccount(false)
.headerName("...my_header_name...")
.headerValue("...my_header_value...")
.paramLocation("query")
.paramName("...my_param_name...")
.paramValue("...my_param_value...")
.build())
.genaiCategory("text/generation")
.llmFormat("openai")
.logging(GatewayPluginAiProxyConfigLoggingArgs.builder()
.logPayloads(false)
.logStatistics(false)
.build())
.maxRequestBodySize(1048576.0)
.model(GatewayPluginAiProxyConfigModelArgs.builder()
.modelAlias("...my_model_alias...")
.name("...my_name...")
.options(GatewayPluginAiProxyConfigModelOptionsArgs.builder()
.anthropicVersion("...my_anthropic_version...")
.azureApiVersion("2023-05-15")
.azureDeploymentId("...my_azure_deployment_id...")
.azureInstance("...my_azure_instance...")
.bedrock(GatewayPluginAiProxyConfigModelOptionsBedrockArgs.builder()
.awsAssumeRoleArn("...my_aws_assume_role_arn...")
.awsRegion("...my_aws_region...")
.awsRoleSessionName("...my_aws_role_session_name...")
.awsStsEndpointUrl("...my_aws_sts_endpoint_url...")
.batchBucketPrefix("...my_batch_bucket_prefix...")
.batchRoleArn("...my_batch_role_arn...")
.embeddingsNormalize(false)
.performanceConfigLatency("...my_performance_config_latency...")
.videoOutputS3Uri("...my_video_output_s3_uri...")
.build())
.cohere(GatewayPluginAiProxyConfigModelOptionsCohereArgs.builder()
.embeddingInputType("classification")
.waitForModel(false)
.build())
.dashscope(GatewayPluginAiProxyConfigModelOptionsDashscopeArgs.builder()
.international(true)
.build())
.databricks(GatewayPluginAiProxyConfigModelOptionsDatabricksArgs.builder()
.workspaceInstanceId("...my_workspace_instance_id...")
.build())
.embeddingsDimensions(6.0)
.gemini(GatewayPluginAiProxyConfigModelOptionsGeminiArgs.builder()
.apiEndpoint("...my_api_endpoint...")
.endpointId("...my_endpoint_id...")
.locationId("...my_location_id...")
.projectId("...my_project_id...")
.build())
.huggingface(GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs.builder()
.useCache(true)
.waitForModel(false)
.build())
.inputCost(7.42)
.llama2Format("openai")
.maxTokens(9.0)
.mistralFormat("ollama")
.outputCost(1.81)
.temperature(2.26)
.topK(359.0)
.topP(0.14)
.upstreamPath("...my_upstream_path...")
.upstreamUrl("...my_upstream_url...")
.build())
.provider("azure")
.build())
.modelNameHeader(true)
.responseStreaming("allow")
.routeType("audio/v1/audio/speech")
.build())
.consumer(GatewayPluginAiProxyConsumerArgs.builder()
.id("...my_id...")
.build())
.consumerGroup(GatewayPluginAiProxyConsumerGroupArgs.builder()
.id("...my_id...")
.build())
.controlPlaneId("9524ec7d-36d9-465d-a8c5-83a3c9390458")
.createdAt(9.0)
.enabled(true)
.gatewayPluginAiProxyId("...my_id...")
.instanceName("...my_instance_name...")
.ordering(GatewayPluginAiProxyOrderingArgs.builder()
.after(GatewayPluginAiProxyOrderingAfterArgs.builder()
.accesses("...")
.build())
.before(GatewayPluginAiProxyOrderingBeforeArgs.builder()
.accesses("...")
.build())
.build())
.partials(GatewayPluginAiProxyPartialArgs.builder()
.id("...my_id...")
.name("...my_name...")
.path("...my_path...")
.build())
.protocols("http")
.route(GatewayPluginAiProxyRouteArgs.builder()
.id("...my_id...")
.build())
.service(GatewayPluginAiProxyServiceArgs.builder()
.id("...my_id...")
.build())
.tags("...")
.updatedAt(3.0)
.build());
}
}
resources:
myGatewaypluginaiproxy:
type: konnect:GatewayPluginAiProxy
name: my_gatewaypluginaiproxy
properties:
condition: '...my_condition...'
config:
auth:
allowOverride: false
awsAccessKeyId: '...my_aws_access_key_id...'
awsSecretAccessKey: '...my_aws_secret_access_key...'
azureClientId: '...my_azure_client_id...'
azureClientSecret: '...my_azure_client_secret...'
azureTenantId: '...my_azure_tenant_id...'
azureUseManagedIdentity: false
gcpMetadataUrl: '...my_gcp_metadata_url...'
gcpOauthTokenUrl: '...my_gcp_oauth_token_url...'
gcpServiceAccountJson: '...my_gcp_service_account_json...'
gcpUseServiceAccount: false
headerName: '...my_header_name...'
headerValue: '...my_header_value...'
paramLocation: query
paramName: '...my_param_name...'
paramValue: '...my_param_value...'
genaiCategory: text/generation
llmFormat: openai
logging:
logPayloads: false
logStatistics: false
maxRequestBodySize: 1.048576e+06
model:
modelAlias: '...my_model_alias...'
name: '...my_name...'
options:
anthropicVersion: '...my_anthropic_version...'
azureApiVersion: 2023-05-15
azureDeploymentId: '...my_azure_deployment_id...'
azureInstance: '...my_azure_instance...'
bedrock:
awsAssumeRoleArn: '...my_aws_assume_role_arn...'
awsRegion: '...my_aws_region...'
awsRoleSessionName: '...my_aws_role_session_name...'
awsStsEndpointUrl: '...my_aws_sts_endpoint_url...'
batchBucketPrefix: '...my_batch_bucket_prefix...'
batchRoleArn: '...my_batch_role_arn...'
embeddingsNormalize: false
performanceConfigLatency: '...my_performance_config_latency...'
videoOutputS3Uri: '...my_video_output_s3_uri...'
cohere:
embeddingInputType: classification
waitForModel: false
dashscope:
international: true
databricks:
workspaceInstanceId: '...my_workspace_instance_id...'
embeddingsDimensions: 6
gemini:
apiEndpoint: '...my_api_endpoint...'
endpointId: '...my_endpoint_id...'
locationId: '...my_location_id...'
projectId: '...my_project_id...'
huggingface:
useCache: true
waitForModel: false
inputCost: 7.42
llama2Format: openai
maxTokens: 9
mistralFormat: ollama
outputCost: 1.81
temperature: 2.26
topK: 359
topP: 0.14
upstreamPath: '...my_upstream_path...'
upstreamUrl: '...my_upstream_url...'
provider: azure
modelNameHeader: true
responseStreaming: allow
routeType: audio/v1/audio/speech
consumer:
id: '...my_id...'
consumerGroup:
id: '...my_id...'
controlPlaneId: 9524ec7d-36d9-465d-a8c5-83a3c9390458
createdAt: 9
enabled: true
gatewayPluginAiProxyId: '...my_id...'
instanceName: '...my_instance_name...'
ordering:
after:
accesses:
- '...'
before:
accesses:
- '...'
partials:
- id: '...my_id...'
name: '...my_name...'
path: '...my_path...'
protocols:
- http
route:
id: '...my_id...'
service:
id: '...my_id...'
tags:
- '...'
updatedAt: 3
Example coming soon!
Create GatewayPluginAiProxy Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new GatewayPluginAiProxy(name: string, args: GatewayPluginAiProxyArgs, opts?: CustomResourceOptions);@overload
def GatewayPluginAiProxy(resource_name: str,
args: GatewayPluginAiProxyArgs,
opts: Optional[ResourceOptions] = None)
@overload
def GatewayPluginAiProxy(resource_name: str,
opts: Optional[ResourceOptions] = None,
control_plane_id: Optional[str] = None,
config: Optional[GatewayPluginAiProxyConfigArgs] = None,
instance_name: Optional[str] = None,
ordering: Optional[GatewayPluginAiProxyOrderingArgs] = None,
consumer: Optional[GatewayPluginAiProxyConsumerArgs] = None,
created_at: Optional[float] = None,
enabled: Optional[bool] = None,
gateway_plugin_ai_proxy_id: Optional[str] = None,
condition: Optional[str] = None,
consumer_group: Optional[GatewayPluginAiProxyConsumerGroupArgs] = None,
partials: Optional[Sequence[GatewayPluginAiProxyPartialArgs]] = None,
protocols: Optional[Sequence[str]] = None,
route: Optional[GatewayPluginAiProxyRouteArgs] = None,
service: Optional[GatewayPluginAiProxyServiceArgs] = None,
tags: Optional[Sequence[str]] = None,
updated_at: Optional[float] = None)func NewGatewayPluginAiProxy(ctx *Context, name string, args GatewayPluginAiProxyArgs, opts ...ResourceOption) (*GatewayPluginAiProxy, error)public GatewayPluginAiProxy(string name, GatewayPluginAiProxyArgs args, CustomResourceOptions? opts = null)
public GatewayPluginAiProxy(String name, GatewayPluginAiProxyArgs args)
public GatewayPluginAiProxy(String name, GatewayPluginAiProxyArgs args, CustomResourceOptions options)
type: konnect:GatewayPluginAiProxy
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
resource "konnect_gatewaypluginaiproxy" "name" {
# resource properties
}Parameters
- name string
- The unique name of the resource.
- args GatewayPluginAiProxyArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args GatewayPluginAiProxyArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args GatewayPluginAiProxyArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args GatewayPluginAiProxyArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args GatewayPluginAiProxyArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var gatewayPluginAiProxyResource = new Konnect.GatewayPluginAiProxy("gatewayPluginAiProxyResource", new()
{
ControlPlaneId = "string",
Config = new Konnect.Inputs.GatewayPluginAiProxyConfigArgs
{
Model = new Konnect.Inputs.GatewayPluginAiProxyConfigModelArgs
{
Provider = "string",
ModelAlias = "string",
Name = "string",
Options = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsArgs
{
AnthropicVersion = "string",
AzureApiVersion = "string",
AzureDeploymentId = "string",
AzureInstance = "string",
Bedrock = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsBedrockArgs
{
AwsAssumeRoleArn = "string",
AwsRegion = "string",
AwsRoleSessionName = "string",
AwsStsEndpointUrl = "string",
BatchBucketPrefix = "string",
BatchRoleArn = "string",
EmbeddingsNormalize = false,
PerformanceConfigLatency = "string",
VideoOutputS3Uri = "string",
},
Cohere = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsCohereArgs
{
EmbeddingInputType = "string",
WaitForModel = false,
},
Dashscope = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsDashscopeArgs
{
International = false,
},
Databricks = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsDatabricksArgs
{
WorkspaceInstanceId = "string",
},
EmbeddingsDimensions = 0,
Gemini = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsGeminiArgs
{
ApiEndpoint = "string",
EndpointId = "string",
LocationId = "string",
ProjectId = "string",
},
Huggingface = new Konnect.Inputs.GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs
{
UseCache = false,
WaitForModel = false,
},
InputCost = 0,
Llama2Format = "string",
MaxTokens = 0,
MistralFormat = "string",
OutputCost = 0,
Temperature = 0,
TopK = 0,
TopP = 0,
UpstreamPath = "string",
UpstreamUrl = "string",
},
},
RouteType = "string",
Auth = new Konnect.Inputs.GatewayPluginAiProxyConfigAuthArgs
{
AllowOverride = false,
AwsAccessKeyId = "string",
AwsSecretAccessKey = "string",
AzureClientId = "string",
AzureClientSecret = "string",
AzureTenantId = "string",
AzureUseManagedIdentity = false,
GcpMetadataUrl = "string",
GcpOauthTokenUrl = "string",
GcpServiceAccountJson = "string",
GcpUseServiceAccount = false,
HeaderName = "string",
HeaderValue = "string",
ParamLocation = "string",
ParamName = "string",
ParamValue = "string",
},
GenaiCategory = "string",
LlmFormat = "string",
Logging = new Konnect.Inputs.GatewayPluginAiProxyConfigLoggingArgs
{
LogPayloads = false,
LogStatistics = false,
},
MaxRequestBodySize = 0,
ModelNameHeader = false,
ResponseStreaming = "string",
},
InstanceName = "string",
Ordering = new Konnect.Inputs.GatewayPluginAiProxyOrderingArgs
{
After = new Konnect.Inputs.GatewayPluginAiProxyOrderingAfterArgs
{
Accesses = new[]
{
"string",
},
},
Before = new Konnect.Inputs.GatewayPluginAiProxyOrderingBeforeArgs
{
Accesses = new[]
{
"string",
},
},
},
Consumer = new Konnect.Inputs.GatewayPluginAiProxyConsumerArgs
{
Id = "string",
},
CreatedAt = 0,
Enabled = false,
GatewayPluginAiProxyId = "string",
Condition = "string",
ConsumerGroup = new Konnect.Inputs.GatewayPluginAiProxyConsumerGroupArgs
{
Id = "string",
},
Partials = new[]
{
new Konnect.Inputs.GatewayPluginAiProxyPartialArgs
{
Id = "string",
Name = "string",
Path = "string",
},
},
Protocols = new[]
{
"string",
},
Route = new Konnect.Inputs.GatewayPluginAiProxyRouteArgs
{
Id = "string",
},
Service = new Konnect.Inputs.GatewayPluginAiProxyServiceArgs
{
Id = "string",
},
Tags = new[]
{
"string",
},
UpdatedAt = 0,
});
example, err := konnect.NewGatewayPluginAiProxy(ctx, "gatewayPluginAiProxyResource", &konnect.GatewayPluginAiProxyArgs{
ControlPlaneId: pulumi.String("string"),
Config: &konnect.GatewayPluginAiProxyConfigArgs{
Model: &konnect.GatewayPluginAiProxyConfigModelArgs{
Provider: pulumi.String("string"),
ModelAlias: pulumi.String("string"),
Name: pulumi.String("string"),
Options: &konnect.GatewayPluginAiProxyConfigModelOptionsArgs{
AnthropicVersion: pulumi.String("string"),
AzureApiVersion: pulumi.String("string"),
AzureDeploymentId: pulumi.String("string"),
AzureInstance: pulumi.String("string"),
Bedrock: &konnect.GatewayPluginAiProxyConfigModelOptionsBedrockArgs{
AwsAssumeRoleArn: pulumi.String("string"),
AwsRegion: pulumi.String("string"),
AwsRoleSessionName: pulumi.String("string"),
AwsStsEndpointUrl: pulumi.String("string"),
BatchBucketPrefix: pulumi.String("string"),
BatchRoleArn: pulumi.String("string"),
EmbeddingsNormalize: pulumi.Bool(false),
PerformanceConfigLatency: pulumi.String("string"),
VideoOutputS3Uri: pulumi.String("string"),
},
Cohere: &konnect.GatewayPluginAiProxyConfigModelOptionsCohereArgs{
EmbeddingInputType: pulumi.String("string"),
WaitForModel: pulumi.Bool(false),
},
Dashscope: &konnect.GatewayPluginAiProxyConfigModelOptionsDashscopeArgs{
International: pulumi.Bool(false),
},
Databricks: &konnect.GatewayPluginAiProxyConfigModelOptionsDatabricksArgs{
WorkspaceInstanceId: pulumi.String("string"),
},
EmbeddingsDimensions: pulumi.Float64(0),
Gemini: &konnect.GatewayPluginAiProxyConfigModelOptionsGeminiArgs{
ApiEndpoint: pulumi.String("string"),
EndpointId: pulumi.String("string"),
LocationId: pulumi.String("string"),
ProjectId: pulumi.String("string"),
},
Huggingface: &konnect.GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs{
UseCache: pulumi.Bool(false),
WaitForModel: pulumi.Bool(false),
},
InputCost: pulumi.Float64(0),
Llama2Format: pulumi.String("string"),
MaxTokens: pulumi.Float64(0),
MistralFormat: pulumi.String("string"),
OutputCost: pulumi.Float64(0),
Temperature: pulumi.Float64(0),
TopK: pulumi.Float64(0),
TopP: pulumi.Float64(0),
UpstreamPath: pulumi.String("string"),
UpstreamUrl: pulumi.String("string"),
},
},
RouteType: pulumi.String("string"),
Auth: &konnect.GatewayPluginAiProxyConfigAuthArgs{
AllowOverride: pulumi.Bool(false),
AwsAccessKeyId: pulumi.String("string"),
AwsSecretAccessKey: pulumi.String("string"),
AzureClientId: pulumi.String("string"),
AzureClientSecret: pulumi.String("string"),
AzureTenantId: pulumi.String("string"),
AzureUseManagedIdentity: pulumi.Bool(false),
GcpMetadataUrl: pulumi.String("string"),
GcpOauthTokenUrl: pulumi.String("string"),
GcpServiceAccountJson: pulumi.String("string"),
GcpUseServiceAccount: pulumi.Bool(false),
HeaderName: pulumi.String("string"),
HeaderValue: pulumi.String("string"),
ParamLocation: pulumi.String("string"),
ParamName: pulumi.String("string"),
ParamValue: pulumi.String("string"),
},
GenaiCategory: pulumi.String("string"),
LlmFormat: pulumi.String("string"),
Logging: &konnect.GatewayPluginAiProxyConfigLoggingArgs{
LogPayloads: pulumi.Bool(false),
LogStatistics: pulumi.Bool(false),
},
MaxRequestBodySize: pulumi.Float64(0),
ModelNameHeader: pulumi.Bool(false),
ResponseStreaming: pulumi.String("string"),
},
InstanceName: pulumi.String("string"),
Ordering: &konnect.GatewayPluginAiProxyOrderingArgs{
After: &konnect.GatewayPluginAiProxyOrderingAfterArgs{
Accesses: pulumi.StringArray{
pulumi.String("string"),
},
},
Before: &konnect.GatewayPluginAiProxyOrderingBeforeArgs{
Accesses: pulumi.StringArray{
pulumi.String("string"),
},
},
},
Consumer: &konnect.GatewayPluginAiProxyConsumerArgs{
Id: pulumi.String("string"),
},
CreatedAt: pulumi.Float64(0),
Enabled: pulumi.Bool(false),
GatewayPluginAiProxyId: pulumi.String("string"),
Condition: pulumi.String("string"),
ConsumerGroup: &konnect.GatewayPluginAiProxyConsumerGroupArgs{
Id: pulumi.String("string"),
},
Partials: konnect.GatewayPluginAiProxyPartialArray{
&konnect.GatewayPluginAiProxyPartialArgs{
Id: pulumi.String("string"),
Name: pulumi.String("string"),
Path: pulumi.String("string"),
},
},
Protocols: pulumi.StringArray{
pulumi.String("string"),
},
Route: &konnect.GatewayPluginAiProxyRouteArgs{
Id: pulumi.String("string"),
},
Service: &konnect.GatewayPluginAiProxyServiceArgs{
Id: pulumi.String("string"),
},
Tags: pulumi.StringArray{
pulumi.String("string"),
},
UpdatedAt: pulumi.Float64(0),
})
resource "konnect_gatewaypluginaiproxy" "gatewayPluginAiProxyResource" {
control_plane_id = "string"
config = {
model = {
provider = "string"
model_alias = "string"
name = "string"
options = {
anthropic_version = "string"
azure_api_version = "string"
azure_deployment_id = "string"
azure_instance = "string"
bedrock = {
aws_assume_role_arn = "string"
aws_region = "string"
aws_role_session_name = "string"
aws_sts_endpoint_url = "string"
batch_bucket_prefix = "string"
batch_role_arn = "string"
embeddings_normalize = false
performance_config_latency = "string"
video_output_s3_uri = "string"
}
cohere = {
embedding_input_type = "string"
wait_for_model = false
}
dashscope = {
international = false
}
databricks = {
workspace_instance_id = "string"
}
embeddings_dimensions = 0
gemini = {
api_endpoint = "string"
endpoint_id = "string"
location_id = "string"
project_id = "string"
}
huggingface = {
use_cache = false
wait_for_model = false
}
input_cost = 0
llama2_format = "string"
max_tokens = 0
mistral_format = "string"
output_cost = 0
temperature = 0
top_k = 0
top_p = 0
upstream_path = "string"
upstream_url = "string"
}
}
route_type = "string"
auth = {
allow_override = false
aws_access_key_id = "string"
aws_secret_access_key = "string"
azure_client_id = "string"
azure_client_secret = "string"
azure_tenant_id = "string"
azure_use_managed_identity = false
gcp_metadata_url = "string"
gcp_oauth_token_url = "string"
gcp_service_account_json = "string"
gcp_use_service_account = false
header_name = "string"
header_value = "string"
param_location = "string"
param_name = "string"
param_value = "string"
}
genai_category = "string"
llm_format = "string"
logging = {
log_payloads = false
log_statistics = false
}
max_request_body_size = 0
model_name_header = false
response_streaming = "string"
}
instance_name = "string"
ordering = {
after = {
accesses = ["string"]
}
before = {
accesses = ["string"]
}
}
consumer = {
id = "string"
}
created_at = 0
enabled = false
gateway_plugin_ai_proxy_id = "string"
condition = "string"
consumer_group = {
id = "string"
}
partials {
id = "string"
name = "string"
path = "string"
}
protocols = ["string"]
route = {
id = "string"
}
service = {
id = "string"
}
tags = ["string"]
updated_at = 0
}
var gatewayPluginAiProxyResource = new GatewayPluginAiProxy("gatewayPluginAiProxyResource", GatewayPluginAiProxyArgs.builder()
.controlPlaneId("string")
.config(GatewayPluginAiProxyConfigArgs.builder()
.model(GatewayPluginAiProxyConfigModelArgs.builder()
.provider("string")
.modelAlias("string")
.name("string")
.options(GatewayPluginAiProxyConfigModelOptionsArgs.builder()
.anthropicVersion("string")
.azureApiVersion("string")
.azureDeploymentId("string")
.azureInstance("string")
.bedrock(GatewayPluginAiProxyConfigModelOptionsBedrockArgs.builder()
.awsAssumeRoleArn("string")
.awsRegion("string")
.awsRoleSessionName("string")
.awsStsEndpointUrl("string")
.batchBucketPrefix("string")
.batchRoleArn("string")
.embeddingsNormalize(false)
.performanceConfigLatency("string")
.videoOutputS3Uri("string")
.build())
.cohere(GatewayPluginAiProxyConfigModelOptionsCohereArgs.builder()
.embeddingInputType("string")
.waitForModel(false)
.build())
.dashscope(GatewayPluginAiProxyConfigModelOptionsDashscopeArgs.builder()
.international(false)
.build())
.databricks(GatewayPluginAiProxyConfigModelOptionsDatabricksArgs.builder()
.workspaceInstanceId("string")
.build())
.embeddingsDimensions(0.0)
.gemini(GatewayPluginAiProxyConfigModelOptionsGeminiArgs.builder()
.apiEndpoint("string")
.endpointId("string")
.locationId("string")
.projectId("string")
.build())
.huggingface(GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs.builder()
.useCache(false)
.waitForModel(false)
.build())
.inputCost(0.0)
.llama2Format("string")
.maxTokens(0.0)
.mistralFormat("string")
.outputCost(0.0)
.temperature(0.0)
.topK(0.0)
.topP(0.0)
.upstreamPath("string")
.upstreamUrl("string")
.build())
.build())
.routeType("string")
.auth(GatewayPluginAiProxyConfigAuthArgs.builder()
.allowOverride(false)
.awsAccessKeyId("string")
.awsSecretAccessKey("string")
.azureClientId("string")
.azureClientSecret("string")
.azureTenantId("string")
.azureUseManagedIdentity(false)
.gcpMetadataUrl("string")
.gcpOauthTokenUrl("string")
.gcpServiceAccountJson("string")
.gcpUseServiceAccount(false)
.headerName("string")
.headerValue("string")
.paramLocation("string")
.paramName("string")
.paramValue("string")
.build())
.genaiCategory("string")
.llmFormat("string")
.logging(GatewayPluginAiProxyConfigLoggingArgs.builder()
.logPayloads(false)
.logStatistics(false)
.build())
.maxRequestBodySize(0.0)
.modelNameHeader(false)
.responseStreaming("string")
.build())
.instanceName("string")
.ordering(GatewayPluginAiProxyOrderingArgs.builder()
.after(GatewayPluginAiProxyOrderingAfterArgs.builder()
.accesses("string")
.build())
.before(GatewayPluginAiProxyOrderingBeforeArgs.builder()
.accesses("string")
.build())
.build())
.consumer(GatewayPluginAiProxyConsumerArgs.builder()
.id("string")
.build())
.createdAt(0.0)
.enabled(false)
.gatewayPluginAiProxyId("string")
.condition("string")
.consumerGroup(GatewayPluginAiProxyConsumerGroupArgs.builder()
.id("string")
.build())
.partials(GatewayPluginAiProxyPartialArgs.builder()
.id("string")
.name("string")
.path("string")
.build())
.protocols("string")
.route(GatewayPluginAiProxyRouteArgs.builder()
.id("string")
.build())
.service(GatewayPluginAiProxyServiceArgs.builder()
.id("string")
.build())
.tags("string")
.updatedAt(0.0)
.build());
gateway_plugin_ai_proxy_resource = konnect.GatewayPluginAiProxy("gatewayPluginAiProxyResource",
control_plane_id="string",
config={
"model": {
"provider": "string",
"model_alias": "string",
"name": "string",
"options": {
"anthropic_version": "string",
"azure_api_version": "string",
"azure_deployment_id": "string",
"azure_instance": "string",
"bedrock": {
"aws_assume_role_arn": "string",
"aws_region": "string",
"aws_role_session_name": "string",
"aws_sts_endpoint_url": "string",
"batch_bucket_prefix": "string",
"batch_role_arn": "string",
"embeddings_normalize": False,
"performance_config_latency": "string",
"video_output_s3_uri": "string",
},
"cohere": {
"embedding_input_type": "string",
"wait_for_model": False,
},
"dashscope": {
"international": False,
},
"databricks": {
"workspace_instance_id": "string",
},
"embeddings_dimensions": float(0),
"gemini": {
"api_endpoint": "string",
"endpoint_id": "string",
"location_id": "string",
"project_id": "string",
},
"huggingface": {
"use_cache": False,
"wait_for_model": False,
},
"input_cost": float(0),
"llama2_format": "string",
"max_tokens": float(0),
"mistral_format": "string",
"output_cost": float(0),
"temperature": float(0),
"top_k": float(0),
"top_p": float(0),
"upstream_path": "string",
"upstream_url": "string",
},
},
"route_type": "string",
"auth": {
"allow_override": False,
"aws_access_key_id": "string",
"aws_secret_access_key": "string",
"azure_client_id": "string",
"azure_client_secret": "string",
"azure_tenant_id": "string",
"azure_use_managed_identity": False,
"gcp_metadata_url": "string",
"gcp_oauth_token_url": "string",
"gcp_service_account_json": "string",
"gcp_use_service_account": False,
"header_name": "string",
"header_value": "string",
"param_location": "string",
"param_name": "string",
"param_value": "string",
},
"genai_category": "string",
"llm_format": "string",
"logging": {
"log_payloads": False,
"log_statistics": False,
},
"max_request_body_size": float(0),
"model_name_header": False,
"response_streaming": "string",
},
instance_name="string",
ordering={
"after": {
"accesses": ["string"],
},
"before": {
"accesses": ["string"],
},
},
consumer={
"id": "string",
},
created_at=float(0),
enabled=False,
gateway_plugin_ai_proxy_id="string",
condition="string",
consumer_group={
"id": "string",
},
partials=[{
"id": "string",
"name": "string",
"path": "string",
}],
protocols=["string"],
route={
"id": "string",
},
service={
"id": "string",
},
tags=["string"],
updated_at=float(0))
const gatewayPluginAiProxyResource = new konnect.GatewayPluginAiProxy("gatewayPluginAiProxyResource", {
controlPlaneId: "string",
config: {
model: {
provider: "string",
modelAlias: "string",
name: "string",
options: {
anthropicVersion: "string",
azureApiVersion: "string",
azureDeploymentId: "string",
azureInstance: "string",
bedrock: {
awsAssumeRoleArn: "string",
awsRegion: "string",
awsRoleSessionName: "string",
awsStsEndpointUrl: "string",
batchBucketPrefix: "string",
batchRoleArn: "string",
embeddingsNormalize: false,
performanceConfigLatency: "string",
videoOutputS3Uri: "string",
},
cohere: {
embeddingInputType: "string",
waitForModel: false,
},
dashscope: {
international: false,
},
databricks: {
workspaceInstanceId: "string",
},
embeddingsDimensions: 0,
gemini: {
apiEndpoint: "string",
endpointId: "string",
locationId: "string",
projectId: "string",
},
huggingface: {
useCache: false,
waitForModel: false,
},
inputCost: 0,
llama2Format: "string",
maxTokens: 0,
mistralFormat: "string",
outputCost: 0,
temperature: 0,
topK: 0,
topP: 0,
upstreamPath: "string",
upstreamUrl: "string",
},
},
routeType: "string",
auth: {
allowOverride: false,
awsAccessKeyId: "string",
awsSecretAccessKey: "string",
azureClientId: "string",
azureClientSecret: "string",
azureTenantId: "string",
azureUseManagedIdentity: false,
gcpMetadataUrl: "string",
gcpOauthTokenUrl: "string",
gcpServiceAccountJson: "string",
gcpUseServiceAccount: false,
headerName: "string",
headerValue: "string",
paramLocation: "string",
paramName: "string",
paramValue: "string",
},
genaiCategory: "string",
llmFormat: "string",
logging: {
logPayloads: false,
logStatistics: false,
},
maxRequestBodySize: 0,
modelNameHeader: false,
responseStreaming: "string",
},
instanceName: "string",
ordering: {
after: {
accesses: ["string"],
},
before: {
accesses: ["string"],
},
},
consumer: {
id: "string",
},
createdAt: 0,
enabled: false,
gatewayPluginAiProxyId: "string",
condition: "string",
consumerGroup: {
id: "string",
},
partials: [{
id: "string",
name: "string",
path: "string",
}],
protocols: ["string"],
route: {
id: "string",
},
service: {
id: "string",
},
tags: ["string"],
updatedAt: 0,
});
type: konnect:GatewayPluginAiProxy
properties:
condition: string
config:
auth:
allowOverride: false
awsAccessKeyId: string
awsSecretAccessKey: string
azureClientId: string
azureClientSecret: string
azureTenantId: string
azureUseManagedIdentity: false
gcpMetadataUrl: string
gcpOauthTokenUrl: string
gcpServiceAccountJson: string
gcpUseServiceAccount: false
headerName: string
headerValue: string
paramLocation: string
paramName: string
paramValue: string
genaiCategory: string
llmFormat: string
logging:
logPayloads: false
logStatistics: false
maxRequestBodySize: 0
model:
modelAlias: string
name: string
options:
anthropicVersion: string
azureApiVersion: string
azureDeploymentId: string
azureInstance: string
bedrock:
awsAssumeRoleArn: string
awsRegion: string
awsRoleSessionName: string
awsStsEndpointUrl: string
batchBucketPrefix: string
batchRoleArn: string
embeddingsNormalize: false
performanceConfigLatency: string
videoOutputS3Uri: string
cohere:
embeddingInputType: string
waitForModel: false
dashscope:
international: false
databricks:
workspaceInstanceId: string
embeddingsDimensions: 0
gemini:
apiEndpoint: string
endpointId: string
locationId: string
projectId: string
huggingface:
useCache: false
waitForModel: false
inputCost: 0
llama2Format: string
maxTokens: 0
mistralFormat: string
outputCost: 0
temperature: 0
topK: 0
topP: 0
upstreamPath: string
upstreamUrl: string
provider: string
modelNameHeader: false
responseStreaming: string
routeType: string
consumer:
id: string
consumerGroup:
id: string
controlPlaneId: string
createdAt: 0
enabled: false
gatewayPluginAiProxyId: string
instanceName: string
ordering:
after:
accesses:
- string
before:
accesses:
- string
partials:
- id: string
name: string
path: string
protocols:
- string
route:
id: string
service:
id: string
tags:
- string
updatedAt: 0
GatewayPluginAiProxy Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The GatewayPluginAiProxy resource accepts the following input properties:
- Config
Gateway
Plugin Ai Proxy Config - Control
Plane stringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- Condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - Consumer
Gateway
Plugin Ai Proxy Consumer - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- Consumer
Group GatewayPlugin Ai Proxy Consumer Group - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- Created
At double - Unix epoch when the resource was created.
- Enabled bool
- Whether the plugin is applied. Default: true
- Gateway
Plugin stringAi Proxy Id - A string representing a UUID (universally unique identifier).
- Instance
Name string - A unique string representing a UTF-8 encoded name.
- Ordering
Gateway
Plugin Ai Proxy Ordering - Partials
List<Gateway
Plugin Ai Proxy Partial> - A list of partials to be used by the plugin.
- Protocols List<string>
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- Route
Gateway
Plugin Ai Proxy Route - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- Service
Gateway
Plugin Ai Proxy Service - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- List<string>
- An optional set of strings associated with the Plugin for grouping and filtering.
- Updated
At double - Unix epoch when the resource was last updated.
- Config
Gateway
Plugin Ai Proxy Config Args - Control
Plane stringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- Condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - Consumer
Gateway
Plugin Ai Proxy Consumer Args - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- Consumer
Group GatewayPlugin Ai Proxy Consumer Group Args - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- Created
At float64 - Unix epoch when the resource was created.
- Enabled bool
- Whether the plugin is applied. Default: true
- Gateway
Plugin stringAi Proxy Id - A string representing a UUID (universally unique identifier).
- Instance
Name string - A unique string representing a UTF-8 encoded name.
- Ordering
Gateway
Plugin Ai Proxy Ordering Args - Partials
[]Gateway
Plugin Ai Proxy Partial Args - A list of partials to be used by the plugin.
- Protocols []string
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- Route
Gateway
Plugin Ai Proxy Route Args - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- Service
Gateway
Plugin Ai Proxy Service Args - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- []string
- An optional set of strings associated with the Plugin for grouping and filtering.
- Updated
At float64 - Unix epoch when the resource was last updated.
- config object
- control_
plane_ stringid - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - consumer object
- If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer_
group object - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- created_
at number - Unix epoch when the resource was created.
- enabled bool
- Whether the plugin is applied. Default: true
- gateway_
plugin_ stringai_ proxy_ id - A string representing a UUID (universally unique identifier).
- instance_
name string - A unique string representing a UTF-8 encoded name.
- ordering object
- partials list(object)
- A list of partials to be used by the plugin.
- protocols list(string)
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route object
- If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service object
- If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- list(string)
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated_
at number - Unix epoch when the resource was last updated.
- config
Gateway
Plugin Ai Proxy Config - control
Plane StringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- condition String
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - consumer
Gateway
Plugin Ai Proxy Consumer - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer
Group GatewayPlugin Ai Proxy Consumer Group - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- created
At Double - Unix epoch when the resource was created.
- enabled Boolean
- Whether the plugin is applied. Default: true
- gateway
Plugin StringAi Proxy Id - A string representing a UUID (universally unique identifier).
- instance
Name String - A unique string representing a UTF-8 encoded name.
- ordering
Gateway
Plugin Ai Proxy Ordering - partials
List<Gateway
Plugin Ai Proxy Partial> - A list of partials to be used by the plugin.
- protocols List<String>
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route
Gateway
Plugin Ai Proxy Route - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service
Gateway
Plugin Ai Proxy Service - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- List<String>
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated
At Double - Unix epoch when the resource was last updated.
- config
Gateway
Plugin Ai Proxy Config - control
Plane stringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - consumer
Gateway
Plugin Ai Proxy Consumer - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer
Group GatewayPlugin Ai Proxy Consumer Group - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- created
At number - Unix epoch when the resource was created.
- enabled boolean
- Whether the plugin is applied. Default: true
- gateway
Plugin stringAi Proxy Id - A string representing a UUID (universally unique identifier).
- instance
Name string - A unique string representing a UTF-8 encoded name.
- ordering
Gateway
Plugin Ai Proxy Ordering - partials
Gateway
Plugin Ai Proxy Partial[] - A list of partials to be used by the plugin.
- protocols string[]
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route
Gateway
Plugin Ai Proxy Route - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service
Gateway
Plugin Ai Proxy Service - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- string[]
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated
At number - Unix epoch when the resource was last updated.
- config
Gateway
Plugin Ai Proxy Config Args - control_
plane_ strid - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- condition str
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - consumer
Gateway
Plugin Ai Proxy Consumer Args - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer_
group GatewayPlugin Ai Proxy Consumer Group Args - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- created_
at float - Unix epoch when the resource was created.
- enabled bool
- Whether the plugin is applied. Default: true
- gateway_
plugin_ strai_ proxy_ id - A string representing a UUID (universally unique identifier).
- instance_
name str - A unique string representing a UTF-8 encoded name.
- ordering
Gateway
Plugin Ai Proxy Ordering Args - partials
Sequence[Gateway
Plugin Ai Proxy Partial Args] - A list of partials to be used by the plugin.
- protocols Sequence[str]
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route
Gateway
Plugin Ai Proxy Route Args - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service
Gateway
Plugin Ai Proxy Service Args - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- Sequence[str]
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated_
at float - Unix epoch when the resource was last updated.
- config Property Map
- control
Plane StringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- condition String
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - consumer Property Map
- If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer
Group Property Map - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- created
At Number - Unix epoch when the resource was created.
- enabled Boolean
- Whether the plugin is applied. Default: true
- gateway
Plugin StringAi Proxy Id - A string representing a UUID (universally unique identifier).
- instance
Name String - A unique string representing a UTF-8 encoded name.
- ordering Property Map
- partials List<Property Map>
- A list of partials to be used by the plugin.
- protocols List<String>
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route Property Map
- If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service Property Map
- If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- List<String>
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated
At Number - Unix epoch when the resource was last updated.
Outputs
All input properties are implicitly available as output properties. Additionally, the GatewayPluginAiProxy resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Id string
- The provider-assigned unique ID for this managed resource.
- id string
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
- id string
- The provider-assigned unique ID for this managed resource.
- id str
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
Look up Existing GatewayPluginAiProxy Resource
Get an existing GatewayPluginAiProxy resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: GatewayPluginAiProxyState, opts?: CustomResourceOptions): GatewayPluginAiProxy@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
condition: Optional[str] = None,
config: Optional[GatewayPluginAiProxyConfigArgs] = None,
consumer: Optional[GatewayPluginAiProxyConsumerArgs] = None,
consumer_group: Optional[GatewayPluginAiProxyConsumerGroupArgs] = None,
control_plane_id: Optional[str] = None,
created_at: Optional[float] = None,
enabled: Optional[bool] = None,
gateway_plugin_ai_proxy_id: Optional[str] = None,
instance_name: Optional[str] = None,
ordering: Optional[GatewayPluginAiProxyOrderingArgs] = None,
partials: Optional[Sequence[GatewayPluginAiProxyPartialArgs]] = None,
protocols: Optional[Sequence[str]] = None,
route: Optional[GatewayPluginAiProxyRouteArgs] = None,
service: Optional[GatewayPluginAiProxyServiceArgs] = None,
tags: Optional[Sequence[str]] = None,
updated_at: Optional[float] = None) -> GatewayPluginAiProxyfunc GetGatewayPluginAiProxy(ctx *Context, name string, id IDInput, state *GatewayPluginAiProxyState, opts ...ResourceOption) (*GatewayPluginAiProxy, error)public static GatewayPluginAiProxy Get(string name, Input<string> id, GatewayPluginAiProxyState? state, CustomResourceOptions? opts = null)public static GatewayPluginAiProxy get(String name, Output<String> id, GatewayPluginAiProxyState state, CustomResourceOptions options)resources: _: type: konnect:GatewayPluginAiProxy get: id: ${id}import {
to = konnect_gatewaypluginaiproxy.example
id = "${id}"
}
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - Config
Gateway
Plugin Ai Proxy Config - Consumer
Gateway
Plugin Ai Proxy Consumer - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- Consumer
Group GatewayPlugin Ai Proxy Consumer Group - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- Control
Plane stringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- Created
At double - Unix epoch when the resource was created.
- Enabled bool
- Whether the plugin is applied. Default: true
- Gateway
Plugin stringAi Proxy Id - A string representing a UUID (universally unique identifier).
- Instance
Name string - A unique string representing a UTF-8 encoded name.
- Ordering
Gateway
Plugin Ai Proxy Ordering - Partials
List<Gateway
Plugin Ai Proxy Partial> - A list of partials to be used by the plugin.
- Protocols List<string>
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- Route
Gateway
Plugin Ai Proxy Route - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- Service
Gateway
Plugin Ai Proxy Service - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- List<string>
- An optional set of strings associated with the Plugin for grouping and filtering.
- Updated
At double - Unix epoch when the resource was last updated.
- Condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - Config
Gateway
Plugin Ai Proxy Config Args - Consumer
Gateway
Plugin Ai Proxy Consumer Args - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- Consumer
Group GatewayPlugin Ai Proxy Consumer Group Args - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- Control
Plane stringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- Created
At float64 - Unix epoch when the resource was created.
- Enabled bool
- Whether the plugin is applied. Default: true
- Gateway
Plugin stringAi Proxy Id - A string representing a UUID (universally unique identifier).
- Instance
Name string - A unique string representing a UTF-8 encoded name.
- Ordering
Gateway
Plugin Ai Proxy Ordering Args - Partials
[]Gateway
Plugin Ai Proxy Partial Args - A list of partials to be used by the plugin.
- Protocols []string
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- Route
Gateway
Plugin Ai Proxy Route Args - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- Service
Gateway
Plugin Ai Proxy Service Args - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- []string
- An optional set of strings associated with the Plugin for grouping and filtering.
- Updated
At float64 - Unix epoch when the resource was last updated.
- condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - config object
- consumer object
- If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer_
group object - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- control_
plane_ stringid - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- created_
at number - Unix epoch when the resource was created.
- enabled bool
- Whether the plugin is applied. Default: true
- gateway_
plugin_ stringai_ proxy_ id - A string representing a UUID (universally unique identifier).
- instance_
name string - A unique string representing a UTF-8 encoded name.
- ordering object
- partials list(object)
- A list of partials to be used by the plugin.
- protocols list(string)
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route object
- If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service object
- If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- list(string)
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated_
at number - Unix epoch when the resource was last updated.
- condition String
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - config
Gateway
Plugin Ai Proxy Config - consumer
Gateway
Plugin Ai Proxy Consumer - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer
Group GatewayPlugin Ai Proxy Consumer Group - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- control
Plane StringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- created
At Double - Unix epoch when the resource was created.
- enabled Boolean
- Whether the plugin is applied. Default: true
- gateway
Plugin StringAi Proxy Id - A string representing a UUID (universally unique identifier).
- instance
Name String - A unique string representing a UTF-8 encoded name.
- ordering
Gateway
Plugin Ai Proxy Ordering - partials
List<Gateway
Plugin Ai Proxy Partial> - A list of partials to be used by the plugin.
- protocols List<String>
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route
Gateway
Plugin Ai Proxy Route - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service
Gateway
Plugin Ai Proxy Service - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- List<String>
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated
At Double - Unix epoch when the resource was last updated.
- condition string
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - config
Gateway
Plugin Ai Proxy Config - consumer
Gateway
Plugin Ai Proxy Consumer - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer
Group GatewayPlugin Ai Proxy Consumer Group - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- control
Plane stringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- created
At number - Unix epoch when the resource was created.
- enabled boolean
- Whether the plugin is applied. Default: true
- gateway
Plugin stringAi Proxy Id - A string representing a UUID (universally unique identifier).
- instance
Name string - A unique string representing a UTF-8 encoded name.
- ordering
Gateway
Plugin Ai Proxy Ordering - partials
Gateway
Plugin Ai Proxy Partial[] - A list of partials to be used by the plugin.
- protocols string[]
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route
Gateway
Plugin Ai Proxy Route - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service
Gateway
Plugin Ai Proxy Service - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- string[]
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated
At number - Unix epoch when the resource was last updated.
- condition str
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - config
Gateway
Plugin Ai Proxy Config Args - consumer
Gateway
Plugin Ai Proxy Consumer Args - If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer_
group GatewayPlugin Ai Proxy Consumer Group Args - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- control_
plane_ strid - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- created_
at float - Unix epoch when the resource was created.
- enabled bool
- Whether the plugin is applied. Default: true
- gateway_
plugin_ strai_ proxy_ id - A string representing a UUID (universally unique identifier).
- instance_
name str - A unique string representing a UTF-8 encoded name.
- ordering
Gateway
Plugin Ai Proxy Ordering Args - partials
Sequence[Gateway
Plugin Ai Proxy Partial Args] - A list of partials to be used by the plugin.
- protocols Sequence[str]
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route
Gateway
Plugin Ai Proxy Route Args - If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service
Gateway
Plugin Ai Proxy Service Args - If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- Sequence[str]
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated_
at float - Unix epoch when the resource was last updated.
- condition String
- An expression used for conditional control over plugin execution. If the expression evaluates to
trueduring the request flow, the plugin is executed; otherwise, it is skipped. - config Property Map
- consumer Property Map
- If set, the plugin will activate only for requests where the specified has been authenticated. (Note that some plugins can not be restricted to consumers this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer.
- consumer
Group Property Map - If set, the plugin will activate only for requests where the specified consumer group has been authenticated. (Note that some plugins can not be restricted to consumers groups this way.). Leave unset for the plugin to activate regardless of the authenticated Consumer Groups
- control
Plane StringId - The UUID of your control plane. This variable is available in the Konnect manager. Requires replacement if changed.
- created
At Number - Unix epoch when the resource was created.
- enabled Boolean
- Whether the plugin is applied. Default: true
- gateway
Plugin StringAi Proxy Id - A string representing a UUID (universally unique identifier).
- instance
Name String - A unique string representing a UTF-8 encoded name.
- ordering Property Map
- partials List<Property Map>
- A list of partials to be used by the plugin.
- protocols List<String>
- A list of the request protocols that will trigger this plugin. The default value, as well as the possible values allowed on this field, may change depending on the plugin type. For example, plugins that only work in stream mode will only support tcp and tls. Default: ["grpc","grpcs","http","https","ws","wss"]
- route Property Map
- If set, the plugin will only activate when receiving requests via the specified route. Leave unset for the plugin to activate regardless of the route being used.
- service Property Map
- If set, the plugin will only activate when receiving requests via one of the routes belonging to the specified Service. Leave unset for the plugin to activate regardless of the Service being matched.
- List<String>
- An optional set of strings associated with the Plugin for grouping and filtering.
- updated
At Number - Unix epoch when the resource was last updated.
Supporting Types
GatewayPluginAiProxyConfig, GatewayPluginAiProxyConfigArgs
- Model
Gateway
Plugin Ai Proxy Config Model - Route
Type string - The model's operation implementation, for this provider. possible known values include one of ["audio/v1/audio/speech", "audio/v1/audio/transcriptions", "audio/v1/audio/translations", "image/v1/images/edits", "image/v1/images/generations", "llm/v1/assistants", "llm/v1/batches", "llm/v1/chat", "llm/v1/completions", "llm/v1/embeddings", "llm/v1/files", "llm/v1/responses", "preserve", "realtime/v1/realtime", "video/v1/videos/generations"]
- Auth
Gateway
Plugin Ai Proxy Config Auth - Genai
Category string - Generative AI category of the request. possible known values include one of ["audio/speech", "audio/transcription", "image/generation", "text/embeddings", "text/generation", "video/generation"]; Default: "text/generation"
- Llm
Format string - LLM input and output format and schema to use. possible known values include one of ["anthropic", "bedrock", "cohere", "gemini", "huggingface", "openai"]; Default: "openai"
- Logging
Gateway
Plugin Ai Proxy Config Logging - Max
Request doubleBody Size - max allowed body size allowed to be introspected. 0 means unlimited, but the size of this body will still be limited by Nginx's clientmaxbody_size. Default: 1048576
- Model
Name boolHeader - Display the model name selected in the X-Kong-LLM-Model response header. Default: true
- Response
Streaming string - Whether to 'optionally allow', 'deny', or 'always' (force) the streaming of answers via server sent events. possible known values include one of ["allow", "always", "deny"]; Default: "allow"
- Model
Gateway
Plugin Ai Proxy Config Model - Route
Type string - The model's operation implementation, for this provider. possible known values include one of ["audio/v1/audio/speech", "audio/v1/audio/transcriptions", "audio/v1/audio/translations", "image/v1/images/edits", "image/v1/images/generations", "llm/v1/assistants", "llm/v1/batches", "llm/v1/chat", "llm/v1/completions", "llm/v1/embeddings", "llm/v1/files", "llm/v1/responses", "preserve", "realtime/v1/realtime", "video/v1/videos/generations"]
- Auth
Gateway
Plugin Ai Proxy Config Auth - Genai
Category string - Generative AI category of the request. possible known values include one of ["audio/speech", "audio/transcription", "image/generation", "text/embeddings", "text/generation", "video/generation"]; Default: "text/generation"
- Llm
Format string - LLM input and output format and schema to use. possible known values include one of ["anthropic", "bedrock", "cohere", "gemini", "huggingface", "openai"]; Default: "openai"
- Logging
Gateway
Plugin Ai Proxy Config Logging - Max
Request float64Body Size - max allowed body size allowed to be introspected. 0 means unlimited, but the size of this body will still be limited by Nginx's clientmaxbody_size. Default: 1048576
- Model
Name boolHeader - Display the model name selected in the X-Kong-LLM-Model response header. Default: true
- Response
Streaming string - Whether to 'optionally allow', 'deny', or 'always' (force) the streaming of answers via server sent events. possible known values include one of ["allow", "always", "deny"]; Default: "allow"
- model object
- route_
type string - The model's operation implementation, for this provider. possible known values include one of ["audio/v1/audio/speech", "audio/v1/audio/transcriptions", "audio/v1/audio/translations", "image/v1/images/edits", "image/v1/images/generations", "llm/v1/assistants", "llm/v1/batches", "llm/v1/chat", "llm/v1/completions", "llm/v1/embeddings", "llm/v1/files", "llm/v1/responses", "preserve", "realtime/v1/realtime", "video/v1/videos/generations"]
- auth object
- genai_
category string - Generative AI category of the request. possible known values include one of ["audio/speech", "audio/transcription", "image/generation", "text/embeddings", "text/generation", "video/generation"]; Default: "text/generation"
- llm_
format string - LLM input and output format and schema to use. possible known values include one of ["anthropic", "bedrock", "cohere", "gemini", "huggingface", "openai"]; Default: "openai"
- logging object
- max_
request_ numberbody_ size - max allowed body size allowed to be introspected. 0 means unlimited, but the size of this body will still be limited by Nginx's clientmaxbody_size. Default: 1048576
- model_
name_ boolheader - Display the model name selected in the X-Kong-LLM-Model response header. Default: true
- response_
streaming string - Whether to 'optionally allow', 'deny', or 'always' (force) the streaming of answers via server sent events. possible known values include one of ["allow", "always", "deny"]; Default: "allow"
- model
Gateway
Plugin Ai Proxy Config Model - route
Type String - The model's operation implementation, for this provider. possible known values include one of ["audio/v1/audio/speech", "audio/v1/audio/transcriptions", "audio/v1/audio/translations", "image/v1/images/edits", "image/v1/images/generations", "llm/v1/assistants", "llm/v1/batches", "llm/v1/chat", "llm/v1/completions", "llm/v1/embeddings", "llm/v1/files", "llm/v1/responses", "preserve", "realtime/v1/realtime", "video/v1/videos/generations"]
- auth
Gateway
Plugin Ai Proxy Config Auth - genai
Category String - Generative AI category of the request. possible known values include one of ["audio/speech", "audio/transcription", "image/generation", "text/embeddings", "text/generation", "video/generation"]; Default: "text/generation"
- llm
Format String - LLM input and output format and schema to use. possible known values include one of ["anthropic", "bedrock", "cohere", "gemini", "huggingface", "openai"]; Default: "openai"
- logging
Gateway
Plugin Ai Proxy Config Logging - max
Request DoubleBody Size - max allowed body size allowed to be introspected. 0 means unlimited, but the size of this body will still be limited by Nginx's clientmaxbody_size. Default: 1048576
- model
Name BooleanHeader - Display the model name selected in the X-Kong-LLM-Model response header. Default: true
- response
Streaming String - Whether to 'optionally allow', 'deny', or 'always' (force) the streaming of answers via server sent events. possible known values include one of ["allow", "always", "deny"]; Default: "allow"
- model
Gateway
Plugin Ai Proxy Config Model - route
Type string - The model's operation implementation, for this provider. possible known values include one of ["audio/v1/audio/speech", "audio/v1/audio/transcriptions", "audio/v1/audio/translations", "image/v1/images/edits", "image/v1/images/generations", "llm/v1/assistants", "llm/v1/batches", "llm/v1/chat", "llm/v1/completions", "llm/v1/embeddings", "llm/v1/files", "llm/v1/responses", "preserve", "realtime/v1/realtime", "video/v1/videos/generations"]
- auth
Gateway
Plugin Ai Proxy Config Auth - genai
Category string - Generative AI category of the request. possible known values include one of ["audio/speech", "audio/transcription", "image/generation", "text/embeddings", "text/generation", "video/generation"]; Default: "text/generation"
- llm
Format string - LLM input and output format and schema to use. possible known values include one of ["anthropic", "bedrock", "cohere", "gemini", "huggingface", "openai"]; Default: "openai"
- logging
Gateway
Plugin Ai Proxy Config Logging - max
Request numberBody Size - max allowed body size allowed to be introspected. 0 means unlimited, but the size of this body will still be limited by Nginx's clientmaxbody_size. Default: 1048576
- model
Name booleanHeader - Display the model name selected in the X-Kong-LLM-Model response header. Default: true
- response
Streaming string - Whether to 'optionally allow', 'deny', or 'always' (force) the streaming of answers via server sent events. possible known values include one of ["allow", "always", "deny"]; Default: "allow"
- model
Gateway
Plugin Ai Proxy Config Model - route_
type str - The model's operation implementation, for this provider. possible known values include one of ["audio/v1/audio/speech", "audio/v1/audio/transcriptions", "audio/v1/audio/translations", "image/v1/images/edits", "image/v1/images/generations", "llm/v1/assistants", "llm/v1/batches", "llm/v1/chat", "llm/v1/completions", "llm/v1/embeddings", "llm/v1/files", "llm/v1/responses", "preserve", "realtime/v1/realtime", "video/v1/videos/generations"]
- auth
Gateway
Plugin Ai Proxy Config Auth - genai_
category str - Generative AI category of the request. possible known values include one of ["audio/speech", "audio/transcription", "image/generation", "text/embeddings", "text/generation", "video/generation"]; Default: "text/generation"
- llm_
format str - LLM input and output format and schema to use. possible known values include one of ["anthropic", "bedrock", "cohere", "gemini", "huggingface", "openai"]; Default: "openai"
- logging
Gateway
Plugin Ai Proxy Config Logging - max_
request_ floatbody_ size - max allowed body size allowed to be introspected. 0 means unlimited, but the size of this body will still be limited by Nginx's clientmaxbody_size. Default: 1048576
- model_
name_ boolheader - Display the model name selected in the X-Kong-LLM-Model response header. Default: true
- response_
streaming str - Whether to 'optionally allow', 'deny', or 'always' (force) the streaming of answers via server sent events. possible known values include one of ["allow", "always", "deny"]; Default: "allow"
- model Property Map
- route
Type String - The model's operation implementation, for this provider. possible known values include one of ["audio/v1/audio/speech", "audio/v1/audio/transcriptions", "audio/v1/audio/translations", "image/v1/images/edits", "image/v1/images/generations", "llm/v1/assistants", "llm/v1/batches", "llm/v1/chat", "llm/v1/completions", "llm/v1/embeddings", "llm/v1/files", "llm/v1/responses", "preserve", "realtime/v1/realtime", "video/v1/videos/generations"]
- auth Property Map
- genai
Category String - Generative AI category of the request. possible known values include one of ["audio/speech", "audio/transcription", "image/generation", "text/embeddings", "text/generation", "video/generation"]; Default: "text/generation"
- llm
Format String - LLM input and output format and schema to use. possible known values include one of ["anthropic", "bedrock", "cohere", "gemini", "huggingface", "openai"]; Default: "openai"
- logging Property Map
- max
Request NumberBody Size - max allowed body size allowed to be introspected. 0 means unlimited, but the size of this body will still be limited by Nginx's clientmaxbody_size. Default: 1048576
- model
Name BooleanHeader - Display the model name selected in the X-Kong-LLM-Model response header. Default: true
- response
Streaming String - Whether to 'optionally allow', 'deny', or 'always' (force) the streaming of answers via server sent events. possible known values include one of ["allow", "always", "deny"]; Default: "allow"
GatewayPluginAiProxyConfigAuth, GatewayPluginAiProxyConfigAuthArgs
- Allow
Override bool - If enabled, the authorization header or parameter can be overridden in the request by the value configured in the plugin. Default: false
- Aws
Access stringKey Id - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSACCESSKEY_ID environment variable for this plugin instance.
- Aws
Secret stringAccess Key - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSSECRETACCESS_KEY environment variable for this plugin instance.
- Azure
Client stringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client ID.
- Azure
Client stringSecret - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client secret.
- Azure
Tenant stringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the tenant ID.
- Azure
Use boolManaged Identity - Set true to use the Azure Cloud Managed Identity (or user-assigned identity) to authenticate with Azure-provider models. Default: false
- Gcp
Metadata stringUrl - Custom metadata URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google metadata endpoint.
- Gcp
Oauth stringToken Url - Custom OAuth token URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google OAuth token endpoint.
- Gcp
Service stringAccount Json - Set this field to the full JSON of the GCP service account to authenticate, if required. If null (and gcpuseservice_account is true), Kong will attempt to read from environment variable
GCP_SERVICE_ACCOUNT. - Gcp
Use boolService Account - Use service account auth for GCP-based providers and models. Default: false
- Header
Name string - If AI model requires authentication via Authorization or API key header, specify its name here.
- Header
Value string - Specify the full auth header value for 'header_name', for example 'Bearer key' or just 'key'.
- Param
Location string - Specify whether the 'paramname' and 'paramvalue' options go in a query string, or the POST form/JSON body. possible known values include one of ["body", "query"]
- Param
Name string - If AI model requires authentication via query parameter, specify its name here.
- Param
Value string - Specify the full parameter value for 'param_name'.
- Allow
Override bool - If enabled, the authorization header or parameter can be overridden in the request by the value configured in the plugin. Default: false
- Aws
Access stringKey Id - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSACCESSKEY_ID environment variable for this plugin instance.
- Aws
Secret stringAccess Key - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSSECRETACCESS_KEY environment variable for this plugin instance.
- Azure
Client stringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client ID.
- Azure
Client stringSecret - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client secret.
- Azure
Tenant stringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the tenant ID.
- Azure
Use boolManaged Identity - Set true to use the Azure Cloud Managed Identity (or user-assigned identity) to authenticate with Azure-provider models. Default: false
- Gcp
Metadata stringUrl - Custom metadata URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google metadata endpoint.
- Gcp
Oauth stringToken Url - Custom OAuth token URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google OAuth token endpoint.
- Gcp
Service stringAccount Json - Set this field to the full JSON of the GCP service account to authenticate, if required. If null (and gcpuseservice_account is true), Kong will attempt to read from environment variable
GCP_SERVICE_ACCOUNT. - Gcp
Use boolService Account - Use service account auth for GCP-based providers and models. Default: false
- Header
Name string - If AI model requires authentication via Authorization or API key header, specify its name here.
- Header
Value string - Specify the full auth header value for 'header_name', for example 'Bearer key' or just 'key'.
- Param
Location string - Specify whether the 'paramname' and 'paramvalue' options go in a query string, or the POST form/JSON body. possible known values include one of ["body", "query"]
- Param
Name string - If AI model requires authentication via query parameter, specify its name here.
- Param
Value string - Specify the full parameter value for 'param_name'.
- allow_
override bool - If enabled, the authorization header or parameter can be overridden in the request by the value configured in the plugin. Default: false
- aws_
access_ stringkey_ id - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSACCESSKEY_ID environment variable for this plugin instance.
- aws_
secret_ stringaccess_ key - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSSECRETACCESS_KEY environment variable for this plugin instance.
- azure_
client_ stringid - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client ID.
- azure_
client_ stringsecret - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client secret.
- azure_
tenant_ stringid - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the tenant ID.
- azure_
use_ boolmanaged_ identity - Set true to use the Azure Cloud Managed Identity (or user-assigned identity) to authenticate with Azure-provider models. Default: false
- gcp_
metadata_ stringurl - Custom metadata URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google metadata endpoint.
- gcp_
oauth_ stringtoken_ url - Custom OAuth token URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google OAuth token endpoint.
- gcp_
service_ stringaccount_ json - Set this field to the full JSON of the GCP service account to authenticate, if required. If null (and gcpuseservice_account is true), Kong will attempt to read from environment variable
GCP_SERVICE_ACCOUNT. - gcp_
use_ boolservice_ account - Use service account auth for GCP-based providers and models. Default: false
- header_
name string - If AI model requires authentication via Authorization or API key header, specify its name here.
- header_
value string - Specify the full auth header value for 'header_name', for example 'Bearer key' or just 'key'.
- param_
location string - Specify whether the 'paramname' and 'paramvalue' options go in a query string, or the POST form/JSON body. possible known values include one of ["body", "query"]
- param_
name string - If AI model requires authentication via query parameter, specify its name here.
- param_
value string - Specify the full parameter value for 'param_name'.
- allow
Override Boolean - If enabled, the authorization header or parameter can be overridden in the request by the value configured in the plugin. Default: false
- aws
Access StringKey Id - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSACCESSKEY_ID environment variable for this plugin instance.
- aws
Secret StringAccess Key - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSSECRETACCESS_KEY environment variable for this plugin instance.
- azure
Client StringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client ID.
- azure
Client StringSecret - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client secret.
- azure
Tenant StringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the tenant ID.
- azure
Use BooleanManaged Identity - Set true to use the Azure Cloud Managed Identity (or user-assigned identity) to authenticate with Azure-provider models. Default: false
- gcp
Metadata StringUrl - Custom metadata URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google metadata endpoint.
- gcp
Oauth StringToken Url - Custom OAuth token URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google OAuth token endpoint.
- gcp
Service StringAccount Json - Set this field to the full JSON of the GCP service account to authenticate, if required. If null (and gcpuseservice_account is true), Kong will attempt to read from environment variable
GCP_SERVICE_ACCOUNT. - gcp
Use BooleanService Account - Use service account auth for GCP-based providers and models. Default: false
- header
Name String - If AI model requires authentication via Authorization or API key header, specify its name here.
- header
Value String - Specify the full auth header value for 'header_name', for example 'Bearer key' or just 'key'.
- param
Location String - Specify whether the 'paramname' and 'paramvalue' options go in a query string, or the POST form/JSON body. possible known values include one of ["body", "query"]
- param
Name String - If AI model requires authentication via query parameter, specify its name here.
- param
Value String - Specify the full parameter value for 'param_name'.
- allow
Override boolean - If enabled, the authorization header or parameter can be overridden in the request by the value configured in the plugin. Default: false
- aws
Access stringKey Id - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSACCESSKEY_ID environment variable for this plugin instance.
- aws
Secret stringAccess Key - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSSECRETACCESS_KEY environment variable for this plugin instance.
- azure
Client stringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client ID.
- azure
Client stringSecret - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client secret.
- azure
Tenant stringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the tenant ID.
- azure
Use booleanManaged Identity - Set true to use the Azure Cloud Managed Identity (or user-assigned identity) to authenticate with Azure-provider models. Default: false
- gcp
Metadata stringUrl - Custom metadata URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google metadata endpoint.
- gcp
Oauth stringToken Url - Custom OAuth token URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google OAuth token endpoint.
- gcp
Service stringAccount Json - Set this field to the full JSON of the GCP service account to authenticate, if required. If null (and gcpuseservice_account is true), Kong will attempt to read from environment variable
GCP_SERVICE_ACCOUNT. - gcp
Use booleanService Account - Use service account auth for GCP-based providers and models. Default: false
- header
Name string - If AI model requires authentication via Authorization or API key header, specify its name here.
- header
Value string - Specify the full auth header value for 'header_name', for example 'Bearer key' or just 'key'.
- param
Location string - Specify whether the 'paramname' and 'paramvalue' options go in a query string, or the POST form/JSON body. possible known values include one of ["body", "query"]
- param
Name string - If AI model requires authentication via query parameter, specify its name here.
- param
Value string - Specify the full parameter value for 'param_name'.
- allow_
override bool - If enabled, the authorization header or parameter can be overridden in the request by the value configured in the plugin. Default: false
- aws_
access_ strkey_ id - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSACCESSKEY_ID environment variable for this plugin instance.
- aws_
secret_ straccess_ key - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSSECRETACCESS_KEY environment variable for this plugin instance.
- azure_
client_ strid - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client ID.
- azure_
client_ strsecret - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client secret.
- azure_
tenant_ strid - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the tenant ID.
- azure_
use_ boolmanaged_ identity - Set true to use the Azure Cloud Managed Identity (or user-assigned identity) to authenticate with Azure-provider models. Default: false
- gcp_
metadata_ strurl - Custom metadata URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google metadata endpoint.
- gcp_
oauth_ strtoken_ url - Custom OAuth token URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google OAuth token endpoint.
- gcp_
service_ straccount_ json - Set this field to the full JSON of the GCP service account to authenticate, if required. If null (and gcpuseservice_account is true), Kong will attempt to read from environment variable
GCP_SERVICE_ACCOUNT. - gcp_
use_ boolservice_ account - Use service account auth for GCP-based providers and models. Default: false
- header_
name str - If AI model requires authentication via Authorization or API key header, specify its name here.
- header_
value str - Specify the full auth header value for 'header_name', for example 'Bearer key' or just 'key'.
- param_
location str - Specify whether the 'paramname' and 'paramvalue' options go in a query string, or the POST form/JSON body. possible known values include one of ["body", "query"]
- param_
name str - If AI model requires authentication via query parameter, specify its name here.
- param_
value str - Specify the full parameter value for 'param_name'.
- allow
Override Boolean - If enabled, the authorization header or parameter can be overridden in the request by the value configured in the plugin. Default: false
- aws
Access StringKey Id - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSACCESSKEY_ID environment variable for this plugin instance.
- aws
Secret StringAccess Key - Set this if you are using an AWS provider (Bedrock) and you are authenticating using static IAM User credentials. Setting this will override the AWSSECRETACCESS_KEY environment variable for this plugin instance.
- azure
Client StringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client ID.
- azure
Client StringSecret - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the client secret.
- azure
Tenant StringId - If azureusemanaged_identity is set to true, and you need to use a different user-assigned identity for this LLM instance, set the tenant ID.
- azure
Use BooleanManaged Identity - Set true to use the Azure Cloud Managed Identity (or user-assigned identity) to authenticate with Azure-provider models. Default: false
- gcp
Metadata StringUrl - Custom metadata URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google metadata endpoint.
- gcp
Oauth StringToken Url - Custom OAuth token URL for GCP authentication. Useful for restricted network environments or custom GCP endpoints. If null, Kong will use the default Google OAuth token endpoint.
- gcp
Service StringAccount Json - Set this field to the full JSON of the GCP service account to authenticate, if required. If null (and gcpuseservice_account is true), Kong will attempt to read from environment variable
GCP_SERVICE_ACCOUNT. - gcp
Use BooleanService Account - Use service account auth for GCP-based providers and models. Default: false
- header
Name String - If AI model requires authentication via Authorization or API key header, specify its name here.
- header
Value String - Specify the full auth header value for 'header_name', for example 'Bearer key' or just 'key'.
- param
Location String - Specify whether the 'paramname' and 'paramvalue' options go in a query string, or the POST form/JSON body. possible known values include one of ["body", "query"]
- param
Name String - If AI model requires authentication via query parameter, specify its name here.
- param
Value String - Specify the full parameter value for 'param_name'.
GatewayPluginAiProxyConfigLogging, GatewayPluginAiProxyConfigLoggingArgs
- Log
Payloads bool - If enabled, will log the request and response body into the Kong log plugin(s) output.Furthermore if Opentelemetry instrumentation is enabled the traces will contain this data as well. Default: false
- Log
Statistics bool - If enabled and supported by the driver, will add model usage and token metrics into the Kong log plugin(s) output. Default: false
- Log
Payloads bool - If enabled, will log the request and response body into the Kong log plugin(s) output.Furthermore if Opentelemetry instrumentation is enabled the traces will contain this data as well. Default: false
- Log
Statistics bool - If enabled and supported by the driver, will add model usage and token metrics into the Kong log plugin(s) output. Default: false
- log_
payloads bool - If enabled, will log the request and response body into the Kong log plugin(s) output.Furthermore if Opentelemetry instrumentation is enabled the traces will contain this data as well. Default: false
- log_
statistics bool - If enabled and supported by the driver, will add model usage and token metrics into the Kong log plugin(s) output. Default: false
- log
Payloads Boolean - If enabled, will log the request and response body into the Kong log plugin(s) output.Furthermore if Opentelemetry instrumentation is enabled the traces will contain this data as well. Default: false
- log
Statistics Boolean - If enabled and supported by the driver, will add model usage and token metrics into the Kong log plugin(s) output. Default: false
- log
Payloads boolean - If enabled, will log the request and response body into the Kong log plugin(s) output.Furthermore if Opentelemetry instrumentation is enabled the traces will contain this data as well. Default: false
- log
Statistics boolean - If enabled and supported by the driver, will add model usage and token metrics into the Kong log plugin(s) output. Default: false
- log_
payloads bool - If enabled, will log the request and response body into the Kong log plugin(s) output.Furthermore if Opentelemetry instrumentation is enabled the traces will contain this data as well. Default: false
- log_
statistics bool - If enabled and supported by the driver, will add model usage and token metrics into the Kong log plugin(s) output. Default: false
- log
Payloads Boolean - If enabled, will log the request and response body into the Kong log plugin(s) output.Furthermore if Opentelemetry instrumentation is enabled the traces will contain this data as well. Default: false
- log
Statistics Boolean - If enabled and supported by the driver, will add model usage and token metrics into the Kong log plugin(s) output. Default: false
GatewayPluginAiProxyConfigModel, GatewayPluginAiProxyConfigModelArgs
- Provider string
- AI provider request format - Kong translates requests to and from the specified backend compatible formats. possible known values include one of ["anthropic", "azure", "bedrock", "cerebras", "cohere", "dashscope", "databricks", "deepseek", "gemini", "huggingface", "llama2", "mistral", "ollama", "openai", "vllm", "xai"]
- Model
Alias string - The model name parameter from the request that this model should map to.
- Name string
- Model name to execute.
- Options
Gateway
Plugin Ai Proxy Config Model Options - Key/value settings for the model
- Provider string
- AI provider request format - Kong translates requests to and from the specified backend compatible formats. possible known values include one of ["anthropic", "azure", "bedrock", "cerebras", "cohere", "dashscope", "databricks", "deepseek", "gemini", "huggingface", "llama2", "mistral", "ollama", "openai", "vllm", "xai"]
- Model
Alias string - The model name parameter from the request that this model should map to.
- Name string
- Model name to execute.
- Options
Gateway
Plugin Ai Proxy Config Model Options - Key/value settings for the model
- provider string
- AI provider request format - Kong translates requests to and from the specified backend compatible formats. possible known values include one of ["anthropic", "azure", "bedrock", "cerebras", "cohere", "dashscope", "databricks", "deepseek", "gemini", "huggingface", "llama2", "mistral", "ollama", "openai", "vllm", "xai"]
- model_
alias string - The model name parameter from the request that this model should map to.
- name string
- Model name to execute.
- options object
- Key/value settings for the model
- provider String
- AI provider request format - Kong translates requests to and from the specified backend compatible formats. possible known values include one of ["anthropic", "azure", "bedrock", "cerebras", "cohere", "dashscope", "databricks", "deepseek", "gemini", "huggingface", "llama2", "mistral", "ollama", "openai", "vllm", "xai"]
- model
Alias String - The model name parameter from the request that this model should map to.
- name String
- Model name to execute.
- options
Gateway
Plugin Ai Proxy Config Model Options - Key/value settings for the model
- provider string
- AI provider request format - Kong translates requests to and from the specified backend compatible formats. possible known values include one of ["anthropic", "azure", "bedrock", "cerebras", "cohere", "dashscope", "databricks", "deepseek", "gemini", "huggingface", "llama2", "mistral", "ollama", "openai", "vllm", "xai"]
- model
Alias string - The model name parameter from the request that this model should map to.
- name string
- Model name to execute.
- options
Gateway
Plugin Ai Proxy Config Model Options - Key/value settings for the model
- provider str
- AI provider request format - Kong translates requests to and from the specified backend compatible formats. possible known values include one of ["anthropic", "azure", "bedrock", "cerebras", "cohere", "dashscope", "databricks", "deepseek", "gemini", "huggingface", "llama2", "mistral", "ollama", "openai", "vllm", "xai"]
- model_
alias str - The model name parameter from the request that this model should map to.
- name str
- Model name to execute.
- options
Gateway
Plugin Ai Proxy Config Model Options - Key/value settings for the model
- provider String
- AI provider request format - Kong translates requests to and from the specified backend compatible formats. possible known values include one of ["anthropic", "azure", "bedrock", "cerebras", "cohere", "dashscope", "databricks", "deepseek", "gemini", "huggingface", "llama2", "mistral", "ollama", "openai", "vllm", "xai"]
- model
Alias String - The model name parameter from the request that this model should map to.
- name String
- Model name to execute.
- options Property Map
- Key/value settings for the model
GatewayPluginAiProxyConfigModelOptions, GatewayPluginAiProxyConfigModelOptionsArgs
- Anthropic
Version string - Defines the schema/API version, if using Anthropic provider.
- Azure
Api stringVersion - 'api-version' for Azure OpenAI instances. Default: "2023-05-15"
- Azure
Deployment stringId - Deployment ID for Azure OpenAI instances.
- Azure
Instance string - Instance name for Azure OpenAI hosted models.
- Bedrock
Gateway
Plugin Ai Proxy Config Model Options Bedrock - Cohere
Gateway
Plugin Ai Proxy Config Model Options Cohere - Dashscope
Gateway
Plugin Ai Proxy Config Model Options Dashscope - Databricks
Gateway
Plugin Ai Proxy Config Model Options Databricks - Embeddings
Dimensions double - If using embeddings models, set the number of dimensions to generate.
- Gemini
Gateway
Plugin Ai Proxy Config Model Options Gemini - Huggingface
Gateway
Plugin Ai Proxy Config Model Options Huggingface - Input
Cost double - Defines the cost per 1M tokens in your prompt.
- Llama2Format string
- If using llama2 provider, select the upstream message format. possible known values include one of ["ollama", "openai", "raw"]
- Max
Tokens double - Defines the max_tokens, if using chat or completion models.
- Mistral
Format string - If using mistral provider, select the upstream message format. possible known values include one of ["ollama", "openai"]
- Output
Cost double - Defines the cost per 1M tokens in the output of the AI.
- Temperature double
- Defines the matching temperature, if using chat or completion models.
- Top
K double - Defines the top-k most likely tokens, if supported.
- Top
P double - Defines the top-p probability mass, if supported.
- Upstream
Path string - Manually specify or override the AI operation path, used when e.g. using the 'preserve' route_type.
- Upstream
Url string - Manually specify or override the full URL to the AI operation endpoints, when calling (self-)hosted models, or for running via a private endpoint.
- Anthropic
Version string - Defines the schema/API version, if using Anthropic provider.
- Azure
Api stringVersion - 'api-version' for Azure OpenAI instances. Default: "2023-05-15"
- Azure
Deployment stringId - Deployment ID for Azure OpenAI instances.
- Azure
Instance string - Instance name for Azure OpenAI hosted models.
- Bedrock
Gateway
Plugin Ai Proxy Config Model Options Bedrock - Cohere
Gateway
Plugin Ai Proxy Config Model Options Cohere - Dashscope
Gateway
Plugin Ai Proxy Config Model Options Dashscope - Databricks
Gateway
Plugin Ai Proxy Config Model Options Databricks - Embeddings
Dimensions float64 - If using embeddings models, set the number of dimensions to generate.
- Gemini
Gateway
Plugin Ai Proxy Config Model Options Gemini - Huggingface
Gateway
Plugin Ai Proxy Config Model Options Huggingface - Input
Cost float64 - Defines the cost per 1M tokens in your prompt.
- Llama2Format string
- If using llama2 provider, select the upstream message format. possible known values include one of ["ollama", "openai", "raw"]
- Max
Tokens float64 - Defines the max_tokens, if using chat or completion models.
- Mistral
Format string - If using mistral provider, select the upstream message format. possible known values include one of ["ollama", "openai"]
- Output
Cost float64 - Defines the cost per 1M tokens in the output of the AI.
- Temperature float64
- Defines the matching temperature, if using chat or completion models.
- Top
K float64 - Defines the top-k most likely tokens, if supported.
- Top
P float64 - Defines the top-p probability mass, if supported.
- Upstream
Path string - Manually specify or override the AI operation path, used when e.g. using the 'preserve' route_type.
- Upstream
Url string - Manually specify or override the full URL to the AI operation endpoints, when calling (self-)hosted models, or for running via a private endpoint.
- anthropic_
version string - Defines the schema/API version, if using Anthropic provider.
- azure_
api_ stringversion - 'api-version' for Azure OpenAI instances. Default: "2023-05-15"
- azure_
deployment_ stringid - Deployment ID for Azure OpenAI instances.
- azure_
instance string - Instance name for Azure OpenAI hosted models.
- bedrock object
- cohere object
- dashscope object
- databricks object
- embeddings_
dimensions number - If using embeddings models, set the number of dimensions to generate.
- gemini object
- huggingface object
- input_
cost number - Defines the cost per 1M tokens in your prompt.
- llama2_
format string - If using llama2 provider, select the upstream message format. possible known values include one of ["ollama", "openai", "raw"]
- max_
tokens number - Defines the max_tokens, if using chat or completion models.
- mistral_
format string - If using mistral provider, select the upstream message format. possible known values include one of ["ollama", "openai"]
- output_
cost number - Defines the cost per 1M tokens in the output of the AI.
- temperature number
- Defines the matching temperature, if using chat or completion models.
- top_
k number - Defines the top-k most likely tokens, if supported.
- top_
p number - Defines the top-p probability mass, if supported.
- upstream_
path string - Manually specify or override the AI operation path, used when e.g. using the 'preserve' route_type.
- upstream_
url string - Manually specify or override the full URL to the AI operation endpoints, when calling (self-)hosted models, or for running via a private endpoint.
- anthropic
Version String - Defines the schema/API version, if using Anthropic provider.
- azure
Api StringVersion - 'api-version' for Azure OpenAI instances. Default: "2023-05-15"
- azure
Deployment StringId - Deployment ID for Azure OpenAI instances.
- azure
Instance String - Instance name for Azure OpenAI hosted models.
- bedrock
Gateway
Plugin Ai Proxy Config Model Options Bedrock - cohere
Gateway
Plugin Ai Proxy Config Model Options Cohere - dashscope
Gateway
Plugin Ai Proxy Config Model Options Dashscope - databricks
Gateway
Plugin Ai Proxy Config Model Options Databricks - embeddings
Dimensions Double - If using embeddings models, set the number of dimensions to generate.
- gemini
Gateway
Plugin Ai Proxy Config Model Options Gemini - huggingface
Gateway
Plugin Ai Proxy Config Model Options Huggingface - input
Cost Double - Defines the cost per 1M tokens in your prompt.
- llama2Format String
- If using llama2 provider, select the upstream message format. possible known values include one of ["ollama", "openai", "raw"]
- max
Tokens Double - Defines the max_tokens, if using chat or completion models.
- mistral
Format String - If using mistral provider, select the upstream message format. possible known values include one of ["ollama", "openai"]
- output
Cost Double - Defines the cost per 1M tokens in the output of the AI.
- temperature Double
- Defines the matching temperature, if using chat or completion models.
- top
K Double - Defines the top-k most likely tokens, if supported.
- top
P Double - Defines the top-p probability mass, if supported.
- upstream
Path String - Manually specify or override the AI operation path, used when e.g. using the 'preserve' route_type.
- upstream
Url String - Manually specify or override the full URL to the AI operation endpoints, when calling (self-)hosted models, or for running via a private endpoint.
- anthropic
Version string - Defines the schema/API version, if using Anthropic provider.
- azure
Api stringVersion - 'api-version' for Azure OpenAI instances. Default: "2023-05-15"
- azure
Deployment stringId - Deployment ID for Azure OpenAI instances.
- azure
Instance string - Instance name for Azure OpenAI hosted models.
- bedrock
Gateway
Plugin Ai Proxy Config Model Options Bedrock - cohere
Gateway
Plugin Ai Proxy Config Model Options Cohere - dashscope
Gateway
Plugin Ai Proxy Config Model Options Dashscope - databricks
Gateway
Plugin Ai Proxy Config Model Options Databricks - embeddings
Dimensions number - If using embeddings models, set the number of dimensions to generate.
- gemini
Gateway
Plugin Ai Proxy Config Model Options Gemini - huggingface
Gateway
Plugin Ai Proxy Config Model Options Huggingface - input
Cost number - Defines the cost per 1M tokens in your prompt.
- llama2Format string
- If using llama2 provider, select the upstream message format. possible known values include one of ["ollama", "openai", "raw"]
- max
Tokens number - Defines the max_tokens, if using chat or completion models.
- mistral
Format string - If using mistral provider, select the upstream message format. possible known values include one of ["ollama", "openai"]
- output
Cost number - Defines the cost per 1M tokens in the output of the AI.
- temperature number
- Defines the matching temperature, if using chat or completion models.
- top
K number - Defines the top-k most likely tokens, if supported.
- top
P number - Defines the top-p probability mass, if supported.
- upstream
Path string - Manually specify or override the AI operation path, used when e.g. using the 'preserve' route_type.
- upstream
Url string - Manually specify or override the full URL to the AI operation endpoints, when calling (self-)hosted models, or for running via a private endpoint.
- anthropic_
version str - Defines the schema/API version, if using Anthropic provider.
- azure_
api_ strversion - 'api-version' for Azure OpenAI instances. Default: "2023-05-15"
- azure_
deployment_ strid - Deployment ID for Azure OpenAI instances.
- azure_
instance str - Instance name for Azure OpenAI hosted models.
- bedrock
Gateway
Plugin Ai Proxy Config Model Options Bedrock - cohere
Gateway
Plugin Ai Proxy Config Model Options Cohere - dashscope
Gateway
Plugin Ai Proxy Config Model Options Dashscope - databricks
Gateway
Plugin Ai Proxy Config Model Options Databricks - embeddings_
dimensions float - If using embeddings models, set the number of dimensions to generate.
- gemini
Gateway
Plugin Ai Proxy Config Model Options Gemini - huggingface
Gateway
Plugin Ai Proxy Config Model Options Huggingface - input_
cost float - Defines the cost per 1M tokens in your prompt.
- llama2_
format str - If using llama2 provider, select the upstream message format. possible known values include one of ["ollama", "openai", "raw"]
- max_
tokens float - Defines the max_tokens, if using chat or completion models.
- mistral_
format str - If using mistral provider, select the upstream message format. possible known values include one of ["ollama", "openai"]
- output_
cost float - Defines the cost per 1M tokens in the output of the AI.
- temperature float
- Defines the matching temperature, if using chat or completion models.
- top_
k float - Defines the top-k most likely tokens, if supported.
- top_
p float - Defines the top-p probability mass, if supported.
- upstream_
path str - Manually specify or override the AI operation path, used when e.g. using the 'preserve' route_type.
- upstream_
url str - Manually specify or override the full URL to the AI operation endpoints, when calling (self-)hosted models, or for running via a private endpoint.
- anthropic
Version String - Defines the schema/API version, if using Anthropic provider.
- azure
Api StringVersion - 'api-version' for Azure OpenAI instances. Default: "2023-05-15"
- azure
Deployment StringId - Deployment ID for Azure OpenAI instances.
- azure
Instance String - Instance name for Azure OpenAI hosted models.
- bedrock Property Map
- cohere Property Map
- dashscope Property Map
- databricks Property Map
- embeddings
Dimensions Number - If using embeddings models, set the number of dimensions to generate.
- gemini Property Map
- huggingface Property Map
- input
Cost Number - Defines the cost per 1M tokens in your prompt.
- llama2Format String
- If using llama2 provider, select the upstream message format. possible known values include one of ["ollama", "openai", "raw"]
- max
Tokens Number - Defines the max_tokens, if using chat or completion models.
- mistral
Format String - If using mistral provider, select the upstream message format. possible known values include one of ["ollama", "openai"]
- output
Cost Number - Defines the cost per 1M tokens in the output of the AI.
- temperature Number
- Defines the matching temperature, if using chat or completion models.
- top
K Number - Defines the top-k most likely tokens, if supported.
- top
P Number - Defines the top-p probability mass, if supported.
- upstream
Path String - Manually specify or override the AI operation path, used when e.g. using the 'preserve' route_type.
- upstream
Url String - Manually specify or override the full URL to the AI operation endpoints, when calling (self-)hosted models, or for running via a private endpoint.
GatewayPluginAiProxyConfigModelOptionsBedrock, GatewayPluginAiProxyConfigModelOptionsBedrockArgs
- Aws
Assume stringRole Arn - If using AWS providers (Bedrock) you can assume a different role after authentication with the current IAM context is successful.
- Aws
Region string - If using AWS providers (Bedrock) you can override the
AWS_REGIONenvironment variable by setting this option. - Aws
Role stringSession Name - If using AWS providers (Bedrock), set the identifier of the assumed role session.
- Aws
Sts stringEndpoint Url - If using AWS providers (Bedrock), override the STS endpoint URL when assuming a different role.
- Batch
Bucket stringPrefix - S3 URI prefix (s3://bucket/prefix/) where Bedrock will get input files from and store results to for native batch API.
- Batch
Role stringArn - AWS role arn used for calling batch API. Try to get the value from request if ommited.
- Embeddings
Normalize bool - If using AWS providers (Bedrock), set to true to normalize the embeddings. Default: false
- Performance
Config stringLatency - Force the client's performance configuration 'latency' for all requests. Leave empty to let the consumer select the performance configuration.
- Video
Output stringS3Uri - S3 URI (s3://bucket/prefix) where Bedrock will store generated video files. Required for video generation.
- Aws
Assume stringRole Arn - If using AWS providers (Bedrock) you can assume a different role after authentication with the current IAM context is successful.
- Aws
Region string - If using AWS providers (Bedrock) you can override the
AWS_REGIONenvironment variable by setting this option. - Aws
Role stringSession Name - If using AWS providers (Bedrock), set the identifier of the assumed role session.
- Aws
Sts stringEndpoint Url - If using AWS providers (Bedrock), override the STS endpoint URL when assuming a different role.
- Batch
Bucket stringPrefix - S3 URI prefix (s3://bucket/prefix/) where Bedrock will get input files from and store results to for native batch API.
- Batch
Role stringArn - AWS role arn used for calling batch API. Try to get the value from request if ommited.
- Embeddings
Normalize bool - If using AWS providers (Bedrock), set to true to normalize the embeddings. Default: false
- Performance
Config stringLatency - Force the client's performance configuration 'latency' for all requests. Leave empty to let the consumer select the performance configuration.
- Video
Output stringS3Uri - S3 URI (s3://bucket/prefix) where Bedrock will store generated video files. Required for video generation.
- aws_
assume_ stringrole_ arn - If using AWS providers (Bedrock) you can assume a different role after authentication with the current IAM context is successful.
- aws_
region string - If using AWS providers (Bedrock) you can override the
AWS_REGIONenvironment variable by setting this option. - aws_
role_ stringsession_ name - If using AWS providers (Bedrock), set the identifier of the assumed role session.
- aws_
sts_ stringendpoint_ url - If using AWS providers (Bedrock), override the STS endpoint URL when assuming a different role.
- batch_
bucket_ stringprefix - S3 URI prefix (s3://bucket/prefix/) where Bedrock will get input files from and store results to for native batch API.
- batch_
role_ stringarn - AWS role arn used for calling batch API. Try to get the value from request if ommited.
- embeddings_
normalize bool - If using AWS providers (Bedrock), set to true to normalize the embeddings. Default: false
- performance_
config_ stringlatency - Force the client's performance configuration 'latency' for all requests. Leave empty to let the consumer select the performance configuration.
- video_
output_ strings3_ uri - S3 URI (s3://bucket/prefix) where Bedrock will store generated video files. Required for video generation.
- aws
Assume StringRole Arn - If using AWS providers (Bedrock) you can assume a different role after authentication with the current IAM context is successful.
- aws
Region String - If using AWS providers (Bedrock) you can override the
AWS_REGIONenvironment variable by setting this option. - aws
Role StringSession Name - If using AWS providers (Bedrock), set the identifier of the assumed role session.
- aws
Sts StringEndpoint Url - If using AWS providers (Bedrock), override the STS endpoint URL when assuming a different role.
- batch
Bucket StringPrefix - S3 URI prefix (s3://bucket/prefix/) where Bedrock will get input files from and store results to for native batch API.
- batch
Role StringArn - AWS role arn used for calling batch API. Try to get the value from request if ommited.
- embeddings
Normalize Boolean - If using AWS providers (Bedrock), set to true to normalize the embeddings. Default: false
- performance
Config StringLatency - Force the client's performance configuration 'latency' for all requests. Leave empty to let the consumer select the performance configuration.
- video
Output StringS3Uri - S3 URI (s3://bucket/prefix) where Bedrock will store generated video files. Required for video generation.
- aws
Assume stringRole Arn - If using AWS providers (Bedrock) you can assume a different role after authentication with the current IAM context is successful.
- aws
Region string - If using AWS providers (Bedrock) you can override the
AWS_REGIONenvironment variable by setting this option. - aws
Role stringSession Name - If using AWS providers (Bedrock), set the identifier of the assumed role session.
- aws
Sts stringEndpoint Url - If using AWS providers (Bedrock), override the STS endpoint URL when assuming a different role.
- batch
Bucket stringPrefix - S3 URI prefix (s3://bucket/prefix/) where Bedrock will get input files from and store results to for native batch API.
- batch
Role stringArn - AWS role arn used for calling batch API. Try to get the value from request if ommited.
- embeddings
Normalize boolean - If using AWS providers (Bedrock), set to true to normalize the embeddings. Default: false
- performance
Config stringLatency - Force the client's performance configuration 'latency' for all requests. Leave empty to let the consumer select the performance configuration.
- video
Output stringS3Uri - S3 URI (s3://bucket/prefix) where Bedrock will store generated video files. Required for video generation.
- aws_
assume_ strrole_ arn - If using AWS providers (Bedrock) you can assume a different role after authentication with the current IAM context is successful.
- aws_
region str - If using AWS providers (Bedrock) you can override the
AWS_REGIONenvironment variable by setting this option. - aws_
role_ strsession_ name - If using AWS providers (Bedrock), set the identifier of the assumed role session.
- aws_
sts_ strendpoint_ url - If using AWS providers (Bedrock), override the STS endpoint URL when assuming a different role.
- batch_
bucket_ strprefix - S3 URI prefix (s3://bucket/prefix/) where Bedrock will get input files from and store results to for native batch API.
- batch_
role_ strarn - AWS role arn used for calling batch API. Try to get the value from request if ommited.
- embeddings_
normalize bool - If using AWS providers (Bedrock), set to true to normalize the embeddings. Default: false
- performance_
config_ strlatency - Force the client's performance configuration 'latency' for all requests. Leave empty to let the consumer select the performance configuration.
- video_
output_ strs3_ uri - S3 URI (s3://bucket/prefix) where Bedrock will store generated video files. Required for video generation.
- aws
Assume StringRole Arn - If using AWS providers (Bedrock) you can assume a different role after authentication with the current IAM context is successful.
- aws
Region String - If using AWS providers (Bedrock) you can override the
AWS_REGIONenvironment variable by setting this option. - aws
Role StringSession Name - If using AWS providers (Bedrock), set the identifier of the assumed role session.
- aws
Sts StringEndpoint Url - If using AWS providers (Bedrock), override the STS endpoint URL when assuming a different role.
- batch
Bucket StringPrefix - S3 URI prefix (s3://bucket/prefix/) where Bedrock will get input files from and store results to for native batch API.
- batch
Role StringArn - AWS role arn used for calling batch API. Try to get the value from request if ommited.
- embeddings
Normalize Boolean - If using AWS providers (Bedrock), set to true to normalize the embeddings. Default: false
- performance
Config StringLatency - Force the client's performance configuration 'latency' for all requests. Leave empty to let the consumer select the performance configuration.
- video
Output StringS3Uri - S3 URI (s3://bucket/prefix) where Bedrock will store generated video files. Required for video generation.
GatewayPluginAiProxyConfigModelOptionsCohere, GatewayPluginAiProxyConfigModelOptionsCohereArgs
- Embedding
Input stringType - The purpose of the input text to calculate embedding vectors. possible known values include one of ["classification", "clustering", "image", "searchdocument", "searchquery"]; Default: "classification"
- Wait
For boolModel - Wait for the model if it is not ready
- Embedding
Input stringType - The purpose of the input text to calculate embedding vectors. possible known values include one of ["classification", "clustering", "image", "searchdocument", "searchquery"]; Default: "classification"
- Wait
For boolModel - Wait for the model if it is not ready
- embedding_
input_ stringtype - The purpose of the input text to calculate embedding vectors. possible known values include one of ["classification", "clustering", "image", "searchdocument", "searchquery"]; Default: "classification"
- wait_
for_ boolmodel - Wait for the model if it is not ready
- embedding
Input StringType - The purpose of the input text to calculate embedding vectors. possible known values include one of ["classification", "clustering", "image", "searchdocument", "searchquery"]; Default: "classification"
- wait
For BooleanModel - Wait for the model if it is not ready
- embedding
Input stringType - The purpose of the input text to calculate embedding vectors. possible known values include one of ["classification", "clustering", "image", "searchdocument", "searchquery"]; Default: "classification"
- wait
For booleanModel - Wait for the model if it is not ready
- embedding_
input_ strtype - The purpose of the input text to calculate embedding vectors. possible known values include one of ["classification", "clustering", "image", "searchdocument", "searchquery"]; Default: "classification"
- wait_
for_ boolmodel - Wait for the model if it is not ready
- embedding
Input StringType - The purpose of the input text to calculate embedding vectors. possible known values include one of ["classification", "clustering", "image", "searchdocument", "searchquery"]; Default: "classification"
- wait
For BooleanModel - Wait for the model if it is not ready
GatewayPluginAiProxyConfigModelOptionsDashscope, GatewayPluginAiProxyConfigModelOptionsDashscopeArgs
- International bool
- Two Dashscope endpoints are available, and the international endpoint will be used when this is set to
true. It is recommended to set this totruewhen using international version of dashscope. Default: true
- International bool
- Two Dashscope endpoints are available, and the international endpoint will be used when this is set to
true. It is recommended to set this totruewhen using international version of dashscope. Default: true
- international bool
- Two Dashscope endpoints are available, and the international endpoint will be used when this is set to
true. It is recommended to set this totruewhen using international version of dashscope. Default: true
- international Boolean
- Two Dashscope endpoints are available, and the international endpoint will be used when this is set to
true. It is recommended to set this totruewhen using international version of dashscope. Default: true
- international boolean
- Two Dashscope endpoints are available, and the international endpoint will be used when this is set to
true. It is recommended to set this totruewhen using international version of dashscope. Default: true
- international bool
- Two Dashscope endpoints are available, and the international endpoint will be used when this is set to
true. It is recommended to set this totruewhen using international version of dashscope. Default: true
- international Boolean
- Two Dashscope endpoints are available, and the international endpoint will be used when this is set to
true. It is recommended to set this totruewhen using international version of dashscope. Default: true
GatewayPluginAiProxyConfigModelOptionsDatabricks, GatewayPluginAiProxyConfigModelOptionsDatabricksArgs
- Workspace
Instance stringId - Workspace Instance ID ('dbc-xxx-yyy') for Databricks model serving.
- Workspace
Instance stringId - Workspace Instance ID ('dbc-xxx-yyy') for Databricks model serving.
- workspace_
instance_ stringid - Workspace Instance ID ('dbc-xxx-yyy') for Databricks model serving.
- workspace
Instance StringId - Workspace Instance ID ('dbc-xxx-yyy') for Databricks model serving.
- workspace
Instance stringId - Workspace Instance ID ('dbc-xxx-yyy') for Databricks model serving.
- workspace_
instance_ strid - Workspace Instance ID ('dbc-xxx-yyy') for Databricks model serving.
- workspace
Instance StringId - Workspace Instance ID ('dbc-xxx-yyy') for Databricks model serving.
GatewayPluginAiProxyConfigModelOptionsGemini, GatewayPluginAiProxyConfigModelOptionsGeminiArgs
- Api
Endpoint string - If running Gemini on Vertex, specify the regional API endpoint (hostname only).
- Endpoint
Id string - If running Gemini on Vertex Model Garden, specify the endpoint ID.
- Location
Id string - If running Gemini on Vertex, specify the location ID.
- Project
Id string - If running Gemini on Vertex, specify the project ID.
- Api
Endpoint string - If running Gemini on Vertex, specify the regional API endpoint (hostname only).
- Endpoint
Id string - If running Gemini on Vertex Model Garden, specify the endpoint ID.
- Location
Id string - If running Gemini on Vertex, specify the location ID.
- Project
Id string - If running Gemini on Vertex, specify the project ID.
- api_
endpoint string - If running Gemini on Vertex, specify the regional API endpoint (hostname only).
- endpoint_
id string - If running Gemini on Vertex Model Garden, specify the endpoint ID.
- location_
id string - If running Gemini on Vertex, specify the location ID.
- project_
id string - If running Gemini on Vertex, specify the project ID.
- api
Endpoint String - If running Gemini on Vertex, specify the regional API endpoint (hostname only).
- endpoint
Id String - If running Gemini on Vertex Model Garden, specify the endpoint ID.
- location
Id String - If running Gemini on Vertex, specify the location ID.
- project
Id String - If running Gemini on Vertex, specify the project ID.
- api
Endpoint string - If running Gemini on Vertex, specify the regional API endpoint (hostname only).
- endpoint
Id string - If running Gemini on Vertex Model Garden, specify the endpoint ID.
- location
Id string - If running Gemini on Vertex, specify the location ID.
- project
Id string - If running Gemini on Vertex, specify the project ID.
- api_
endpoint str - If running Gemini on Vertex, specify the regional API endpoint (hostname only).
- endpoint_
id str - If running Gemini on Vertex Model Garden, specify the endpoint ID.
- location_
id str - If running Gemini on Vertex, specify the location ID.
- project_
id str - If running Gemini on Vertex, specify the project ID.
- api
Endpoint String - If running Gemini on Vertex, specify the regional API endpoint (hostname only).
- endpoint
Id String - If running Gemini on Vertex Model Garden, specify the endpoint ID.
- location
Id String - If running Gemini on Vertex, specify the location ID.
- project
Id String - If running Gemini on Vertex, specify the project ID.
GatewayPluginAiProxyConfigModelOptionsHuggingface, GatewayPluginAiProxyConfigModelOptionsHuggingfaceArgs
- Use
Cache bool - Use the cache layer on the inference API
- Wait
For boolModel - Wait for the model if it is not ready
- Use
Cache bool - Use the cache layer on the inference API
- Wait
For boolModel - Wait for the model if it is not ready
- use_
cache bool - Use the cache layer on the inference API
- wait_
for_ boolmodel - Wait for the model if it is not ready
- use
Cache Boolean - Use the cache layer on the inference API
- wait
For BooleanModel - Wait for the model if it is not ready
- use
Cache boolean - Use the cache layer on the inference API
- wait
For booleanModel - Wait for the model if it is not ready
- use_
cache bool - Use the cache layer on the inference API
- wait_
for_ boolmodel - Wait for the model if it is not ready
- use
Cache Boolean - Use the cache layer on the inference API
- wait
For BooleanModel - Wait for the model if it is not ready
GatewayPluginAiProxyConsumer, GatewayPluginAiProxyConsumerArgs
- Id string
- Id string
- id string
- id String
- id string
- id str
- id String
GatewayPluginAiProxyConsumerGroup, GatewayPluginAiProxyConsumerGroupArgs
- Id string
- Id string
- id string
- id String
- id string
- id str
- id String
GatewayPluginAiProxyOrdering, GatewayPluginAiProxyOrderingArgs
GatewayPluginAiProxyOrderingAfter, GatewayPluginAiProxyOrderingAfterArgs
- Accesses List<string>
- Accesses []string
- accesses list(string)
- accesses List<String>
- accesses string[]
- accesses Sequence[str]
- accesses List<String>
GatewayPluginAiProxyOrderingBefore, GatewayPluginAiProxyOrderingBeforeArgs
- Accesses List<string>
- Accesses []string
- accesses list(string)
- accesses List<String>
- accesses string[]
- accesses Sequence[str]
- accesses List<String>
GatewayPluginAiProxyPartial, GatewayPluginAiProxyPartialArgs
GatewayPluginAiProxyRoute, GatewayPluginAiProxyRouteArgs
- Id string
- Id string
- id string
- id String
- id string
- id str
- id String
GatewayPluginAiProxyService, GatewayPluginAiProxyServiceArgs
- Id string
- Id string
- id string
- id String
- id string
- id str
- id String
Import
In Terraform v1.5.0 and later, the import block can be used with the id attribute, for example:
terraform
import {
to = konnect_gateway_plugin_ai_proxy.my_konnect_gateway_plugin_ai_proxy
id = jsonencode({
control_plane_id = "9524ec7d-36d9-465d-a8c5-83a3c9390458"
id = "3473c251-5b6c-4f45-b1ff-7ede735a366d"
})
}
The pulumi import command can be used, for example:
$ pulumi import konnect:index/gatewayPluginAiProxy:GatewayPluginAiProxy my_konnect_gateway_plugin_ai_proxy '{"control_plane_id": "9524ec7d-36d9-465d-a8c5-83a3c9390458", "id": "3473c251-5b6c-4f45-b1ff-7ede735a366d"}'
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- konnect kong/terraform-provider-konnect
- License
- Notes
- This Pulumi package is based on the
konnectTerraform Provider.
published on Tuesday, May 12, 2026 by kong
