1. Packages
  2. Hsdp Provider
  3. API Docs
  4. AiInferenceModel
hsdp 0.65.3 published on Tuesday, Apr 15, 2025 by philips-software

hsdp.AiInferenceModel

Explore with Pulumi AI

hsdp logo
hsdp 0.65.3 published on Tuesday, Apr 15, 2025 by philips-software

    Manages HSDP AI Inference models.

    Example Usage

    import * as pulumi from "@pulumi/pulumi";
    import * as hsdp from "@pulumi/hsdp";
    
    const inferenceConfig = hsdp.getConfig({
        service: "inference",
    });
    const inferenceAiInferenceServiceInstance = inferenceConfig.then(inferenceConfig => hsdp.getAiInferenceServiceInstance({
        baseUrl: inferenceConfig.url,
        organizationId: _var.inference_tenant_org_id,
    }));
    const model = new hsdp.AiInferenceModel("model", {
        endpoint: inferenceAiInferenceServiceInstance.then(inferenceAiInferenceServiceInstance => inferenceAiInferenceServiceInstance.endpoint),
        version: "v1",
        description: "Test model",
        computeEnvironment: {
            reference: "foo",
            identifier: "bar",
        },
        sourceCode: {
            url: "git@github.com:testuser/source.git",
            branch: "main",
            commitId: "e1f9366",
            sshKey: "...",
        },
        artifactPath: "git@github.com:testuser/example.git",
        entryCommands: ["python main/train.py -s 134786"],
        environment: {
            FOO: "bar",
            BAR: "baz",
        },
        labels: ["CNN"],
        additionalConfiguration: "{\"Tags\": [ { \"Key\": \"name\",\"Value\": \"hsp\"}]}",
    });
    
    import pulumi
    import pulumi_hsdp as hsdp
    
    inference_config = hsdp.get_config(service="inference")
    inference_ai_inference_service_instance = hsdp.get_ai_inference_service_instance(base_url=inference_config.url,
        organization_id=var["inference_tenant_org_id"])
    model = hsdp.AiInferenceModel("model",
        endpoint=inference_ai_inference_service_instance.endpoint,
        version="v1",
        description="Test model",
        compute_environment={
            "reference": "foo",
            "identifier": "bar",
        },
        source_code={
            "url": "git@github.com:testuser/source.git",
            "branch": "main",
            "commit_id": "e1f9366",
            "ssh_key": "...",
        },
        artifact_path="git@github.com:testuser/example.git",
        entry_commands=["python main/train.py -s 134786"],
        environment={
            "FOO": "bar",
            "BAR": "baz",
        },
        labels=["CNN"],
        additional_configuration="{\"Tags\": [ { \"Key\": \"name\",\"Value\": \"hsp\"}]}")
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-terraform-provider/sdks/go/hsdp/hsdp"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		inferenceConfig, err := hsdp.GetConfig(ctx, &hsdp.GetConfigArgs{
    			Service: "inference",
    		}, nil)
    		if err != nil {
    			return err
    		}
    		inferenceAiInferenceServiceInstance, err := hsdp.GetAiInferenceServiceInstance(ctx, &hsdp.GetAiInferenceServiceInstanceArgs{
    			BaseUrl:        inferenceConfig.Url,
    			OrganizationId: _var.Inference_tenant_org_id,
    		}, nil)
    		if err != nil {
    			return err
    		}
    		_, err = hsdp.NewAiInferenceModel(ctx, "model", &hsdp.AiInferenceModelArgs{
    			Endpoint:    pulumi.String(inferenceAiInferenceServiceInstance.Endpoint),
    			Version:     pulumi.String("v1"),
    			Description: pulumi.String("Test model"),
    			ComputeEnvironment: &hsdp.AiInferenceModelComputeEnvironmentArgs{
    				Reference:  pulumi.String("foo"),
    				Identifier: pulumi.String("bar"),
    			},
    			SourceCode: &hsdp.AiInferenceModelSourceCodeArgs{
    				Url:      pulumi.String("git@github.com:testuser/source.git"),
    				Branch:   pulumi.String("main"),
    				CommitId: pulumi.String("e1f9366"),
    				SshKey:   pulumi.String("..."),
    			},
    			ArtifactPath: pulumi.String("git@github.com:testuser/example.git"),
    			EntryCommands: pulumi.StringArray{
    				pulumi.String("python main/train.py -s 134786"),
    			},
    			Environment: pulumi.StringMap{
    				"FOO": pulumi.String("bar"),
    				"BAR": pulumi.String("baz"),
    			},
    			Labels: pulumi.StringArray{
    				pulumi.String("CNN"),
    			},
    			AdditionalConfiguration: pulumi.String("{\"Tags\": [ { \"Key\": \"name\",\"Value\": \"hsp\"}]}"),
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Hsdp = Pulumi.Hsdp;
    
    return await Deployment.RunAsync(() => 
    {
        var inferenceConfig = Hsdp.GetConfig.Invoke(new()
        {
            Service = "inference",
        });
    
        var inferenceAiInferenceServiceInstance = Hsdp.GetAiInferenceServiceInstance.Invoke(new()
        {
            BaseUrl = inferenceConfig.Apply(getConfigResult => getConfigResult.Url),
            OrganizationId = @var.Inference_tenant_org_id,
        });
    
        var model = new Hsdp.AiInferenceModel("model", new()
        {
            Endpoint = inferenceAiInferenceServiceInstance.Apply(getAiInferenceServiceInstanceResult => getAiInferenceServiceInstanceResult.Endpoint),
            Version = "v1",
            Description = "Test model",
            ComputeEnvironment = new Hsdp.Inputs.AiInferenceModelComputeEnvironmentArgs
            {
                Reference = "foo",
                Identifier = "bar",
            },
            SourceCode = new Hsdp.Inputs.AiInferenceModelSourceCodeArgs
            {
                Url = "git@github.com:testuser/source.git",
                Branch = "main",
                CommitId = "e1f9366",
                SshKey = "...",
            },
            ArtifactPath = "git@github.com:testuser/example.git",
            EntryCommands = new[]
            {
                "python main/train.py -s 134786",
            },
            Environment = 
            {
                { "FOO", "bar" },
                { "BAR", "baz" },
            },
            Labels = new[]
            {
                "CNN",
            },
            AdditionalConfiguration = "{\"Tags\": [ { \"Key\": \"name\",\"Value\": \"hsp\"}]}",
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.hsdp.HsdpFunctions;
    import com.pulumi.hsdp.inputs.GetConfigArgs;
    import com.pulumi.hsdp.inputs.GetAiInferenceServiceInstanceArgs;
    import com.pulumi.hsdp.AiInferenceModel;
    import com.pulumi.hsdp.AiInferenceModelArgs;
    import com.pulumi.hsdp.inputs.AiInferenceModelComputeEnvironmentArgs;
    import com.pulumi.hsdp.inputs.AiInferenceModelSourceCodeArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            final var inferenceConfig = HsdpFunctions.getConfig(GetConfigArgs.builder()
                .service("inference")
                .build());
    
            final var inferenceAiInferenceServiceInstance = HsdpFunctions.getAiInferenceServiceInstance(GetAiInferenceServiceInstanceArgs.builder()
                .baseUrl(inferenceConfig.applyValue(getConfigResult -> getConfigResult.url()))
                .organizationId(var_.inference_tenant_org_id())
                .build());
    
            var model = new AiInferenceModel("model", AiInferenceModelArgs.builder()
                .endpoint(inferenceAiInferenceServiceInstance.applyValue(getAiInferenceServiceInstanceResult -> getAiInferenceServiceInstanceResult.endpoint()))
                .version("v1")
                .description("Test model")
                .computeEnvironment(AiInferenceModelComputeEnvironmentArgs.builder()
                    .reference("foo")
                    .identifier("bar")
                    .build())
                .sourceCode(AiInferenceModelSourceCodeArgs.builder()
                    .url("git@github.com:testuser/source.git")
                    .branch("main")
                    .commitId("e1f9366")
                    .sshKey("...")
                    .build())
                .artifactPath("git@github.com:testuser/example.git")
                .entryCommands("python main/train.py -s 134786")
                .environment(Map.ofEntries(
                    Map.entry("FOO", "bar"),
                    Map.entry("BAR", "baz")
                ))
                .labels("CNN")
                .additionalConfiguration("{\"Tags\": [ { \"Key\": \"name\",\"Value\": \"hsp\"}]}")
                .build());
    
        }
    }
    
    resources:
      model:
        type: hsdp:AiInferenceModel
        properties:
          endpoint: ${inferenceAiInferenceServiceInstance.endpoint}
          version: v1
          description: Test model
          computeEnvironment:
            reference: foo
            identifier: bar
          sourceCode:
            url: git@github.com:testuser/source.git
            branch: main
            commitId: e1f9366
            sshKey: '...'
          artifactPath: git@github.com:testuser/example.git
          entryCommands:
            - python main/train.py -s 134786
          environment:
            FOO: bar
            BAR: baz
          labels:
            - CNN
          additionalConfiguration: '{"Tags": [ { "Key": "name","Value": "hsp"}]}'
    variables:
      inferenceConfig:
        fn::invoke:
          function: hsdp:getConfig
          arguments:
            service: inference
      inferenceAiInferenceServiceInstance:
        fn::invoke:
          function: hsdp:getAiInferenceServiceInstance
          arguments:
            baseUrl: ${inferenceConfig.url}
            organizationId: ${var.inference_tenant_org_id}
    

    Create AiInferenceModel Resource

    Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

    Constructor syntax

    new AiInferenceModel(name: string, args: AiInferenceModelArgs, opts?: CustomResourceOptions);
    @overload
    def AiInferenceModel(resource_name: str,
                         args: AiInferenceModelArgs,
                         opts: Optional[ResourceOptions] = None)
    
    @overload
    def AiInferenceModel(resource_name: str,
                         opts: Optional[ResourceOptions] = None,
                         endpoint: Optional[str] = None,
                         version: Optional[str] = None,
                         additional_configuration: Optional[str] = None,
                         ai_inference_model_id: Optional[str] = None,
                         artifact_path: Optional[str] = None,
                         compute_environment: Optional[AiInferenceModelComputeEnvironmentArgs] = None,
                         description: Optional[str] = None,
                         entry_commands: Optional[Sequence[str]] = None,
                         environment: Optional[Mapping[str, str]] = None,
                         labels: Optional[Sequence[str]] = None,
                         name: Optional[str] = None,
                         source_code: Optional[AiInferenceModelSourceCodeArgs] = None)
    func NewAiInferenceModel(ctx *Context, name string, args AiInferenceModelArgs, opts ...ResourceOption) (*AiInferenceModel, error)
    public AiInferenceModel(string name, AiInferenceModelArgs args, CustomResourceOptions? opts = null)
    public AiInferenceModel(String name, AiInferenceModelArgs args)
    public AiInferenceModel(String name, AiInferenceModelArgs args, CustomResourceOptions options)
    
    type: hsdp:AiInferenceModel
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    

    Parameters

    name string
    The unique name of the resource.
    args AiInferenceModelArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args AiInferenceModelArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args AiInferenceModelArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args AiInferenceModelArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args AiInferenceModelArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    Constructor example

    The following reference example uses placeholder values for all input properties.

    var aiInferenceModelResource = new Hsdp.AiInferenceModel("aiInferenceModelResource", new()
    {
        Endpoint = "string",
        Version = "string",
        AdditionalConfiguration = "string",
        AiInferenceModelId = "string",
        ArtifactPath = "string",
        ComputeEnvironment = new Hsdp.Inputs.AiInferenceModelComputeEnvironmentArgs
        {
            Identifier = "string",
            Reference = "string",
        },
        Description = "string",
        EntryCommands = new[]
        {
            "string",
        },
        Environment = 
        {
            { "string", "string" },
        },
        Labels = new[]
        {
            "string",
        },
        Name = "string",
        SourceCode = new Hsdp.Inputs.AiInferenceModelSourceCodeArgs
        {
            Url = "string",
            Branch = "string",
            CommitId = "string",
            SshKey = "string",
        },
    });
    
    example, err := hsdp.NewAiInferenceModel(ctx, "aiInferenceModelResource", &hsdp.AiInferenceModelArgs{
    	Endpoint:                pulumi.String("string"),
    	Version:                 pulumi.String("string"),
    	AdditionalConfiguration: pulumi.String("string"),
    	AiInferenceModelId:      pulumi.String("string"),
    	ArtifactPath:            pulumi.String("string"),
    	ComputeEnvironment: &hsdp.AiInferenceModelComputeEnvironmentArgs{
    		Identifier: pulumi.String("string"),
    		Reference:  pulumi.String("string"),
    	},
    	Description: pulumi.String("string"),
    	EntryCommands: pulumi.StringArray{
    		pulumi.String("string"),
    	},
    	Environment: pulumi.StringMap{
    		"string": pulumi.String("string"),
    	},
    	Labels: pulumi.StringArray{
    		pulumi.String("string"),
    	},
    	Name: pulumi.String("string"),
    	SourceCode: &hsdp.AiInferenceModelSourceCodeArgs{
    		Url:      pulumi.String("string"),
    		Branch:   pulumi.String("string"),
    		CommitId: pulumi.String("string"),
    		SshKey:   pulumi.String("string"),
    	},
    })
    
    var aiInferenceModelResource = new AiInferenceModel("aiInferenceModelResource", AiInferenceModelArgs.builder()
        .endpoint("string")
        .version("string")
        .additionalConfiguration("string")
        .aiInferenceModelId("string")
        .artifactPath("string")
        .computeEnvironment(AiInferenceModelComputeEnvironmentArgs.builder()
            .identifier("string")
            .reference("string")
            .build())
        .description("string")
        .entryCommands("string")
        .environment(Map.of("string", "string"))
        .labels("string")
        .name("string")
        .sourceCode(AiInferenceModelSourceCodeArgs.builder()
            .url("string")
            .branch("string")
            .commitId("string")
            .sshKey("string")
            .build())
        .build());
    
    ai_inference_model_resource = hsdp.AiInferenceModel("aiInferenceModelResource",
        endpoint="string",
        version="string",
        additional_configuration="string",
        ai_inference_model_id="string",
        artifact_path="string",
        compute_environment={
            "identifier": "string",
            "reference": "string",
        },
        description="string",
        entry_commands=["string"],
        environment={
            "string": "string",
        },
        labels=["string"],
        name="string",
        source_code={
            "url": "string",
            "branch": "string",
            "commit_id": "string",
            "ssh_key": "string",
        })
    
    const aiInferenceModelResource = new hsdp.AiInferenceModel("aiInferenceModelResource", {
        endpoint: "string",
        version: "string",
        additionalConfiguration: "string",
        aiInferenceModelId: "string",
        artifactPath: "string",
        computeEnvironment: {
            identifier: "string",
            reference: "string",
        },
        description: "string",
        entryCommands: ["string"],
        environment: {
            string: "string",
        },
        labels: ["string"],
        name: "string",
        sourceCode: {
            url: "string",
            branch: "string",
            commitId: "string",
            sshKey: "string",
        },
    });
    
    type: hsdp:AiInferenceModel
    properties:
        additionalConfiguration: string
        aiInferenceModelId: string
        artifactPath: string
        computeEnvironment:
            identifier: string
            reference: string
        description: string
        endpoint: string
        entryCommands:
            - string
        environment:
            string: string
        labels:
            - string
        name: string
        sourceCode:
            branch: string
            commitId: string
            sshKey: string
            url: string
        version: string
    

    AiInferenceModel Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.

    The AiInferenceModel resource accepts the following input properties:

    Endpoint string
    The AI Inference instance endpoint
    Version string
    AdditionalConfiguration string
    AiInferenceModelId string
    The GUID of the Model
    ArtifactPath string
    ComputeEnvironment AiInferenceModelComputeEnvironment
    Description string
    Description of the Compute Target
    EntryCommands List<string>
    ) Commands to execute
    Environment Dictionary<string, string>
    List of environment variables to set
    Labels List<string>
    Name string
    The name of the Model
    SourceCode AiInferenceModelSourceCode
    Endpoint string
    The AI Inference instance endpoint
    Version string
    AdditionalConfiguration string
    AiInferenceModelId string
    The GUID of the Model
    ArtifactPath string
    ComputeEnvironment AiInferenceModelComputeEnvironmentArgs
    Description string
    Description of the Compute Target
    EntryCommands []string
    ) Commands to execute
    Environment map[string]string
    List of environment variables to set
    Labels []string
    Name string
    The name of the Model
    SourceCode AiInferenceModelSourceCodeArgs
    endpoint String
    The AI Inference instance endpoint
    version String
    additionalConfiguration String
    aiInferenceModelId String
    The GUID of the Model
    artifactPath String
    computeEnvironment AiInferenceModelComputeEnvironment
    description String
    Description of the Compute Target
    entryCommands List<String>
    ) Commands to execute
    environment Map<String,String>
    List of environment variables to set
    labels List<String>
    name String
    The name of the Model
    sourceCode AiInferenceModelSourceCode
    endpoint string
    The AI Inference instance endpoint
    version string
    additionalConfiguration string
    aiInferenceModelId string
    The GUID of the Model
    artifactPath string
    computeEnvironment AiInferenceModelComputeEnvironment
    description string
    Description of the Compute Target
    entryCommands string[]
    ) Commands to execute
    environment {[key: string]: string}
    List of environment variables to set
    labels string[]
    name string
    The name of the Model
    sourceCode AiInferenceModelSourceCode
    endpoint str
    The AI Inference instance endpoint
    version str
    additional_configuration str
    ai_inference_model_id str
    The GUID of the Model
    artifact_path str
    compute_environment AiInferenceModelComputeEnvironmentArgs
    description str
    Description of the Compute Target
    entry_commands Sequence[str]
    ) Commands to execute
    environment Mapping[str, str]
    List of environment variables to set
    labels Sequence[str]
    name str
    The name of the Model
    source_code AiInferenceModelSourceCodeArgs
    endpoint String
    The AI Inference instance endpoint
    version String
    additionalConfiguration String
    aiInferenceModelId String
    The GUID of the Model
    artifactPath String
    computeEnvironment Property Map
    description String
    Description of the Compute Target
    entryCommands List<String>
    ) Commands to execute
    environment Map<String>
    List of environment variables to set
    labels List<String>
    name String
    The name of the Model
    sourceCode Property Map

    Outputs

    All input properties are implicitly available as output properties. Additionally, the AiInferenceModel resource produces the following output properties:

    Created string
    The date this Model was created
    CreatedBy string
    Who created the Model
    Id string
    The provider-assigned unique ID for this managed resource.
    Reference string
    The reference of this Model
    Created string
    The date this Model was created
    CreatedBy string
    Who created the Model
    Id string
    The provider-assigned unique ID for this managed resource.
    Reference string
    The reference of this Model
    created String
    The date this Model was created
    createdBy String
    Who created the Model
    id String
    The provider-assigned unique ID for this managed resource.
    reference String
    The reference of this Model
    created string
    The date this Model was created
    createdBy string
    Who created the Model
    id string
    The provider-assigned unique ID for this managed resource.
    reference string
    The reference of this Model
    created str
    The date this Model was created
    created_by str
    Who created the Model
    id str
    The provider-assigned unique ID for this managed resource.
    reference str
    The reference of this Model
    created String
    The date this Model was created
    createdBy String
    Who created the Model
    id String
    The provider-assigned unique ID for this managed resource.
    reference String
    The reference of this Model

    Look up Existing AiInferenceModel Resource

    Get an existing AiInferenceModel resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: AiInferenceModelState, opts?: CustomResourceOptions): AiInferenceModel
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            additional_configuration: Optional[str] = None,
            ai_inference_model_id: Optional[str] = None,
            artifact_path: Optional[str] = None,
            compute_environment: Optional[AiInferenceModelComputeEnvironmentArgs] = None,
            created: Optional[str] = None,
            created_by: Optional[str] = None,
            description: Optional[str] = None,
            endpoint: Optional[str] = None,
            entry_commands: Optional[Sequence[str]] = None,
            environment: Optional[Mapping[str, str]] = None,
            labels: Optional[Sequence[str]] = None,
            name: Optional[str] = None,
            reference: Optional[str] = None,
            source_code: Optional[AiInferenceModelSourceCodeArgs] = None,
            version: Optional[str] = None) -> AiInferenceModel
    func GetAiInferenceModel(ctx *Context, name string, id IDInput, state *AiInferenceModelState, opts ...ResourceOption) (*AiInferenceModel, error)
    public static AiInferenceModel Get(string name, Input<string> id, AiInferenceModelState? state, CustomResourceOptions? opts = null)
    public static AiInferenceModel get(String name, Output<String> id, AiInferenceModelState state, CustomResourceOptions options)
    resources:  _:    type: hsdp:AiInferenceModel    get:      id: ${id}
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    AdditionalConfiguration string
    AiInferenceModelId string
    The GUID of the Model
    ArtifactPath string
    ComputeEnvironment AiInferenceModelComputeEnvironment
    Created string
    The date this Model was created
    CreatedBy string
    Who created the Model
    Description string
    Description of the Compute Target
    Endpoint string
    The AI Inference instance endpoint
    EntryCommands List<string>
    ) Commands to execute
    Environment Dictionary<string, string>
    List of environment variables to set
    Labels List<string>
    Name string
    The name of the Model
    Reference string
    The reference of this Model
    SourceCode AiInferenceModelSourceCode
    Version string
    AdditionalConfiguration string
    AiInferenceModelId string
    The GUID of the Model
    ArtifactPath string
    ComputeEnvironment AiInferenceModelComputeEnvironmentArgs
    Created string
    The date this Model was created
    CreatedBy string
    Who created the Model
    Description string
    Description of the Compute Target
    Endpoint string
    The AI Inference instance endpoint
    EntryCommands []string
    ) Commands to execute
    Environment map[string]string
    List of environment variables to set
    Labels []string
    Name string
    The name of the Model
    Reference string
    The reference of this Model
    SourceCode AiInferenceModelSourceCodeArgs
    Version string
    additionalConfiguration String
    aiInferenceModelId String
    The GUID of the Model
    artifactPath String
    computeEnvironment AiInferenceModelComputeEnvironment
    created String
    The date this Model was created
    createdBy String
    Who created the Model
    description String
    Description of the Compute Target
    endpoint String
    The AI Inference instance endpoint
    entryCommands List<String>
    ) Commands to execute
    environment Map<String,String>
    List of environment variables to set
    labels List<String>
    name String
    The name of the Model
    reference String
    The reference of this Model
    sourceCode AiInferenceModelSourceCode
    version String
    additionalConfiguration string
    aiInferenceModelId string
    The GUID of the Model
    artifactPath string
    computeEnvironment AiInferenceModelComputeEnvironment
    created string
    The date this Model was created
    createdBy string
    Who created the Model
    description string
    Description of the Compute Target
    endpoint string
    The AI Inference instance endpoint
    entryCommands string[]
    ) Commands to execute
    environment {[key: string]: string}
    List of environment variables to set
    labels string[]
    name string
    The name of the Model
    reference string
    The reference of this Model
    sourceCode AiInferenceModelSourceCode
    version string
    additional_configuration str
    ai_inference_model_id str
    The GUID of the Model
    artifact_path str
    compute_environment AiInferenceModelComputeEnvironmentArgs
    created str
    The date this Model was created
    created_by str
    Who created the Model
    description str
    Description of the Compute Target
    endpoint str
    The AI Inference instance endpoint
    entry_commands Sequence[str]
    ) Commands to execute
    environment Mapping[str, str]
    List of environment variables to set
    labels Sequence[str]
    name str
    The name of the Model
    reference str
    The reference of this Model
    source_code AiInferenceModelSourceCodeArgs
    version str
    additionalConfiguration String
    aiInferenceModelId String
    The GUID of the Model
    artifactPath String
    computeEnvironment Property Map
    created String
    The date this Model was created
    createdBy String
    Who created the Model
    description String
    Description of the Compute Target
    endpoint String
    The AI Inference instance endpoint
    entryCommands List<String>
    ) Commands to execute
    environment Map<String>
    List of environment variables to set
    labels List<String>
    name String
    The name of the Model
    reference String
    The reference of this Model
    sourceCode Property Map
    version String

    Supporting Types

    AiInferenceModelComputeEnvironment, AiInferenceModelComputeEnvironmentArgs

    Identifier string
    Reference string
    The reference of this Model
    Identifier string
    Reference string
    The reference of this Model
    identifier String
    reference String
    The reference of this Model
    identifier string
    reference string
    The reference of this Model
    identifier str
    reference str
    The reference of this Model
    identifier String
    reference String
    The reference of this Model

    AiInferenceModelSourceCode, AiInferenceModelSourceCodeArgs

    Url string
    Branch string
    CommitId string
    SshKey string
    Url string
    Branch string
    CommitId string
    SshKey string
    url String
    branch String
    commitId String
    sshKey String
    url string
    branch string
    commitId string
    sshKey string
    url String
    branch String
    commitId String
    sshKey String

    Package Details

    Repository
    hsdp philips-software/terraform-provider-hsdp
    License
    Notes
    This Pulumi package is based on the hsdp Terraform Provider.
    hsdp logo
    hsdp 0.65.3 published on Tuesday, Apr 15, 2025 by philips-software