1. Packages
  2. Scaleway
  3. API Docs
  4. inference
  5. Model
Scaleway v1.29.0 published on Tuesday, May 27, 2025 by pulumiverse

scaleway.inference.Model

Explore with Pulumi AI

scaleway logo
Scaleway v1.29.0 published on Tuesday, May 27, 2025 by pulumiverse

    The scaleway.inference.Model resource allows you to upload and manage inference models in the Scaleway Inference ecosystem. Once registered, a model can be used in any scaleway.inference.Deployment resource.

    Example Usage

    Basic

    import * as pulumi from "@pulumi/pulumi";
    import * as scaleway from "@pulumiverse/scaleway";
    
    const test = new scaleway.inference.Model("test", {
        name: "my-awesome-model",
        url: "https://huggingface.co/agentica-org/DeepCoder-14B-Preview",
        secret: "my-secret-token",
    });
    
    import pulumi
    import pulumiverse_scaleway as scaleway
    
    test = scaleway.inference.Model("test",
        name="my-awesome-model",
        url="https://huggingface.co/agentica-org/DeepCoder-14B-Preview",
        secret="my-secret-token")
    
    package main
    
    import (
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    	"github.com/pulumiverse/pulumi-scaleway/sdk/go/scaleway/inference"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		_, err := inference.NewModel(ctx, "test", &inference.ModelArgs{
    			Name:   pulumi.String("my-awesome-model"),
    			Url:    pulumi.String("https://huggingface.co/agentica-org/DeepCoder-14B-Preview"),
    			Secret: pulumi.String("my-secret-token"),
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Scaleway = Pulumiverse.Scaleway;
    
    return await Deployment.RunAsync(() => 
    {
        var test = new Scaleway.Inference.Model("test", new()
        {
            Name = "my-awesome-model",
            Url = "https://huggingface.co/agentica-org/DeepCoder-14B-Preview",
            Secret = "my-secret-token",
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.scaleway.inference.Model;
    import com.pulumi.scaleway.inference.ModelArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var test = new Model("test", ModelArgs.builder()
                .name("my-awesome-model")
                .url("https://huggingface.co/agentica-org/DeepCoder-14B-Preview")
                .secret("my-secret-token")
                .build());
    
        }
    }
    
    resources:
      test:
        type: scaleway:inference:Model
        properties:
          name: my-awesome-model
          url: https://huggingface.co/agentica-org/DeepCoder-14B-Preview
          secret: my-secret-token
    

    Deploy your own model on your managed inference

    import * as pulumi from "@pulumi/pulumi";
    import * as scaleway from "@pulumiverse/scaleway";
    
    const myModel = new scaleway.inference.Model("my_model", {
        name: "my-awesome-model",
        url: "https://huggingface.co/agentica-org/DeepCoder-14B-Preview",
        secret: "my-secret-token",
    });
    const myDeployment = new scaleway.inference.Deployment("my_deployment", {
        name: "test-inference-deployment-basic",
        nodeType: "H100",
        modelId: myModel.id,
        publicEndpoint: {
            isEnabled: true,
        },
        acceptEula: true,
    });
    
    import pulumi
    import pulumiverse_scaleway as scaleway
    
    my_model = scaleway.inference.Model("my_model",
        name="my-awesome-model",
        url="https://huggingface.co/agentica-org/DeepCoder-14B-Preview",
        secret="my-secret-token")
    my_deployment = scaleway.inference.Deployment("my_deployment",
        name="test-inference-deployment-basic",
        node_type="H100",
        model_id=my_model.id,
        public_endpoint={
            "is_enabled": True,
        },
        accept_eula=True)
    
    package main
    
    import (
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    	"github.com/pulumiverse/pulumi-scaleway/sdk/go/scaleway/inference"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		myModel, err := inference.NewModel(ctx, "my_model", &inference.ModelArgs{
    			Name:   pulumi.String("my-awesome-model"),
    			Url:    pulumi.String("https://huggingface.co/agentica-org/DeepCoder-14B-Preview"),
    			Secret: pulumi.String("my-secret-token"),
    		})
    		if err != nil {
    			return err
    		}
    		_, err = inference.NewDeployment(ctx, "my_deployment", &inference.DeploymentArgs{
    			Name:     pulumi.String("test-inference-deployment-basic"),
    			NodeType: pulumi.String("H100"),
    			ModelId:  myModel.ID(),
    			PublicEndpoint: &inference.DeploymentPublicEndpointArgs{
    				IsEnabled: pulumi.Bool(true),
    			},
    			AcceptEula: pulumi.Bool(true),
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Scaleway = Pulumiverse.Scaleway;
    
    return await Deployment.RunAsync(() => 
    {
        var myModel = new Scaleway.Inference.Model("my_model", new()
        {
            Name = "my-awesome-model",
            Url = "https://huggingface.co/agentica-org/DeepCoder-14B-Preview",
            Secret = "my-secret-token",
        });
    
        var myDeployment = new Scaleway.Inference.Deployment("my_deployment", new()
        {
            Name = "test-inference-deployment-basic",
            NodeType = "H100",
            ModelId = myModel.Id,
            PublicEndpoint = new Scaleway.Inference.Inputs.DeploymentPublicEndpointArgs
            {
                IsEnabled = true,
            },
            AcceptEula = true,
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.scaleway.inference.Model;
    import com.pulumi.scaleway.inference.ModelArgs;
    import com.pulumi.scaleway.inference.Deployment;
    import com.pulumi.scaleway.inference.DeploymentArgs;
    import com.pulumi.scaleway.inference.inputs.DeploymentPublicEndpointArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var myModel = new Model("myModel", ModelArgs.builder()
                .name("my-awesome-model")
                .url("https://huggingface.co/agentica-org/DeepCoder-14B-Preview")
                .secret("my-secret-token")
                .build());
    
            var myDeployment = new Deployment("myDeployment", DeploymentArgs.builder()
                .name("test-inference-deployment-basic")
                .nodeType("H100")
                .modelId(myModel.id())
                .publicEndpoint(DeploymentPublicEndpointArgs.builder()
                    .isEnabled(true)
                    .build())
                .acceptEula(true)
                .build());
    
        }
    }
    
    resources:
      myModel:
        type: scaleway:inference:Model
        name: my_model
        properties:
          name: my-awesome-model
          url: https://huggingface.co/agentica-org/DeepCoder-14B-Preview
          secret: my-secret-token
      myDeployment:
        type: scaleway:inference:Deployment
        name: my_deployment
        properties:
          name: test-inference-deployment-basic
          nodeType: H100
          modelId: ${myModel.id}
          publicEndpoint:
            isEnabled: true
          acceptEula: true
    

    Create Model Resource

    Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

    Constructor syntax

    new Model(name: string, args: ModelArgs, opts?: CustomResourceOptions);
    @overload
    def Model(resource_name: str,
              args: ModelArgs,
              opts: Optional[ResourceOptions] = None)
    
    @overload
    def Model(resource_name: str,
              opts: Optional[ResourceOptions] = None,
              url: Optional[str] = None,
              name: Optional[str] = None,
              project_id: Optional[str] = None,
              region: Optional[str] = None,
              secret: Optional[str] = None)
    func NewModel(ctx *Context, name string, args ModelArgs, opts ...ResourceOption) (*Model, error)
    public Model(string name, ModelArgs args, CustomResourceOptions? opts = null)
    public Model(String name, ModelArgs args)
    public Model(String name, ModelArgs args, CustomResourceOptions options)
    
    type: scaleway:inference:Model
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    

    Parameters

    name string
    The unique name of the resource.
    args ModelArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args ModelArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args ModelArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args ModelArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args ModelArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    Constructor example

    The following reference example uses placeholder values for all input properties.

    var modelResource = new Scaleway.Inference.Model("modelResource", new()
    {
        Url = "string",
        Name = "string",
        ProjectId = "string",
        Region = "string",
        Secret = "string",
    });
    
    example, err := inference.NewModel(ctx, "modelResource", &inference.ModelArgs{
    	Url:       pulumi.String("string"),
    	Name:      pulumi.String("string"),
    	ProjectId: pulumi.String("string"),
    	Region:    pulumi.String("string"),
    	Secret:    pulumi.String("string"),
    })
    
    var modelResource = new Model("modelResource", ModelArgs.builder()
        .url("string")
        .name("string")
        .projectId("string")
        .region("string")
        .secret("string")
        .build());
    
    model_resource = scaleway.inference.Model("modelResource",
        url="string",
        name="string",
        project_id="string",
        region="string",
        secret="string")
    
    const modelResource = new scaleway.inference.Model("modelResource", {
        url: "string",
        name: "string",
        projectId: "string",
        region: "string",
        secret: "string",
    });
    
    type: scaleway:inference:Model
    properties:
        name: string
        projectId: string
        region: string
        secret: string
        url: string
    

    Model Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.

    The Model resource accepts the following input properties:

    Url string
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    Name string
    The name of the model. This must be unique within the project.
    ProjectId string
    project_id) The ID of the project the deployment is associated with.
    Region string
    region) The region in which the deployment is created.
    Secret string
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    Url string
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    Name string
    The name of the model. This must be unique within the project.
    ProjectId string
    project_id) The ID of the project the deployment is associated with.
    Region string
    region) The region in which the deployment is created.
    Secret string
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    url String
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    name String
    The name of the model. This must be unique within the project.
    projectId String
    project_id) The ID of the project the deployment is associated with.
    region String
    region) The region in which the deployment is created.
    secret String
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    url string
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    name string
    The name of the model. This must be unique within the project.
    projectId string
    project_id) The ID of the project the deployment is associated with.
    region string
    region) The region in which the deployment is created.
    secret string
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    url str
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    name str
    The name of the model. This must be unique within the project.
    project_id str
    project_id) The ID of the project the deployment is associated with.
    region str
    region) The region in which the deployment is created.
    secret str
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    url String
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    name String
    The name of the model. This must be unique within the project.
    projectId String
    project_id) The ID of the project the deployment is associated with.
    region String
    region) The region in which the deployment is created.
    secret String
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).

    Outputs

    All input properties are implicitly available as output properties. Additionally, the Model resource produces the following output properties:

    CreatedAt string
    The date and time of the creation of the model
    Description string
    A textual description of the model (if available).
    HasEula bool
    Whether the model requires end-user license agreement acceptance before use.
    Id string
    The provider-assigned unique ID for this managed resource.
    NodesSupports List<Pulumiverse.Scaleway.Inference.Outputs.ModelNodesSupport>
    List of supported node types and their quantization options. Each entry contains:
    ParameterSizeBits int
    Size, in bits, of the model parameters.
    SizeBytes int
    Total size, in bytes, of the model archive.
    Status string
    The current status of the model (e.g., ready, error, etc.).
    Tags List<string>
    Tags associated with the model.
    UpdatedAt string
    The date and time of the last update of the model
    CreatedAt string
    The date and time of the creation of the model
    Description string
    A textual description of the model (if available).
    HasEula bool
    Whether the model requires end-user license agreement acceptance before use.
    Id string
    The provider-assigned unique ID for this managed resource.
    NodesSupports []ModelNodesSupport
    List of supported node types and their quantization options. Each entry contains:
    ParameterSizeBits int
    Size, in bits, of the model parameters.
    SizeBytes int
    Total size, in bytes, of the model archive.
    Status string
    The current status of the model (e.g., ready, error, etc.).
    Tags []string
    Tags associated with the model.
    UpdatedAt string
    The date and time of the last update of the model
    createdAt String
    The date and time of the creation of the model
    description String
    A textual description of the model (if available).
    hasEula Boolean
    Whether the model requires end-user license agreement acceptance before use.
    id String
    The provider-assigned unique ID for this managed resource.
    nodesSupports List<ModelNodesSupport>
    List of supported node types and their quantization options. Each entry contains:
    parameterSizeBits Integer
    Size, in bits, of the model parameters.
    sizeBytes Integer
    Total size, in bytes, of the model archive.
    status String
    The current status of the model (e.g., ready, error, etc.).
    tags List<String>
    Tags associated with the model.
    updatedAt String
    The date and time of the last update of the model
    createdAt string
    The date and time of the creation of the model
    description string
    A textual description of the model (if available).
    hasEula boolean
    Whether the model requires end-user license agreement acceptance before use.
    id string
    The provider-assigned unique ID for this managed resource.
    nodesSupports ModelNodesSupport[]
    List of supported node types and their quantization options. Each entry contains:
    parameterSizeBits number
    Size, in bits, of the model parameters.
    sizeBytes number
    Total size, in bytes, of the model archive.
    status string
    The current status of the model (e.g., ready, error, etc.).
    tags string[]
    Tags associated with the model.
    updatedAt string
    The date and time of the last update of the model
    created_at str
    The date and time of the creation of the model
    description str
    A textual description of the model (if available).
    has_eula bool
    Whether the model requires end-user license agreement acceptance before use.
    id str
    The provider-assigned unique ID for this managed resource.
    nodes_supports Sequence[ModelNodesSupport]
    List of supported node types and their quantization options. Each entry contains:
    parameter_size_bits int
    Size, in bits, of the model parameters.
    size_bytes int
    Total size, in bytes, of the model archive.
    status str
    The current status of the model (e.g., ready, error, etc.).
    tags Sequence[str]
    Tags associated with the model.
    updated_at str
    The date and time of the last update of the model
    createdAt String
    The date and time of the creation of the model
    description String
    A textual description of the model (if available).
    hasEula Boolean
    Whether the model requires end-user license agreement acceptance before use.
    id String
    The provider-assigned unique ID for this managed resource.
    nodesSupports List<Property Map>
    List of supported node types and their quantization options. Each entry contains:
    parameterSizeBits Number
    Size, in bits, of the model parameters.
    sizeBytes Number
    Total size, in bytes, of the model archive.
    status String
    The current status of the model (e.g., ready, error, etc.).
    tags List<String>
    Tags associated with the model.
    updatedAt String
    The date and time of the last update of the model

    Look up Existing Model Resource

    Get an existing Model resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: ModelState, opts?: CustomResourceOptions): Model
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            created_at: Optional[str] = None,
            description: Optional[str] = None,
            has_eula: Optional[bool] = None,
            name: Optional[str] = None,
            nodes_supports: Optional[Sequence[ModelNodesSupportArgs]] = None,
            parameter_size_bits: Optional[int] = None,
            project_id: Optional[str] = None,
            region: Optional[str] = None,
            secret: Optional[str] = None,
            size_bytes: Optional[int] = None,
            status: Optional[str] = None,
            tags: Optional[Sequence[str]] = None,
            updated_at: Optional[str] = None,
            url: Optional[str] = None) -> Model
    func GetModel(ctx *Context, name string, id IDInput, state *ModelState, opts ...ResourceOption) (*Model, error)
    public static Model Get(string name, Input<string> id, ModelState? state, CustomResourceOptions? opts = null)
    public static Model get(String name, Output<String> id, ModelState state, CustomResourceOptions options)
    resources:  _:    type: scaleway:inference:Model    get:      id: ${id}
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    CreatedAt string
    The date and time of the creation of the model
    Description string
    A textual description of the model (if available).
    HasEula bool
    Whether the model requires end-user license agreement acceptance before use.
    Name string
    The name of the model. This must be unique within the project.
    NodesSupports List<Pulumiverse.Scaleway.Inference.Inputs.ModelNodesSupport>
    List of supported node types and their quantization options. Each entry contains:
    ParameterSizeBits int
    Size, in bits, of the model parameters.
    ProjectId string
    project_id) The ID of the project the deployment is associated with.
    Region string
    region) The region in which the deployment is created.
    Secret string
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    SizeBytes int
    Total size, in bytes, of the model archive.
    Status string
    The current status of the model (e.g., ready, error, etc.).
    Tags List<string>
    Tags associated with the model.
    UpdatedAt string
    The date and time of the last update of the model
    Url string
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    CreatedAt string
    The date and time of the creation of the model
    Description string
    A textual description of the model (if available).
    HasEula bool
    Whether the model requires end-user license agreement acceptance before use.
    Name string
    The name of the model. This must be unique within the project.
    NodesSupports []ModelNodesSupportArgs
    List of supported node types and their quantization options. Each entry contains:
    ParameterSizeBits int
    Size, in bits, of the model parameters.
    ProjectId string
    project_id) The ID of the project the deployment is associated with.
    Region string
    region) The region in which the deployment is created.
    Secret string
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    SizeBytes int
    Total size, in bytes, of the model archive.
    Status string
    The current status of the model (e.g., ready, error, etc.).
    Tags []string
    Tags associated with the model.
    UpdatedAt string
    The date and time of the last update of the model
    Url string
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    createdAt String
    The date and time of the creation of the model
    description String
    A textual description of the model (if available).
    hasEula Boolean
    Whether the model requires end-user license agreement acceptance before use.
    name String
    The name of the model. This must be unique within the project.
    nodesSupports List<ModelNodesSupport>
    List of supported node types and their quantization options. Each entry contains:
    parameterSizeBits Integer
    Size, in bits, of the model parameters.
    projectId String
    project_id) The ID of the project the deployment is associated with.
    region String
    region) The region in which the deployment is created.
    secret String
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    sizeBytes Integer
    Total size, in bytes, of the model archive.
    status String
    The current status of the model (e.g., ready, error, etc.).
    tags List<String>
    Tags associated with the model.
    updatedAt String
    The date and time of the last update of the model
    url String
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    createdAt string
    The date and time of the creation of the model
    description string
    A textual description of the model (if available).
    hasEula boolean
    Whether the model requires end-user license agreement acceptance before use.
    name string
    The name of the model. This must be unique within the project.
    nodesSupports ModelNodesSupport[]
    List of supported node types and their quantization options. Each entry contains:
    parameterSizeBits number
    Size, in bits, of the model parameters.
    projectId string
    project_id) The ID of the project the deployment is associated with.
    region string
    region) The region in which the deployment is created.
    secret string
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    sizeBytes number
    Total size, in bytes, of the model archive.
    status string
    The current status of the model (e.g., ready, error, etc.).
    tags string[]
    Tags associated with the model.
    updatedAt string
    The date and time of the last update of the model
    url string
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    created_at str
    The date and time of the creation of the model
    description str
    A textual description of the model (if available).
    has_eula bool
    Whether the model requires end-user license agreement acceptance before use.
    name str
    The name of the model. This must be unique within the project.
    nodes_supports Sequence[ModelNodesSupportArgs]
    List of supported node types and their quantization options. Each entry contains:
    parameter_size_bits int
    Size, in bits, of the model parameters.
    project_id str
    project_id) The ID of the project the deployment is associated with.
    region str
    region) The region in which the deployment is created.
    secret str
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    size_bytes int
    Total size, in bytes, of the model archive.
    status str
    The current status of the model (e.g., ready, error, etc.).
    tags Sequence[str]
    Tags associated with the model.
    updated_at str
    The date and time of the last update of the model
    url str
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret
    createdAt String
    The date and time of the creation of the model
    description String
    A textual description of the model (if available).
    hasEula Boolean
    Whether the model requires end-user license agreement acceptance before use.
    name String
    The name of the model. This must be unique within the project.
    nodesSupports List<Property Map>
    List of supported node types and their quantization options. Each entry contains:
    parameterSizeBits Number
    Size, in bits, of the model parameters.
    projectId String
    project_id) The ID of the project the deployment is associated with.
    region String
    region) The region in which the deployment is created.
    secret String
    Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
    sizeBytes Number
    Total size, in bytes, of the model archive.
    status String
    The current status of the model (e.g., ready, error, etc.).
    tags List<String>
    Tags associated with the model.
    updatedAt String
    The date and time of the last update of the model
    url String
    The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via secret

    Supporting Types

    ModelNodesSupport, ModelNodesSupportArgs

    NodeTypeName string
    The type of node supported.
    Quantizations List<Pulumiverse.Scaleway.Inference.Inputs.ModelNodesSupportQuantization>
    A list of supported quantization options, including:
    NodeTypeName string
    The type of node supported.
    Quantizations []ModelNodesSupportQuantization
    A list of supported quantization options, including:
    nodeTypeName String
    The type of node supported.
    quantizations List<ModelNodesSupportQuantization>
    A list of supported quantization options, including:
    nodeTypeName string
    The type of node supported.
    quantizations ModelNodesSupportQuantization[]
    A list of supported quantization options, including:
    node_type_name str
    The type of node supported.
    quantizations Sequence[ModelNodesSupportQuantization]
    A list of supported quantization options, including:
    nodeTypeName String
    The type of node supported.
    quantizations List<Property Map>
    A list of supported quantization options, including:

    ModelNodesSupportQuantization, ModelNodesSupportQuantizationArgs

    Allowed bool
    Whether this quantization is allowed.
    MaxContextSize int
    Maximum context length supported by this quantization.
    QuantizationBits int
    Number of bits used for quantization (e.g., 8, 16).
    Allowed bool
    Whether this quantization is allowed.
    MaxContextSize int
    Maximum context length supported by this quantization.
    QuantizationBits int
    Number of bits used for quantization (e.g., 8, 16).
    allowed Boolean
    Whether this quantization is allowed.
    maxContextSize Integer
    Maximum context length supported by this quantization.
    quantizationBits Integer
    Number of bits used for quantization (e.g., 8, 16).
    allowed boolean
    Whether this quantization is allowed.
    maxContextSize number
    Maximum context length supported by this quantization.
    quantizationBits number
    Number of bits used for quantization (e.g., 8, 16).
    allowed bool
    Whether this quantization is allowed.
    max_context_size int
    Maximum context length supported by this quantization.
    quantization_bits int
    Number of bits used for quantization (e.g., 8, 16).
    allowed Boolean
    Whether this quantization is allowed.
    maxContextSize Number
    Maximum context length supported by this quantization.
    quantizationBits Number
    Number of bits used for quantization (e.g., 8, 16).

    Import

    Models can be imported using, {region}/{id}, as shown below:

    bash

    $ pulumi import scaleway:inference/model:Model my_model fr-par/11111111-1111-1111-1111-111111111111
    

    To learn more about importing existing cloud resources, see Importing resources.

    Package Details

    Repository
    scaleway pulumiverse/pulumi-scaleway
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the scaleway Terraform Provider.
    scaleway logo
    Scaleway v1.29.0 published on Tuesday, May 27, 2025 by pulumiverse