databricks.MetastoreAssignment
Explore with Pulumi AI
A single databricks.Metastore can be shared across Databricks workspaces, and each linked workspace has a consistent view of the data and a single set of access policies. You can only create a single metastore for each region in which your organization operates.
Example Usage
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Databricks = Pulumi.Databricks;
return await Deployment.RunAsync(() =>
{
var thisMetastore = new Databricks.Metastore("thisMetastore", new()
{
StorageRoot = $"s3://{aws_s3_bucket.Metastore.Id}/metastore",
Owner = "uc admins",
ForceDestroy = true,
});
var thisMetastoreAssignment = new Databricks.MetastoreAssignment("thisMetastoreAssignment", new()
{
MetastoreId = thisMetastore.Id,
WorkspaceId = local.Workspace_id,
});
});
package main
import (
"fmt"
"github.com/pulumi/pulumi-databricks/sdk/go/databricks"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
thisMetastore, err := databricks.NewMetastore(ctx, "thisMetastore", &databricks.MetastoreArgs{
StorageRoot: pulumi.String(fmt.Sprintf("s3://%v/metastore", aws_s3_bucket.Metastore.Id)),
Owner: pulumi.String("uc admins"),
ForceDestroy: pulumi.Bool(true),
})
if err != nil {
return err
}
_, err = databricks.NewMetastoreAssignment(ctx, "thisMetastoreAssignment", &databricks.MetastoreAssignmentArgs{
MetastoreId: thisMetastore.ID(),
WorkspaceId: pulumi.Any(local.Workspace_id),
})
if err != nil {
return err
}
return nil
})
}
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.databricks.Metastore;
import com.pulumi.databricks.MetastoreArgs;
import com.pulumi.databricks.MetastoreAssignment;
import com.pulumi.databricks.MetastoreAssignmentArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var thisMetastore = new Metastore("thisMetastore", MetastoreArgs.builder()
.storageRoot(String.format("s3://%s/metastore", aws_s3_bucket.metastore().id()))
.owner("uc admins")
.forceDestroy(true)
.build());
var thisMetastoreAssignment = new MetastoreAssignment("thisMetastoreAssignment", MetastoreAssignmentArgs.builder()
.metastoreId(thisMetastore.id())
.workspaceId(local.workspace_id())
.build());
}
}
import pulumi
import pulumi_databricks as databricks
this_metastore = databricks.Metastore("thisMetastore",
storage_root=f"s3://{aws_s3_bucket['metastore']['id']}/metastore",
owner="uc admins",
force_destroy=True)
this_metastore_assignment = databricks.MetastoreAssignment("thisMetastoreAssignment",
metastore_id=this_metastore.id,
workspace_id=local["workspace_id"])
import * as pulumi from "@pulumi/pulumi";
import * as databricks from "@pulumi/databricks";
const thisMetastore = new databricks.Metastore("thisMetastore", {
storageRoot: `s3://${aws_s3_bucket.metastore.id}/metastore`,
owner: "uc admins",
forceDestroy: true,
});
const thisMetastoreAssignment = new databricks.MetastoreAssignment("thisMetastoreAssignment", {
metastoreId: thisMetastore.id,
workspaceId: local.workspace_id,
});
resources:
thisMetastore:
type: databricks:Metastore
properties:
storageRoot: s3://${aws_s3_bucket.metastore.id}/metastore
owner: uc admins
forceDestroy: true
thisMetastoreAssignment:
type: databricks:MetastoreAssignment
properties:
metastoreId: ${thisMetastore.id}
workspaceId: ${local.workspace_id}
Create MetastoreAssignment Resource
new MetastoreAssignment(name: string, args: MetastoreAssignmentArgs, opts?: CustomResourceOptions);
@overload
def MetastoreAssignment(resource_name: str,
opts: Optional[ResourceOptions] = None,
default_catalog_name: Optional[str] = None,
metastore_id: Optional[str] = None,
workspace_id: Optional[int] = None)
@overload
def MetastoreAssignment(resource_name: str,
args: MetastoreAssignmentArgs,
opts: Optional[ResourceOptions] = None)
func NewMetastoreAssignment(ctx *Context, name string, args MetastoreAssignmentArgs, opts ...ResourceOption) (*MetastoreAssignment, error)
public MetastoreAssignment(string name, MetastoreAssignmentArgs args, CustomResourceOptions? opts = null)
public MetastoreAssignment(String name, MetastoreAssignmentArgs args)
public MetastoreAssignment(String name, MetastoreAssignmentArgs args, CustomResourceOptions options)
type: databricks:MetastoreAssignment
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args MetastoreAssignmentArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args MetastoreAssignmentArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args MetastoreAssignmentArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args MetastoreAssignmentArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args MetastoreAssignmentArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
MetastoreAssignment Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
The MetastoreAssignment resource accepts the following input properties:
- Metastore
Id string Unique identifier of the parent Metastore
- Workspace
Id int id of the workspace for the assignment
- Default
Catalog stringName Default catalog used for this assignment, default to
hive_metastore
- Metastore
Id string Unique identifier of the parent Metastore
- Workspace
Id int id of the workspace for the assignment
- Default
Catalog stringName Default catalog used for this assignment, default to
hive_metastore
- metastore
Id String Unique identifier of the parent Metastore
- workspace
Id Integer id of the workspace for the assignment
- default
Catalog StringName Default catalog used for this assignment, default to
hive_metastore
- metastore
Id string Unique identifier of the parent Metastore
- workspace
Id number id of the workspace for the assignment
- default
Catalog stringName Default catalog used for this assignment, default to
hive_metastore
- metastore_
id str Unique identifier of the parent Metastore
- workspace_
id int id of the workspace for the assignment
- default_
catalog_ strname Default catalog used for this assignment, default to
hive_metastore
- metastore
Id String Unique identifier of the parent Metastore
- workspace
Id Number id of the workspace for the assignment
- default
Catalog StringName Default catalog used for this assignment, default to
hive_metastore
Outputs
All input properties are implicitly available as output properties. Additionally, the MetastoreAssignment resource produces the following output properties:
- Id string
The provider-assigned unique ID for this managed resource.
- Id string
The provider-assigned unique ID for this managed resource.
- id String
The provider-assigned unique ID for this managed resource.
- id string
The provider-assigned unique ID for this managed resource.
- id str
The provider-assigned unique ID for this managed resource.
- id String
The provider-assigned unique ID for this managed resource.
Look up Existing MetastoreAssignment Resource
Get an existing MetastoreAssignment resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: MetastoreAssignmentState, opts?: CustomResourceOptions): MetastoreAssignment
@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
default_catalog_name: Optional[str] = None,
metastore_id: Optional[str] = None,
workspace_id: Optional[int] = None) -> MetastoreAssignment
func GetMetastoreAssignment(ctx *Context, name string, id IDInput, state *MetastoreAssignmentState, opts ...ResourceOption) (*MetastoreAssignment, error)
public static MetastoreAssignment Get(string name, Input<string> id, MetastoreAssignmentState? state, CustomResourceOptions? opts = null)
public static MetastoreAssignment get(String name, Output<String> id, MetastoreAssignmentState state, CustomResourceOptions options)
Resource lookup is not supported in YAML
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Default
Catalog stringName Default catalog used for this assignment, default to
hive_metastore
- Metastore
Id string Unique identifier of the parent Metastore
- Workspace
Id int id of the workspace for the assignment
- Default
Catalog stringName Default catalog used for this assignment, default to
hive_metastore
- Metastore
Id string Unique identifier of the parent Metastore
- Workspace
Id int id of the workspace for the assignment
- default
Catalog StringName Default catalog used for this assignment, default to
hive_metastore
- metastore
Id String Unique identifier of the parent Metastore
- workspace
Id Integer id of the workspace for the assignment
- default
Catalog stringName Default catalog used for this assignment, default to
hive_metastore
- metastore
Id string Unique identifier of the parent Metastore
- workspace
Id number id of the workspace for the assignment
- default_
catalog_ strname Default catalog used for this assignment, default to
hive_metastore
- metastore_
id str Unique identifier of the parent Metastore
- workspace_
id int id of the workspace for the assignment
- default
Catalog StringName Default catalog used for this assignment, default to
hive_metastore
- metastore
Id String Unique identifier of the parent Metastore
- workspace
Id Number id of the workspace for the assignment
Package Details
- Repository
- databricks pulumi/pulumi-databricks
- License
- Apache-2.0
- Notes
This Pulumi package is based on the
databricks
Terraform Provider.