databricks logo
Databricks v1.14.0, May 23 23

databricks.MwsLogDelivery

Explore with Pulumi AI

Note Initialize provider with alias = "mws", host = "https://accounts.cloud.databricks.com" and use provider = databricks.mws for all databricks_mws_* resources.

Note This resource has an evolving API, which will change in the upcoming versions of the provider in order to simplify user experience.

Make sure you have authenticated with username and password for Accounts Console. This resource configures the delivery of the two supported log types from Databricks workspaces: billable usage logs and audit logs.

You cannot delete a log delivery configuration, but you can disable it when you no longer need it. This fact is important because there is a limit to the number of enabled log delivery configurations that you can create for an account. You can create a maximum of two enabled configurations that use the account level (no workspace filter) and two enabled configurations for every specific workspace (a workspaceId can occur in the workspace filter for two configurations). You can re-enable a disabled configuration, but the request fails if it violates the limits previously described.

Billable Usage

CSV files are delivered to <delivery_path_prefix>/billable-usage/csv/ and are named workspaceId=<workspace-id>-usageMonth=<month>.csv, which are delivered daily by overwriting the month’s CSV file for each workspace. Format of CSV file, as well as some usage examples, can be found here.

Common processing scenario is to apply cost allocation tags, that could be enforced by setting custom_tags on a cluster or through cluster policy. Report contains clusterId field, that could be joined with data from AWS cost and usage reports, that can be joined with user:ClusterId tag from AWS usage report.

import * as pulumi from "@pulumi/pulumi";
import * as databricks from "@pulumi/databricks";

const usageLogs = new databricks.MwsLogDelivery("usageLogs", {
    accountId: _var.databricks_account_id,
    credentialsId: databricks_mws_credentials.log_writer.credentials_id,
    storageConfigurationId: databricks_mws_storage_configurations.log_bucket.storage_configuration_id,
    deliveryPathPrefix: "billable-usage",
    configName: "Usage Logs",
    logType: "BILLABLE_USAGE",
    outputFormat: "CSV",
});
import pulumi
import pulumi_databricks as databricks

usage_logs = databricks.MwsLogDelivery("usageLogs",
    account_id=var["databricks_account_id"],
    credentials_id=databricks_mws_credentials["log_writer"]["credentials_id"],
    storage_configuration_id=databricks_mws_storage_configurations["log_bucket"]["storage_configuration_id"],
    delivery_path_prefix="billable-usage",
    config_name="Usage Logs",
    log_type="BILLABLE_USAGE",
    output_format="CSV")
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Databricks = Pulumi.Databricks;

return await Deployment.RunAsync(() => 
{
    var usageLogs = new Databricks.MwsLogDelivery("usageLogs", new()
    {
        AccountId = @var.Databricks_account_id,
        CredentialsId = databricks_mws_credentials.Log_writer.Credentials_id,
        StorageConfigurationId = databricks_mws_storage_configurations.Log_bucket.Storage_configuration_id,
        DeliveryPathPrefix = "billable-usage",
        ConfigName = "Usage Logs",
        LogType = "BILLABLE_USAGE",
        OutputFormat = "CSV",
    });

});
package main

import (
	"github.com/pulumi/pulumi-databricks/sdk/go/databricks"
	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)

func main() {
	pulumi.Run(func(ctx *pulumi.Context) error {
		_, err := databricks.NewMwsLogDelivery(ctx, "usageLogs", &databricks.MwsLogDeliveryArgs{
			AccountId:              pulumi.Any(_var.Databricks_account_id),
			CredentialsId:          pulumi.Any(databricks_mws_credentials.Log_writer.Credentials_id),
			StorageConfigurationId: pulumi.Any(databricks_mws_storage_configurations.Log_bucket.Storage_configuration_id),
			DeliveryPathPrefix:     pulumi.String("billable-usage"),
			ConfigName:             pulumi.String("Usage Logs"),
			LogType:                pulumi.String("BILLABLE_USAGE"),
			OutputFormat:           pulumi.String("CSV"),
		})
		if err != nil {
			return err
		}
		return nil
	})
}
package generated_program;

import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.databricks.MwsLogDelivery;
import com.pulumi.databricks.MwsLogDeliveryArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;

public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }

    public static void stack(Context ctx) {
        var usageLogs = new MwsLogDelivery("usageLogs", MwsLogDeliveryArgs.builder()        
            .accountId(var_.databricks_account_id())
            .credentialsId(databricks_mws_credentials.log_writer().credentials_id())
            .storageConfigurationId(databricks_mws_storage_configurations.log_bucket().storage_configuration_id())
            .deliveryPathPrefix("billable-usage")
            .configName("Usage Logs")
            .logType("BILLABLE_USAGE")
            .outputFormat("CSV")
            .build());

    }
}
resources:
  usageLogs:
    type: databricks:MwsLogDelivery
    properties:
      accountId: ${var.databricks_account_id}
      credentialsId: ${databricks_mws_credentials.log_writer.credentials_id}
      storageConfigurationId: ${databricks_mws_storage_configurations.log_bucket.storage_configuration_id}
      deliveryPathPrefix: billable-usage
      configName: Usage Logs
      logType: BILLABLE_USAGE
      outputFormat: CSV

Audit Logs

JSON files with static schema are delivered to <delivery_path_prefix>/workspaceId=<workspaceId>/date=<yyyy-mm-dd>/auditlogs_<internal-id>.json. Logs are available within 15 minutes of activation for audit logs. New JSON files are delivered every few minutes, potentially overwriting existing files for each workspace. Sometimes data may arrive later than 15 minutes. Databricks can overwrite the delivered log files in your bucket at any time. If a file is overwritten, the existing content remains, but there may be additional lines for more auditable events. Overwriting ensures exactly-once semantics without requiring read or delete access to your account.

import * as pulumi from "@pulumi/pulumi";
import * as databricks from "@pulumi/databricks";

const auditLogs = new databricks.MwsLogDelivery("auditLogs", {
    accountId: _var.databricks_account_id,
    credentialsId: databricks_mws_credentials.log_writer.credentials_id,
    storageConfigurationId: databricks_mws_storage_configurations.log_bucket.storage_configuration_id,
    deliveryPathPrefix: "audit-logs",
    configName: "Audit Logs",
    logType: "AUDIT_LOGS",
    outputFormat: "JSON",
});
import pulumi
import pulumi_databricks as databricks

audit_logs = databricks.MwsLogDelivery("auditLogs",
    account_id=var["databricks_account_id"],
    credentials_id=databricks_mws_credentials["log_writer"]["credentials_id"],
    storage_configuration_id=databricks_mws_storage_configurations["log_bucket"]["storage_configuration_id"],
    delivery_path_prefix="audit-logs",
    config_name="Audit Logs",
    log_type="AUDIT_LOGS",
    output_format="JSON")
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Databricks = Pulumi.Databricks;

return await Deployment.RunAsync(() => 
{
    var auditLogs = new Databricks.MwsLogDelivery("auditLogs", new()
    {
        AccountId = @var.Databricks_account_id,
        CredentialsId = databricks_mws_credentials.Log_writer.Credentials_id,
        StorageConfigurationId = databricks_mws_storage_configurations.Log_bucket.Storage_configuration_id,
        DeliveryPathPrefix = "audit-logs",
        ConfigName = "Audit Logs",
        LogType = "AUDIT_LOGS",
        OutputFormat = "JSON",
    });

});
package main

import (
	"github.com/pulumi/pulumi-databricks/sdk/go/databricks"
	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)

func main() {
	pulumi.Run(func(ctx *pulumi.Context) error {
		_, err := databricks.NewMwsLogDelivery(ctx, "auditLogs", &databricks.MwsLogDeliveryArgs{
			AccountId:              pulumi.Any(_var.Databricks_account_id),
			CredentialsId:          pulumi.Any(databricks_mws_credentials.Log_writer.Credentials_id),
			StorageConfigurationId: pulumi.Any(databricks_mws_storage_configurations.Log_bucket.Storage_configuration_id),
			DeliveryPathPrefix:     pulumi.String("audit-logs"),
			ConfigName:             pulumi.String("Audit Logs"),
			LogType:                pulumi.String("AUDIT_LOGS"),
			OutputFormat:           pulumi.String("JSON"),
		})
		if err != nil {
			return err
		}
		return nil
	})
}
package generated_program;

import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.databricks.MwsLogDelivery;
import com.pulumi.databricks.MwsLogDeliveryArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;

public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }

    public static void stack(Context ctx) {
        var auditLogs = new MwsLogDelivery("auditLogs", MwsLogDeliveryArgs.builder()        
            .accountId(var_.databricks_account_id())
            .credentialsId(databricks_mws_credentials.log_writer().credentials_id())
            .storageConfigurationId(databricks_mws_storage_configurations.log_bucket().storage_configuration_id())
            .deliveryPathPrefix("audit-logs")
            .configName("Audit Logs")
            .logType("AUDIT_LOGS")
            .outputFormat("JSON")
            .build());

    }
}
resources:
  auditLogs:
    type: databricks:MwsLogDelivery
    properties:
      accountId: ${var.databricks_account_id}
      credentialsId: ${databricks_mws_credentials.log_writer.credentials_id}
      storageConfigurationId: ${databricks_mws_storage_configurations.log_bucket.storage_configuration_id}
      deliveryPathPrefix: audit-logs
      configName: Audit Logs
      logType: AUDIT_LOGS
      outputFormat: JSON

The following resources are used in the same context:

  • Provisioning Databricks on AWS guide.
  • databricks.MwsCredentials to configure the cross-account role for creation of new workspaces within AWS.
  • databricks.MwsCustomerManagedKeys to configure KMS keys for new workspaces within AWS.
  • databricks.MwsNetworks to configure VPC & subnets for new workspaces within AWS.
  • databricks.MwsStorageConfigurations to configure root bucket new workspaces within AWS.
  • databricks.MwsWorkspaces to set up workspaces in E2 architecture on AWS.

Create MwsLogDelivery Resource

new MwsLogDelivery(name: string, args: MwsLogDeliveryArgs, opts?: CustomResourceOptions);
@overload
def MwsLogDelivery(resource_name: str,
                   opts: Optional[ResourceOptions] = None,
                   account_id: Optional[str] = None,
                   config_id: Optional[str] = None,
                   config_name: Optional[str] = None,
                   credentials_id: Optional[str] = None,
                   delivery_path_prefix: Optional[str] = None,
                   delivery_start_time: Optional[str] = None,
                   log_type: Optional[str] = None,
                   output_format: Optional[str] = None,
                   status: Optional[str] = None,
                   storage_configuration_id: Optional[str] = None,
                   workspace_ids_filters: Optional[Sequence[int]] = None)
@overload
def MwsLogDelivery(resource_name: str,
                   args: MwsLogDeliveryArgs,
                   opts: Optional[ResourceOptions] = None)
func NewMwsLogDelivery(ctx *Context, name string, args MwsLogDeliveryArgs, opts ...ResourceOption) (*MwsLogDelivery, error)
public MwsLogDelivery(string name, MwsLogDeliveryArgs args, CustomResourceOptions? opts = null)
public MwsLogDelivery(String name, MwsLogDeliveryArgs args)
public MwsLogDelivery(String name, MwsLogDeliveryArgs args, CustomResourceOptions options)
type: databricks:MwsLogDelivery
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.

name string
The unique name of the resource.
args MwsLogDeliveryArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
resource_name str
The unique name of the resource.
args MwsLogDeliveryArgs
The arguments to resource properties.
opts ResourceOptions
Bag of options to control resource's behavior.
ctx Context
Context object for the current deployment.
name string
The unique name of the resource.
args MwsLogDeliveryArgs
The arguments to resource properties.
opts ResourceOption
Bag of options to control resource's behavior.
name string
The unique name of the resource.
args MwsLogDeliveryArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
name String
The unique name of the resource.
args MwsLogDeliveryArgs
The arguments to resource properties.
options CustomResourceOptions
Bag of options to control resource's behavior.

MwsLogDelivery Resource Properties

To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

Inputs

The MwsLogDelivery resource accepts the following input properties:

AccountId string

Account Id that could be found in the bottom left corner of Accounts Console.

CredentialsId string

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

LogType string

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

OutputFormat string

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

StorageConfigurationId string

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

ConfigId string

Databricks log delivery configuration ID.

ConfigName string

The optional human-readable name of the log delivery configuration. Defaults to empty.

DeliveryPathPrefix string

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

DeliveryStartTime string

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

Status string

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

WorkspaceIdsFilters List<int>

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

AccountId string

Account Id that could be found in the bottom left corner of Accounts Console.

CredentialsId string

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

LogType string

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

OutputFormat string

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

StorageConfigurationId string

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

ConfigId string

Databricks log delivery configuration ID.

ConfigName string

The optional human-readable name of the log delivery configuration. Defaults to empty.

DeliveryPathPrefix string

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

DeliveryStartTime string

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

Status string

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

WorkspaceIdsFilters []int

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

accountId String

Account Id that could be found in the bottom left corner of Accounts Console.

credentialsId String

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

logType String

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

outputFormat String

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

storageConfigurationId String

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

configId String

Databricks log delivery configuration ID.

configName String

The optional human-readable name of the log delivery configuration. Defaults to empty.

deliveryPathPrefix String

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

deliveryStartTime String

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

status String

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

workspaceIdsFilters List<Integer>

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

accountId string

Account Id that could be found in the bottom left corner of Accounts Console.

credentialsId string

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

logType string

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

outputFormat string

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

storageConfigurationId string

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

configId string

Databricks log delivery configuration ID.

configName string

The optional human-readable name of the log delivery configuration. Defaults to empty.

deliveryPathPrefix string

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

deliveryStartTime string

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

status string

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

workspaceIdsFilters number[]

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

account_id str

Account Id that could be found in the bottom left corner of Accounts Console.

credentials_id str

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

log_type str

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

output_format str

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

storage_configuration_id str

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

config_id str

Databricks log delivery configuration ID.

config_name str

The optional human-readable name of the log delivery configuration. Defaults to empty.

delivery_path_prefix str

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

delivery_start_time str

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

status str

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

workspace_ids_filters Sequence[int]

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

accountId String

Account Id that could be found in the bottom left corner of Accounts Console.

credentialsId String

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

logType String

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

outputFormat String

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

storageConfigurationId String

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

configId String

Databricks log delivery configuration ID.

configName String

The optional human-readable name of the log delivery configuration. Defaults to empty.

deliveryPathPrefix String

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

deliveryStartTime String

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

status String

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

workspaceIdsFilters List<Number>

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

Outputs

All input properties are implicitly available as output properties. Additionally, the MwsLogDelivery resource produces the following output properties:

Id string

The provider-assigned unique ID for this managed resource.

Id string

The provider-assigned unique ID for this managed resource.

id String

The provider-assigned unique ID for this managed resource.

id string

The provider-assigned unique ID for this managed resource.

id str

The provider-assigned unique ID for this managed resource.

id String

The provider-assigned unique ID for this managed resource.

Look up Existing MwsLogDelivery Resource

Get an existing MwsLogDelivery resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

public static get(name: string, id: Input<ID>, state?: MwsLogDeliveryState, opts?: CustomResourceOptions): MwsLogDelivery
@staticmethod
def get(resource_name: str,
        id: str,
        opts: Optional[ResourceOptions] = None,
        account_id: Optional[str] = None,
        config_id: Optional[str] = None,
        config_name: Optional[str] = None,
        credentials_id: Optional[str] = None,
        delivery_path_prefix: Optional[str] = None,
        delivery_start_time: Optional[str] = None,
        log_type: Optional[str] = None,
        output_format: Optional[str] = None,
        status: Optional[str] = None,
        storage_configuration_id: Optional[str] = None,
        workspace_ids_filters: Optional[Sequence[int]] = None) -> MwsLogDelivery
func GetMwsLogDelivery(ctx *Context, name string, id IDInput, state *MwsLogDeliveryState, opts ...ResourceOption) (*MwsLogDelivery, error)
public static MwsLogDelivery Get(string name, Input<string> id, MwsLogDeliveryState? state, CustomResourceOptions? opts = null)
public static MwsLogDelivery get(String name, Output<String> id, MwsLogDeliveryState state, CustomResourceOptions options)
Resource lookup is not supported in YAML
name
The unique name of the resulting resource.
id
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
resource_name
The unique name of the resulting resource.
id
The unique provider ID of the resource to lookup.
name
The unique name of the resulting resource.
id
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
name
The unique name of the resulting resource.
id
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
name
The unique name of the resulting resource.
id
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
The following state arguments are supported:
AccountId string

Account Id that could be found in the bottom left corner of Accounts Console.

ConfigId string

Databricks log delivery configuration ID.

ConfigName string

The optional human-readable name of the log delivery configuration. Defaults to empty.

CredentialsId string

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

DeliveryPathPrefix string

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

DeliveryStartTime string

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

LogType string

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

OutputFormat string

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

Status string

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

StorageConfigurationId string

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

WorkspaceIdsFilters List<int>

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

AccountId string

Account Id that could be found in the bottom left corner of Accounts Console.

ConfigId string

Databricks log delivery configuration ID.

ConfigName string

The optional human-readable name of the log delivery configuration. Defaults to empty.

CredentialsId string

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

DeliveryPathPrefix string

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

DeliveryStartTime string

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

LogType string

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

OutputFormat string

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

Status string

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

StorageConfigurationId string

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

WorkspaceIdsFilters []int

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

accountId String

Account Id that could be found in the bottom left corner of Accounts Console.

configId String

Databricks log delivery configuration ID.

configName String

The optional human-readable name of the log delivery configuration. Defaults to empty.

credentialsId String

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

deliveryPathPrefix String

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

deliveryStartTime String

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

logType String

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

outputFormat String

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

status String

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

storageConfigurationId String

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

workspaceIdsFilters List<Integer>

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

accountId string

Account Id that could be found in the bottom left corner of Accounts Console.

configId string

Databricks log delivery configuration ID.

configName string

The optional human-readable name of the log delivery configuration. Defaults to empty.

credentialsId string

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

deliveryPathPrefix string

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

deliveryStartTime string

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

logType string

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

outputFormat string

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

status string

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

storageConfigurationId string

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

workspaceIdsFilters number[]

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

account_id str

Account Id that could be found in the bottom left corner of Accounts Console.

config_id str

Databricks log delivery configuration ID.

config_name str

The optional human-readable name of the log delivery configuration. Defaults to empty.

credentials_id str

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

delivery_path_prefix str

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

delivery_start_time str

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

log_type str

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

output_format str

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

status str

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

storage_configuration_id str

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

workspace_ids_filters Sequence[int]

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

accountId String

Account Id that could be found in the bottom left corner of Accounts Console.

configId String

Databricks log delivery configuration ID.

configName String

The optional human-readable name of the log delivery configuration. Defaults to empty.

credentialsId String

The ID for a Databricks credential configuration that represents the AWS IAM role with policy and trust relationship as described in the main billable usage documentation page.

deliveryPathPrefix String

Defaults to empty, which means that logs are delivered to the root of the bucket. The value must be a valid S3 object key. It must not start or end with a slash character.

deliveryStartTime String

The optional start month and year for delivery, specified in YYYY-MM format. Defaults to current year and month. Usage is not available before 2019-03.

logType String

The type of log delivery. BILLABLE_USAGE and AUDIT_LOGS are supported.

outputFormat String

The file type of log delivery. Currently CSV (for BILLABLE_USAGE) and JSON (for AUDIT_LOGS) are supported.

status String

Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update.

storageConfigurationId String

The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page.

workspaceIdsFilters List<Number>

By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces.

Import

-> Note Importing this resource is not currently supported.

Package Details

Repository
databricks pulumi/pulumi-databricks
License
Apache-2.0
Notes

This Pulumi package is based on the databricks Terraform Provider.