1. Packages
  2. Google Cloud (GCP) Classic
  3. API Docs
  4. bigquery
  5. Connection
Google Cloud Classic v7.16.0 published on Wednesday, Mar 27, 2024 by Pulumi

gcp.bigquery.Connection

Explore with Pulumi AI

gcp logo
Google Cloud Classic v7.16.0 published on Wednesday, Mar 27, 2024 by Pulumi

    A connection allows BigQuery connections to external data sources..

    To get more information about Connection, see:

    Example Usage

    Bigquery Connection Cloud Resource

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    
    const connection = new gcp.bigquery.Connection("connection", {
        connectionId: "my-connection",
        location: "US",
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        cloudResource: {},
    });
    
    import pulumi
    import pulumi_gcp as gcp
    
    connection = gcp.bigquery.Connection("connection",
        connection_id="my-connection",
        location="US",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        cloud_resource=gcp.bigquery.ConnectionCloudResourceArgs())
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		_, err := bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			ConnectionId:  pulumi.String("my-connection"),
    			Location:      pulumi.String("US"),
    			FriendlyName:  pulumi.String("πŸ‘‹"),
    			Description:   pulumi.String("a riveting description"),
    			CloudResource: nil,
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    
    return await Deployment.RunAsync(() => 
    {
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            ConnectionId = "my-connection",
            Location = "US",
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            CloudResource = null,
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionCloudResourceArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .connectionId("my-connection")
                .location("US")
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .cloudResource()
                .build());
    
        }
    }
    
    resources:
      connection:
        type: gcp:bigquery:Connection
        properties:
          connectionId: my-connection
          location: US
          friendlyName: "\U0001F44B"
          description: a riveting description
          cloudResource: {}
    

    Bigquery Connection Basic

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    import * as random from "@pulumi/random";
    
    const instance = new gcp.sql.DatabaseInstance("instance", {
        name: "my-database-instance",
        databaseVersion: "POSTGRES_11",
        region: "us-central1",
        settings: {
            tier: "db-f1-micro",
        },
        deletionProtection: true,
    });
    const db = new gcp.sql.Database("db", {
        instance: instance.name,
        name: "db",
    });
    const pwd = new random.RandomPassword("pwd", {
        length: 16,
        special: false,
    });
    const user = new gcp.sql.User("user", {
        name: "user",
        instance: instance.name,
        password: pwd.result,
    });
    const connection = new gcp.bigquery.Connection("connection", {
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        location: "US",
        cloudSql: {
            instanceId: instance.connectionName,
            database: db.name,
            type: "POSTGRES",
            credential: {
                username: user.name,
                password: user.password,
            },
        },
    });
    
    import pulumi
    import pulumi_gcp as gcp
    import pulumi_random as random
    
    instance = gcp.sql.DatabaseInstance("instance",
        name="my-database-instance",
        database_version="POSTGRES_11",
        region="us-central1",
        settings=gcp.sql.DatabaseInstanceSettingsArgs(
            tier="db-f1-micro",
        ),
        deletion_protection=True)
    db = gcp.sql.Database("db",
        instance=instance.name,
        name="db")
    pwd = random.RandomPassword("pwd",
        length=16,
        special=False)
    user = gcp.sql.User("user",
        name="user",
        instance=instance.name,
        password=pwd.result)
    connection = gcp.bigquery.Connection("connection",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        location="US",
        cloud_sql=gcp.bigquery.ConnectionCloudSqlArgs(
            instance_id=instance.connection_name,
            database=db.name,
            type="POSTGRES",
            credential=gcp.bigquery.ConnectionCloudSqlCredentialArgs(
                username=user.name,
                password=user.password,
            ),
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/sql"
    	"github.com/pulumi/pulumi-random/sdk/v4/go/random"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		instance, err := sql.NewDatabaseInstance(ctx, "instance", &sql.DatabaseInstanceArgs{
    			Name:            pulumi.String("my-database-instance"),
    			DatabaseVersion: pulumi.String("POSTGRES_11"),
    			Region:          pulumi.String("us-central1"),
    			Settings: &sql.DatabaseInstanceSettingsArgs{
    				Tier: pulumi.String("db-f1-micro"),
    			},
    			DeletionProtection: pulumi.Bool(true),
    		})
    		if err != nil {
    			return err
    		}
    		db, err := sql.NewDatabase(ctx, "db", &sql.DatabaseArgs{
    			Instance: instance.Name,
    			Name:     pulumi.String("db"),
    		})
    		if err != nil {
    			return err
    		}
    		pwd, err := random.NewRandomPassword(ctx, "pwd", &random.RandomPasswordArgs{
    			Length:  pulumi.Int(16),
    			Special: pulumi.Bool(false),
    		})
    		if err != nil {
    			return err
    		}
    		user, err := sql.NewUser(ctx, "user", &sql.UserArgs{
    			Name:     pulumi.String("user"),
    			Instance: instance.Name,
    			Password: pwd.Result,
    		})
    		if err != nil {
    			return err
    		}
    		_, err = bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			FriendlyName: pulumi.String("πŸ‘‹"),
    			Description:  pulumi.String("a riveting description"),
    			Location:     pulumi.String("US"),
    			CloudSql: &bigquery.ConnectionCloudSqlArgs{
    				InstanceId: instance.ConnectionName,
    				Database:   db.Name,
    				Type:       pulumi.String("POSTGRES"),
    				Credential: &bigquery.ConnectionCloudSqlCredentialArgs{
    					Username: user.Name,
    					Password: user.Password,
    				},
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    using Random = Pulumi.Random;
    
    return await Deployment.RunAsync(() => 
    {
        var instance = new Gcp.Sql.DatabaseInstance("instance", new()
        {
            Name = "my-database-instance",
            DatabaseVersion = "POSTGRES_11",
            Region = "us-central1",
            Settings = new Gcp.Sql.Inputs.DatabaseInstanceSettingsArgs
            {
                Tier = "db-f1-micro",
            },
            DeletionProtection = true,
        });
    
        var db = new Gcp.Sql.Database("db", new()
        {
            Instance = instance.Name,
            Name = "db",
        });
    
        var pwd = new Random.RandomPassword("pwd", new()
        {
            Length = 16,
            Special = false,
        });
    
        var user = new Gcp.Sql.User("user", new()
        {
            Name = "user",
            Instance = instance.Name,
            Password = pwd.Result,
        });
    
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            Location = "US",
            CloudSql = new Gcp.BigQuery.Inputs.ConnectionCloudSqlArgs
            {
                InstanceId = instance.ConnectionName,
                Database = db.Name,
                Type = "POSTGRES",
                Credential = new Gcp.BigQuery.Inputs.ConnectionCloudSqlCredentialArgs
                {
                    Username = user.Name,
                    Password = user.Password,
                },
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.sql.DatabaseInstance;
    import com.pulumi.gcp.sql.DatabaseInstanceArgs;
    import com.pulumi.gcp.sql.inputs.DatabaseInstanceSettingsArgs;
    import com.pulumi.gcp.sql.Database;
    import com.pulumi.gcp.sql.DatabaseArgs;
    import com.pulumi.random.RandomPassword;
    import com.pulumi.random.RandomPasswordArgs;
    import com.pulumi.gcp.sql.User;
    import com.pulumi.gcp.sql.UserArgs;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionCloudSqlArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionCloudSqlCredentialArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var instance = new DatabaseInstance("instance", DatabaseInstanceArgs.builder()        
                .name("my-database-instance")
                .databaseVersion("POSTGRES_11")
                .region("us-central1")
                .settings(DatabaseInstanceSettingsArgs.builder()
                    .tier("db-f1-micro")
                    .build())
                .deletionProtection("true")
                .build());
    
            var db = new Database("db", DatabaseArgs.builder()        
                .instance(instance.name())
                .name("db")
                .build());
    
            var pwd = new RandomPassword("pwd", RandomPasswordArgs.builder()        
                .length(16)
                .special(false)
                .build());
    
            var user = new User("user", UserArgs.builder()        
                .name("user")
                .instance(instance.name())
                .password(pwd.result())
                .build());
    
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .location("US")
                .cloudSql(ConnectionCloudSqlArgs.builder()
                    .instanceId(instance.connectionName())
                    .database(db.name())
                    .type("POSTGRES")
                    .credential(ConnectionCloudSqlCredentialArgs.builder()
                        .username(user.name())
                        .password(user.password())
                        .build())
                    .build())
                .build());
    
        }
    }
    
    resources:
      instance:
        type: gcp:sql:DatabaseInstance
        properties:
          name: my-database-instance
          databaseVersion: POSTGRES_11
          region: us-central1
          settings:
            tier: db-f1-micro
          deletionProtection: 'true'
      db:
        type: gcp:sql:Database
        properties:
          instance: ${instance.name}
          name: db
      pwd:
        type: random:RandomPassword
        properties:
          length: 16
          special: false
      user:
        type: gcp:sql:User
        properties:
          name: user
          instance: ${instance.name}
          password: ${pwd.result}
      connection:
        type: gcp:bigquery:Connection
        properties:
          friendlyName: "\U0001F44B"
          description: a riveting description
          location: US
          cloudSql:
            instanceId: ${instance.connectionName}
            database: ${db.name}
            type: POSTGRES
            credential:
              username: ${user.name}
              password: ${user.password}
    

    Bigquery Connection Full

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    import * as random from "@pulumi/random";
    
    const instance = new gcp.sql.DatabaseInstance("instance", {
        name: "my-database-instance",
        databaseVersion: "POSTGRES_11",
        region: "us-central1",
        settings: {
            tier: "db-f1-micro",
        },
        deletionProtection: true,
    });
    const db = new gcp.sql.Database("db", {
        instance: instance.name,
        name: "db",
    });
    const pwd = new random.RandomPassword("pwd", {
        length: 16,
        special: false,
    });
    const user = new gcp.sql.User("user", {
        name: "user",
        instance: instance.name,
        password: pwd.result,
    });
    const connection = new gcp.bigquery.Connection("connection", {
        connectionId: "my-connection",
        location: "US",
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        cloudSql: {
            instanceId: instance.connectionName,
            database: db.name,
            type: "POSTGRES",
            credential: {
                username: user.name,
                password: user.password,
            },
        },
    });
    
    import pulumi
    import pulumi_gcp as gcp
    import pulumi_random as random
    
    instance = gcp.sql.DatabaseInstance("instance",
        name="my-database-instance",
        database_version="POSTGRES_11",
        region="us-central1",
        settings=gcp.sql.DatabaseInstanceSettingsArgs(
            tier="db-f1-micro",
        ),
        deletion_protection=True)
    db = gcp.sql.Database("db",
        instance=instance.name,
        name="db")
    pwd = random.RandomPassword("pwd",
        length=16,
        special=False)
    user = gcp.sql.User("user",
        name="user",
        instance=instance.name,
        password=pwd.result)
    connection = gcp.bigquery.Connection("connection",
        connection_id="my-connection",
        location="US",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        cloud_sql=gcp.bigquery.ConnectionCloudSqlArgs(
            instance_id=instance.connection_name,
            database=db.name,
            type="POSTGRES",
            credential=gcp.bigquery.ConnectionCloudSqlCredentialArgs(
                username=user.name,
                password=user.password,
            ),
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/sql"
    	"github.com/pulumi/pulumi-random/sdk/v4/go/random"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		instance, err := sql.NewDatabaseInstance(ctx, "instance", &sql.DatabaseInstanceArgs{
    			Name:            pulumi.String("my-database-instance"),
    			DatabaseVersion: pulumi.String("POSTGRES_11"),
    			Region:          pulumi.String("us-central1"),
    			Settings: &sql.DatabaseInstanceSettingsArgs{
    				Tier: pulumi.String("db-f1-micro"),
    			},
    			DeletionProtection: pulumi.Bool(true),
    		})
    		if err != nil {
    			return err
    		}
    		db, err := sql.NewDatabase(ctx, "db", &sql.DatabaseArgs{
    			Instance: instance.Name,
    			Name:     pulumi.String("db"),
    		})
    		if err != nil {
    			return err
    		}
    		pwd, err := random.NewRandomPassword(ctx, "pwd", &random.RandomPasswordArgs{
    			Length:  pulumi.Int(16),
    			Special: pulumi.Bool(false),
    		})
    		if err != nil {
    			return err
    		}
    		user, err := sql.NewUser(ctx, "user", &sql.UserArgs{
    			Name:     pulumi.String("user"),
    			Instance: instance.Name,
    			Password: pwd.Result,
    		})
    		if err != nil {
    			return err
    		}
    		_, err = bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			ConnectionId: pulumi.String("my-connection"),
    			Location:     pulumi.String("US"),
    			FriendlyName: pulumi.String("πŸ‘‹"),
    			Description:  pulumi.String("a riveting description"),
    			CloudSql: &bigquery.ConnectionCloudSqlArgs{
    				InstanceId: instance.ConnectionName,
    				Database:   db.Name,
    				Type:       pulumi.String("POSTGRES"),
    				Credential: &bigquery.ConnectionCloudSqlCredentialArgs{
    					Username: user.Name,
    					Password: user.Password,
    				},
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    using Random = Pulumi.Random;
    
    return await Deployment.RunAsync(() => 
    {
        var instance = new Gcp.Sql.DatabaseInstance("instance", new()
        {
            Name = "my-database-instance",
            DatabaseVersion = "POSTGRES_11",
            Region = "us-central1",
            Settings = new Gcp.Sql.Inputs.DatabaseInstanceSettingsArgs
            {
                Tier = "db-f1-micro",
            },
            DeletionProtection = true,
        });
    
        var db = new Gcp.Sql.Database("db", new()
        {
            Instance = instance.Name,
            Name = "db",
        });
    
        var pwd = new Random.RandomPassword("pwd", new()
        {
            Length = 16,
            Special = false,
        });
    
        var user = new Gcp.Sql.User("user", new()
        {
            Name = "user",
            Instance = instance.Name,
            Password = pwd.Result,
        });
    
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            ConnectionId = "my-connection",
            Location = "US",
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            CloudSql = new Gcp.BigQuery.Inputs.ConnectionCloudSqlArgs
            {
                InstanceId = instance.ConnectionName,
                Database = db.Name,
                Type = "POSTGRES",
                Credential = new Gcp.BigQuery.Inputs.ConnectionCloudSqlCredentialArgs
                {
                    Username = user.Name,
                    Password = user.Password,
                },
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.sql.DatabaseInstance;
    import com.pulumi.gcp.sql.DatabaseInstanceArgs;
    import com.pulumi.gcp.sql.inputs.DatabaseInstanceSettingsArgs;
    import com.pulumi.gcp.sql.Database;
    import com.pulumi.gcp.sql.DatabaseArgs;
    import com.pulumi.random.RandomPassword;
    import com.pulumi.random.RandomPasswordArgs;
    import com.pulumi.gcp.sql.User;
    import com.pulumi.gcp.sql.UserArgs;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionCloudSqlArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionCloudSqlCredentialArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var instance = new DatabaseInstance("instance", DatabaseInstanceArgs.builder()        
                .name("my-database-instance")
                .databaseVersion("POSTGRES_11")
                .region("us-central1")
                .settings(DatabaseInstanceSettingsArgs.builder()
                    .tier("db-f1-micro")
                    .build())
                .deletionProtection("true")
                .build());
    
            var db = new Database("db", DatabaseArgs.builder()        
                .instance(instance.name())
                .name("db")
                .build());
    
            var pwd = new RandomPassword("pwd", RandomPasswordArgs.builder()        
                .length(16)
                .special(false)
                .build());
    
            var user = new User("user", UserArgs.builder()        
                .name("user")
                .instance(instance.name())
                .password(pwd.result())
                .build());
    
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .connectionId("my-connection")
                .location("US")
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .cloudSql(ConnectionCloudSqlArgs.builder()
                    .instanceId(instance.connectionName())
                    .database(db.name())
                    .type("POSTGRES")
                    .credential(ConnectionCloudSqlCredentialArgs.builder()
                        .username(user.name())
                        .password(user.password())
                        .build())
                    .build())
                .build());
    
        }
    }
    
    resources:
      instance:
        type: gcp:sql:DatabaseInstance
        properties:
          name: my-database-instance
          databaseVersion: POSTGRES_11
          region: us-central1
          settings:
            tier: db-f1-micro
          deletionProtection: 'true'
      db:
        type: gcp:sql:Database
        properties:
          instance: ${instance.name}
          name: db
      pwd:
        type: random:RandomPassword
        properties:
          length: 16
          special: false
      user:
        type: gcp:sql:User
        properties:
          name: user
          instance: ${instance.name}
          password: ${pwd.result}
      connection:
        type: gcp:bigquery:Connection
        properties:
          connectionId: my-connection
          location: US
          friendlyName: "\U0001F44B"
          description: a riveting description
          cloudSql:
            instanceId: ${instance.connectionName}
            database: ${db.name}
            type: POSTGRES
            credential:
              username: ${user.name}
              password: ${user.password}
    

    Bigquery Connection Aws

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    
    const connection = new gcp.bigquery.Connection("connection", {
        connectionId: "my-connection",
        location: "aws-us-east-1",
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        aws: {
            accessRole: {
                iamRoleId: "arn:aws:iam::999999999999:role/omnirole",
            },
        },
    });
    
    import pulumi
    import pulumi_gcp as gcp
    
    connection = gcp.bigquery.Connection("connection",
        connection_id="my-connection",
        location="aws-us-east-1",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        aws=gcp.bigquery.ConnectionAwsArgs(
            access_role=gcp.bigquery.ConnectionAwsAccessRoleArgs(
                iam_role_id="arn:aws:iam::999999999999:role/omnirole",
            ),
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		_, err := bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			ConnectionId: pulumi.String("my-connection"),
    			Location:     pulumi.String("aws-us-east-1"),
    			FriendlyName: pulumi.String("πŸ‘‹"),
    			Description:  pulumi.String("a riveting description"),
    			Aws: &bigquery.ConnectionAwsArgs{
    				AccessRole: &bigquery.ConnectionAwsAccessRoleArgs{
    					IamRoleId: pulumi.String("arn:aws:iam::999999999999:role/omnirole"),
    				},
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    
    return await Deployment.RunAsync(() => 
    {
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            ConnectionId = "my-connection",
            Location = "aws-us-east-1",
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            Aws = new Gcp.BigQuery.Inputs.ConnectionAwsArgs
            {
                AccessRole = new Gcp.BigQuery.Inputs.ConnectionAwsAccessRoleArgs
                {
                    IamRoleId = "arn:aws:iam::999999999999:role/omnirole",
                },
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionAwsArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionAwsAccessRoleArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .connectionId("my-connection")
                .location("aws-us-east-1")
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .aws(ConnectionAwsArgs.builder()
                    .accessRole(ConnectionAwsAccessRoleArgs.builder()
                        .iamRoleId("arn:aws:iam::999999999999:role/omnirole")
                        .build())
                    .build())
                .build());
    
        }
    }
    
    resources:
      connection:
        type: gcp:bigquery:Connection
        properties:
          connectionId: my-connection
          location: aws-us-east-1
          friendlyName: "\U0001F44B"
          description: a riveting description
          aws:
            accessRole:
              iamRoleId: arn:aws:iam::999999999999:role/omnirole
    

    Bigquery Connection Azure

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    
    const connection = new gcp.bigquery.Connection("connection", {
        connectionId: "my-connection",
        location: "azure-eastus2",
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        azure: {
            customerTenantId: "customer-tenant-id",
            federatedApplicationClientId: "b43eeeee-eeee-eeee-eeee-a480155501ce",
        },
    });
    
    import pulumi
    import pulumi_gcp as gcp
    
    connection = gcp.bigquery.Connection("connection",
        connection_id="my-connection",
        location="azure-eastus2",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        azure=gcp.bigquery.ConnectionAzureArgs(
            customer_tenant_id="customer-tenant-id",
            federated_application_client_id="b43eeeee-eeee-eeee-eeee-a480155501ce",
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		_, err := bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			ConnectionId: pulumi.String("my-connection"),
    			Location:     pulumi.String("azure-eastus2"),
    			FriendlyName: pulumi.String("πŸ‘‹"),
    			Description:  pulumi.String("a riveting description"),
    			Azure: &bigquery.ConnectionAzureArgs{
    				CustomerTenantId:             pulumi.String("customer-tenant-id"),
    				FederatedApplicationClientId: pulumi.String("b43eeeee-eeee-eeee-eeee-a480155501ce"),
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    
    return await Deployment.RunAsync(() => 
    {
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            ConnectionId = "my-connection",
            Location = "azure-eastus2",
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            Azure = new Gcp.BigQuery.Inputs.ConnectionAzureArgs
            {
                CustomerTenantId = "customer-tenant-id",
                FederatedApplicationClientId = "b43eeeee-eeee-eeee-eeee-a480155501ce",
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionAzureArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .connectionId("my-connection")
                .location("azure-eastus2")
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .azure(ConnectionAzureArgs.builder()
                    .customerTenantId("customer-tenant-id")
                    .federatedApplicationClientId("b43eeeee-eeee-eeee-eeee-a480155501ce")
                    .build())
                .build());
    
        }
    }
    
    resources:
      connection:
        type: gcp:bigquery:Connection
        properties:
          connectionId: my-connection
          location: azure-eastus2
          friendlyName: "\U0001F44B"
          description: a riveting description
          azure:
            customerTenantId: customer-tenant-id
            federatedApplicationClientId: b43eeeee-eeee-eeee-eeee-a480155501ce
    

    Bigquery Connection Cloudspanner

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    
    const connection = new gcp.bigquery.Connection("connection", {
        connectionId: "my-connection",
        location: "US",
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        cloudSpanner: {
            database: "projects/project/instances/instance/databases/database",
            databaseRole: "database_role",
        },
    });
    
    import pulumi
    import pulumi_gcp as gcp
    
    connection = gcp.bigquery.Connection("connection",
        connection_id="my-connection",
        location="US",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        cloud_spanner=gcp.bigquery.ConnectionCloudSpannerArgs(
            database="projects/project/instances/instance/databases/database",
            database_role="database_role",
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		_, err := bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			ConnectionId: pulumi.String("my-connection"),
    			Location:     pulumi.String("US"),
    			FriendlyName: pulumi.String("πŸ‘‹"),
    			Description:  pulumi.String("a riveting description"),
    			CloudSpanner: &bigquery.ConnectionCloudSpannerArgs{
    				Database:     pulumi.String("projects/project/instances/instance/databases/database"),
    				DatabaseRole: pulumi.String("database_role"),
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    
    return await Deployment.RunAsync(() => 
    {
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            ConnectionId = "my-connection",
            Location = "US",
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            CloudSpanner = new Gcp.BigQuery.Inputs.ConnectionCloudSpannerArgs
            {
                Database = "projects/project/instances/instance/databases/database",
                DatabaseRole = "database_role",
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionCloudSpannerArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .connectionId("my-connection")
                .location("US")
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .cloudSpanner(ConnectionCloudSpannerArgs.builder()
                    .database("projects/project/instances/instance/databases/database")
                    .databaseRole("database_role")
                    .build())
                .build());
    
        }
    }
    
    resources:
      connection:
        type: gcp:bigquery:Connection
        properties:
          connectionId: my-connection
          location: US
          friendlyName: "\U0001F44B"
          description: a riveting description
          cloudSpanner:
            database: projects/project/instances/instance/databases/database
            databaseRole: database_role
    

    Bigquery Connection Cloudspanner Databoost

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    
    const connection = new gcp.bigquery.Connection("connection", {
        connectionId: "my-connection",
        location: "US",
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        cloudSpanner: {
            database: "projects/project/instances/instance/databases/database",
            useParallelism: true,
            useDataBoost: true,
            maxParallelism: 100,
        },
    });
    
    import pulumi
    import pulumi_gcp as gcp
    
    connection = gcp.bigquery.Connection("connection",
        connection_id="my-connection",
        location="US",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        cloud_spanner=gcp.bigquery.ConnectionCloudSpannerArgs(
            database="projects/project/instances/instance/databases/database",
            use_parallelism=True,
            use_data_boost=True,
            max_parallelism=100,
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		_, err := bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			ConnectionId: pulumi.String("my-connection"),
    			Location:     pulumi.String("US"),
    			FriendlyName: pulumi.String("πŸ‘‹"),
    			Description:  pulumi.String("a riveting description"),
    			CloudSpanner: &bigquery.ConnectionCloudSpannerArgs{
    				Database:       pulumi.String("projects/project/instances/instance/databases/database"),
    				UseParallelism: pulumi.Bool(true),
    				UseDataBoost:   pulumi.Bool(true),
    				MaxParallelism: pulumi.Int(100),
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    
    return await Deployment.RunAsync(() => 
    {
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            ConnectionId = "my-connection",
            Location = "US",
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            CloudSpanner = new Gcp.BigQuery.Inputs.ConnectionCloudSpannerArgs
            {
                Database = "projects/project/instances/instance/databases/database",
                UseParallelism = true,
                UseDataBoost = true,
                MaxParallelism = 100,
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionCloudSpannerArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .connectionId("my-connection")
                .location("US")
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .cloudSpanner(ConnectionCloudSpannerArgs.builder()
                    .database("projects/project/instances/instance/databases/database")
                    .useParallelism(true)
                    .useDataBoost(true)
                    .maxParallelism(100)
                    .build())
                .build());
    
        }
    }
    
    resources:
      connection:
        type: gcp:bigquery:Connection
        properties:
          connectionId: my-connection
          location: US
          friendlyName: "\U0001F44B"
          description: a riveting description
          cloudSpanner:
            database: projects/project/instances/instance/databases/database
            useParallelism: true
            useDataBoost: true
            maxParallelism: 100
    

    Bigquery Connection Spark

    import * as pulumi from "@pulumi/pulumi";
    import * as gcp from "@pulumi/gcp";
    
    const basic = new gcp.dataproc.Cluster("basic", {
        name: "my-connection",
        region: "us-central1",
        clusterConfig: {
            softwareConfig: {
                overrideProperties: {
                    "dataproc:dataproc.allow.zero.workers": "true",
                },
            },
            masterConfig: {
                numInstances: 1,
                machineType: "e2-standard-2",
                diskConfig: {
                    bootDiskSizeGb: 35,
                },
            },
        },
    });
    const connection = new gcp.bigquery.Connection("connection", {
        connectionId: "my-connection",
        location: "US",
        friendlyName: "πŸ‘‹",
        description: "a riveting description",
        spark: {
            sparkHistoryServerConfig: {
                dataprocCluster: basic.id,
            },
        },
    });
    
    import pulumi
    import pulumi_gcp as gcp
    
    basic = gcp.dataproc.Cluster("basic",
        name="my-connection",
        region="us-central1",
        cluster_config=gcp.dataproc.ClusterClusterConfigArgs(
            software_config=gcp.dataproc.ClusterClusterConfigSoftwareConfigArgs(
                override_properties={
                    "dataproc:dataproc.allow.zero.workers": "true",
                },
            ),
            master_config=gcp.dataproc.ClusterClusterConfigMasterConfigArgs(
                num_instances=1,
                machine_type="e2-standard-2",
                disk_config=gcp.dataproc.ClusterClusterConfigMasterConfigDiskConfigArgs(
                    boot_disk_size_gb=35,
                ),
            ),
        ))
    connection = gcp.bigquery.Connection("connection",
        connection_id="my-connection",
        location="US",
        friendly_name="πŸ‘‹",
        description="a riveting description",
        spark=gcp.bigquery.ConnectionSparkArgs(
            spark_history_server_config=gcp.bigquery.ConnectionSparkSparkHistoryServerConfigArgs(
                dataproc_cluster=basic.id,
            ),
        ))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/bigquery"
    	"github.com/pulumi/pulumi-gcp/sdk/v7/go/gcp/dataproc"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		basic, err := dataproc.NewCluster(ctx, "basic", &dataproc.ClusterArgs{
    			Name:   pulumi.String("my-connection"),
    			Region: pulumi.String("us-central1"),
    			ClusterConfig: &dataproc.ClusterClusterConfigArgs{
    				SoftwareConfig: &dataproc.ClusterClusterConfigSoftwareConfigArgs{
    					OverrideProperties: pulumi.StringMap{
    						"dataproc:dataproc.allow.zero.workers": pulumi.String("true"),
    					},
    				},
    				MasterConfig: &dataproc.ClusterClusterConfigMasterConfigArgs{
    					NumInstances: pulumi.Int(1),
    					MachineType:  pulumi.String("e2-standard-2"),
    					DiskConfig: &dataproc.ClusterClusterConfigMasterConfigDiskConfigArgs{
    						BootDiskSizeGb: pulumi.Int(35),
    					},
    				},
    			},
    		})
    		if err != nil {
    			return err
    		}
    		_, err = bigquery.NewConnection(ctx, "connection", &bigquery.ConnectionArgs{
    			ConnectionId: pulumi.String("my-connection"),
    			Location:     pulumi.String("US"),
    			FriendlyName: pulumi.String("πŸ‘‹"),
    			Description:  pulumi.String("a riveting description"),
    			Spark: &bigquery.ConnectionSparkArgs{
    				SparkHistoryServerConfig: &bigquery.ConnectionSparkSparkHistoryServerConfigArgs{
    					DataprocCluster: basic.ID(),
    				},
    			},
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Gcp = Pulumi.Gcp;
    
    return await Deployment.RunAsync(() => 
    {
        var basic = new Gcp.Dataproc.Cluster("basic", new()
        {
            Name = "my-connection",
            Region = "us-central1",
            ClusterConfig = new Gcp.Dataproc.Inputs.ClusterClusterConfigArgs
            {
                SoftwareConfig = new Gcp.Dataproc.Inputs.ClusterClusterConfigSoftwareConfigArgs
                {
                    OverrideProperties = 
                    {
                        { "dataproc:dataproc.allow.zero.workers", "true" },
                    },
                },
                MasterConfig = new Gcp.Dataproc.Inputs.ClusterClusterConfigMasterConfigArgs
                {
                    NumInstances = 1,
                    MachineType = "e2-standard-2",
                    DiskConfig = new Gcp.Dataproc.Inputs.ClusterClusterConfigMasterConfigDiskConfigArgs
                    {
                        BootDiskSizeGb = 35,
                    },
                },
            },
        });
    
        var connection = new Gcp.BigQuery.Connection("connection", new()
        {
            ConnectionId = "my-connection",
            Location = "US",
            FriendlyName = "πŸ‘‹",
            Description = "a riveting description",
            Spark = new Gcp.BigQuery.Inputs.ConnectionSparkArgs
            {
                SparkHistoryServerConfig = new Gcp.BigQuery.Inputs.ConnectionSparkSparkHistoryServerConfigArgs
                {
                    DataprocCluster = basic.Id,
                },
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.gcp.dataproc.Cluster;
    import com.pulumi.gcp.dataproc.ClusterArgs;
    import com.pulumi.gcp.dataproc.inputs.ClusterClusterConfigArgs;
    import com.pulumi.gcp.dataproc.inputs.ClusterClusterConfigSoftwareConfigArgs;
    import com.pulumi.gcp.dataproc.inputs.ClusterClusterConfigMasterConfigArgs;
    import com.pulumi.gcp.dataproc.inputs.ClusterClusterConfigMasterConfigDiskConfigArgs;
    import com.pulumi.gcp.bigquery.Connection;
    import com.pulumi.gcp.bigquery.ConnectionArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionSparkArgs;
    import com.pulumi.gcp.bigquery.inputs.ConnectionSparkSparkHistoryServerConfigArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var basic = new Cluster("basic", ClusterArgs.builder()        
                .name("my-connection")
                .region("us-central1")
                .clusterConfig(ClusterClusterConfigArgs.builder()
                    .softwareConfig(ClusterClusterConfigSoftwareConfigArgs.builder()
                        .overrideProperties(Map.of("dataproc:dataproc.allow.zero.workers", "true"))
                        .build())
                    .masterConfig(ClusterClusterConfigMasterConfigArgs.builder()
                        .numInstances(1)
                        .machineType("e2-standard-2")
                        .diskConfig(ClusterClusterConfigMasterConfigDiskConfigArgs.builder()
                            .bootDiskSizeGb(35)
                            .build())
                        .build())
                    .build())
                .build());
    
            var connection = new Connection("connection", ConnectionArgs.builder()        
                .connectionId("my-connection")
                .location("US")
                .friendlyName("πŸ‘‹")
                .description("a riveting description")
                .spark(ConnectionSparkArgs.builder()
                    .sparkHistoryServerConfig(ConnectionSparkSparkHistoryServerConfigArgs.builder()
                        .dataprocCluster(basic.id())
                        .build())
                    .build())
                .build());
    
        }
    }
    
    resources:
      connection:
        type: gcp:bigquery:Connection
        properties:
          connectionId: my-connection
          location: US
          friendlyName: "\U0001F44B"
          description: a riveting description
          spark:
            sparkHistoryServerConfig:
              dataprocCluster: ${basic.id}
      basic:
        type: gcp:dataproc:Cluster
        properties:
          name: my-connection
          region: us-central1
          clusterConfig:
            softwareConfig:
              overrideProperties:
                dataproc:dataproc.allow.zero.workers: 'true'
            masterConfig:
              numInstances: 1
              machineType: e2-standard-2
              diskConfig:
                bootDiskSizeGb: 35
    

    Create Connection Resource

    new Connection(name: string, args?: ConnectionArgs, opts?: CustomResourceOptions);
    @overload
    def Connection(resource_name: str,
                   opts: Optional[ResourceOptions] = None,
                   aws: Optional[ConnectionAwsArgs] = None,
                   azure: Optional[ConnectionAzureArgs] = None,
                   cloud_resource: Optional[ConnectionCloudResourceArgs] = None,
                   cloud_spanner: Optional[ConnectionCloudSpannerArgs] = None,
                   cloud_sql: Optional[ConnectionCloudSqlArgs] = None,
                   connection_id: Optional[str] = None,
                   description: Optional[str] = None,
                   friendly_name: Optional[str] = None,
                   location: Optional[str] = None,
                   project: Optional[str] = None,
                   spark: Optional[ConnectionSparkArgs] = None)
    @overload
    def Connection(resource_name: str,
                   args: Optional[ConnectionArgs] = None,
                   opts: Optional[ResourceOptions] = None)
    func NewConnection(ctx *Context, name string, args *ConnectionArgs, opts ...ResourceOption) (*Connection, error)
    public Connection(string name, ConnectionArgs? args = null, CustomResourceOptions? opts = null)
    public Connection(String name, ConnectionArgs args)
    public Connection(String name, ConnectionArgs args, CustomResourceOptions options)
    
    type: gcp:bigquery:Connection
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    
    name string
    The unique name of the resource.
    args ConnectionArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args ConnectionArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args ConnectionArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args ConnectionArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args ConnectionArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    Connection Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    The Connection resource accepts the following input properties:

    Aws ConnectionAws
    Connection properties specific to Amazon Web Services. Structure is documented below.
    Azure ConnectionAzure
    Container for connection properties specific to Azure. Structure is documented below.
    CloudResource ConnectionCloudResource
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    CloudSpanner ConnectionCloudSpanner
    Connection properties specific to Cloud Spanner Structure is documented below.
    CloudSql ConnectionCloudSql
    Connection properties specific to the Cloud SQL. Structure is documented below.
    ConnectionId string
    Optional connection id that should be assigned to the created connection.
    Description string
    A descriptive description for the connection
    FriendlyName string
    A descriptive name for the connection
    Location string
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    Project string
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    Spark ConnectionSpark
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    Aws ConnectionAwsArgs
    Connection properties specific to Amazon Web Services. Structure is documented below.
    Azure ConnectionAzureArgs
    Container for connection properties specific to Azure. Structure is documented below.
    CloudResource ConnectionCloudResourceArgs
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    CloudSpanner ConnectionCloudSpannerArgs
    Connection properties specific to Cloud Spanner Structure is documented below.
    CloudSql ConnectionCloudSqlArgs
    Connection properties specific to the Cloud SQL. Structure is documented below.
    ConnectionId string
    Optional connection id that should be assigned to the created connection.
    Description string
    A descriptive description for the connection
    FriendlyName string
    A descriptive name for the connection
    Location string
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    Project string
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    Spark ConnectionSparkArgs
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws ConnectionAws
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure ConnectionAzure
    Container for connection properties specific to Azure. Structure is documented below.
    cloudResource ConnectionCloudResource
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloudSpanner ConnectionCloudSpanner
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloudSql ConnectionCloudSql
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connectionId String
    Optional connection id that should be assigned to the created connection.
    description String
    A descriptive description for the connection
    friendlyName String
    A descriptive name for the connection
    location String
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    project String
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark ConnectionSpark
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws ConnectionAws
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure ConnectionAzure
    Container for connection properties specific to Azure. Structure is documented below.
    cloudResource ConnectionCloudResource
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloudSpanner ConnectionCloudSpanner
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloudSql ConnectionCloudSql
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connectionId string
    Optional connection id that should be assigned to the created connection.
    description string
    A descriptive description for the connection
    friendlyName string
    A descriptive name for the connection
    location string
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    project string
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark ConnectionSpark
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws ConnectionAwsArgs
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure ConnectionAzureArgs
    Container for connection properties specific to Azure. Structure is documented below.
    cloud_resource ConnectionCloudResourceArgs
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloud_spanner ConnectionCloudSpannerArgs
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloud_sql ConnectionCloudSqlArgs
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connection_id str
    Optional connection id that should be assigned to the created connection.
    description str
    A descriptive description for the connection
    friendly_name str
    A descriptive name for the connection
    location str
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    project str
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark ConnectionSparkArgs
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws Property Map
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure Property Map
    Container for connection properties specific to Azure. Structure is documented below.
    cloudResource Property Map
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloudSpanner Property Map
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloudSql Property Map
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connectionId String
    Optional connection id that should be assigned to the created connection.
    description String
    A descriptive description for the connection
    friendlyName String
    A descriptive name for the connection
    location String
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    project String
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark Property Map
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.

    Outputs

    All input properties are implicitly available as output properties. Additionally, the Connection resource produces the following output properties:

    HasCredential bool
    True if the connection has credential assigned.
    Id string
    The provider-assigned unique ID for this managed resource.
    Name string
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    HasCredential bool
    True if the connection has credential assigned.
    Id string
    The provider-assigned unique ID for this managed resource.
    Name string
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    hasCredential Boolean
    True if the connection has credential assigned.
    id String
    The provider-assigned unique ID for this managed resource.
    name String
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    hasCredential boolean
    True if the connection has credential assigned.
    id string
    The provider-assigned unique ID for this managed resource.
    name string
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    has_credential bool
    True if the connection has credential assigned.
    id str
    The provider-assigned unique ID for this managed resource.
    name str
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    hasCredential Boolean
    True if the connection has credential assigned.
    id String
    The provider-assigned unique ID for this managed resource.
    name String
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"

    Look up Existing Connection Resource

    Get an existing Connection resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: ConnectionState, opts?: CustomResourceOptions): Connection
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            aws: Optional[ConnectionAwsArgs] = None,
            azure: Optional[ConnectionAzureArgs] = None,
            cloud_resource: Optional[ConnectionCloudResourceArgs] = None,
            cloud_spanner: Optional[ConnectionCloudSpannerArgs] = None,
            cloud_sql: Optional[ConnectionCloudSqlArgs] = None,
            connection_id: Optional[str] = None,
            description: Optional[str] = None,
            friendly_name: Optional[str] = None,
            has_credential: Optional[bool] = None,
            location: Optional[str] = None,
            name: Optional[str] = None,
            project: Optional[str] = None,
            spark: Optional[ConnectionSparkArgs] = None) -> Connection
    func GetConnection(ctx *Context, name string, id IDInput, state *ConnectionState, opts ...ResourceOption) (*Connection, error)
    public static Connection Get(string name, Input<string> id, ConnectionState? state, CustomResourceOptions? opts = null)
    public static Connection get(String name, Output<String> id, ConnectionState state, CustomResourceOptions options)
    Resource lookup is not supported in YAML
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    Aws ConnectionAws
    Connection properties specific to Amazon Web Services. Structure is documented below.
    Azure ConnectionAzure
    Container for connection properties specific to Azure. Structure is documented below.
    CloudResource ConnectionCloudResource
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    CloudSpanner ConnectionCloudSpanner
    Connection properties specific to Cloud Spanner Structure is documented below.
    CloudSql ConnectionCloudSql
    Connection properties specific to the Cloud SQL. Structure is documented below.
    ConnectionId string
    Optional connection id that should be assigned to the created connection.
    Description string
    A descriptive description for the connection
    FriendlyName string
    A descriptive name for the connection
    HasCredential bool
    True if the connection has credential assigned.
    Location string
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    Name string
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    Project string
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    Spark ConnectionSpark
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    Aws ConnectionAwsArgs
    Connection properties specific to Amazon Web Services. Structure is documented below.
    Azure ConnectionAzureArgs
    Container for connection properties specific to Azure. Structure is documented below.
    CloudResource ConnectionCloudResourceArgs
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    CloudSpanner ConnectionCloudSpannerArgs
    Connection properties specific to Cloud Spanner Structure is documented below.
    CloudSql ConnectionCloudSqlArgs
    Connection properties specific to the Cloud SQL. Structure is documented below.
    ConnectionId string
    Optional connection id that should be assigned to the created connection.
    Description string
    A descriptive description for the connection
    FriendlyName string
    A descriptive name for the connection
    HasCredential bool
    True if the connection has credential assigned.
    Location string
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    Name string
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    Project string
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    Spark ConnectionSparkArgs
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws ConnectionAws
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure ConnectionAzure
    Container for connection properties specific to Azure. Structure is documented below.
    cloudResource ConnectionCloudResource
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloudSpanner ConnectionCloudSpanner
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloudSql ConnectionCloudSql
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connectionId String
    Optional connection id that should be assigned to the created connection.
    description String
    A descriptive description for the connection
    friendlyName String
    A descriptive name for the connection
    hasCredential Boolean
    True if the connection has credential assigned.
    location String
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    name String
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    project String
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark ConnectionSpark
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws ConnectionAws
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure ConnectionAzure
    Container for connection properties specific to Azure. Structure is documented below.
    cloudResource ConnectionCloudResource
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloudSpanner ConnectionCloudSpanner
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloudSql ConnectionCloudSql
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connectionId string
    Optional connection id that should be assigned to the created connection.
    description string
    A descriptive description for the connection
    friendlyName string
    A descriptive name for the connection
    hasCredential boolean
    True if the connection has credential assigned.
    location string
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    name string
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    project string
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark ConnectionSpark
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws ConnectionAwsArgs
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure ConnectionAzureArgs
    Container for connection properties specific to Azure. Structure is documented below.
    cloud_resource ConnectionCloudResourceArgs
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloud_spanner ConnectionCloudSpannerArgs
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloud_sql ConnectionCloudSqlArgs
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connection_id str
    Optional connection id that should be assigned to the created connection.
    description str
    A descriptive description for the connection
    friendly_name str
    A descriptive name for the connection
    has_credential bool
    True if the connection has credential assigned.
    location str
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    name str
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    project str
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark ConnectionSparkArgs
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.
    aws Property Map
    Connection properties specific to Amazon Web Services. Structure is documented below.
    azure Property Map
    Container for connection properties specific to Azure. Structure is documented below.
    cloudResource Property Map
    Container for connection properties for delegation of access to GCP resources. Structure is documented below.
    cloudSpanner Property Map
    Connection properties specific to Cloud Spanner Structure is documented below.
    cloudSql Property Map
    Connection properties specific to the Cloud SQL. Structure is documented below.
    connectionId String
    Optional connection id that should be assigned to the created connection.
    description String
    A descriptive description for the connection
    friendlyName String
    A descriptive name for the connection
    hasCredential Boolean
    True if the connection has credential assigned.
    location String
    The geographic location where the connection should reside. Cloud SQL instance must be in the same location as the connection with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU. Examples: US, EU, asia-northeast1, us-central1, europe-west1. Spanner Connections same as spanner region AWS allowed regions are aws-us-east-1 Azure allowed regions are azure-eastus2
    name String
    The resource name of the connection in the form of: "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
    project String
    The ID of the project in which the resource belongs. If it is not provided, the provider project is used.
    spark Property Map
    Container for connection properties to execute stored procedures for Apache Spark. resources. Structure is documented below.

    Supporting Types

    ConnectionAws, ConnectionAwsArgs

    AccessRole ConnectionAwsAccessRole
    Authentication using Google owned service account to assume into customer's AWS IAM Role. Structure is documented below.
    AccessRole ConnectionAwsAccessRole
    Authentication using Google owned service account to assume into customer's AWS IAM Role. Structure is documented below.
    accessRole ConnectionAwsAccessRole
    Authentication using Google owned service account to assume into customer's AWS IAM Role. Structure is documented below.
    accessRole ConnectionAwsAccessRole
    Authentication using Google owned service account to assume into customer's AWS IAM Role. Structure is documented below.
    access_role ConnectionAwsAccessRole
    Authentication using Google owned service account to assume into customer's AWS IAM Role. Structure is documented below.
    accessRole Property Map
    Authentication using Google owned service account to assume into customer's AWS IAM Role. Structure is documented below.

    ConnectionAwsAccessRole, ConnectionAwsAccessRoleArgs

    IamRoleId string
    The user’s AWS IAM Role that trusts the Google-owned AWS IAM user Connection.
    Identity string
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's AWS IAM Role.
    IamRoleId string
    The user’s AWS IAM Role that trusts the Google-owned AWS IAM user Connection.
    Identity string
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's AWS IAM Role.
    iamRoleId String
    The user’s AWS IAM Role that trusts the Google-owned AWS IAM user Connection.
    identity String
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's AWS IAM Role.
    iamRoleId string
    The user’s AWS IAM Role that trusts the Google-owned AWS IAM user Connection.
    identity string
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's AWS IAM Role.
    iam_role_id str
    The user’s AWS IAM Role that trusts the Google-owned AWS IAM user Connection.
    identity str
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's AWS IAM Role.
    iamRoleId String
    The user’s AWS IAM Role that trusts the Google-owned AWS IAM user Connection.
    identity String
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's AWS IAM Role.

    ConnectionAzure, ConnectionAzureArgs

    CustomerTenantId string
    The id of customer's directory that host the data.
    Application string
    (Output) The name of the Azure Active Directory Application.
    ClientId string
    (Output) The client id of the Azure Active Directory Application.
    FederatedApplicationClientId string
    The Azure Application (client) ID where the federated credentials will be hosted.
    Identity string
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's Azure Active Directory Application.
    ObjectId string
    (Output) The object id of the Azure Active Directory Application.
    RedirectUri string
    (Output) The URL user will be redirected to after granting consent during connection setup.
    CustomerTenantId string
    The id of customer's directory that host the data.
    Application string
    (Output) The name of the Azure Active Directory Application.
    ClientId string
    (Output) The client id of the Azure Active Directory Application.
    FederatedApplicationClientId string
    The Azure Application (client) ID where the federated credentials will be hosted.
    Identity string
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's Azure Active Directory Application.
    ObjectId string
    (Output) The object id of the Azure Active Directory Application.
    RedirectUri string
    (Output) The URL user will be redirected to after granting consent during connection setup.
    customerTenantId String
    The id of customer's directory that host the data.
    application String
    (Output) The name of the Azure Active Directory Application.
    clientId String
    (Output) The client id of the Azure Active Directory Application.
    federatedApplicationClientId String
    The Azure Application (client) ID where the federated credentials will be hosted.
    identity String
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's Azure Active Directory Application.
    objectId String
    (Output) The object id of the Azure Active Directory Application.
    redirectUri String
    (Output) The URL user will be redirected to after granting consent during connection setup.
    customerTenantId string
    The id of customer's directory that host the data.
    application string
    (Output) The name of the Azure Active Directory Application.
    clientId string
    (Output) The client id of the Azure Active Directory Application.
    federatedApplicationClientId string
    The Azure Application (client) ID where the federated credentials will be hosted.
    identity string
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's Azure Active Directory Application.
    objectId string
    (Output) The object id of the Azure Active Directory Application.
    redirectUri string
    (Output) The URL user will be redirected to after granting consent during connection setup.
    customer_tenant_id str
    The id of customer's directory that host the data.
    application str
    (Output) The name of the Azure Active Directory Application.
    client_id str
    (Output) The client id of the Azure Active Directory Application.
    federated_application_client_id str
    The Azure Application (client) ID where the federated credentials will be hosted.
    identity str
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's Azure Active Directory Application.
    object_id str
    (Output) The object id of the Azure Active Directory Application.
    redirect_uri str
    (Output) The URL user will be redirected to after granting consent during connection setup.
    customerTenantId String
    The id of customer's directory that host the data.
    application String
    (Output) The name of the Azure Active Directory Application.
    clientId String
    (Output) The client id of the Azure Active Directory Application.
    federatedApplicationClientId String
    The Azure Application (client) ID where the federated credentials will be hosted.
    identity String
    (Output) A unique Google-owned and Google-generated identity for the Connection. This identity will be used to access the user's Azure Active Directory Application.
    objectId String
    (Output) The object id of the Azure Active Directory Application.
    redirectUri String
    (Output) The URL user will be redirected to after granting consent during connection setup.

    ConnectionCloudResource, ConnectionCloudResourceArgs

    ServiceAccountId string
    (Output) The account ID of the service created for the purpose of this connection.
    ServiceAccountId string
    (Output) The account ID of the service created for the purpose of this connection.
    serviceAccountId String
    (Output) The account ID of the service created for the purpose of this connection.
    serviceAccountId string
    (Output) The account ID of the service created for the purpose of this connection.
    service_account_id str
    (Output) The account ID of the service created for the purpose of this connection.
    serviceAccountId String
    (Output) The account ID of the service created for the purpose of this connection.

    ConnectionCloudSpanner, ConnectionCloudSpannerArgs

    Database string
    Cloud Spanner database in the form `project/instance/database'.
    DatabaseRole string
    Cloud Spanner database role for fine-grained access control. The Cloud Spanner admin should have provisioned the database role with appropriate permissions, such as SELECT and INSERT. Other users should only use roles provided by their Cloud Spanner admins. The database role name must start with a letter, and can only contain letters, numbers, and underscores. For more details, see https://cloud.google.com/spanner/docs/fgac-about.
    MaxParallelism int
    Allows setting max parallelism per query when executing on Spanner independent compute resources. If unspecified, default values of parallelism are chosen that are dependent on the Cloud Spanner instance configuration. useParallelism and useDataBoost must be set when setting max parallelism.
    UseDataBoost bool
    If set, the request will be executed via Spanner independent compute resources. use_parallelism must be set when using data boost.
    UseParallelism bool
    If parallelism should be used when reading from Cloud Spanner.
    UseServerlessAnalytics bool

    (Optional, Deprecated) If the serverless analytics service should be used to read data from Cloud Spanner. useParallelism must be set when using serverless analytics.

    Warning: useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    Deprecated:useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    Database string
    Cloud Spanner database in the form `project/instance/database'.
    DatabaseRole string
    Cloud Spanner database role for fine-grained access control. The Cloud Spanner admin should have provisioned the database role with appropriate permissions, such as SELECT and INSERT. Other users should only use roles provided by their Cloud Spanner admins. The database role name must start with a letter, and can only contain letters, numbers, and underscores. For more details, see https://cloud.google.com/spanner/docs/fgac-about.
    MaxParallelism int
    Allows setting max parallelism per query when executing on Spanner independent compute resources. If unspecified, default values of parallelism are chosen that are dependent on the Cloud Spanner instance configuration. useParallelism and useDataBoost must be set when setting max parallelism.
    UseDataBoost bool
    If set, the request will be executed via Spanner independent compute resources. use_parallelism must be set when using data boost.
    UseParallelism bool
    If parallelism should be used when reading from Cloud Spanner.
    UseServerlessAnalytics bool

    (Optional, Deprecated) If the serverless analytics service should be used to read data from Cloud Spanner. useParallelism must be set when using serverless analytics.

    Warning: useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    Deprecated:useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    database String
    Cloud Spanner database in the form `project/instance/database'.
    databaseRole String
    Cloud Spanner database role for fine-grained access control. The Cloud Spanner admin should have provisioned the database role with appropriate permissions, such as SELECT and INSERT. Other users should only use roles provided by their Cloud Spanner admins. The database role name must start with a letter, and can only contain letters, numbers, and underscores. For more details, see https://cloud.google.com/spanner/docs/fgac-about.
    maxParallelism Integer
    Allows setting max parallelism per query when executing on Spanner independent compute resources. If unspecified, default values of parallelism are chosen that are dependent on the Cloud Spanner instance configuration. useParallelism and useDataBoost must be set when setting max parallelism.
    useDataBoost Boolean
    If set, the request will be executed via Spanner independent compute resources. use_parallelism must be set when using data boost.
    useParallelism Boolean
    If parallelism should be used when reading from Cloud Spanner.
    useServerlessAnalytics Boolean

    (Optional, Deprecated) If the serverless analytics service should be used to read data from Cloud Spanner. useParallelism must be set when using serverless analytics.

    Warning: useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    Deprecated:useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    database string
    Cloud Spanner database in the form `project/instance/database'.
    databaseRole string
    Cloud Spanner database role for fine-grained access control. The Cloud Spanner admin should have provisioned the database role with appropriate permissions, such as SELECT and INSERT. Other users should only use roles provided by their Cloud Spanner admins. The database role name must start with a letter, and can only contain letters, numbers, and underscores. For more details, see https://cloud.google.com/spanner/docs/fgac-about.
    maxParallelism number
    Allows setting max parallelism per query when executing on Spanner independent compute resources. If unspecified, default values of parallelism are chosen that are dependent on the Cloud Spanner instance configuration. useParallelism and useDataBoost must be set when setting max parallelism.
    useDataBoost boolean
    If set, the request will be executed via Spanner independent compute resources. use_parallelism must be set when using data boost.
    useParallelism boolean
    If parallelism should be used when reading from Cloud Spanner.
    useServerlessAnalytics boolean

    (Optional, Deprecated) If the serverless analytics service should be used to read data from Cloud Spanner. useParallelism must be set when using serverless analytics.

    Warning: useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    Deprecated:useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    database str
    Cloud Spanner database in the form `project/instance/database'.
    database_role str
    Cloud Spanner database role for fine-grained access control. The Cloud Spanner admin should have provisioned the database role with appropriate permissions, such as SELECT and INSERT. Other users should only use roles provided by their Cloud Spanner admins. The database role name must start with a letter, and can only contain letters, numbers, and underscores. For more details, see https://cloud.google.com/spanner/docs/fgac-about.
    max_parallelism int
    Allows setting max parallelism per query when executing on Spanner independent compute resources. If unspecified, default values of parallelism are chosen that are dependent on the Cloud Spanner instance configuration. useParallelism and useDataBoost must be set when setting max parallelism.
    use_data_boost bool
    If set, the request will be executed via Spanner independent compute resources. use_parallelism must be set when using data boost.
    use_parallelism bool
    If parallelism should be used when reading from Cloud Spanner.
    use_serverless_analytics bool

    (Optional, Deprecated) If the serverless analytics service should be used to read data from Cloud Spanner. useParallelism must be set when using serverless analytics.

    Warning: useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    Deprecated:useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    database String
    Cloud Spanner database in the form `project/instance/database'.
    databaseRole String
    Cloud Spanner database role for fine-grained access control. The Cloud Spanner admin should have provisioned the database role with appropriate permissions, such as SELECT and INSERT. Other users should only use roles provided by their Cloud Spanner admins. The database role name must start with a letter, and can only contain letters, numbers, and underscores. For more details, see https://cloud.google.com/spanner/docs/fgac-about.
    maxParallelism Number
    Allows setting max parallelism per query when executing on Spanner independent compute resources. If unspecified, default values of parallelism are chosen that are dependent on the Cloud Spanner instance configuration. useParallelism and useDataBoost must be set when setting max parallelism.
    useDataBoost Boolean
    If set, the request will be executed via Spanner independent compute resources. use_parallelism must be set when using data boost.
    useParallelism Boolean
    If parallelism should be used when reading from Cloud Spanner.
    useServerlessAnalytics Boolean

    (Optional, Deprecated) If the serverless analytics service should be used to read data from Cloud Spanner. useParallelism must be set when using serverless analytics.

    Warning: useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    Deprecated:useServerlessAnalytics is deprecated and will be removed in a future major release. Use useDataBoost instead.

    ConnectionCloudSql, ConnectionCloudSqlArgs

    Credential ConnectionCloudSqlCredential
    Cloud SQL properties. Structure is documented below.
    Database string
    Database name.
    InstanceId string
    Cloud SQL instance ID in the form project:location:instance.
    Type string
    Type of the Cloud SQL database. Possible values are: DATABASE_TYPE_UNSPECIFIED, POSTGRES, MYSQL.
    ServiceAccountId string
    (Output) When the connection is used in the context of an operation in BigQuery, this service account will serve as the identity being used for connecting to the CloudSQL instance specified in this connection.
    Credential ConnectionCloudSqlCredential
    Cloud SQL properties. Structure is documented below.
    Database string
    Database name.
    InstanceId string
    Cloud SQL instance ID in the form project:location:instance.
    Type string
    Type of the Cloud SQL database. Possible values are: DATABASE_TYPE_UNSPECIFIED, POSTGRES, MYSQL.
    ServiceAccountId string
    (Output) When the connection is used in the context of an operation in BigQuery, this service account will serve as the identity being used for connecting to the CloudSQL instance specified in this connection.
    credential ConnectionCloudSqlCredential
    Cloud SQL properties. Structure is documented below.
    database String
    Database name.
    instanceId String
    Cloud SQL instance ID in the form project:location:instance.
    type String
    Type of the Cloud SQL database. Possible values are: DATABASE_TYPE_UNSPECIFIED, POSTGRES, MYSQL.
    serviceAccountId String
    (Output) When the connection is used in the context of an operation in BigQuery, this service account will serve as the identity being used for connecting to the CloudSQL instance specified in this connection.
    credential ConnectionCloudSqlCredential
    Cloud SQL properties. Structure is documented below.
    database string
    Database name.
    instanceId string
    Cloud SQL instance ID in the form project:location:instance.
    type string
    Type of the Cloud SQL database. Possible values are: DATABASE_TYPE_UNSPECIFIED, POSTGRES, MYSQL.
    serviceAccountId string
    (Output) When the connection is used in the context of an operation in BigQuery, this service account will serve as the identity being used for connecting to the CloudSQL instance specified in this connection.
    credential ConnectionCloudSqlCredential
    Cloud SQL properties. Structure is documented below.
    database str
    Database name.
    instance_id str
    Cloud SQL instance ID in the form project:location:instance.
    type str
    Type of the Cloud SQL database. Possible values are: DATABASE_TYPE_UNSPECIFIED, POSTGRES, MYSQL.
    service_account_id str
    (Output) When the connection is used in the context of an operation in BigQuery, this service account will serve as the identity being used for connecting to the CloudSQL instance specified in this connection.
    credential Property Map
    Cloud SQL properties. Structure is documented below.
    database String
    Database name.
    instanceId String
    Cloud SQL instance ID in the form project:location:instance.
    type String
    Type of the Cloud SQL database. Possible values are: DATABASE_TYPE_UNSPECIFIED, POSTGRES, MYSQL.
    serviceAccountId String
    (Output) When the connection is used in the context of an operation in BigQuery, this service account will serve as the identity being used for connecting to the CloudSQL instance specified in this connection.

    ConnectionCloudSqlCredential, ConnectionCloudSqlCredentialArgs

    Password string
    Password for database. Note: This property is sensitive and will not be displayed in the plan.
    Username string
    Username for database.
    Password string
    Password for database. Note: This property is sensitive and will not be displayed in the plan.
    Username string
    Username for database.
    password String
    Password for database. Note: This property is sensitive and will not be displayed in the plan.
    username String
    Username for database.
    password string
    Password for database. Note: This property is sensitive and will not be displayed in the plan.
    username string
    Username for database.
    password str
    Password for database. Note: This property is sensitive and will not be displayed in the plan.
    username str
    Username for database.
    password String
    Password for database. Note: This property is sensitive and will not be displayed in the plan.
    username String
    Username for database.

    ConnectionSpark, ConnectionSparkArgs

    MetastoreServiceConfig ConnectionSparkMetastoreServiceConfig
    Dataproc Metastore Service configuration for the connection. Structure is documented below.
    ServiceAccountId string
    (Output) The account ID of the service created for the purpose of this connection.
    SparkHistoryServerConfig ConnectionSparkSparkHistoryServerConfig
    Spark History Server configuration for the connection. Structure is documented below.
    MetastoreServiceConfig ConnectionSparkMetastoreServiceConfig
    Dataproc Metastore Service configuration for the connection. Structure is documented below.
    ServiceAccountId string
    (Output) The account ID of the service created for the purpose of this connection.
    SparkHistoryServerConfig ConnectionSparkSparkHistoryServerConfig
    Spark History Server configuration for the connection. Structure is documented below.
    metastoreServiceConfig ConnectionSparkMetastoreServiceConfig
    Dataproc Metastore Service configuration for the connection. Structure is documented below.
    serviceAccountId String
    (Output) The account ID of the service created for the purpose of this connection.
    sparkHistoryServerConfig ConnectionSparkSparkHistoryServerConfig
    Spark History Server configuration for the connection. Structure is documented below.
    metastoreServiceConfig ConnectionSparkMetastoreServiceConfig
    Dataproc Metastore Service configuration for the connection. Structure is documented below.
    serviceAccountId string
    (Output) The account ID of the service created for the purpose of this connection.
    sparkHistoryServerConfig ConnectionSparkSparkHistoryServerConfig
    Spark History Server configuration for the connection. Structure is documented below.
    metastore_service_config ConnectionSparkMetastoreServiceConfig
    Dataproc Metastore Service configuration for the connection. Structure is documented below.
    service_account_id str
    (Output) The account ID of the service created for the purpose of this connection.
    spark_history_server_config ConnectionSparkSparkHistoryServerConfig
    Spark History Server configuration for the connection. Structure is documented below.
    metastoreServiceConfig Property Map
    Dataproc Metastore Service configuration for the connection. Structure is documented below.
    serviceAccountId String
    (Output) The account ID of the service created for the purpose of this connection.
    sparkHistoryServerConfig Property Map
    Spark History Server configuration for the connection. Structure is documented below.

    ConnectionSparkMetastoreServiceConfig, ConnectionSparkMetastoreServiceConfigArgs

    MetastoreService string
    Resource name of an existing Dataproc Metastore service in the form of projects/[projectId]/locations/[region]/services/[serviceId].
    MetastoreService string
    Resource name of an existing Dataproc Metastore service in the form of projects/[projectId]/locations/[region]/services/[serviceId].
    metastoreService String
    Resource name of an existing Dataproc Metastore service in the form of projects/[projectId]/locations/[region]/services/[serviceId].
    metastoreService string
    Resource name of an existing Dataproc Metastore service in the form of projects/[projectId]/locations/[region]/services/[serviceId].
    metastore_service str
    Resource name of an existing Dataproc Metastore service in the form of projects/[projectId]/locations/[region]/services/[serviceId].
    metastoreService String
    Resource name of an existing Dataproc Metastore service in the form of projects/[projectId]/locations/[region]/services/[serviceId].

    ConnectionSparkSparkHistoryServerConfig, ConnectionSparkSparkHistoryServerConfigArgs

    DataprocCluster string
    Resource name of an existing Dataproc Cluster to act as a Spark History Server for the connection if the form of projects/[projectId]/regions/[region]/clusters/[cluster_name].
    DataprocCluster string
    Resource name of an existing Dataproc Cluster to act as a Spark History Server for the connection if the form of projects/[projectId]/regions/[region]/clusters/[cluster_name].
    dataprocCluster String
    Resource name of an existing Dataproc Cluster to act as a Spark History Server for the connection if the form of projects/[projectId]/regions/[region]/clusters/[cluster_name].
    dataprocCluster string
    Resource name of an existing Dataproc Cluster to act as a Spark History Server for the connection if the form of projects/[projectId]/regions/[region]/clusters/[cluster_name].
    dataproc_cluster str
    Resource name of an existing Dataproc Cluster to act as a Spark History Server for the connection if the form of projects/[projectId]/regions/[region]/clusters/[cluster_name].
    dataprocCluster String
    Resource name of an existing Dataproc Cluster to act as a Spark History Server for the connection if the form of projects/[projectId]/regions/[region]/clusters/[cluster_name].

    Import

    Connection can be imported using any of these accepted formats:

    • projects/{{project}}/locations/{{location}}/connections/{{connection_id}}

    • {{project}}/{{location}}/{{connection_id}}

    • {{location}}/{{connection_id}}

    When using the pulumi import command, Connection can be imported using one of the formats above. For example:

    $ pulumi import gcp:bigquery/connection:Connection default projects/{{project}}/locations/{{location}}/connections/{{connection_id}}
    
    $ pulumi import gcp:bigquery/connection:Connection default {{project}}/{{location}}/{{connection_id}}
    
    $ pulumi import gcp:bigquery/connection:Connection default {{location}}/{{connection_id}}
    

    Package Details

    Repository
    Google Cloud (GCP) Classic pulumi/pulumi-gcp
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the google-beta Terraform Provider.
    gcp logo
    Google Cloud Classic v7.16.0 published on Wednesday, Mar 27, 2024 by Pulumi