1. Packages
  2. MongoDB Atlas
  3. API Docs
  4. getDataLakePipelineRun
MongoDB Atlas v3.14.3 published on Monday, Apr 1, 2024 by Pulumi

mongodbatlas.getDataLakePipelineRun

Explore with Pulumi AI

mongodbatlas logo
MongoDB Atlas v3.14.3 published on Monday, Apr 1, 2024 by Pulumi

    mongodbatlas.getDataLakePipelineRun describe a Data Lake Pipeline Run.

    NOTE: Groups and projects are synonymous terms. You may find groupId in the official documentation.

    Example Usage

    Coming soon!
    
    Coming soon!
    
    Coming soon!
    
    Coming soon!
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.mongodbatlas.DataLakePipeline;
    import com.pulumi.mongodbatlas.DataLakePipelineArgs;
    import com.pulumi.mongodbatlas.inputs.DataLakePipelineSinkArgs;
    import com.pulumi.mongodbatlas.inputs.DataLakePipelineSourceArgs;
    import com.pulumi.mongodbatlas.inputs.DataLakePipelineTransformationArgs;
    import com.pulumi.mongodbatlas.MongodbatlasFunctions;
    import com.pulumi.mongodbatlas.inputs.GetDataLakePipelineRunsArgs;
    import com.pulumi.mongodbatlas.inputs.GetDataLakePipelineRunArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var pipeline = new DataLakePipeline("pipeline", DataLakePipelineArgs.builder()        
                .projectId(mongodbatlas_project.projectTest().project_id())
                .sink(DataLakePipelineSinkArgs.builder()
                    .type("DLS")
                    .partitionFields(DataLakePipelineSinkPartitionFieldArgs.builder()
                        .name("access")
                        .order(0)
                        .build())
                    .build())
                .source(DataLakePipelineSourceArgs.builder()
                    .type("ON_DEMAND_CPS")
                    .clusterName(mongodbatlas_cluster.clusterTest().name())
                    .databaseName("sample_airbnb")
                    .collectionName("listingsAndReviews")
                    .build())
                .transformations(            
                    DataLakePipelineTransformationArgs.builder()
                        .field("test")
                        .type("EXCLUDE")
                        .build(),
                    DataLakePipelineTransformationArgs.builder()
                        .field("test22")
                        .type("EXCLUDE")
                        .build())
                .build());
    
            final var pipelineRun = MongodbatlasFunctions.getDataLakePipelineRuns(GetDataLakePipelineRunsArgs.builder()
                .projectId(mongodbatlas_project.projectTest().project_id())
                .name(pipeline.name())
                .build());
    
            final var test = MongodbatlasFunctions.getDataLakePipelineRun(GetDataLakePipelineRunArgs.builder()
                .projectId(mongodbatlas_project.projectTest().project_id())
                .pipelineName(pipeline.name())
                .pipelineRunId(mongodbatlas_data_lake_pipeline_runs.pipeline_run().results()[0].pipeline_run_id())
                .build());
    
        }
    }
    
    resources:
      pipeline:
        type: mongodbatlas:DataLakePipeline
        properties:
          # assuming we've already setup project and cluster in another block
          projectId: ${mongodbatlas_project.projectTest.project_id}
          sink:
            type: DLS
            partitionFields:
              - name: access
                order: 0
          source:
            type: ON_DEMAND_CPS
            clusterName: ${mongodbatlas_cluster.clusterTest.name}
            databaseName: sample_airbnb
            collectionName: listingsAndReviews
          transformations:
            - field: test
              type: EXCLUDE
            - field: test22
              type: EXCLUDE
    variables:
      pipelineRun:
        fn::invoke:
          Function: mongodbatlas:getDataLakePipelineRuns
          Arguments:
            projectId: ${mongodbatlas_project.projectTest.project_id}
            name: ${pipeline.name}
      test:
        fn::invoke:
          Function: mongodbatlas:getDataLakePipelineRun
          Arguments:
            projectId: ${mongodbatlas_project.projectTest.project_id}
            pipelineName: ${pipeline.name}
            pipelineRunId: ${mongodbatlas_data_lake_pipeline_runs.pipeline_run.results[0].pipeline_run_id}
    

    Using getDataLakePipelineRun

    Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.

    function getDataLakePipelineRun(args: GetDataLakePipelineRunArgs, opts?: InvokeOptions): Promise<GetDataLakePipelineRunResult>
    function getDataLakePipelineRunOutput(args: GetDataLakePipelineRunOutputArgs, opts?: InvokeOptions): Output<GetDataLakePipelineRunResult>
    def get_data_lake_pipeline_run(pipeline_name: Optional[str] = None,
                                   pipeline_run_id: Optional[str] = None,
                                   project_id: Optional[str] = None,
                                   opts: Optional[InvokeOptions] = None) -> GetDataLakePipelineRunResult
    def get_data_lake_pipeline_run_output(pipeline_name: Optional[pulumi.Input[str]] = None,
                                   pipeline_run_id: Optional[pulumi.Input[str]] = None,
                                   project_id: Optional[pulumi.Input[str]] = None,
                                   opts: Optional[InvokeOptions] = None) -> Output[GetDataLakePipelineRunResult]
    func GetDataLakePipelineRun(ctx *Context, args *GetDataLakePipelineRunArgs, opts ...InvokeOption) (*GetDataLakePipelineRunResult, error)
    func GetDataLakePipelineRunOutput(ctx *Context, args *GetDataLakePipelineRunOutputArgs, opts ...InvokeOption) GetDataLakePipelineRunResultOutput

    > Note: This function is named GetDataLakePipelineRun in the Go SDK.

    public static class GetDataLakePipelineRun 
    {
        public static Task<GetDataLakePipelineRunResult> InvokeAsync(GetDataLakePipelineRunArgs args, InvokeOptions? opts = null)
        public static Output<GetDataLakePipelineRunResult> Invoke(GetDataLakePipelineRunInvokeArgs args, InvokeOptions? opts = null)
    }
    public static CompletableFuture<GetDataLakePipelineRunResult> getDataLakePipelineRun(GetDataLakePipelineRunArgs args, InvokeOptions options)
    // Output-based functions aren't available in Java yet
    
    fn::invoke:
      function: mongodbatlas:index/getDataLakePipelineRun:getDataLakePipelineRun
      arguments:
        # arguments dictionary

    The following arguments are supported:

    PipelineName string
    Human-readable label that identifies the Data Lake Pipeline.
    PipelineRunId string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    ProjectId string
    Unique 24-hexadecimal digit string that identifies your project.
    PipelineName string
    Human-readable label that identifies the Data Lake Pipeline.
    PipelineRunId string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    ProjectId string
    Unique 24-hexadecimal digit string that identifies your project.
    pipelineName String
    Human-readable label that identifies the Data Lake Pipeline.
    pipelineRunId String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    projectId String
    Unique 24-hexadecimal digit string that identifies your project.
    pipelineName string
    Human-readable label that identifies the Data Lake Pipeline.
    pipelineRunId string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    projectId string
    Unique 24-hexadecimal digit string that identifies your project.
    pipeline_name str
    Human-readable label that identifies the Data Lake Pipeline.
    pipeline_run_id str
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    project_id str
    Unique 24-hexadecimal digit string that identifies your project.
    pipelineName String
    Human-readable label that identifies the Data Lake Pipeline.
    pipelineRunId String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    projectId String
    Unique 24-hexadecimal digit string that identifies your project.

    getDataLakePipelineRun Result

    The following output properties are available:

    BackupFrequencyType string
    Backup schedule interval of the Data Lake Pipeline.
    CreatedDate string
    Timestamp that indicates when the pipeline run was created.
    DatasetName string
    Human-readable label that identifies the dataset that Atlas generates during this pipeline run.
    Id string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    LastUpdatedDate string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    Phase string
    Processing phase of the Data Lake Pipeline.
    PipelineId string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline.
    PipelineName string
    PipelineRunId string
    ProjectId string
    SnapshotId string
    Unique 24-hexadecimal character string that identifies the snapshot of a cluster.
    State string
    State of the pipeline run.
    Stats List<GetDataLakePipelineRunStat>
    Runtime statistics for this Data Lake Pipeline run.
    BackupFrequencyType string
    Backup schedule interval of the Data Lake Pipeline.
    CreatedDate string
    Timestamp that indicates when the pipeline run was created.
    DatasetName string
    Human-readable label that identifies the dataset that Atlas generates during this pipeline run.
    Id string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    LastUpdatedDate string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    Phase string
    Processing phase of the Data Lake Pipeline.
    PipelineId string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline.
    PipelineName string
    PipelineRunId string
    ProjectId string
    SnapshotId string
    Unique 24-hexadecimal character string that identifies the snapshot of a cluster.
    State string
    State of the pipeline run.
    Stats []GetDataLakePipelineRunStat
    Runtime statistics for this Data Lake Pipeline run.
    backupFrequencyType String
    Backup schedule interval of the Data Lake Pipeline.
    createdDate String
    Timestamp that indicates when the pipeline run was created.
    datasetName String
    Human-readable label that identifies the dataset that Atlas generates during this pipeline run.
    id String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    lastUpdatedDate String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    phase String
    Processing phase of the Data Lake Pipeline.
    pipelineId String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline.
    pipelineName String
    pipelineRunId String
    projectId String
    snapshotId String
    Unique 24-hexadecimal character string that identifies the snapshot of a cluster.
    state String
    State of the pipeline run.
    stats List<GetDataLakePipelineRunStat>
    Runtime statistics for this Data Lake Pipeline run.
    backupFrequencyType string
    Backup schedule interval of the Data Lake Pipeline.
    createdDate string
    Timestamp that indicates when the pipeline run was created.
    datasetName string
    Human-readable label that identifies the dataset that Atlas generates during this pipeline run.
    id string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    lastUpdatedDate string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    phase string
    Processing phase of the Data Lake Pipeline.
    pipelineId string
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline.
    pipelineName string
    pipelineRunId string
    projectId string
    snapshotId string
    Unique 24-hexadecimal character string that identifies the snapshot of a cluster.
    state string
    State of the pipeline run.
    stats GetDataLakePipelineRunStat[]
    Runtime statistics for this Data Lake Pipeline run.
    backup_frequency_type str
    Backup schedule interval of the Data Lake Pipeline.
    created_date str
    Timestamp that indicates when the pipeline run was created.
    dataset_name str
    Human-readable label that identifies the dataset that Atlas generates during this pipeline run.
    id str
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    last_updated_date str
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    phase str
    Processing phase of the Data Lake Pipeline.
    pipeline_id str
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline.
    pipeline_name str
    pipeline_run_id str
    project_id str
    snapshot_id str
    Unique 24-hexadecimal character string that identifies the snapshot of a cluster.
    state str
    State of the pipeline run.
    stats Sequence[GetDataLakePipelineRunStat]
    Runtime statistics for this Data Lake Pipeline run.
    backupFrequencyType String
    Backup schedule interval of the Data Lake Pipeline.
    createdDate String
    Timestamp that indicates when the pipeline run was created.
    datasetName String
    Human-readable label that identifies the dataset that Atlas generates during this pipeline run.
    id String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    lastUpdatedDate String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline run.
    phase String
    Processing phase of the Data Lake Pipeline.
    pipelineId String
    Unique 24-hexadecimal character string that identifies a Data Lake Pipeline.
    pipelineName String
    pipelineRunId String
    projectId String
    snapshotId String
    Unique 24-hexadecimal character string that identifies the snapshot of a cluster.
    state String
    State of the pipeline run.
    stats List<Property Map>
    Runtime statistics for this Data Lake Pipeline run.

    Supporting Types

    GetDataLakePipelineRunStat

    BytesExported int
    Total data size in bytes exported for this pipeline run.
    NumDocs int
    Number of docs ingested for a this pipeline run.
    BytesExported int
    Total data size in bytes exported for this pipeline run.
    NumDocs int
    Number of docs ingested for a this pipeline run.
    bytesExported Integer
    Total data size in bytes exported for this pipeline run.
    numDocs Integer
    Number of docs ingested for a this pipeline run.
    bytesExported number
    Total data size in bytes exported for this pipeline run.
    numDocs number
    Number of docs ingested for a this pipeline run.
    bytes_exported int
    Total data size in bytes exported for this pipeline run.
    num_docs int
    Number of docs ingested for a this pipeline run.
    bytesExported Number
    Total data size in bytes exported for this pipeline run.
    numDocs Number
    Number of docs ingested for a this pipeline run.

    Package Details

    Repository
    MongoDB Atlas pulumi/pulumi-mongodbatlas
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the mongodbatlas Terraform Provider.
    mongodbatlas logo
    MongoDB Atlas v3.14.3 published on Monday, Apr 1, 2024 by Pulumi