DP-203 Exam Question 126

You plan to build a structured streaming solution in Azure Databricks. The solution will count new events in five-minute intervals and report only events that arrive during the interval. The output will be sent to a Delta Lake table.
Which output mode should you use?
  • DP-203 Exam Question 127

    You build an Azure Data Factory pipeline to move data from an Azure Data Lake Storage Gen2 container to a database in an Azure Synapse Analytics dedicated SQL pool.
    Data in the container is stored in the following folder structure.
    /in/{YYYY}/{MM}/{DD}/{HH}/{mm}
    The earliest folder is /in/2021/01/01/00/00. The latest folder is /in/2021/01/15/01/45.
    You need to configure a pipeline trigger to meet the following requirements:
    Existing data must be loaded.
    Data must be loaded every 30 minutes.
    Late-arriving data of up to two minutes must he included in the load for the time at which the data should have arrived.
    How should you configure the pipeline trigger? To answer, select the appropriate options in the answer area.
    NOTE: Each correct selection is worth one point.

    DP-203 Exam Question 128

    You have an Azure Synapse Analytics dedicated SQL pool that contains a table named Table1. Table1 contains the following:
    One billion rows
    A clustered columnstore index
    A hash-distributed column named Product Key
    A column named Sales Date that is of the date data type and cannot be null Thirty million rows will be added to Table1 each month.
    You need to partition Table1 based on the Sales Date column. The solution must optimize query performance and data loading.
    How often should you create a partition?
  • DP-203 Exam Question 129

    You have an Azure Synapse Analytics dedicated SQL pool named Pool1. Pool1 contains a table named table1.
    You load 5 TB of data intotable1.
    You need to ensure that columnstore compression is maximized for table1.
    Which statement should you execute?