Exam Details

  • Exam Code
    :DP-203
  • Exam Name
    :Data Engineering on Microsoft Azure
  • Certification
    :Microsoft Certifications
  • Vendor
    :Microsoft
  • Total Questions
    :398 Q&As
  • Last Updated
    :Mar 30, 2025

Microsoft Microsoft Certifications DP-203 Questions & Answers

  • Question 121:

    HOTSPOT

    You are designing a monitoring solution for a fleet of 500 vehicles. Each vehicle has a GPS tracking device that sends data to an Azure event hub once per minute.

    You have a CSV file in an Azure Data Lake Storage Gen2 container. The file maintains the expected geographical area in which each vehicle should be.

    You need to ensure that when a GPS position is outside the expected area, a message is added to another event hub for processing within 30 seconds. The solution must minimize cost.

    What should you include in the solution? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point.

    Hot Area:

  • Question 122:

    HOTSPOT

    You are building an Azure Analytics query that will receive input data from Azure IoT Hub and write the results to Azure Blob storage.

    You need to calculate the difference in readings per sensor per hour.

    How should you complete the query? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point.

    Hot Area:

  • Question 123:

    HOTSPOT

    You are designing a real-time dashboard solution that will visualize streaming data from remote sensors that connect to the internet. The streaming data must be aggregated to show the average value of each 10-second interval. The data will

    be discarded after being displayed in the dashboard.

    The solution will use Azure Stream Analytics and must meet the following requirements:

    1.

    Minimize latency from an Azure Event hub to the dashboard.

    2.

    Minimize the required storage.

    3.

    Minimize development effort.

    What should you include in the solution? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point

    Hot Area:

  • Question 124:

    HOTSPOT

    You have an Azure SQL database named Database1 and two Azure event hubs named HubA and HubB. The data consumed from each source is shown in the following table.

    You need to implement Azure Stream Analytics to calculate the average fare per mile by driver.

    How should you configure the Stream Analytics input for each source? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point.

    Hot Area:

  • Question 125:

    HOTSPOT

    You build an Azure Data Factory pipeline to move data from an Azure Data Lake Storage Gen2 container to a database in an Azure Synapse Analytics dedicated SQL pool.

    Data in the container is stored in the following folder structure.

    /in/{YYYY}/{MM}/{DD}/{HH}/{mm}

    The earliest folder is /in/2021/01/01/00/00. The latest folder is /in/2021/01/15/01/45.

    You need to configure a pipeline trigger to meet the following requirements:

    Existing data must be loaded.

    Data must be loaded every 30 minutes.

    Late-arriving data of up to two minutes must he included in the load for the time at which the data should have arrived.

    How should you configure the pipeline trigger? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point.

    Hot Area:

  • Question 126:

    HOTSPOT

    You have a data model that you plan to implement in a data warehouse in Azure Synapse Analytics as shown in the following exhibit.

    All the dimension tables will be less than 2 GB after compression, and the fact table will be approximately 6 TB. Which type of table should you use for each table? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

    Hot Area:

  • Question 127:

    HOTSPOT

    You have an Azure Data Lake Storage Gen2 container.

    Data is ingested into the container, and then transformed by a data integration application. The data is NOT modified after that. Users can read files in the container but cannot modify the files.

    You need to design a data archiving solution that meets the following requirements:

    1.

    New data is accessed frequently and must be available as quickly as possible.

    2.

    Data that is older than five years is accessed infrequently but must be available within one second when requested.

    3.

    Data that is older than seven years is NOT accessed. After seven years, the data must be persisted at the lowest cost possible.

    4.

    Costs must be minimized while maintaining the required availability.

    How should you manage the data? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point

    Hot Area:

  • Question 128:

    HOTSPOT

    You are building an Azure Stream Analytics job to identify how much time a user spends interacting with a feature on a webpage.

    The job receives events based on user actions on the webpage. Each row of data represents an event. Each event has a type of either 'start' or 'end'.

    You need to calculate the duration between start and end events.

    How should you complete the query? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point.

    Hot Area:

  • Question 129:

    HOTSPOT

    You use Azure Data Factory to prepare data to be queried by Azure Synapse Analytics serverless SQL pools.

    Files are initially ingested into an Azure Data Lake Storage Gen2 account as 10 small JSON files. Each file contains the same data attributes and data from a subsidiary of your company.

    You need to move the files to a different folder and transform the data to meet the following requirements:

    1.

    Provide the fastest possible query times.

    2.

    Automatically infer the schema from the underlying files.

    How should you configure the Data Factory copy activity? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point.

    Hot Area:

  • Question 130:

    HOTSPOT

    You are planning the deployment of Azure Data Lake Storage Gen2.

    You have the following two reports that will access the data lake:

    1.

    Report1: Reads three columns from a file that contains 50 columns.

    2.

    Report2: Queries a single record based on a timestamp.

    You need to recommend in which format to store the data in the data lake to support the reports. The solution must minimize read times.

    What should you recommend for each report? To answer, select the appropriate options in the answer area.

    NOTE: Each correct selection is worth one point.

    Hot Area:

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Microsoft exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your DP-203 exam preparations and Microsoft certification application, do not hesitate to visit our Vcedump.com to find your solutions here.