DON'T WANT TO MISS A THING?

Certification Exam Passing Tips

Latest exam news and discount info

Curated and up-to-date by our experts

Yes, send me the newsletter

Master Microsoft DP-203 Certification Questions & Study Resources, Microsoft Azure Data Engineering Associate | SPOTO

Prepare for success with our comprehensive collection of practice tests and study materials tailored for the Microsoft DP-203 certification exam. Our platform offers free tests, exam practice, online exam questions, sample questions, exam dumps, exam questions and answers, mock exams, and more. With our latest practice tests, you'll be equipped to excel and pass the certification exam with confidence. Earning the Microsoft Certified: Azure Data Engineer Associate certification by passing the DP-203 exam demonstrates proficiency in four key subject areas: designing and implementing data storage, designing and developing data processing, designing and implementing data security, and monitoring and optimizing data storage and data processing. Trust our exam materials to guide you towards success in your certification journey.
Take other online exams

Question #1
You manage an enterprise data warehouse in Azure Synapse Analytics.Users report slow performance when they run commonly used queries. Users do not report performance changes forinfrequently used queries.You need to monitor resource utilization to determine the source of the performance issues.Which metric should you monitor?
A. DWU percentage
B. Cache hit percentage
C. DWU limit
D. Data IO percentage
View answer
Correct Answer: B
Question #2
You configure monitoring for a Microsoft Azure SQL Data Warehouse implementation. The implementation uses PolyBase to load data from comma-separated value (CSV) files stored in Azure Data Lake Gen 2 using an external table. Files with an invalid schema cause errors to occur. You need to monitor for an invalid schema error. For which error should you monitor?
A. EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error[com
B. EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error [No FileSystem for scheme: wasbs] occurred while accessing external file
C. Cannot execute the query "Remote Query" against OLE DB provider "SQLNCLI11": for linked server "(null)", Query aborted- the maximum reject threshold (orows) was reached while regarding from an external source: 1 rows rejected out of total 1 rows processed
D. EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error [Unable to instantiate LoginClass] occurredwhile accessing external files
View answer
Correct Answer: A
Question #3
You have an Azure subscription that contains an Azure Synapse Analytics workspace named ws1 and an Azure Cosmos D6 database account named Cosmos1 Costmos1 contains a container named container 1 and ws1 contains a serverless1 SQL pool. you need to ensure that you can Query the data in container by using the serverless1 SQL pool. Which three actions should you perform? Each correct answer presents part of the solution NOTE: Each correct selection is worth one point.
A. nable Azure Synapse Link for Cosmos1
B. isable the analytical store for container1
C. n ws1
D. nable the analytical store for container1
E. isable indexing for container1
View answer
Correct Answer: ACD
Question #4
You have a Microsoft SQL Server database that uses a third normal form schema. You plan to migrate the data in the database to a star schema in an A?\ire Synapse Analytics dedicated SQI pool. You need to design the dimension tables. The solution must optimize read operations. What should you include in the solution? to answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
View answer
Correct Answer: D
Question #5
You have an Azure subscription that contains an Azure Data Factory data pipeline named Pipeline1, a Log Analytics workspace named LA1, and a storage account named account1. You need to retain pipeline-run data for 90 days. The solution must meet the following requirements: * The pipeline-run data must be removed automatically after 90 days. * Ongoing costs must be minimized. Which two actions should you perform? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one
A. onfigure Pipeline1 to send logs to LA1
B. rom the Diagnostic settings (classic) settings of account1
C. onfigure Pipeline1 to send logs to account1
D. rom the Data Retention settings of LA1, set the data retention period to 90 days
View answer
Correct Answer: AB
Question #6
You have an Azure Synapse Analytics dedicated SQL pod. You need to create a pipeline that will execute a stored procedure in the dedicated SQL pool and use the returned result set as the input (or a downstream activity. The solution must minimize development effort. Which Type of activity should you use in the pipeline?
A. otebook
B. -SQL
C. cript
D. tored Procedure
View answer
Correct Answer: D
Question #7
You have an Azure Data Factory pipeline named pipeline1 that is invoked by a tumbling window trigger named Trigger1. Trigger1 has a recurrence of 60 minutes. You need to ensure that pipeline1 will execute only if the previous execution completes successfully. How should you configure the self-dependency for Trigger1? Tumbling window self-dependency properties In scenarios where the trigger shouldn't proceed to the next window until the preceding window is successfully completed, build a self-dependency. A s
A. ffset: '-00:01:00' size: '00:01:00'
B. ffset: '01:00:00' size: '-01:00:00'
C. ffset: '01:00:00' size: '01:00:00'
D. ffset: '-01:00:00' size: '01:00:00'
View answer
Correct Answer: D
Question #8
You have a SQL pool in Azure Synapse.A user reports that queries against the pool take longer than expected to complete. You determine that the issue relates toqueried columnstore segments.You need to add monitoring to the underlying storage to help diagnose the issue.Which two metrics should you monitor? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point.
A. Snapshot Storage Size
B. Cache used percentage
C. DWU Limit
D. Cache hit percentage
View answer
Correct Answer: BD
Question #9
You have a data model that you plan to implement in a data warehouse in Azure Synapse Analytics as shown in the following exhibit. All the dimension tables will be less than 2 GB after compression, and the fact table will be approximately 6 TB. Which type of table should you use for each table? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
View answer
Correct Answer: A
Question #10
You create an Azure Databricks cluster and specify an additional library to install.When you attempt to load the library to a notebook, the library in not found.You need to identify the cause of the issue.What should you review?
A. notebook logs
B. cluster event logs
C. global init scripts logs
D. workspace logs
View answer
Correct Answer: C
Question #11
You plan to implement an Azure Data Lake Gen2 storage account. You need to ensure that the data lake will remain available if a data center fails in the primary Azure region. The solution must minimize costs. Which type of replication should you use for the storage account?
A. geo-redundant storage (GRS)
B. zone-redundant storage (ZRS)
C. locally-redundant storage (LRS)
D. geo-zone-redundant storage (GZRS)
View answer
Correct Answer: A
Question #12
A company purchases IoT devices to monitor manufacturing machinery. The company uses an Azure IoT Hub tocommunicate with the IoT devices.The company must be able to monitor the devices in real-time.You need to design the solution.What should you recommend?
A. Azure Analysis Services using Azure Portal
B. Azure Analysis Services using Azure PowerShell
C. Azure Stream Analytics cloud job using Azure Portal
D. Azure Data Factory instance using Microsoft Visual Studio
View answer
Correct Answer: C
Question #13
You have an Azure Synapse Analytics dedicated SQL pool that contains a large fact table. The table contains 50 columnsand 5 billion rows and is a heap.Most queries against the table aggregate values from approximately 100 million rows and return only two columns.You discover that the queries against the fact table are very slow.Which type of index should you add to provide the fastest query times?
A. nonclustered columnstore
B. clustered columnstore
C. nonclustered
D. clustered
View answer
Correct Answer: B
Question #14
You have an Azure data solution that contains an enterprise data warehouse in Azure Synapse Analytics named DW1.Several users execute ad hoc queries to DW1 concurrently.You regularly perform automated data loads to DW1.You need to ensure that the automated data loads have enough memory available to complete quickly and successfullywhen the adhoc queries run.What should you do?
A. Hash distribute the large fact tables in DW1 before performing the automated data loads
B. Assign a smaller resource class to the automated data load queries
C. Assign a larger resource class to the automated data load queries
D. Create sampled statistics for every column in each table of DW1
View answer
Correct Answer: C
Question #15
You have an Azure data factory.You need to examine the pipeline failures from the last 60 days.What should you use?
A. the Activity log blade for the Data Factory resource
B. the Monitor & Manage app in Data Factory
C. the Resource health blade for the Data Factory resource
D. Azure Monitor
View answer
Correct Answer: D
Question #16
You are monitoring an Azure Stream Analytics job.You discover that the Backlogged Input Events metric is increasing slowly and is consistently non-zero.You need to ensure that the job can handle all the events.What should you do?
A. Change the compatibility level of the Stream Analytics job
B. Increase the number of streaming units (SUs)
C. Remove any named consumer groups from the connection and use $default
D. Create an additional output stream for the existing input stream
View answer
Correct Answer: B
Question #17
You have an Azure Synapse Analytics dedicated SQL pool named Pool1 and a database named DB1. DB1 contains a facttable named Table1.You need to identify the extent of the data skew in Table1.What should you do in Synapse Studio?
A. Connect to the built-in pool and run DBCC PDW_SHOWSPACEUSED
B. Connect to the built-in pool and run DBCC CHECKALLOC
C. Connect to Pool1 and query sys
D. Connect to Pool1 and query sys
View answer
Correct Answer: D
Question #18
You are designing an Azure Stream Analytics job to process incoming events from sensors in retail environments. You need to process the events to produce a running average of shopper counts during the previous 15 minutes, calculated at five-minute intervals. Which type of window should you use?
A. snapshot
B. tumbling
C. hopping
D. sliding
View answer
Correct Answer: A
Question #19
You have an Azure Synapse Analytics dedicated SQL pool that contains a table named Table1. You have files that are ingested and loaded into an Azure Data Lake Storage Gen2 container named container1. You plan to insert data from the files into Table1 and azure Data Lake Storage Gen2 container named container1. You plan to insert data from the files into Table1 and transform the data. Each row of data in the files will produce one row in the serving layer of Table1. You need to ensure that when the source da
A. Yes
B. No
View answer
Correct Answer: B
Question #20
You are monitoring an Azure Stream Analytics job.The Backlogged Input Events count has been 20 for the last hour.You need to reduce the Backlogged Input Events count.What should you do?
A. Drop late arriving events from the job
B. Add an Azure Storage account to the job
C. Increase the streaming units for the job
D. Stop the job
View answer
Correct Answer: C
Question #21
You have a SQL pool in Azure Synapse.You discover that some queries fail or take a long time to complete.You need to monitor for transactions that have rolled back.Which dynamic management view should you query?
A. sys
B. sys
C. sys
D. sys
View answer
Correct Answer: B
Question #22
You plan to create an Azure Synapse Analytics dedicated SQL pool. You need to minimize the time it takes to identify queries that return confidential information as defined by the company's data privacy regulations and the users who executed the queues. Which two components should you include in the solution? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.
A. sensitivity-classification labels applied to columns that contain confidential information
B. resource tags for databases that contain confidential information
C. audit logs sent to a Log Analytics workspace
D. dynamic data masking for columns that contain confidential information
View answer
Correct Answer: C
Question #23
You have an Azure Active Directory (Azure AD) tenant that contains a security group named Group1. You have an Azure Synapse Analytics dedicated SQL pool named dw1 that contains a schema named schema1. You need to grant Group1 read-only permissions to all the tables and views in schema1. The solution must use the principle of least privilege. Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct
A. Mastered
B. Not Mastered
View answer
Correct Answer: B
Question #24
You are designing an Azure Databricks interactive cluster. The cluster will be used infrequently and will be configured forauto-termination.You need to use that the cluster configuration is retained indefinitely after the cluster is terminated. The solution mustminimize costs.What should you do?
A. Pin the cluster
B. Create an Azure runbook that starts the cluster every 90 days
C. Terminate the cluster manually when processing completes
D. Clone the cluster after it is terminated
View answer
Correct Answer: A
Question #25
You are designing an inventory updates table in an Azure Synapse Analytics dedicated SQL pool. The table will have aclustered columnstore index and will include the following columns:You identify the following usage patterns:Analysts will most commonly analyze transactions for a warehouse.Queries will summarize by product category type, date, and/or inventory event type.You need to recommend a partition strategy for the table to minimize query times.On which column should you partition the table?
A. EventTypeID
B. ProductCategoryTypeID
C. EventDate
D. WarehouseID
View answer
Correct Answer: D
Question #26
You are designing a star schema for a dataset that contains records of online orders. Each record includes an order date, anorder due date, and an order ship date.You need to ensure that the design provides the fastest query times of the records when querying for arbitrary date rangesand aggregating by fiscal calendar attributes.Which two actions should you perform? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point.
A. Create a date dimension table that has a DateTime key
B. Use built-in SQL functions to extract date attributes
C. Create a date dimension table that has an integer key in the format of YYYYMMDD
D. In the fact table, use integer columns for the date fields
E. Use DateTime columns for the date fields
View answer
Correct Answer: BD
Question #27
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution. After you answer a question in this scenario, you will NOT be able to return to it. As a result, these questions will not appear in the review screen. You have an Azure Storage account that contains 100 GB of files. The files co
A. Yes
B. No
View answer
Correct Answer: D
Question #28
You have an Azure SQL database named DB1 and an Azure Data Factory data pipeline named pipeline. From Data Factory, you configure a linked service to DB1. In DB1, you create a stored procedure named SP1. SP1 returns a single row of data that has four columns. You need to add an activity to pipeline to execute SP1. The solution must ensure that the values in the columns are stored as pipeline variables. Which two types of activities can you use to execute SP1? (Refer to Data Engineering on Microsoft Azure d
A. tored Procedure
B. ookup
C. cript
D. opy
View answer
Correct Answer: AB
Question #29
You need to design an Azure Synapse Analytics dedicated SQL pool that meets the following requirements: Can return an employee record from a given point in time. Maintains the latest employee information. Minimizes query complexity. How should you model the employee data?
A. as a temporal table
B. as a SQL graph table
C. as a degenerate dimension table
D. as a Type 2 slowly changing dimension (SCD) table
View answer
Correct Answer: A

View Answers after Submission

Please submit your email and WhatsApp to get the answers of questions.

Note: Please make sure your email ID and Whatsapp are valid so that you can get the correct exam results.

Email:
Whatsapp/phone number: