Weekend Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: suredis

Microsoft DP-600 Implementing Analytics Solutions Using Microsoft Fabric (beta) Exam Practice Test

Page: 1 / 8
Total 80 questions

Implementing Analytics Solutions Using Microsoft Fabric (beta) Questions and Answers

Testing Engine

  • Product Type: Testing Engine
$47.25  $134.99

PDF Study Guide

  • Product Type: PDF Study Guide
$40.25  $114.99
Question 1

You need to ensure the data loading activities in the AnalyticsPOC workspace are executed in the appropriate sequence. The solution must meet the technical requirements.

What should you do?

Options:

A.

Create a pipeline that has dependencies between activities and schedule the pipeline.

B.

Create and schedule a Spark job definition.

C.

Create a dataflow that has multiple steps and schedule the dataflow.

D.

Create and schedule a Spark notebook.

Question 2

You to need assign permissions for the data store in the AnalyticsPOC workspace. The solution must meet the security requirements.

Which additional permissions should you assign when you share the data store? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 3

Which type of data store should you recommend in the AnalyticsPOC workspace?

Options:

A.

a data lake

B.

a warehouse

C.

a lakehouse

D.

an external Hive metaStore

Question 4

You need to resolve the issue with the pricing group classification.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 5

What should you recommend using to ingest the customer data into the data store in the AnatyticsPOC workspace?

Options:

A.

a stored procedure

B.

a pipeline that contains a KQL activity

C.

a Spark notebook

D.

a dataflow

Question 6

You need to implement the date dimension in the data store. The solution must meet the technical requirements.

What are two ways to achieve the goal? Each correct answer presents a complete solution.

NOTE: Each correct selection is worth one point.

Options:

A.

Populate the date dimension table by using a dataflow.

B.

Populate the date dimension table by using a Stored procedure activity in a pipeline.

C.

Populate the date dimension view by using T-SQL.

D.

Populate the date dimension table by using a Copy activity in a pipeline.

Question 7

You need to design a semantic model for the customer satisfaction report.

Which data source authentication method and mode should you use? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 8

You need to create a DAX measure to calculate the average overall satisfaction score.

How should you complete the DAX code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 9

You need to recommend a solution to prepare the tenant for the PoC.

Which two actions should you recommend performing from the Fabric Admin portal? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

Options:

A.

Enable the Users can try Microsoft Fabric paid features option for specific security groups.

B.

Enable the Allow Azure Active Directory guest users to access Microsoft Fabric option for specific security groups.

C.

Enable the Users can create Fabric items option and exclude specific security groups.

D.

Enable the Users can try Microsoft Fabric paid features option for the entire organization.

E.

Enable the Users can create Fabric items option for specific security groups.

Question 10

You need to create a data loading pattern for a Type 1 slowly changing dimension (SCD).

Which two actions should you include in the process? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

Options:

A.

Update rows when the non-key attributes have changed.

B.

Insert new rows when the natural key exists in the dimension table, and the non-key attribute values have changed.

C.

Update the effective end date of rows when the non-key attribute values have changed.

D.

Insert new records when the natural key is a new value in the table.

Question 11

You have a Fabric tenant that contains a lakehouse named lakehouse1. Lakehouse1 contains a table named Table1.

You are creating a new data pipeline.

You plan to copy external data to Table1. The schema of the external data changes regularly.

You need the copy operation to meet the following requirements:

• Replace Table1 with the schema of the external data.

• Replace all the data in Table1 with the rows in the external data.

You add a Copy data activity to the pipeline. What should you do for the Copy data activity?

Options:

A.

From the Source tab, add additional columns.

B.

From the Destination tab, set Table action to Overwrite.

C.

From the Settings tab, select Enable staging

D.

From the Source tab, select Enable partition discovery

E.

From the Source tab, select Recursively

Question 12

You have a Fabric tenant that contains a warehouse. The warehouse uses row-level security (RLS). You create a Direct Lake semantic model that uses the Delta tables and RLS of the warehouse. When users interact with a report built from the model, which mode will be used by the DAX queries?

Options:

A.

DirectQuery

B.

Dual

C.

Direct Lake

D.

Import

Page: 1 / 8
Total 80 questions