Summer Special Flat 65% Limited Time Discount offer - Ends in 0d 00h 00m 00s - Coupon code: netdisc

Microsoft DP-600 Implementing Analytics Solutions Using Microsoft Fabric Exam Practice Test

Page: 1 / 12
Total 117 questions

Implementing Analytics Solutions Using Microsoft Fabric Questions and Answers

Testing Engine

  • Product Type: Testing Engine
$49  $139.99

PDF Study Guide

  • Product Type: PDF Study Guide
$42  $119.99
Question 1

You have a Fabric tenant that contains a semantic model named Model1. Model1 uses Import mode. Model1 contains a table named Orders. Orders has 100 million rows and the following fields.

You need to reduce the memory used by Model! and the time it takes to refresh the model. Which two actions should you perform? Each correct answer presents part of the solution. NOTE: Each correct answer is worth one point.

Options:

A.

Split OrderDateTime into separate date and time columns.

B.

Replace TotalQuantity with a calculated column.

C.

Convert Quantity into the Text data type.

D.

Replace TotalSalesAmount with a measure.

Question 2

You have a Fabric tenant that contains a semantic model. The model contains data about retail stores.

You need to write a DAX query that will be executed by using the XMLA endpoint The query must return a table of stores that have opened since December 1,2023.

How should you complete the DAX expression? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.

NOTE: Each correct selection is worth one point.

Options:

Question 3

You have a Fabric tenant that contains a warehouse.

You are designing a star schema model that will contain a customer dimension. The customer dimension table will be a Type 2 slowly changing dimension (SCD).

You need to recommend which columns to add to the table. The columns must NOT already exist in the source.

Which three types of columns should you recommend? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

Options:

A.

an effective end date and time

B.

a foreign key

C.

a surrogate key

D.

a natural key

E.

an effective start date and time

Question 4

You have a Microsoft Power B1 report and a semantic model that uses Direct Lake mode. From Power Si Desktop, you open Performance analyzer as shown in the following exhibit.

Use the drop-down menus to select the answer choice that completes each statement based on the information presented in the graphic. NOTE: Each correct selection is worth one point.

Options:

Question 5

You have a Fabric tenant tha1 contains a takehouse named Lakehouse1. Lakehouse1 contains a Delta table named Customer.

When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT performed on the table.

You need to identify whether maintenance tasks were performed on Customer.

Solution: You run the following Spark SQL statement:

REFRESH TABLE customer

Does this meet the goal?

Options:

A.

Yes

B.

No

Question 6

You have a Fabric workspace named Workspace1 that contains a dataflow named Dataflow1. Dataflow1 returns 500 rows of data.

You need to identify the min and max values for each column in the query results.

Which three Data view options should you select? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

Options:

A.

Show column quality details

B.

Show column profile in details pane

C.

Enable details pane

D.

Show column value distribution

E.

Enable column profile

Question 7

You need to provide Power Bl developers with access to the pipeline. The solution must meet the following requirements:

• Ensure that the developers can deploy items to the workspaces for Development and Test.

• Prevent the developers from deploying items to the workspace for Production.

• Follow the principle of least privilege.

Which three levels of access should you assign to the developers? Each correct answer presents part of the solution. NOTE: Each correct answer is worth one point.

Options:

A.

Build permission to the production semantic models

B.

Admin access to the deployment pipeline

C.

Viewer access to the Development and Test workspaces

D.

Viewer access to the Production workspace

E.

Contributor access to the Development and Test workspaces

F.

Contributor access to the Production workspace

Question 8

You have a Fabric tenant tha1 contains a takehouse named Lakehouse1. Lakehouse1 contains a Delta table named Customer.

When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT performed on the table.

You need to identify whether maintenance tasks were performed on Customer.

Solution: You run the following Spark SQL statement:

EXPLAIN TABLE customer

Does this meet the goal?

Options:

A.

Yes

B.

No

Question 9

You have a Fabric workspace named Workspace1 and an Azure SQL database.

You plan to create a dataflow that will read data from the database, and then transform the data by performing an inner join. You need to ignore spaces in the values when performing the inner join The solution must minimize development effort. What should you do?

Options:

A.

Append the queries by using fuzzy matching.

B.

Merge the queries by using fuzzy matching.

C.

Append the queries by using a lookup table.

D.

Merge the queries by using a lookup table.

Question 10

You have a Fabric tenant that contains a lakehouse.

You are using a Fabric notebook to save a large DataFrame by using the following code.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Options:

Question 11

You have a Fabric tenant that contains JSON files in OneLake. The files have one billion items.

You plan to perform time series analysis of the items.

You need to transform the data, visualize the data to find insights, perform anomaly detection, and share the insights with other business users. The solution must meet the following requirements:

♦ Use parallel processing.

♦ Minimize the duplication of data.

♦ Minimize how long it takes to load the data.

What should you use to transform and visualize the data?

Options:

A.

a Microsoft Power B1 report that uses core visuals

B.

the PySpark library in a Fabric notebook

C.

the pandas library in a Fabric notebook

Question 12

Note: This section contains one or more sets of questions with the same scenario and problem. Each question presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More than one solution in the set might solve the problem. It is also possible that none of the solutions in the set solve the problem.

After you answer a question in this section, you will NOT be able to return. As a result, these questions do not appear on the Review Screen.

Your network contains an on-premises Active Directory Domain Services (AD DS) domain named contoso.com that syncs with a Microsoft Entra tenant by using Microsoft Entra Connect.

You have a Fabric tenant that contains a semantic model.

You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.

You query a measure that includes the username () function, and the query returns a blank result.

You need to ensure that the measure returns the user principal name (UPN) of a user.

Solution: You update the measure to use the USEROBJECT () function.

Does this meet the goal?

Options:

A.

Yes

B.

No

Question 13

You have a Fabric tenant that contains a warehouse named Warehouse1. Warehouse1 contains three schemas named schemaA, schemaB. and schemaC

You need to ensure that a user named User1 can truncate tables in schemaA only.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 14

You have a Fabric tenant that contains a data pipeline.

You need to ensure that the pipeline runs every four hours on Mondays and Fridays.

To what should you set Repeat for the schedule?

Options:

A.

Daily

B.

By the minute

C.

Weekly

D.

Hourly

Question 15

You need to recommend a solution to group the Research division workspaces.

What should you include in the recommendation? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 16

You need to ensure that Contoso can use version control to meet the data analytics requirements and the general requirements. What should you do?

Options:

A.

Store all the semantic models and reports in Data Lake Gen2 storage.

B.

Modify the settings of the Research workspaces to use a GitHub repository.

C.

Store all the semantic models and reports in Microsoft OneDrive.

D.

Modify the settings of the Research division workspaces to use an Azure Repos repository.

Question 17

You need to refresh the Orders table of the Online Sales department. The solution must meet the semantic model requirements. What should you include in the solution?

Options:

A.

an Azure Data Factory pipeline that executes a dataflow to retrieve the minimum value of the OrderlD column in the destination lakehouse

B.

an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the maximum value of the OrderlD column in the destination lakehouse

C.

an Azure Data Factory pipeline that executes a dataflow to retrieve the maximum value of the OrderlD column in the destination lakehouse

D.

an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the minimum value of the OrderiD column m the

destination lakehouse

Question 18

Which workspace rote assignments should you recommend for ResearchReviewersGroupl and ResearchReviewersGroupZ? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 19

Which syntax should you use in a notebook to access the Research division data for Productlinel?

A)

B)

C)

D)

Options:

A.

Option A

B.

Option B

C.

Option C

D.

Option D

Question 20

You need to migrate the Research division data for Productline2. The solution must meet the data preparation requirements. How should you complete the code? To answer, select the appropriate options in the answer area

NOTE: Each correct selection is worth one point.

Options:

Question 21

What should you use to implement calculation groups for the Research division semantic models?

Options:

A.

DAX Studio

B.

Microsoft Power Bl Desktop

C.

the Power Bl service

D.

Tabular Editor

Question 22

You need to recommend which type of fabric capacity SKU meets the data analytics requirements for the Research division. What should you recommend?

Options:

A.

EM

B.

F

C.

P

D.

A

Question 23

You need to create a DAX measure to calculate the average overall satisfaction score.

How should you complete the DAX code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 24

You need to resolve the issue with the pricing group classification.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 25

You to need assign permissions for the data store in the AnalyticsPOC workspace. The solution must meet the security requirements.

Which additional permissions should you assign when you share the data store? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 26

You need to design a semantic model for the customer satisfaction report.

Which data source authentication method and mode should you use? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Options:

Question 27

Which type of data store should you recommend in the AnalyticsPOC workspace?

Options:

A.

a data lake

B.

a warehouse

C.

a lakehouse

D.

an external Hive metaStore

Question 28

You need to ensure the data loading activities in the AnalyticsPOC workspace are executed in the appropriate sequence. The solution must meet the technical requirements.

What should you do?

Options:

A.

Create a pipeline that has dependencies between activities and schedule the pipeline.

B.

Create and schedule a Spark job definition.

C.

Create a dataflow that has multiple steps and schedule the dataflow.

D.

Create and schedule a Spark notebook.

Page: 1 / 12
Total 117 questions