Weekend Special Limited Time Flat 70% Discount offer - Ends in 0d 00h 00m 00s - Coupon code: 70spcl

Microsoft DP-600 Implementing Analytics Solutions Using Microsoft Fabric Exam Practice Test

Page: 1 / 12
Total 117 questions

Implementing Analytics Solutions Using Microsoft Fabric Questions and Answers

Question 1

You have the source data model shown in the following exhibit.

Question # 1

The primary keys of the tables are indicated by a key symbol beside the columns involved in each key.

You need to create a dimensional data model that will enable the analysis of order items by date, product, and customer.

What should you include in the solution? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 1

Options:

Question 2

You have a Fabric tenant that contains a warehouse.

You use a dataflow to load a new dataset from OneLake to the warehouse.

You need to add a Power Query step to identify the maximum values for the numeric columns.

Which function should you include in the step?

Options:

A.

Table. MaxN

B.

Table.Max

C.

Table.Range

D.

Table.Profile

Question 3

You have a Microsoft Power Bl semantic model.

You need to identify any surrogate key columns in the model that have the Summarize By property set to a value other than to None. The solution must minimize effort.

What should you use?

Options:

A.

DAX Formatter in DAX Studio

B.

Model view in Microsoft Power Bl Desktop

C.

Model explorer in Microsoft Power Bl Desktop

D.

Best Practice Analyzer in Tabular Editor

Question 4

You have a Fabric workspace named Workspace1 that contains a dataflow named Dataflow1. Dataflow1 returns 500 rows of data.

You need to identify the min and max values for each column in the query results.

Which three Data view options should you select? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

Options:

A.

Show column quality details

B.

Show column profile in details pane

C.

Enable details pane

D.

Show column value distribution

E.

Enable column profile

Question 5

You have a Fabric tenant.

You are creating a Fabric Data Factory pipeline.

You have a stored procedure that returns the number of active customers and their average sales for the current month.

You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be available to the downstream activities of the pipeline.

Which type of activity should you add?

Options:

A.

Script

B.

Copy data

C.

Switch

D.

Append variable

Question 6

You have a Fabric tenant tha1 contains a takehouse named Lakehouse1. Lakehouse1 contains a Delta table named Customer.

When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT performed on the table.

You need to identify whether maintenance tasks were performed on Customer.

Solution: You run the following Spark SQL statement:

EXPLAIN TABLE customer

Does this meet the goal?

Options:

A.

Yes

B.

No

Question 7

You have a Fabric tenant that contains 30 CSV files in OneLake. The files are updated daily.

You create a Microsoft Power Bl semantic model named Modell that uses the CSV files as a data source. You configure incremental refresh for Model 1 and publish the model to a Premium capacity in the Fabric tenant.

When you initiate a refresh of Model1, the refresh fails after running out of resources.

What is a possible cause of the failure?

Options:

A.

Query folding is occurring.

B.

Only refresh complete days is selected.

C.

XMLA Endpoint is set to Read Only.

D.

Query folding is NOT occurring.

E.

The data type of the column used to partition the data has changed.

Question 8

You have a Fabric tenant that contains a Microsoft Power Bl report.

You are exploring a new semantic model.

You need to display the following column statistics:

• Count

• Average

• Null count

• Distinct count

• Standard deviation

Which Power Query function should you run?

Options:

A.

Table. FuzzyGroup

B.

Table.Profile

C.

Table.View

D.

Table.Schema

Question 9

You have a Fabric notebook that has the Python code and output shown in the following exhibit.

Question # 9

Question # 9

Which type of analytics are you performing?

Options:

A.

predictive

B.

descriptive

C.

prescriptive

D.

diagnostic

Question 10

You have a Fabric tenant that contains a new semantic model in OneLake.

You use a Fabric notebook to read the data into a Spark DataFrame.

You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and numeric columns.

Solution: You use the following PySpark expression:

df .sumary ()

Does this meet the goal?

Options:

A.

Yes

B.

No

Question 11

You have a Fabric tenant that contains a warehouse named WH1. You run the following T-SQL query against WH1.

Question # 11

For each of the following statements, select Yes if the statement is true. Otherwise, select No. NOTE: Each correct selection is worth one point.

Question # 11

Options:

Question 12

You need to create a data loading pattern for a Type 1 slowly changing dimension (SCD).

Which two actions should you include in the process? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

Options:

A.

Update rows when the non-key attributes have changed.

B.

Insert new rows when the natural key exists in the dimension table, and the non-key attribute values have changed.

C.

Update the effective end date of rows when the non-key attribute values have changed.

D.

Insert new records when the natural key is a new value in the table.

Question 13

You need to ensure the data loading activities in the AnalyticsPOC workspace are executed in the appropriate sequence. The solution must meet the technical requirements.

What should you do?

Options:

A.

Create a pipeline that has dependencies between activities and schedule the pipeline.

B.

Create and schedule a Spark job definition.

C.

Create a dataflow that has multiple steps and schedule the dataflow.

D.

Create and schedule a Spark notebook.

Question 14

Which type of data store should you recommend in the AnalyticsPOC workspace?

Options:

A.

a data lake

B.

a warehouse

C.

a lakehouse

D.

an external Hive metaStore

Question 15

You to need assign permissions for the data store in the AnalyticsPOC workspace. The solution must meet the security requirements.

Which additional permissions should you assign when you share the data store? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 15

Options:

Question 16

You need to recommend a solution to prepare the tenant for the PoC.

Which two actions should you recommend performing from the Fabric Admin portal? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

Options:

A.

Enable the Users can try Microsoft Fabric paid features option for specific security groups.

B.

Enable the Allow Azure Active Directory guest users to access Microsoft Fabric option for specific security groups.

C.

Enable the Users can create Fabric items option and exclude specific security groups.

D.

Enable the Users can try Microsoft Fabric paid features option for the entire organization.

E.

Enable the Users can create Fabric items option for specific security groups.

Question 17

You need to design a semantic model for the customer satisfaction report.

Which data source authentication method and mode should you use? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 17

Options:

Question 18

NO: 8

You need to implement the date dimension in the data store. The solution must meet the technical requirements.

What are two ways to achieve the goal? Each correct answer presents a complete solution.

NOTE: Each correct selection is worth one point.

Options:

A.

Populate the date dimension table by using a dataflow.

B.

Populate the date dimension table by using a Stored procedure activity in a pipeline.

C.

Populate the date dimension view by using T-SQL.

D.

Populate the date dimension table by using a Copy activity in a pipeline.

Question 19

You need to resolve the issue with the pricing group classification.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 19

Question # 19

Options:

Question 20

What should you recommend using to ingest the customer data into the data store in the AnatyticsPOC workspace?

Options:

A.

a stored procedure

B.

a pipeline that contains a KQL activity

C.

a Spark notebook

D.

a dataflow

Question 21

You need to create a DAX measure to calculate the average overall satisfaction score.

How should you complete the DAX code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 21

Question # 21

Options:

Question 22

Which workspace rote assignments should you recommend for ResearchReviewersGroupl and ResearchReviewersGroupZ? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 22

Options:

Question 23

You need to recommend which type of fabric capacity SKU meets the data analytics requirements for the Research division. What should you recommend?

Options:

A.

EM

B.

F

C.

P

D.

A

Question 24

You need to refresh the Orders table of the Online Sales department. The solution must meet the semantic model requirements. What should you include in the solution?

Options:

A.

an Azure Data Factory pipeline that executes a dataflow to retrieve the minimum value of the OrderlD column in the destination lakehouse

B.

an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the maximum value of the OrderlD column in the destination lakehouse

C.

an Azure Data Factory pipeline that executes a dataflow to retrieve the maximum value of the OrderlD column in the destination lakehouse

D.

an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the minimum value of the OrderiD column m the

destination lakehouse

Page: 1 / 12
Total 117 questions