0% found this document useful (0 votes)
757 views

DP-600

Uploaded by

John basha Shaik
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
757 views

DP-600

Uploaded by

John basha Shaik
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 212

Question: 1

Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -


Contoso identifies the following data analytics requirements:
All the workspaces for the Sales division and the Research division must support all Fabric experiences.
The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
The Research division workspaces must be grouped together logically to support OneLake data hub filtering based
on the department name.
For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -


Contoso identifies the following data preparation requirements:
The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -


Contoso identifies the following requirements for implementing and managing semantic models:
The number of rows added to the Orders table during refreshes must be minimized.
The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
You need to ensure that Contoso can use version control to meet the data analytics requirements and the general
requirements.
What should you do?

A.Store at the semantic models and reports in Data Lake Gen2 storage.
B.Modify the settings of the Research workspaces to use a GitHub repository.
C.Modify the settings of the Research division workspaces to use an Azure Repos repository.
D.Store all the semantic models and reports in Microsoft OneDrive.

Answer: C

Explanation:

C. Modify the settings of the Research division workspaces to use an Azure Repos repository.

Contoso needs version control for its data analytics and general requirements. In Microsoft Fabric and Power
BI, version control is best managed through Git integration using Azure Repos or GitHub.

Azure Repos is a Git-based repository within Azure DevOps, providing versioning, tracking, and collaboration
for code, reports, and semantic models.

This is the most appropriate choice for version-controlling Power BI artifacts like datasets, reports, and
semantic models in a workspace.

Reference:

https://learn.microsoft.com/en-us/fabric/cicd/git-integration/intro-to-git-integration

Question: 2
HOTSPOT -

Case study -

To start the case study -

the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -


Contoso identifies the following data analytics requirements:
All the workspaces for the Sales division and the Research division must support all Fabric experiences.
The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
The Research division workspaces must be grouped together logically to support OneLake data hub filtering based
on the department name.
For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -


Contoso identifies the following data preparation requirements:
The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -


Contoso identifies the following requirements for implementing and managing semantic models:
The number of rows added to the Orders table during refreshes must be minimized.
The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
You need to recommend a solution to group the Research division workspaces.
What should you include in the recommendation? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:

Explanation:

Domain: Group the Research division workspaces based on their departmental context.

Domain allows you to group workspaces based on their purpose or business context. In this case, grouping by
department (Research division) aligns with the requirement.

The Fabric Admin Portal is the primary management interface for Microsoft Fabric.

Reference:

https://learn.microsoft.com/en-us/fabric/governance/domains#configure-domain-settings

Question: 3
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -


Contoso identifies the following data analytics requirements:
All the workspaces for the Sales division and the Research division must support all Fabric experiences.
The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
The Research division workspaces must be grouped together logically to support OneLake data hub filtering based
on the department name.
For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -


Contoso identifies the following data preparation requirements:
The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -


Contoso identifies the following requirements for implementing and managing semantic models:
The number of rows added to the Orders table during refreshes must be minimized.
The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
You need to refresh the Orders table of the Online Sales department. The solution must meet the semantic model
requirements.
What should you include in the solution?

A.an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the maximum value of
the OrderID column in the destination lakehouse
B.an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the minimum value of the
OrderID column in the destination lakehouse
C.an Azure Data Factory pipeline that executes a dataflow to retrieve the minimum value of the OrderID column
in the destination lakehouse
D.an Azure Data Factory pipeline that executes a dataflow to retrieve the maximum value of the OrderID
column in the destination lakehouse

Answer: D
Explanation:

D. an Azure Data Factory pipeline that executes a dataflow to retrieve the maximum value of the OrderID
column in the destination lakehouse. A dataflow can be used to retrieve the max OrderID number (stored in
the destination table - OrderID is a sequencial number). This number can be used to set from which row data
must be added to the destination table (implementing an incremental load).

Question: 4
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -


Contoso identifies the following data analytics requirements:
All the workspaces for the Sales division and the Research division must support all Fabric experiences.
The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
The Research division workspaces must be grouped together logically to support OneLake data hub filtering based
on the department name.
For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -


Contoso identifies the following data preparation requirements:
The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -


Contoso identifies the following requirements for implementing and managing semantic models:
The number of rows added to the Orders table during refreshes must be minimized.
The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
Which syntax should you use in a notebook to access the Research division data for Productline1?

A.spark.read.format(“delta”).load(“Tables/productline1/ResearchProduct”)
B.spark.sql(“SELECT * FROM Lakehouse1.ResearchProduct ”)
C.external_table(‘Tables/ResearchProduct)
D.external_table(ResearchProduct)

Answer: B

Explanation:

spark.sql(“SELECT * FROM Lakehouse1.ResearchProduct ”)


The syntax of C and D is correct for KQL databases (incorrect in this use-case). When the shortcut is created,
no additional folders have been added to the Tables section, therefore answer A is incorrect. Once created,
the line of answer B can be used to access data correctly.

https://learn.microsoft.com/en-us/fabric/onelake/onelake-shortcuts

Question: 5
HOTSPOT -

Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -
Litware has data that must be analyzed as shown in the following table.

The Product data contains a single table and the following columns.
The customer satisfaction data contains the following tables:

Survey -

Question -

Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -

Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)

A custom semantic model -

A default semantic model -

Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
You need to assign permissions for the data store in the AnalyticsPOC workspace. The solution must meet the
security requirements.
Which additional permissions should you assign when you share the data store? To answer, select the appropriate
options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:

Data Engineers: Read all Apache Spark - because they need to be able to work with Spark for Data curation.

Data Analysts: Build Reports on the default dataset - because they are report builders

Data Scientists: Read All SQL analytics Endpoints data_ They leverage curated data (by engineers) to do
predictive analytics.

Question: 6
HOTSPOT -

Case study -

To start the case study -

as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -
Litware has data that must be analyzed as shown in the following table.

The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

Survey -

Question -

Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -

Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)
A custom semantic model -

A default semantic model -

Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.

Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
You need to create a DAX measure to calculate the average overall satisfaction score.
How should you complete the DAX code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.

Answer:
Explanation:

1. Average(Survey[Response Value]) as this one is the only value that can be aggregated.

Customer Key Does not make any sense , as it just a key to identify a customer, this can never provide us the
score.

2. Period : Variable is defined as to select 1 year of date range. Can directly be passed in the Filter context of
Calculate formula.

Question: 7
HOTSPOT -

Case study -

To start the case study -

as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -
Litware has data that must be analyzed as shown in the following table.

The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

Survey -

Question -

Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -

Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)
A custom semantic model -

A default semantic model -

Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.

Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
You need to resolve the issue with the pricing group classification.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.

Answer:
Explanation:

VIEW: from an existing table.

* CASE: correct syntax before the WHENs.

* WHEN ListPrice BETWEEN 50 AND 1000 THEN 'medium': the other two options miss value 1000; on the other
hand, the BETWEEN includes both 50 and 1000.

Question: 8
Case study -
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -
Litware has data that must be analyzed as shown in the following table.

The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

Survey -

Question -

Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -

Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)

A custom semantic model -

A default semantic model -

Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.

Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
What should you recommend using to ingest the customer data into the data store in the AnalyticsPOC
workspace?

A.a stored procedure


B.a pipeline that contains a KQL activity
C.a Spark notebook
D.a dataflow

Answer: D
Explanation:

D. a dataflow.

Even though the text reads "Data will be loaded without transformation in one area of the AnalyticsPOC data
store": in general, dataflows are used when data transformations are involved after ingestion. As suggested
by user BHARAT, the Copy Activity should be the optimal solution.

Question: 9
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

Survey -

Question -

Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -

Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)

A custom semantic model -

A default semantic model -

Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.

Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
Which type of data store should you recommend in the AnalyticsPOC workspace?

A.a data lake


B.a warehouse
C.a lakehouse
D.an external Hive metastore

Answer: C
Explanation:

A lakehouse The data store must handle semi-structured and unstructured data, therefore a Lakehouse
should be the optimal solution supporting read access with T-SQL and Python.

Question: 10
You have a Fabric warehouse that contains a table named Staging.Sales. Staging.Sales contains the following
columns.

You need to write a T-SQL query that will return data for the year 2023 that displays ProductID and ProductName
and has a summarized Amount that is higher than 10,000.
Which query should you use?

A.

B.

C.

D.

Answer: A

Explanation:

SELECT ProductID, ProductName, SUM(Amount) AS TotalAmount


FROM Staging.Sales

WHERE DATEPART(YEAR, SaleDate) = '2023'

GROUP BY ProductID, ProductName

Answer B is excluded because the filter on year should be included in the WHERE-clause. It is the result of a
function, but it is not an aggregate.

Answer D is excluded because the HAVING-clause does not accept aliases.

Question: 11
HOTSPOT -
You have a data warehouse that contains a table named Stage.Customers. Stage.Customers contains all the
customer record updates from a customer relationship management (CRM) system. There can be multiple updates
per customer.
You need to write a T-SQL query that will return the customer ID, name. postal code, and the last updated time of
the most recent row for each customer ID.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.

Answer:
Explanation:

Box 1: ROW_NUMBER()

ROW_NUMBER() is a window function that assigns a unique sequential number to each row within a partition,
based on the ORDER BY clause.

In this case, it is used to assign a row number to each CustomerID, ordering by LastUpdated DESC.

This ensures that the most recent record for each CustomerID gets ROW_NUMBER() = 1.

Box 2: WHERE X = 1

The alias X stores the row number assigned by ROW_NUMBER().

Since we only want the most recent row per CustomerID, we filter for X = 1.

This approach ensures we get only the latest entry for each customer, based on LastUpdated.

Question: 12
HOTSPOT -
You have a Fabric tenant.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:

Explanation:

"The code embeds an existing Power BI report."

Answer: No

Embedding a Power BI report typically involves using the Power BI Embedded API, which is not indicated here.

"The code creates a Power BI report."

Answer: Yes

If the code generates a new Power BI report (e.g., using Python, Power BI REST API, or Power BI Desktop
automation), then this is correct.

"The code displays a summary of the DataFrame."

Answer: No

Displaying a summary of a DataFrame would involve functions like df.describe(), df.info(), or df.head() in
Pandas. Since this is about Power BI, this function is not relevant.
Question: 13
You are the administrator of a Fabric workspace that contains a lakehouse named Lakehouse1. Lakehouse1
contains the following tables:
Table1: A Delta table created by using a shortcut
Table2: An external table created by using Spark

Table3: A managed table -


You plan to connect to Lakehouse1 by using its SQL endpoint.
What will you be able to do after connecting to Lakehouse1?

A.Read Table3.
B.Update the data Table3.
C.Read Table2.
D.Update the data in Table1.

Answer: A

Explanation:

The right answer is A. A managed table, it is stored within the Fabric storage and becomes immediately
accessible through the SQL endpoint upon connection.D is not right becouse tables created using shortcuts
might not be immediately accessible through the SQL endpoint for updates.

Question: 14
You have a Fabric tenant that contains a warehouse.
You use a dataflow to load a new dataset from OneLake to the warehouse.
You need to add a PowerQuery step to identify the maximum values for the numeric columns.
Which function should you include in the step?

A.Table.MaxN
B.Table.Max
C.Table.Range
D.Table.Profile

Answer: D

Explanation:

The Table.Profile function in PowerQuery is specifically designed to provide statistical information about a
table's columns, including the maximum values for numeric columns. When you apply Table.Profile to your
dataset, it generates a summary of each column, which includes the minimum, maximum, average, standard
deviation, count of distinct values, and other useful statistics.

Reference:

https://learn.microsoft.com/en-us/powerquery-m/table-profile

Question: 15

You need to use the model to generate predictions by using the PREDICT function in a Fabric notebook.
NOTE: Each correct answer is worth one point.

A.T-SQL
B.DAX
C.Spark SQL
D.PySpark

Answer: CD

Explanation:

Spark SQL: This language leverages the Spark framework, which is commonly used for big data processing
and machine learning. Spark SQL allows you to run SQL queries over data and includes support for machine
learning model scoring with the PREDICT function.

PySpark: This is the Python API for Spark. PySpark allows you to interact with Spark using Python, which is
one of the most popular programming languages for data analysis and machine learning. PySpark includes
capabilities for integrating with machine learning models and running predictions using the PREDICT function.

Why Not the Other Options:

T-SQL: This is Transact-SQL, primarily used for managing and querying data in SQL Server and Azure SQL
Database. It doesn't support the PREDICT function for machine learning model scoring in Fabric notebooks.

DAX: This is Data Analysis Expressions, a language used in Power BI, Power Pivot, and Analysis Services for
data modeling and calculations. It doesn't support the PREDICT function for machine learning model scoring
in Fabric notebooks.

Question: 16

You have a Spark DataFrame assigned to a variable named df.


You need to use the Chart view in the notebook to explore the data manually.
Which function should you run to make the data available in the Chart view?

A.displayHTML
B.show

Answer: D

Explanation:

Display: This function is specifically designed for rendering visual representations of data within interactive
notebooks. When you use the display function on a Spark DataFrame, it enables the data to be shown in a

Question: 17
You need all rows to appear in the visual.
What should you do?

Modify the Sort Column By property for all columns.


Add a unique field to each row.
Modify the Summarize By property for all columns.

Answer: C
Explanation:

Question: 18
DRAG DROP -
You have a Fabric tenant that contains a semantic model. The model contains data about retail stores.
You need to write a DAX query that will be executed by using the XMLA endpoint. The query must return a table of
stores that have opened since December 1, 2023.
How should you complete the DAX expression? To answer, drag the appropriate values to the correct targets.
Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or
scroll to view content.
NOTE: Each correct selection is worth one point.

Answer:
Explanation:

DEFINE

This section defines a variable (VAR _SalesSince) that stores a specific date (2023-12-01).

EVALUATE

This keyword is required for executing and returning a table result in DAX queries.

SUMMARIZE

Used to create a summary table with specific columns.

Here, it summarizes data from the Store table, including Store[Name] and Store[OpenDate].

Question: 19
What can you identify about the pickupLongitude column?

A. The column has duplicate values.


B. All the table rows are profiled.
C. The column has missing values.
D. There are 935 values that occur only once.

Answer: A
Explanation:

Answer A

Distinct Value: This refers to all different values present in a dataset. When you retrieve distinct values from a
column, you eliminate duplicate values so that each value is shown once. For example, if a column contains
the values [1, 2, 2, 3, 3, 3], the distinct values would be [1, 2, 3].

Unique Value: This usually refers to values that appear only once in the dataset. Unlike distinct values, a
unique value will only be considered if it has no duplicates at all. For example, if a column contains the values
[1, 2, 2, 3, 3, 3], the unique values would be [1], since only 1 appears without repetition.

Question: 20
You have a Fabric tenant named Tenant1 that contains a workspace named WS1. WS1 uses a capacity named C1
and contains a dataset named DS1.

the DS1 settings


the WS1 settings
the C1 settings
the Tenant1 settings

Answer: C

Explanation:

As XMLA is set to Read-Only first, you must go to the capacity settings to enable read-write.

https://learn.microsoft.com/en-us/power-bi/enterprise/service-premium-connect-tools#enable-xmla-read-

Question: 21

You need to recommend a solution to provide users with the ability to create and publish custom Direct Lake

NOTE: Each correct answer is worth one point.


A. From the Tenant settings, set Allow XMLA Endpoints and Analyze in Excel with on-premises datasets to
Enabled.
B. From the Tenant settings, set Allow Azure Active Directory guest users to access Microsoft Fabric to
Enabled.
C. From the Tenant settings, select Users can edit data model in the Power BI service.
D. From the Capacity settings, set XMLA Endpoint to Read Write.
E. From the Tenant settings, set Users can create Fabric items to Enabled.
F. From the Tenant settings, enable Publish to Web.

Answer: ADE
Explanation:

A Allow XMLA Endpoints and Analyze in Excel with on-premises datasets:

From the Tenant settings, set Allow XMLA Endpoints and Analyze in Excel with on-premises datasets to
Enabled. This allows users to interact with the dataset via XMLA endpoints and analyze data using Excel.

D. Set XMLA Endpoint to Read Write:

In the Capacity settings, configure the XMLA Endpoint to Read Write. This ensures that users have the
necessary permissions to create and modify semantic models through external tools.

E Enable Users to Create Fabric Items:

From the Tenant settings, set Users can create Fabric items to Enabled. This grants users the ability to create
custom semantic models within the Fabric workspace.

Question: 22
You are creating a semantic model in Microsoft Power BI Desktop.
You plan to make bulk changes to the model by using the Tabular Model Definition Language (TMDL) extension for
Microsoft Visual Studio Code.
You need to save the semantic model to a file.
Which file format should you use?

A.PBIP
B.PBIX
C.PBIT
D.PBIDS

Answer: A

Explanation:

A. PBIP.

PBIP (Power BI Project) is a file format that supports the open-source TMDL format and is designed for
integrating Power BI with external development environments like Visual Studio Code.

PBIX is the common Power BI report file format but is not intended for bulk edits through TMDL.

PBIT is a Power BI template file, used for creating new reports based on an existing structure but not for bulk
editing in Visual Studio Code.

PBIDS is for creating Power BI dataset connections and is unrelated to TMDL editing.
Question: 23
HOTSPOT -
You have a Fabric tenant that contains a warehouse named Warehouse1. Warehouse1 contains three schemas
named schemaA, schemaB, and schemaC.
You need to ensure that a user named User1 can truncate tables in schemaA only.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.

Answer:

Explanation:

ALTER: Allows modification of schema objects .

SCHEMA::schemaA: Grants permissions on an entire schema instead of an individual database or object .

Question: 24

Ensure that the developers can deploy items to the workspaces for Development and Test.

Follow the principle of least privilege.


Which three levels of access should you assign to the developers? Each correct answer presents part of the

NOTE: Each correct answer is worth one point.

Build permission to the production semantic models


Admin access to the deployment pipeline
C.Viewer access to the Development and Test workspaces
D.Viewer access to the Production workspace
E.Contributor access to the Development and Test workspaces
F.Contributor access to the Production workspace

Answer: BDE
Explanation:

Admin access to the deployment pipeline : Admin access to the deployment pipeline is necessary to manage
the pipeline stages and deployment processes. This allows developers to configure deployments between
stages but does not grant direct access to the workspaces themselves.

Viewer access to the Production workspace : Providing Viewer access to the Production workspace ensures
that developers can see the items in the Production stage but cannot make changes or deploy items to the
Production workspace. This restriction aligns with the requirement to prevent developers from deploying to
the Production workspace.

Contributor access to the Development and Test workspaces : Contributor access to the Development and
Test workspaces allows developers to deploy and modify items within these stages. This access is necessary
to enable developers to work on and test their items before they move to production.

Why Other Options Are Not Selected:

A. Build permission to the production semantic models: This does not relate directly to the deployment
pipeline access requirements and does not provide the needed permissions for deployment.

C. Viewer access to the Development and Test workspaces: Viewer access would only allow developers to
see the items without making any changes or deployments, which does not meet the requirement to deploy
items to these workspaces.

F. Contributor access to the Production workspace: This would grant developers the ability to deploy and
modify items in the Production workspace, which contradicts the requirement to prevent developers from
deploying to the Production stage.

Question: 25
You have a Fabric workspace that contains a DirectQuery semantic model. The model queries a data source that
has 500 million rows.
You have a Microsoft Power Bi report named Report1 that uses the model. Report1 contains visuals on multiple
pages.
You need to reduce the query execution time for the visuals on all the pages.
What are two features that you can use? Each correct answer presents a complete solution,
NOTE: Each correct answer is worth one point.

A.user-defined aggregations
B.automatic aggregation
C.query caching
D.OneLake integration

Answer: AB
Explanation:

A. User-defined aggregations.
Users define aggregations based on expected queries, reducing the need to scan massive datasets.

Automatic aggregations optimize query performance by learning from query patterns and storing frequently
used aggregate results in memory.

Question: 26
You have a Fabric tenant that contains 30 CSV files in OneLake. The files are updated daily.
You create a Microsoft Power BI semantic model named Model1 that uses the CSV files as a data source. You
configure incremental refresh for Model1 and publish the model to a Premium capacity in the Fabric tenant.
When you initiate a refresh of Model1, the refresh fails after running out of resources.
What is a possible cause of the failure?

A. Query folding is occurring.


B. Only refresh complete days is selected.
C. XMLA Endpoint is set to Read Only.
D. Query folding is NOT occurring.
E. The delta type of the column used to partition the data has changed.

Answer: D
Explanation:

D. Query folding is NOT occurring.

As described in Incremental refresh and real-time data for models - Requirements, incremental refresh is
designed for data sources that support query folding. Make sure data source queries are being folded in
Power BI Desktop before publishing to the service, where query folding issues can be significantly
compounded.

So having said this, csv files is not a data source that support query folding.

Reference:

https://learn.microsoft.com/en-us/power-bi/connect-data/incremental-refresh-troubleshoot#problem-
loading-data-takes-too-long

Question: 27
You have a Fabric tenant that uses a Microsoft Power BI Premium capacity.

What should you do first?

At the semantic model level, set Large dataset storage format to Off.
At the tenant level, set Create and use Metrics to Enabled.
At the semantic model level, set Large dataset storage format to On.
At the tenant level, set Data Activator to Enabled.
Answer: C
Explanation:

At the semantic model level, set Large dataset storage format to On.

Reference:

https://learn.microsoft.com/en-us/power-bi/enterprise/service-premium-scale-out-configure

Question: 28
You have a Fabric tenant that contains a warehouse. The warehouse uses row-level security (RLS).
You create a Direct Lake semantic model that uses the Delta tables and RLS of the warehouse.
When users interact with a report built from the model, which mode will be used by the DAX queries?

A.DirectQuery
B.Dual
C.Direct Lake
D.Import

Answer: A

Explanation:

A. Direct Query "Row-level security only applies to queries on a Warehouse or SQL analytics endpoint in
Fabric. Power BI queries on a warehouse in Direct Lake mode will fall back to Direct Query mode to abide by
row-level security."

When users interact with a report built from a Direct Lake semantic model, which leverages row-level security
(RLS) and Delta tables from a warehouse, the DAX queries will operate in DirectQuery mode.

Reference:

https://learn.microsoft.com/en-us/fabric/data-warehouse/row-level-security

Question: 29

You need to create a diagram of the model. The diagram must contain only the Sales table and related tables.

Data view

Answer: C

Explanation:

C. Model view.
Model view in Microsoft Power BI Desktop allows you to visualize and manage relationships between tables in
a semantic model.

Since the requirement is to create a diagram that includes only the Sales table and its related tables, Model
view is the best choice.

In Model view, you can filter and arrange tables to focus only on relevant parts of the data model, making it
easier to analyze relationships.

Why not the other options?

Question: 30
You have a Fabric tenant that contains a semantic model. The model uses Direct Lake mode.
You suspect that some DAX queries load unnecessary columns into memory.
You need to identify the frequently used columns that are loaded into memory.
What are two ways to achieve the goal? Each correct answer presents a complete solution.
NOTE: Each correct answer is worth one point.

A. Use the Analyze in Excel feature.


B. Use the Vertipaq Analyzer tool.
C. Query the $System.DISCOVER_STORAGE_TABLE_COLUMN_SEGMENTS dynamic management view (DMV).
D. Query the DISCOVER_MEMORYGRANT dynamic management view (DMV).

Answer: BC

Explanation:

B. Use the Vertipaq Analyzer tool.

Vertipaq Analyzer: This tool helps analyze the internal structure of your Power BI model. It provides detailed
information about the storage and memory usage of your model, including which columns are frequently
accessed and loaded into memory. This can help you identify unnecessary columns that are consuming
resources.

Steps:

Export your Power BI model to a .pbix file.

Open the .pbix file in Power BI Desktop.

Use the Vertipaq Analyzer tool to analyze the model and review the column usage statistics.

C. Query the $System.DISCOVER_STORAGE_TABLE_COLUMN_SEGMENTS dynamic management view


(DMV).

DMVs: Dynamic Management Views (DMVs) provide detailed information about the operations of your Power
BI models. Specifically, the $System.DISCOVER_STORAGE_TABLE_COLUMN_SEGMENTS DMV can give you
insights into the storage and usage patterns of individual columns within your model.
Question: 31
HOTSPOT -
You have the source data model shown in the following exhibit.

The primary keys of the tables are indicated by a key symbol beside the columns involved in each key.
You need to create a dimensional data model that will enable the analysis of order items by date, product, and
customer.
What should you include in the solution? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.

Answer:
Explanation:

1. "The relationship between OrderItem and Product must be based on: The ProductID column.

This selection indicates that the relationship between OrderItem and Product is based only on ProductID.

This implies that a ProductID is unique across the system or that OrderItem only needs to reference a product
without considering other attributes like CompanyID.

If products can exist independently and are not company-specific, then this choice is valid.

However, if multiple companies sell the same ProductID, it may be necessary to include CompanyID to avoid
ambiguity.

2. "The Company entity must be: Denormalized into the Customer and Product entities.

Denormalization means storing redundant data in multiple places to improve performance, at the cost of
increased storage and potential update anomalies.

Here, the Company entity is not kept as a separate table but instead integrated into both the Customer and
Product tables.

This can reduce the need for joins when querying customer and product data but might introduce redundancy.

Question: 32
You have a Fabric tenant that contains a semantic model named Model1. Model1 uses Import mode. Model1
contains a table named Orders. Orders has 100 million rows and the following fields.

You need to reduce the memory used by Model1 and the time it takes to refresh the model.
Which two actions should you perform? Each correct answer presents part of the solution.

A.Split OrderDateTime into separate date and time columns.


B.Replace TotalQuantity with a calculated column.
C.Convert Quantity into the Text data type.
D.Replace TotalSalesAmount with a measure.

Answer: AD
Explanation:

A. Split OrderDateTime into separate date and time columns .

In Import mode, Power BI compresses data more efficiently when Date and Time are stored separately instead
of a single DateTime column.

DateTime values have higher cardinality (more unique values), which increases memory usage.

Splitting into Date and Time reduces the number of unique values, improving compression and reducing
memory usage.

D. Replace TotalSalesAmount with a measure .

Measures in Power BI are calculated at query time and do not consume additional memory, unlike calculated
columns, which take up storage.

TotalSalesAmount is likely derived from Quantity * UnitPrice, so instead of storing it as a column, it should be
computed dynamically using a DAX measure.

This reduces model size and improves performance.

Question: 33
You have a Fabric tenant that contains a semantic model.
You need to prevent report creators from populating visuals by using implicit measures.
What are two tools that you can use to achieve the goal? Each correct answer presents a complete solution.
NOTE: Each correct answer is worth one point.

A.Microsoft Power BI Desktop


B.Tabular Editor
C.Microsoft SQL Server Management Studio (SSMS)
D.DAX Studio

Answer: AB
Explanation:

A. Microsoft Power BI Desktop: Power BI Desktop allows you to control and manage how measures are used
within your reports. By carefully defining and using explicit measures within your data model, you can ensure
that report creators use only these predefined measures instead of creating implicit measures automatically.

B. Tabular Editor: Tabular Editor is a powerful tool for managing and editing Power BI and Analysis Services
tabular models. It allows you to enforce best practices, such as disabling implicit measures, by modifying the
model's properties and ensuring that only explicit measures are available for use in reports.

Question: 34
HOTSPOT -
You have a Fabric tenant that contains two lakehouses.
You are building a dataflow that will combine data from the lakehouses. The applied steps from one of the queries
in the dataflow is shown in the following exhibit.

Use the drop-down menus to select the answer choice that completes each statement based on the information
presented in the graphic.
NOTE: Each correct selection is worth one point.
Answer:

Explanation:

1. Some of the transformation steps in the query will fold."

Query folding refers to the ability of Power Query to push transformations back to the data source rather than
processing them in Power Query itself.

The answer "Some".

Some transformations are foldable (executed at the source).

Some transformations are not foldable and will be executed locally in Power Query.

This is common because basic transformations (filtering, joins, aggregations, etc.) often fold, but more
complex operations (custom columns, advanced functions, etc.) may not.

2. The Added custom step will be performed in the Microsoft Power Query engine.

When a transformation cannot be folded, it is processed in the Power Query engine instead of being pushed to
the source.

Since the question mentions an added custom step, it likely includes:

Custom calculations

Complex transformations (like M code or certain joins)

Operations that cannot be executed at the data source

Such transformations are executed locally within Power Query rather than at the source (lakehouse query
engine).

Question: 35
You have a Fabric tenant that contains a lakehouse named Lakehouse’. Lakehouse1 contains a table named Tablet.
You are creating a new data pipeline.
You plan to copy external data to Table’. The schema of the external data changes regularly.
You need the copy operation to meet the following requirements:
Replace Table1 with the schema of the external data.
Replace all the data in Table1 with the rows in the external data.
You add a Copy data activity to the pipeline.
What should you do for the Copy data activity?

A. From the Source tab, add additional columns.


B. From the Destination tab, set Table action to Overwrite.
C. From the Settings tab, select Enable staging.
D. From the Source tab, select Enable partition discovery.
E. From the Source tab, select Recursively.

Answer: B

Explanation:

B. From the Destination tab, set Table action to Overwrite.

You need the copy operation to:

Replace Table1 with the schema of the external data

Replace all data in Table1 with the external data

The best way to achieve this is by setting Table action to "Overwrite" in the Destination tab of the Copy Data
activity.

"Overwrite" ensures that the table is completely replaced, including the schema and data, which meets both
requirements.

This means that every time the pipeline runs, Table1 will have the latest schema and data from the external
source.

Why not the other options?

A. From the Source tab, add additional columns

Adding columns does not replace the schema; it only appends extra fields.

This does not ensure that the table structure matches the external data.

C. From the Settings tab, select Enable staging

Staging is useful for performance improvements in large data transfers, but it does not affect schema
replacement.

D. From the Source tab, select Enable partition discovery

Partition discovery helps when reading partitioned data, but it does not affect replacing the table schema or
data.
E. From the Source tab, select Recursively

Recursion is used for reading multiple files in hierarchical folders but does not replace the schema or data in
the destination table.

Question: 36
You have a Fabric tenant that contains a lakehouse.
You plan to query sales data files by using the SQL endpoint. The files will be in an Amazon Simple Storage Service
(Amazon S3) storage bucket.
You need to recommend which file format to use and where to create a shortcut.
Which two actions should you include in the recommendation? Each correct answer presents part of the solution.
NOTE: Each correct answer is worth one point.

A. Create a shortcut in the Files section.


B. Use the Parquet format
C. Use the CSV format.
D. Create a shortcut in the Tables section.
E. Use the delta format.

Answer: BD

Explanation:

B. Use the Parquet format.

D. Create a shortcut in the Tables section.

Creating a shortcut in the Files section would treat the data more like raw files, which might not leverage the
full potential of SQL querying capabilities. Creating the shortcut in the Tables section aligns more closely with
the structured query requirements and provides a better-integrated experience.

Question: 37
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a subfolder named
Subfolder1 that contains CSV files.
You need to convert the CSV files into the delta format that has V-Order optimization enabled.
What should you do from Lakehouse explorer?

A.Use the Load to Tables feature.


B.Create a new shortcut in the Files section.
C.Create a new shortcut in the Tables section.
D.Use the Optimize feature.

Answer: A

Explanation:

A. Use the "Load to Tables" feature.

To convert CSV files into Delta format with V-Order optimization, you need to load the files into tables in the
Lakehouse.

"Load to Tables" is a feature in Microsoft Fabric's Lakehouse Explorer that allows you to convert files (such as
CSV) into managed Delta tables.

During this process, the data is converted into the Delta format, which supports V-Order optimization for
better performance.

V-Order is an advanced columnar storage format that enhances query performance and compression for
Delta tables.

https://learn.microsoft.com/en-us/fabric/data-engineering/load-to-tables#load-to-table-capabilities-overview

Question: 38
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains an unpartitioned table
named Table1.
You plan to copy data to Table1 and partition the table based on a date column in the source data.
You create a Copy activity to copy the data to Table1.
You need to specify the partition column in the Destination settings of the Copy activity.
What should you do first?

A.From the Destination tab, set Mode to Append.


B.From the Destination tab, select the partition column.
C.From the Source tab, select Enable partition discovery.
D.From the Destination tabs, set Mode to Overwrite.

Answer: D

Explanation:

D. From the Destination tabs, set Mode to Overwrite.

Expand Advanced, in Table action, select Overwrite, and then select Enable partition, under Partition columns,
select Add column, and choose the column you want to use as the partition column. You can choose to use a
single column or multiple columns as the partition column.

Reference:

https://learn.microsoft.com/en-us/fabric/data-factory/tutorial-lakehouse-partition#load-data-to-lakehouse-
using-partition-columns

Question: 39
HOTSPOT -

You run the following T-SQL statement.


CREATE TABLE test.FactSales AS CLONE OF Dbo.FactSales;
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:

Explanation:

1. Yes

A replica of dbo.Sales is created in the test schema by copying the metadata only.

If metadata-only copying is used, it means the structure (schema) is copied but not the data.

This is possible with views, synonyms, or schema-only copies (e.g., CREATE TABLE ... AS NO DATA).

If the intent is to create an empty table with the same structure, this would be true.

2. No

Additional schema changes to dbo.FactSales will also apply to test.FactSales.

If test.FactSales was independently created (i.e., copied separately from dbo.FactSales), then any schema
changes to dbo.FactSales will NOT automatically reflect in test.FactSales.

Schema changes would only propagate if test.FactSales was a synonym, view, or linked table.

Since it's likely an independent table, changes to dbo.FactSales do not apply automatically.

3.No

Additional schema changes to dbo.FactSales will also apply to test.FactSales.

This statement is identical to statement 2, so the answer remains the same.

Changes to dbo.FactSales do not automatically apply to test.FactSales unless explicitly linked (e.g., via a view
or a synonym.
Question: 40
You have source data in a folder on a local computer.
You need to create a solution that will use Fabric to populate a data store. The solution must meet the following
requirements:
Support the use of dataflows to load and append data to the data store.
Ensure that Delta tables are V-Order optimized and compacted automatically.
Which type of data store should you use?

A.a lakehouse
B.an Azure SQL database
C.a warehouse
D.a KQL database

Answer: A

Explanation:

The Lakehouse and the Delta Lake table format are central to Microsoft Fabric, assuring that tables are
optimized for analytics is a key requirement.

To meet the requirements of supporting dataflows to load and append data to the data store while ensuring
that Delta tables are V-Order optimized and compacted automatically, you should use a lakehouse in Fabric as
your solution.

Reference:

https://learn.microsoft.com/en-us/fabric/data-engineering/delta-optimization-and-v-order?tabs=sparksql

Question: 41
HOTSPOT -

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

Answer:
Explanation:

1. "The results will form a hierarchy of folders for each partition key."

Yes

Partitioning creates a folder structure where each folder represents a partition key value.

For example, in a data lake, partitioned data may be stored as:

bash

Copy

Edit

/sales/year=2023/month=01/

/sales/year=2023/month=02/

This improves query performance by allowing queries to skip unnecessary partitions.

2. "The resulting file partitions can be read in parallel across multiple nodes."

Yes

Partitioned files are designed for parallel processing.

In distributed environments like Spark, Synapse, or Fabric Lakehouse, each partition can be processed
independently across multiple compute nodes, improving performance.

This is especially true for columnar formats like Parquet and Delta tables, which support distributed query
execution.

3. "The resulting file partitions will use file compression."

Yes

File compression is commonly used in partitioned storage to reduce storage costs and improve read
performance.

Formats like Parquet, ORC, and Delta support built-in compression (e.g., Snappy, Gzip, or ZSTD).

Compressed partitions reduce I/O costs and improve query performance, making them a best practice.
Question: 42
You have a Fabric workspace named Workspace1 that contains a data flow named Dataflow1 contains a query that
returns the data shown in the following exhibit.

You need to transform the data columns into attribute-value pairs, where columns become rows.
You select the VendorID column.
Which transformation should you select from the context menu of the VendorID column?

A. Group by
B. Unpivot columns
C. Unpivot other columns
D. Split column
E. Remove other columns

Answer: C

Explanation:

C. Unpivot Other Columns.

You need to transform data columns into attribute-value pairs, meaning that columns will be converted into
rows. This is achieved using Unpivoting in Power Query.

"Unpivot Columns" (Option B) → Converts selected columns into rows.

"Unpivot Other Columns" (Option C) → Converts all columns except the selected column into rows.

Since you are selecting the VendorID column and want to transform the other columns into attribute-value
pairs, you should choose "Unpivot Other Columns."

Question: 43

You need to ensure that the pipeline runs every four hours on Mondays and Fridays.

By the minute
Answer: C

Explanation:

Answer C : Weekly.

The only way to do this is to set the schedule to ''Weekly'', set the days on Monday and Friday and add
manually 6 Time of 4 hour intervals.

You need the pipeline to run every four hours on Mondays and Fridays.

Question: 44
You have a Fabric tenant that contains a warehouse.
Several times a day, the performance of all warehouse queries degrades. You suspect that Fabric is throttling the
compute used by the warehouse.
What should you use to identify whether throttling is occurring?

A. the Capacity settings


B. the Monitoring hub
C. dynamic management views (DMVs)
D. the Microsoft Fabric Capacity Metrics app

Answer: D

Explanation:

The Microsoft Capacity Metrics app, also known as the metrics app, serves as a monitoring tool within the
Microsoft Power BI service. It offers functionalities to track and analyze the resource utilization.

Reference:

https://learn.microsoft.com/en-us/fabric/data-warehouse/compute-capacity-smoothing-throttling

Question: 45
HOTSPOT -
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.

Answer:

Explanation:

No: The Spark engine will initially read all columns from the CSV file because the .select() transformation is
applied after the data has been read into memory. Therefore, all 12 columns from Sales_raw.csv are read
before the selection of specific columns is applied.

Yes: Adding inferSchema = 'true' will increase the execution time of the query because Spark will need to
read through the entire dataset to determine the data types of each column. This extra pass over the data
adds to the initial read time.

Question: 46
You have a Fabric tenant that contains a warehouse.
A user discovers that a report that usually takes two minutes to render has been running for 45 minutes and has
still not rendered.
You need to identify what is preventing the report query from completing.
Which dynamic management view (DMV) should you use?

A.sys.dm_exec_requests
B.sys.dm_exec_sessions
C.sys.dm_exec_connections
D.sys.dm_pdw_exec_requests

Answer: A

Explanation:

The sys.dm_exec_requests dynamic management view provides real-time information about all requests
currently executing on the SQL Server. This view helps you understand the status of each request, including
whether it is blocked, waiting, or running.

Reference:

https://learn.microsoft.com/en-us/fabric/data-warehouse/monitor-using-dmv

Question: 47
DRAG DROP -

How should you complete the code? To answer, drag the appropriate values to the correct targets. Each value may
be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view
Answer:

Explanation:

Value.

Value is the function used to store the result of the query.

NativeQuery.

NativeQuery allows executing a direct SQL query within Power Query M code.

EnableFolding .

This enables query folding, ensuring that transformations are executed in SQL Server rather than in Power
Query.

Question: 48
DRAG DROP -

Readings from 100 IoT devices are appended to a Delta table in Lakehouse1. Each set of readings is approximately
All the table and SparkSession settings are set to the default.
You discover that queries are slow to execute. In addition, the lakehouse storage contains data and log files that
are no longer used.
You need to remove the files that are no longer used and combine small files into larger files with a target size of 1
GB per file.
What should you do? To answer, drag the appropriate actions to the correct requirements. Each action may be
used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view
content.
NOTE: Each correct selection is worth one point.

Answer:
Explanation:

Remove the files that are no longer used:

Run the VACUUM command on a schedule: The VACUUM command cleans up old files and log files that are
no longer needed by the Delta table, helping to free up storage and potentially improve performance by
reducing the number of files the query engine needs to consider.

Combine small files into larger files:

Run the OPTIMIZE command on a schedule: The OPTIMIZE command compacts small files into larger ones,
improving read performance by reducing the overhead associated with opening many small files. This can be
particularly useful when you have a large number of small files due to frequent appends of small data sets.

Question: 49
You need to create a data loading pattern for a Type 1 slowly changing dimension (SCD).
Which two actions should you include in the process? Each correct answer presents part of the solution.
NOTE: Each correct answer is worth one point.

A.Update rows when the non-key attributes have changed.


B.Insert new rows when the natural key exists in the dimension table, and the non-key attribute values have
changed.
C.Update the effective end date of rows when the non-key attribute values have changed.
D.Insert new records when the natural key is a new value in the table.

Answer: AD

Explanation:

A. Update rows when the non-key attributes have changed: In a Type 1 SCD, when a change is detected in any
of the non-key attributes of an existing row, the current row is updated with the new values. This type of SCD
does not keep any historical data; it simply overwrites the old data with the new data.
D. Insert new records when the natural key is a new value in the table: When a new record (with a new natural
key) is encountered that does not already exist in the dimension table, it is inserted as a new row. This is
necessary to ensure that all new entities are captured in the dimension.

Question: 50
HOTSPOT -
You have a Fabric workspace named Workspace1 and an Azure Data Lake Storage Gen2 account named storage1.
Workspace1 contains a lakehouse named Lakehouse1.
You need to create a shortcut to storage1 in Lakehouse1.
Which connection and endpoint should you specify? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.

Answer:

Explanation:

1. Connection Type: https


The selection of https means that the connection is being made using secure HTTPS endpoints rather than
using Azure Blob File System (ABFS/ABFSS) protocols.

2. Endpoint Type: dfs

The selection of dfs (Distributed File System) means that the connection is using Azure Data Lake Storage
Gen2.

dfs endpoints enable Hierarchical Namespace (HNS), which allows for file system-like directory structures
and optimized big data analytics.

Question: 51
You are analyzing customer purchases in a Fabric notebook by using PySpark.
You have the following DataFrames:
transactions: Contains five columns named transaction_id, customer_id, product_id, amount, and date and has 10
million rows, with each row representing a transaction. customers: Contains customer details in 1,000 rows and
three columns named customer_id, name, and country.
You need to join the DataFrames on the customer_id column. The solution must minimize data shuffling.
You write the following code.
from pyspark.sql import functions as F
results =
Which code should you run to populate the results DataFrame?

A.transactions.join(F.broadcast(customers), transactions.customer_id == customers.customer_id)


B.transactions.join(customers, transactions.customer_id == customers.customer_id).distinct()
C.transactions.join(customers, transactions.customer_id == customers.customer_id)
D.transactions.crossJoin(customers).where(transactions.customer_id == customers.customer_id)

Answer: A

Explanation:

A. transactions.join(F.broadcast(customers), transactions.customer_id == customers.customer_id)Optimized


method to perform a join between a very large table and a smaller one.

In Apache Spark, broadcasting refers to an optimization technique for join operations. When you join two
DataFrames or RDDs and one of them is significantly smaller than the other, Spark can "broadcast" the
smaller table to all nodes in the cluster. This approach avoids the need for network shuffles for each row of
the larger table, significantly reducing the execution time of the join operation.

https://sparkbyexamples.com/spark/broadcast-join-in-spark/"

Question: 52
HOTSPOT -
Use the drop-down menus to select the answer choice that completes each statement based on the information
presented in the graphic.
NOTE: Each correct selection is worth one point.

Answer:
Explanation:

1. Direct Lake Fallback Behavior: Automatic.

Direct Lake Mode enables real-time querying of Delta tables in a Fabric lakehouse without requiring data
import or DirectQuery.

When fallback behavior is set to Automatic, Power BI will automatically switch between Direct Lake Mode and
DirectQuery when necessary (e.g., if a feature isn't supported in Direct Lake).

This provides the best balance between performance and flexibility.

2. Query Execution: Direct Lake.

Direct Lake Mode allows Power BI to load data directly from Delta tables in a Fabric lakehouse without
preloading or sending queries to an external database.

It provides better performance than DirectQuery by eliminating query overhead and improving response times.

Question: 53
HOTSPOT -
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a table named
Nyctaxi_raw. Nyctaxi_row contains the following table:

You create a Fabric notebook and attach it to Lakehouse1.


You need to use PySpark code to transform the data. The solution must meet the following requirements:
Add a column named pickupDate that will contain only the date portion of pickupDateTime.
Filter the DataFrame to include only rows where fareAmount is a positive number that is less than 100.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:

1. df.withColumn.

The withColumn function is used to add or replace a column in a DataFrame.

It is appropriate here because we are creating a new column (pickupDate) based on an existing column
(tpepPickupDateTime).

Alternative options like df.select or df.withColumnsRenamed would not modify a column directly.

2 .cast('date')

The .cast('date') function is used to convert a column’s data type to date.

Since tpepPickupDateTime is likely a timestamp, converting it to a date ensures it only contains the date
portion.

Other options like .alias('date') only rename the column without changing its type.

3.filter("fareAmount > 0 AND fareAmount < 100")

The .filter() function in PySpark is used to apply SQL-like filtering conditions to DataFrames.

The selected condition ensures that only rows where fareAmount is greater than 0 and less than 100 are
retained.

Other options like .where(df.fareAmount.isin([1,100])) filter for specific values rather than a range.

Question: 54
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a new semantic model in OneLake.
You use a Fabric notebook to read the data into a Spark DataFrame.
You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
numeric columns.
Solution: You use the following PySpark expression:
df.explain()
Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

The df.explain() method in PySpark is used to print the logical and physical plans of a DataFrame, which helps
in understanding how Spark plans to execute the query. It does not compute any statistical values like min,
max, mean, or standard deviation.

To achieve the goal, you should use: df.describe().show()


Question: 55
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a new semantic model in OneLake.
You use a Fabric notebook to read the data into a Spark DataFrame.
You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
numeric columns.
Solution: You use the following PySpark expression:
df.show()
Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

Correct methods: Use df.describe().show() for basic statistics and df.agg() with appropriate functions (min,
max, mean, stddev) for detailed statistics.

The correct syntax is df.describe().Sources:* describe -->


https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.DataFrame.describe.html*
show -->
https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.DataFrame.show.html

Question: 56
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a new semantic model in OneLake.
You use a Fabric notebook to read the data into a Spark DataFrame.
You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
numeric columns.
Solution: You use the following PySpark expression:
df.summary()
Does this meet the goal?

A.Yes
B.No

Answer: A

Explanation:

A. Yes

The df.summary() function in PySpark provides summary statistics for both numeric and string columns in a
DataFrame, including:

For numeric columns: count, mean, stddev, min, max


Question: 57
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table named
Customer.
When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT
performed on the table.
You need to identify whether maintenance tasks were performed on Customer.
Solution: You run the following Spark SQL statement:

DESCRIBE HISTORY customer -


Does this meet the goal?

A.Yes
B.No

Answer: A

Explanation:

Yes, running DESCRIBE HISTORY customer meets the goal of identifying whether maintenance tasks were
performed on the Delta table.

Question: 58
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table named
Customer.
When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT
performed on the table.
You need to identify whether maintenance tasks were performed on Customer.
Solution: You run the following Spark SQL statement:

REFRESH TABLE customer -


Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

No, running REFRESH TABLE customer does not meet the goal of identifying whether maintenance tasks
were performed on the Delta table. This Spark SQL command is used to refresh the metadata of a table. It
ensures that the latest schema and data are available for queries but does not give any historical information
Question: 59
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table named
Customer.
When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT
performed on the table.
You need to identify whether maintenance tasks were performed on Customer.
Solution: You run the following Spark SQL statement:

EXPLAIN TABLE customer -


Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

The EXPLAIN statement in Spark SQL is used to display the execution plan of a query. This plan shows how
Spark will execute the query, including details about the operations and stages involved. While it is useful for
understanding and optimizing query performance, it does not provide historical information about
maintenance tasks like optimization, compaction, or vacuuming performed on the table.

Question: 60
Case study -

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -

Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -

Litware has data that must be analyzed as shown in the following table.

The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

• Survey
• Question
• Response

For each survey submitted, the following occurs:

• One row is added to the Survey table.


• One row is added to the Response table for each question in the survey.

The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems -

The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.

Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -
Planned Changes -

Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity

The following three workspaces will be created:

• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists

The following will be created in the AnalyticsPOC workspace:

• A data store (type to be decided)


• A custom semantic model
• A default semantic model
• Interactive reports

The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.

All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements -

The data store must support the following:

• Read access by using T-SQL or Python


• Semi-structured and unstructured data
• Row-level security (RLS) for users executing T-SQL queries

Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.

Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model

The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model

The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.

The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:

• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.

Security Requirements -

Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.

Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
• Fabric administrators will be the workspace administrators.
• The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
• The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
• The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
• The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
• The date dimension must be available to all users of the data store.
• The principle of least privilege must be followed.

Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:

• FabricAdmins: Fabric administrators


• AnalyticsTeam: All the members of the analytics team
• DataAnalysts: The data analysts on the analytics team
• DataScientists: The data scientists on the analytics team
• DataEngineers: The data engineers on the analytics team
• AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements -

The data analysts must create a customer satisfaction report that meets the following requirements:

• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.

You need to recommend a solution to prepare the tenant for the PoC.

Which two actions should you recommend performing from the Fabric Admin portal? Each correct answer presents
part of the solution.

NOTE: Each correct answer is worth one point.

A. Enable the Users can try Microsoft Fabric paid features option for the entire organization.
B. Enable the Users can try Microsoft Fabric paid features option for specific security groups.
C. Enable the Allow Azure Active Directory guest users to access Microsoft Fabric option for specific security
groups.
D. Enable the Users can create Fabric items option and exclude specific security groups.
E. Enable the Users can create Fabric items option for specific security groups.

Answer: BE
Explanation:

Enable the Users can try Microsoft Fabric paid features option for specific security groups: This will allow
specific security groups (like the AnalyticsTeam, DataAnalysts, DataScientists, DataEngineers, and
AnalyticsEngineers) to access and use the paid features of Microsoft Fabric necessary for the PoC. This is
important to ensure that only the relevant team members can utilize these advanced features while
preventing unnecessary access for other users.

Enable the Users can create Fabric items option for specific security groups: This will allow only specific
security groups to create Fabric items, ensuring that the creation of these items is controlled and managed by
the appropriate team members. This helps maintain the principle of least privilege and ensures that only
authorized personnel can create and manage Fabric items during the PoC.

Question: 61
HOTSPOT
-

Case study
-

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study


-
To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview
-

Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment
-

Fabric Environment
-

Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data
-

Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

• Survey
• Question
• Response

For each survey submitted, the following occurs:

• One row is added to the Survey table.


• One row is added to the Response table for each question in the survey.

The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems
-

The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.

Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements
-

Planned Changes
-

Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity

The following three workspaces will be created:

• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:

• A data store (type to be decided)


• A custom semantic model
• A default semantic model
• Interactive reports

The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.

All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements
-

The data store must support the following:

• Read access by using T-SQL or Python


• Semi-structured and unstructured data
• Row-level security (RLS) for users executing T-SQL queries

Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.

Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model

The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model

The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.

The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:

• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.

Security Requirements
-

Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.

Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:

• Fabric administrators will be the workspace administrators.


• The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
• The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
• The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
• The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
• The date dimension must be available to all users of the data store.
• The principle of least privilege must be followed.

Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:

• FabricAdmins: Fabric administrators


• AnalyticsTeam: All the members of the analytics team
• DataAnalysts: The data analysts on the analytics team
• DataScientists: The data scientists on the analytics team
• DataEngineers: The data engineers on the analytics team
• AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements
-

The data analysts must create a customer satisfaction report that meets the following requirements:

• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.

You need to design a semantic model for the customer satisfaction report.

Which data source authentication method and mode should you use? To answer, select the appropriate options in
the answer area.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

1. Authentication Method: "Single sign-on (SSO) authentication"

SSO (Single Sign-On) allows users to authenticate once and gain access to multiple applications without
needing to log in separately.

This is a secure and seamless authentication method commonly used in enterprise environments for Microsoft
Fabric, Power BI, and other services.

Compared to Basic authentication (which requires username/password) or Service principal authentication


(used for automated access), SSO provides better user experience and security.

2. Mode: "Direct Lake"

Direct Lake mode is a new hybrid query mode that combines the performance of Import mode with the real-
time access of DirectQuery.

Unlike Import mode, which loads data into memory, or DirectQuery, which queries the source live, Direct Lake
queries Delta tables directly from OneLake without the need for pre-loading.

This mode is optimized for high performance and scalability, making it an ideal choice for Power BI and Fabric
scenarios.

Question: 62
Case study -

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
To start the case study -
To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -

Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -

Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -

Litware has data that must be analyzed as shown in the following table.

The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

• Survey
• Question
• Response

For each survey submitted, the following occurs:

• One row is added to the Survey table.


• One row is added to the Response table for each question in the survey.

The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -

The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.

Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -

Planned Changes -

Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity

The following three workspaces will be created:

• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists

The following will be created in the AnalyticsPOC workspace:

• A data store (type to be decided)


• A custom semantic model
• A default semantic model
• Interactive reports

The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.

All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements -

The data store must support the following:

• Read access by using T-SQL or Python


• Semi-structured and unstructured data
• Row-level security (RLS) for users executing T-SQL queries

Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.

Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model

The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model

The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.

The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:

• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.

Security Requirements -

Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.

Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:

• Fabric administrators will be the workspace administrators.


• The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
• The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
• The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
• The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
• The date dimension must be available to all users of the data store.
• The principle of least privilege must be followed.

Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:

• FabricAdmins: Fabric administrators


• AnalyticsTeam: All the members of the analytics team
• DataAnalysts: The data analysts on the analytics team
• DataScientists: The data scientists on the analytics team
• DataEngineers: The data engineers on the analytics team
• AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements -

The data analysts must create a customer satisfaction report that meets the following requirements:

• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.

You need to implement the date dimension in the data store. The solution must meet the technical requirements.

What are two ways to achieve the goal? Each correct answer presents a complete solution.

NOTE: Each correct selection is worth one point.

A. Populate the date dimension table by using a dataflow.


B. Populate the date dimension table by using a Copy activity in a pipeline.
C. Populate the date dimension view by using T-SQL.
D. Populate the date dimension table by using a Stored procedure activity in a pipeline.

Answer: AD
Explanation:

Dataflows in Microsoft Fabric (or Power BI) allow ETL (Extract, Transform, Load) operations to create and
populate tables.

Stored procedures allow for predefined logic and transformations to be applied while inserting or updating
records in the date dimension table.

Question: 63
Case study -

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -

Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.

Existing Environment -

Fabric Environment -

Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.

Available Data -

Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.

The customer satisfaction data contains the following tables:

• Survey
• Question
• Response

For each survey submitted, the following occurs:

• One row is added to the Survey table.


• One row is added to the Response table for each question in the survey.

The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.

User Problems -

The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.

Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.

Requirements -

Planned Changes -

Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity

The following three workspaces will be created:

• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists

The following will be created in the AnalyticsPOC workspace:

• A data store (type to be decided)


• A custom semantic model
• A default semantic model
• Interactive reports

The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.

All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.

Technical Requirements -

The data store must support the following:

• Read access by using T-SQL or Python


• Semi-structured and unstructured data
• Row-level security (RLS) for users executing T-SQL queries

Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.

Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model

The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model

The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.

The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:

• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.

Security Requirements -

Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.

Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:

• Fabric administrators will be the workspace administrators.


• The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
• The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
• The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
• The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
• The date dimension must be available to all users of the data store.
• The principle of least privilege must be followed.

Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
• FabricAdmins: Fabric administrators
• AnalyticsTeam: All the members of the analytics team
• DataAnalysts: The data analysts on the analytics team
• DataScientists: The data scientists on the analytics team
• DataEngineers: The data engineers on the analytics team
• AnalyticsEngineers: The analytics engineers on the analytics team

Report Requirements -

The data analysts must create a customer satisfaction report that meets the following requirements:

• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.

You need to ensure the data loading activities in the AnalyticsPOC workspace are executed in the appropriate
sequence. The solution must meet the technical requirements.

What should you do?

A.Create a dataflow that has multiple steps and schedule the dataflow.
B.Create and schedule a Spark notebook.
C.Create and schedule a Spark job definition.
D.Create a pipeline that has dependencies between activities and schedule the pipeline.

Answer: D

Explanation:

D. Create a pipeline that has dependencies between activities and schedule the pipeline.

To ensure that data loading activities in the AnalyticsPOC workspace are executed in the appropriate
sequence while meeting technical requirements, the best approach is to use a data pipeline with
dependencies between activities.

Pipelines in Microsoft Fabric (or Azure Data Factory) allow you to orchestrate data movement, transformation,
and processing across multiple steps.

You can define activity dependencies, ensuring that each step executes only after the previous step
completes successfully.

You can schedule the pipeline to run at specific times, making it automated and reliable.

Question: 64
Case study -

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -

Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -

Contoso identifies the following data preparation requirements:

• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -

Contoso identifies the following requirements for implementing and managing semantic models:

• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -

Contoso identifies the following high-level requirements that must be considered for all solutions:

• Follow the principle of least privilege when applicable.


• Minimize implementation and maintenance effort when possible.

You need to recommend which type of Fabric capacity SKU meets the data analytics requirements for the
Research division.

What should you recommend?

A.A
B.EM
C.P
D.F

Answer: D

Explanation:

Fabric capacity SKUs all start with F, they allows for Pay as you go by the minute. Reserved capacity is also
available.

Reference:

https://azure.microsoft.com/en-us/pricing/details/microsoft-fabric/

Question: 65
HOTSPOT

Case
-

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study


-
To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview
-

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment
-

Identity Environment
-

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment
-

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements
-

Planned Changes
-
Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements


-

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements


-

Contoso identifies the following data preparation requirements:

• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements


-

Contoso identifies the following requirements for implementing and managing semantic models:

• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements
-

Contoso identifies the following high-level requirements that must be considered for all solutions:

• Follow the principle of least privilege when applicable.


• Minimize implementation and maintenance effort when possible.

You need to migrate the Research division data for Productline1. The solution must meet the data preparation
requirements.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.


Answer:

Explanation:

"Tables/productline1"

Requirements: Use managed tables.

If you use saveAsTable() you don't need to specify the path "Table/"

If you you save() you specify the full path

Question: 66
Case study -

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -

Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -

Contoso identifies the following data preparation requirements:

• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -

Contoso identifies the following requirements for implementing and managing semantic models:

• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -

Contoso identifies the following high-level requirements that must be considered for all solutions:
• Follow the principle of least privilege when applicable.
• Minimize implementation and maintenance effort when possible.

What should you use to implement calculation groups for the Research division semantic models?

A.Microsoft Power BI Desktop


B.the Power BI service
C.DAX Studio
D.Tabular Editor

Answer: D
Explanation:

D .Tabular Editor

The semantic models in the Research division workspaces must use Direct Lake mode.

Considering Contoso's requirement to minimize implementation and maintenance effort (general


requirement), Tabular Editor offers a more efficient way to define calculation groups compared to manually
writing DAX code. Additionally, since calculation groups are part of the semantic model itself, they can be
deployed and managed alongside the model, simplifying maintenance.

Question: 67
HOTSPOT
-

Case study
-

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study


-
To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview
-

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment
-

Identity Environment
-

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment
-

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements
-

Planned Changes
-

Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements


-

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements


-

Contoso identifies the following data preparation requirements:


• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements


-

Contoso identifies the following requirements for implementing and managing semantic models:

The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements
-

Contoso identifies the following high-level requirements that must be considered for all solutions:

• Follow the principle of least privilege when applicable.


• Minimize implementation and maintenance effort when possible.

Which workspace role assignments should you recommend for ResearchReviewersGroup1 and
ResearchReviewersGroup2? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

1. ResearchReviewersGroup1: Viewer.

Members of this group can only view resources but cannot edit, add, or delete them.

Suitable for users who need read-only access to reports, datasets, or workspaces.

2.ResearchReviewersGroup2: Contributor.

Members can edit and add content but cannot manage access or settings.

Suitable for users who need to modify or update reports and datasets but should not have full administrative
privileges.

Question: 68
HOTSPOT

You have a Fabric tenant.

You need to configure OneLake security for users shown in the following table.

The solution must follow the principle of least privilege.

Which permission should you assign to each user? To answer, select the appropriate options in the answer area.
Answer:

Explanation:

User1: "ReadAll"

This permission allows access to all data and metadata within the system.

Suitable for users who need full read access across multiple datasets, reports, or tables.

Often assigned to users who perform comprehensive analysis.

User2: "ReadData"

This permission allows access to the data only, but not the metadata or system-level configurations.

Suitable for users who need to query and analyze specific datasets but should not see system-wide metadata.

Often assigned to data analysts or restricted users.

Reference:
https://learn.microsoft.com/en-us/fabric/data-engineering/lakehouse-sharing#sharing-and-permissions

Question: 69
You have an Azure Repos repository named Repo1 and a Fabric-enabled Microsoft Power BI Premium capacity. The
capacity contains two workspaces named Workspace1 and Workspace2. Git integration is enabled at the
workspace level.

You plan to use Microsoft Power BI Desktop and Workspace1 to make version-controlled changes to a semantic
model stored in Repo1. The changes will be built and deployed to Workspace2 by using Azure Pipelines.

You need to ensure that report and semantic model definitions are saved as individual text files in a folder
hierarchy. The solution must minimize development and maintenance effort.

In which file format should you save the changes?

A.PBIP
B.PBIDS
C.PBIT
D.PBIX

Answer: A

Explanation:

PBIP (Power BI Project)

PBIP format is designed to work with version control systems like Azure Repos. It breaks down Power BI
artifacts into individual files that can be managed and versioned separately, facilitating better collaboration
and change tracking.

Folder Hierarchy: It saves the project structure in a folder hierarchy, where each component of the Power BI
project (like datasets, reports, data sources) is stored in separate files.

Text-Based: Being a text-based format, it integrates well with Git repositories and supports diff and merge
operations.

Question: 70
DRAG DROP

You have a lakehouse that contains the Bronze and Silver layers and a warehouse that contains the Gold layer.
You need to ensure that the layers are populated daily in sequential order such that Silver is populated only after
Bronze is complete, and Gold is populated only after Silver is complete. The solution must minimize development
effort and complexity.

What should you use to execute each set of items? To answer, drag the appropriate options to the correct items.
Each option may be used once, more than once, or not at all. You may need to drag the split bar between panes or
scroll to view content.

NOTE: Each correct selection is worth one point.

Answer:

Explanation:

Orchestration Pipeline → A Schedule.

The orchestration pipeline ensures that the entire data pipeline runs on a set schedule.

Bronze Layer → An Invoke Pipeline Activity.

The Bronze layer typically ingests raw data from various sources.

An Invoke Pipeline Activity is used to trigger data ingestion pipelines dynamically.

Silver Layer → A Pipeline Dataflow Activity.


Gold Layer → A Pipeline Stored Procedure Activity.

The Gold layer is responsible for aggregating and modeling data for analytics.

A Stored Procedure Activity is used to execute complex SQL logic on the transformed data.

Question: 71
DRAG DROP
-

You are building a solution by using a Fabric notebook.

You have a Spark DataFrame assigned to a variable named df. The DataFrame returns four columns.

You need to change the data type of a string column named Age to integer. The solution must return a DataFrame
that includes all the columns.

How should you complete the code? To answer, drag the appropriate values to the correct targets. Each value may
be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view
content.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

The cast function is applied to convert the "age" column to an integer type.

Question: 72
HOTSPOT

Using a notebook in Workspace1, you need to load the content of the file to the default lakehouse. The solution
must ensure that the content will display automatically as a table named Sales in Lakehouse explorer.

How should you complete the code? To answer, select the appropriate options in the answer area.
Answer:

Explanation:

df = spark.read.parquet("abfss://fs1@storage1.dfs.core.windows.net/files/sales.parquet")

df.write.mode("overwrite").format("delta").saveAsTable("sales")

https://learn.microsoft.com/en-us/fabric/data-engineering/lakehouse-notebook-load-data

Question: 73
You have a Fabric workspace named Workspace1 that contains a lakehouse named Lakehouse1.

In Workspace1, you create a data pipeline named Pipeline1.

You have CSV files stored in an Azure Storage account.

Which type of activity should you add?

D.Copy data

Answer: A

Explanation:

A. Dataflow.

To copy data from CSV files in Azure Storage to Lakehouse1 while supporting Power Query M formula

Question: 74
HOTSPOT
You have a Fabric tenant that contains lakehouse named Lakehouse1. Lakehouse1 contains a Delta table with eight
columns.

You receive new data that contains the same eight columns and two additional columns.

You create a Spark DataFrame and assign the DataFrame to a variable named df. The DataFrame contains the new
data.

You need to add the new data to the Delta table to meet the following requirements:

• Keep all the existing rows.


• Ensure that all the new data is added to the table.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

1. Mode: "append"

Selected Option: "append"

The "append" mode adds new data to the existing Delta table without overwriting it.

This ensures that previously stored data is retained, and only new data is added.

2. Option: ("mergeSchema", "true")

Selected Option: ("mergeSchema", "true")

mergeSchema=true allows automatic schema evolution when new columns are introduced in the DataFrame
but are not present in the existing Delta table.

This prevents schema mismatch errors and enables seamless schema updates.

Question: 75
HOTSPOT
You need to write a T-SQL query that will return the following columns.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.


Answer:

Explanation:

1. COALESCE (Weight, Quantity, 1) AS ChargedQuantity

The COALESCE function returns the first non-null value from the provided arguments.

In this case, it checks Weight first. If it is NULL, it checks Quantity. If both are NULL, it defaults to 1.

This ensures that there is always a value for ChargedQuantity.

2. LEAST (ListPrice, SalePrice) AS OrderPrice.

The LEAST function returns the smallest value among the provided arguments.

It selects the lowest price between ListPrice and SalePrice, ensuring that the minimum price is used for
Question: 76

You plan to use a visual query to merge two tables.

You need to ensure that the query returns all the rows in both tables.

Answer: B

Explanation:

Full Outer Join returns all rows from both tables.

If there is a match between the tables, it merges the data.

If there is no match, it still includes the row, filling missing values with NULLs.

Question: 77
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table that has
one million Parquet files.

You need to remove files that were NOT referenced by the table during the past 30 days. The solution must ensure
that the transaction log remains consistent, and the ACID properties of the table are maintained.

What should you do?

A.From OneLake file explorer, delete the files.


B.Run the OPTIMIZE command and specify the Z-order parameter.
C.Run the OPTIMIZE command and specify the V-order parameter.
D.Run the VACUUM command.

Answer: D

Explanation:

D. Run the VACUUM command.

The VACUUM command is used in Delta tables to remove old, unreferenced Parquet files while ensuring that:
The transaction log remains consistent

By default, Delta Lake retains all historical data files for seven days to allow for time travel and rollback. The
VACUUM command permanently deletes these unreferenced files, but you can specify a retention period
(e.g., 30 days) to control the cleanup.

Question: 78
DRAG DROP
-

You are implementing two dimension tables named Customers and Products in a Fabric warehouse.

You need to use slowly changing dimension (SCD) to manage the versioning of data. The solution must meet the
requirements shown in the following table.

Which type of SCD should you use for each table? To answer, drag the appropriate SCD types to the correct tables.
Each SCD type may be used once, more than once, or not at all. You may need to drag the split bar between panes
or scroll to view content.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

Customers :Type 2

Customer records often require history tracking (e.g., change of address, email, or preferences).

Question: 79
You have a Fabric workspace named Workspace1 and an Azure SQL database.

You plan to create a dataflow that will read data from the database, and then transform the data by performing an
inner join.

You need to ignore spaces in the values when performing the inner join. The solution must minimize development
effort.

What should you do?

A.Append the queries by using fuzzy matching.


B.Merge the queries by using fuzzy matching.
C.Append the queries by using a lookup table.
D.Merge the queries by using a lookup table.

Answer: B

Explanation:

B. Merge the queries by using fuzzy matching.

You need to perform an inner join while ignoring spaces in the values. The best way to achieve this with
minimal development effort in a Fabric dataflow is by using fuzzy matching in a merge query.
Fuzzy Matching → Allows approximate matching, which can ignore leading, trailing, or extra spaces.

https://learn.microsoft.com/en-us/power-query/merge-queries-fuzzy-match

Question: 80
You have a Fabric tenant that contains a warehouse named Warehouse1. Warehouse1 contains two schemas name
schema1 and schema2 and a table named schema1.city.

You need to make a copy of schema1.city in schema2. The solution must minimize the copying of data.

Which T-SQL statement should you run?

A.INSERT INTO schema2.city SELECT * FROM schema1.city;


B.SELECT * INTO schema2.city FROM schema1.city;
C.CREATE TABLE schema2.city AS CLONE OF schema1.city;
D.CREATE TABLE schema2.city AS SELECT * FROM schema1.city;

Answer: C

Explanation:

C. CREATE TABLE schema2.city AS CLONE OF schema1.city;

This statement creates a new table named city in schema2 that has the same structure as the city table in
schema1 without copying any data. It essentially creates a metadata reference to the original table, which
minimizes the data copying.

Reference:

https://learn.microsoft.com/en-us/fabric/data-warehouse/clone-table

Question: 81

You need to prevent new tables added to Lakehouse1 from being added automatically to the default semantic

What should you configure?

A.the SQL analytics endpoint settings


B.the semantic model settings
C.the workspace settings
D.the Lakehouse1 settings

Answer: A

Explanation:

analytics endpoint settings is indeed the correct approach to prevent new tables added to Lakehouse1 from

Question: 82
HOTSPOT
-

You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains a lakehouse named
Lakehouse1 and a warehouse named Warehouse1.

You need to create a new table in Warehouse1 named POSCustomers by querying the customer table in
Lakehouse1.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.


Answer:
Explanation:

CREATE TABLE dbo.POSCustomers AS SELECT

FROM lakehouse1.dbo.customer.

This specifies the source table (customer) from lakehouse1.dbo.

Question: 83
You have a Fabric tenant.

You have a stored procedure that returns the number of active customers and their average sales for the current

You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be

Which type of activity should you add?

Copy data

Answer: D

Explanation:

executing a query or stored procedure. The output can be a singleton value or an array of attributes, which
can be consumed in a subsequent copy, transformation, or control flow activities like ForEach activity.

https://learn.microsoft.com/en-us/azure/data-factory/control-flow-lookup-activity

Question: 84
A.a shortcut
B.a dataflow
C.a view
D.a managed table

Answer: A
Explanation:

Shortcuts are specifically designed for cross-lakehouse referencing and provide a seamless and efficient way

Views typically reference tables within the same database or lakehouse. While you can create views that
reference tables in different schemas or databases, they generally do not support cross-workspace
references directly. Creating views for cross-workspace data might require additional configuration or data
duplication.

Question: 85
You have a Fabric tenant that contains a warehouse.

You are designing a star schema model that will contain a customer dimension. The customer dimension table will
be a Type 2 slowly changing dimension (SCD).

You need to recommend which columns to add to the table. The columns must NOT already exist in the source.

Which three types of columns should you recommend? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

A.a foreign key


B. a natural key
C. an effective end date and time
D. a surrogate key
E. an effective start date and time

Answer: CDE

Explanation:

Surrogate keys are typically used in dimension tables rather than fact tables. In a data warehouse, a
surrogate key is a unique identifier assigned to each record in a dimension table, usually for internal
processing and joining purposes. It provides a stable reference to the dimension record, regardless of any
changes in the natural key or other attributes.

Question: 86
You have a Fabric tenant.
You plan to create a data pipeline named Pipeline1. Pipeline1 will include two activities that will execute in
sequence.

You need to ensure that a failure of the first activity will NOT block the second activity.

Which conditional path should you configure between the first activity and the second activity?

Upon Failure
Upon Completion
Upon Skip
Upon Skip

Answer: B

Explanation:

Upon Completion ensures that the second activity executes regardless of whether the first activity succeeds
or fails.

Question: 87
You have a Microsoft Power BI semantic model.

You need to identify any surrogate key columns in the model that have the Summarize By property set to a value
other than to None. The solution must minimize effort.

What should you use?

A.DAX Formatter in DAX Studio


B.Model explorer in Microsoft Power BI Desktop
C.Model view in Microsoft Power BI Desktop
D.Best Practice Analyzer in Tabular Editor

Answer: D

Explanation:

The Best Practice Analyzer (BPA) in Tabular Editor can be configured to check for specific properties and
configurations in your model. This includes identifying columns with certain "Summarize By" settings.

You can create or use existing BPA rules to quickly identify any surrogate key columns that do not have the
"Summarize By" property set to "None," thus minimizing the manual effort required.

Question: 88
DRAG DROP
Which five actions should you perform in sequence? To answer, move the appropriate actions from the list of
actions to the answer area and arrange them in the correct order.

Answer:

Explanation:

Capture a Recording in Performance Analyzer.

Before analyzing query performance, you must first capture a recording in the Performance Analyzer.

This helps in identifying which queries take the most time.

Sort the Duration (ms) Column in Descending Order by DAX Query Time.

Sorting the DAX query time in descending order helps in identifying the slowest queries.

Copy the First Query to DAX Studio.

After identifying the slowest query, copy it into DAX Studio for further analysis and debugging.

Enable Query Timings and Server Timings, Then Run the Query.

In DAX Studio, enabling Query Timings and Server Timings helps to distinguish between:

Formula Engine (FE) execution time

Storage Engine (SE) execution time

View the Server Timings Tab.

The final step is to analyze the query's execution at the server level to identify bottlenecks and optimize
performance.

Question: 89

You need to programmatically change each column that ends in the word Key to meet the following requirements:
Set Nullable to False

Set Available in MDX to False.

A.Microsoft Power BI Desktop


B.ALM Toolkit
C.Tabular Editor
D.DAX Studio

Answer: C
Explanation:

C. Tabular Editor.

Tabular Editor is a powerful tool for modifying and managing Power BI semantic models (Tabular models)

Question: 90
HOTSPOT

You need to create a calculation item that will change the context from the selected date to month-to-date (MTD).

How should you complete the DAX expression? To answer, select the appropriate options in the answer area.
Answer:

Explanation:

CALCULATE:

CALCULATE is one of the most powerful DAX functions.

It evaluates an expression in a modified filter context.

Used to apply filters dynamically to an existing measure.

SELECTEDMEASURE:

This function is used in calculation items within a calculation group.


Helps create dynamic calculations without hardcoding a specific measure.

Question: 91
You have a Microsoft Power BI report named Report1 that uses a Fabric semantic model.

Users discover that Report1 renders slowly.

You open Performance analyzer and identify that a visual named Orders By Date is the slowest to render. The
duration breakdown for Orders By Date is shown in the following table.

What will provide the greatest reduction in the rendering duration of Report1?

A.Enable automatic page refresh.


B.Optimize the DAX query of Orders By Date by using DAX Studio.
C.Change the visual type of Orders By Date.
D.Reduce the number of visuals in Report1.

Answer: D

Explanation:

D. Reduce the number of visuals in Report1.

Rendering duration in Power BI is significantly impacted by the number of visuals on a report page.

Each visual requires query execution, data processing, and rendering time.

Reducing the number of visuals will lower the computational load, leading to faster rendering.

Question: 92
You have a custom Direct Lake semantic model named Model1 that has one billion rows of data.
You use Tabular Editor to connect to Model1 by using the XMLA endpoint.

You need to ensure that when users interact with reports based on Model1, their queries always use Direct Lake

From Model, configure the Default Mode option.


From Partitions, configure the Mode option.
From Model, configure the Storage Location option.
From Model, configure the Direct Lake Behavior option.
Answer: D

Explanation:

From Model, configure the Direct Lake Behavior option.

https://learn.microsoft.com/en-us/power-bi/enterprise/directlake-overview#fallback-behavior

The DirectLakeBehavior property can be configured by using Tabular Object Model (TOM) or Tabular Model
Scripting Language (TMSL).

The following example specifies all queries use Direct Lake mode only:

C#

Copy

// Disable fallback to DirectQuery mode.

//

database.Model.DirectLakeBehavior = DirectLakeBehavior.DirectLakeOnly = 1;

database.Model.SaveChanges();

Additional Question

Answer:

1 Open the model in Tabular Editor

2 Select the address column in SalesAddress

3 Set the Object Level Security to None for the sales manager role.

Question: 93
DRAG DROP
-

You create a semantic model by using Microsoft Power BI Desktop. The model contains one security role named
SalesRegionManager and the following tables:

• Sales
• SalesRegion
• SalesAddress

You need to modify the model to ensure that users assigned the SalesRegionManager role cannot see a column
named Address in SalesAddress.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of
actions to the answer area and arrange them in the correct order.

Answer:

Explanation:

1. "Open the model in Tabular Editor."

Since Tabular Editor is used for advanced metadata management, including setting OLS, this is the correct
tool.

Power BI Desktop does not provide direct OLS configuration.

2. Select the Address column in SalesAddress.

To modify OLS settings, the relevant column (Address) must be selected in SalesAddress.

3. Set Object Level Security to None for SalesRegionManager.

This ensures that SalesRegionManager has no restrictions on accessing the Address column.

If security were set to Default, the column could be hidden or restricted based on predefined roles.

Question: 94
functions and a FILTER function.

You are evaluating the performance of the measures.

In which use case will replacing the FILTER function with the KEEPFILTERS function reduce execution time?

A.when the FILTER function uses a nested calculate function


B.when the FILTER function references a measure
C.when the FILTER function references columns from multiple tables
D.when the FILTER function references a column from a single table that uses Import mode

Answer: D

Explanation:

D. When the FILTER function references a column from a single table that uses Import mode.

The FILTER function creates a new row context, which can be slow when iterating over large tables.

KEEPFILTERS applies additional filters without overriding the existing ones, improving query performance by
avoiding unnecessary row context transitions.

When FILTER references a column from a single Import mode table, replacing it with KEEPFILTERS reduces
the overhead of row context evaluation and improves execution speed.

Question: 95
You have a semantic model named Model1. Model1 contains five tables that all use Import mode. Model1 contains a
dynamic row-level security (RLS) role named HR. The HR role filters employee data so that HR managers only see
the data of the department to which they are assigned.

You publish Model1 to a Fabric tenant and configure RLS role membership. You share the model and related
reports to users.

An HR manager reports that the data they see in a report is incomplete.

What should you do to validate the data seen by the HR Manager?

A.Select Test as role to view the data as the HR role.


B.Filter the data in the report to match the intended logic of the filter for the HR department.
C.Select Test as role to view the report as the HR manager.
D.Ask the HR manager to open the report in Microsoft Power BI Desktop.

Answer: C

Explanation:

Option C (Select "Test as role" to view the report as the HR manager) is the best approach as it directly
validates what the specific HR manager sees under the dynamic RLS conditions, ensuring the completeness
and accuracy of the data.

A. Select "Test as role" to view the data as the HR role: This option is useful, but it doesn't specify viewing the
report as the specific HR manager, which is crucial to identify user-specific issues.
Question: 96
You have a Microsoft Fabric tenant that contains a dataflow.

You are exploring a new semantic model.

From Power Query, you need to view column information as shown in the following exhibit.

Which three Data view options should you select? Each correct answer presents part of the solution.

A. Show column value distribution


B. Enable details pane
C. Enable column profile
D. Show column quality details
E. Show column profile in details pane

Answer: ACD

Explanation:

Show column value distribution: This option provides a visual representation of the distribution of values in
each column, which is visible in the exhibit.

Enable column profile: This option displays statistics and other detailed information about each column,
including value distribution, which aligns with the data shown in the exhibit.

Show column quality details: This option shows the quality of the data in each column, indicating valid, error,
and empty values, as displayed in the exhibit.

Question: 97
HOTSPOT
You need to write a T-SQL query that will return the following columns.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct answer is worth one point.

Answer:
Explanation:

GREATEST(ListPrice, WholesalePrice, AgentPrice) → HighestSellingPrice

The GREATEST function returns the highest value from the given columns.

This is useful for finding the maximum price among different price categories.

If any value is NULL, it does not affect the result unless all values are NULL.

COALESCE(ListPrice, WholesalePrice, AgentPrice) → TradePrice

The COALESCE function returns the first non-null value from the list of columns.

This ensures that TradePrice takes the first available price instead of leaving it NULL.

If all values are NULL, the result is NULL.

Question: 98
You have a Fabric notebook that has the Python code and output shown in the following exhibit.
Which type of analytics are you performing?

A.descriptive
B.diagnostic
C.prescriptive
D.predictive

Answer: A

Explanation:

Descriptive Analytics: This type of analytics focuses on summarizing historical data to understand what has
happened. For example, generating reports and dashboards to show past sales performance.

Reference:

https://azure.microsoft.com/es-es/blog/answering-whats-happening-whys-happening-and-what-will-happen-
with-iot-analytics/
Question: 99
HOTSPOT
-

You have a Fabric warehouse that contains a table named Sales.Orders. Sales.Orders contains the following
columns.

You need to write a T-SQL query that will return the following columns.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct answer is worth one point.


Answer:

Explanation:

DATETRUNC

Weekday

The DATETRUNC function returns an input date truncated to a specified datepart.

Reference:
https://learn.microsoft.com/en-us/sql/t-sql/functions/datetrunc-transact-sql?view=sql-server-ver16

Question: 100
You have a Fabric tenant that contains JSON files in OneLake. The files have one billion items.

You plan to perform time series analysis of the items.

You need to transform the data, visualize the data to find insights, perform anomaly detection, and share the
insights with other business users. The solution must meet the following requirements:

• Use parallel processing.


• Minimize the duplication of data.
• Minimize how long it takes to load the data.

What should you use to transform and visualize the data?

A. the PySpark library in a Fabric notebook


B. the pandas library in a Fabric notebook
C. a Microsoft Power BI report that uses core visuals

Answer: A

Explanation:

PySpark provides robust capabilities for data transformation and manipulation, making it well-suited for
preparing data for time series analysis. You can then use libraries like pandas for further data manipulation if
needed and leverage Spark's machine learning capabilities for anomaly detection.

Question: 101
You have a Fabric tenant that contains customer churn data stored as Parquet files in OneLake. The data contains
details about customer demographics and product usage.

You create a Fabric notebook to read the data into a Spark DataFrame. You then create column charts in the
notebook that show the distribution of retained customers as compared to lost customers based on geography, the
number of products purchased, age, and customer tenure.

Which type of analytics are you performing?

A.diagnostic
B.descriptive
C.prescriptive
D.predictive

Answer: B

Explanation:

Descriptive - it's just describing the customers, not saying why they stayed or left

Descriptive analytics tells what happened in the past, presenting it as numbers and visuals in reports and
dashboards.Diagnostic analytics gives the reason why something happened.Predictive analytics determines
the potential outcomes of present and past actions and trends.Prescriptive analytics offers decision support
for the best course of action.Given the scenario in the question where data is read into a Spark DataFrame and
column charts are created to show the distribution of retained customers compared to lost customers based
on various factors, this falls under the definition of descriptive analytics. No future predictions or
prescriptions are made, nor are reasons for the past events provided.Final Answer:The type of analytics being
performed in the scenario described is descriptive analytics.

https://www.selecthub.com/business-intelligence/predictive-descriptive-prescriptive-analytics/

Question: 102
HOTSPOT
-

You have a Fabric tenant that contains a semantic model. The model contains data about retail stores.

You need to write a DAX query that will be executed by using the XMLA endpoint. The query must return the total
amount of sales from the same period last year.

How should you complete the DAX expression? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

CALCULATE.

-LY Sales.

CALCULATE([Total Sales], SAMEPERIODLASTYEAR('Order'[Order Date]))

The entire expression calculates the sales amount for the same period last year based on the current date
context.

The SUMMARIZE function creates a summary table based on grouping and aggregation.

RETURN

_ LYSales

The RETURN statement outputs the _LYSales variable.

Question: 103

You need to identify the min and max values for each column in the query results.

Which three Data view options should you select? Each correct answer presents part of the solution.

Show column value distribution


Enable column profile
Show column profile in details pane
Show column quality details

Answer: BCE
Explanation:

B. Enable column profile.

E. Enable details pane.

Required to view the detailed column profile statistics.mn profile in details pane E - Enable details pane; this
needs to be enabled to be able to see the output of (C).

Question: 104
You have a Fabric tenant that contains a Microsoft Power BI report.

You need to display the following column statistics:

A.Table.schema
B.Table.view
C.Table.FuzzyGroup
D.Table.Profile

Answer: D

Explanation:

https://learn.microsoft.com/en-us/powerquery-m/table-profile

Question: 105

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not

You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table named
DESCRIBE DETAIL customer -

Yes

Answer: B
Explanation:

It should be DESCRIBE HISTORY customer

DESCRIBE HISTORY will show maintenance history of the table.

Question: 106

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not

You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and

Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

No. describe() and summary() provide the summary statistics.


Question: 107
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.

You have a Fabric tenant that contains a semantic model named Model1.

You discover that the following query performs slowly against Model1.

You need to reduce the execution time of the query.

Solution: You replace line 4 by using the following code:

ISEMPTY ( RELATEDTABLE ( 'Order Item' ) )

Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

B is correct (NO) because the suggested change does not meet the goal because it reverses the logic of the
query.

Question: 108

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not

You have a Fabric tenant that contains a semantic model named Model1.
You need to reduce the execution time of the query.

Solution: You replace line 4 by using the following code:

NOT ISEMPTY ( CALCULATETABLE ( 'Order Item ' ) )

Does this meet the goal?

A.Yes
B.No

Answer: A

Explanation:

A — The proposed solution improves efficiency by reducing the number of calculations required. Instead of
counting all the rows for each customer and then checking if the count is greater than zero, it simply checks if
there are any rows at all, which requires fewer computational resources and execution time

Yes, replacing CALCULATE ( COUNTROWS( 'Order Item' ) ) > 0 with NOT ISEMPTY ( CALCULATETABLE (
'Order Item ' ) ) should reduce the execution time of the query. It is a simpler, more meaningful, and faster way
to check if a table is empty.

https://www.sqlbi.com/articles/check-empty-table-condition-with-dax/

Question: 109

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not

You have a Fabric tenant that contains a semantic model named Model1.

You need to reduce the execution time of the query.

Solution: You replace line 4 by using the following code:


CALCULATE ( COUNTROWS ( 'Order Item' ) ) >= 0

Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

Question: 110
HOTSPOT
-

Case study
-

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study


-
To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview
-

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment
-

Identity Environment
-

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment
-

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements
-

Planned Changes
-

Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements


-

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements


-

Contoso identifies the following data preparation requirements:

• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements


-

Contoso identifies the following requirements for implementing and managing semantic models:

• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements
-

Contoso identifies the following high-level requirements that must be considered for all solutions:

• Follow the principle of least privilege when applicable.


• Minimize implementation and maintenance effort when possible.

You need to migrate the Research division data for Productline2. The solution must meet the data preparation
requirements.

How should you complete the code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:

Explanation:

format("delta") → Saves the file in Delta Lake format, which supports ACID transactions and versioning.

save("Tables/productline2") → Saves the Delta table in the default location for managed tables.

Case study -

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -

Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
Data Preparation Requirements -

Contoso identifies the following data preparation requirements:

• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -

Contoso identifies the following requirements for implementing and managing semantic models:

• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -

Contoso identifies the following high-level requirements that must be considered for all solutions:

• Follow the principle of least privilege when applicable.


• Minimize implementation and maintenance effort when possible.

Which syntax should you use in a notebook to access the Research division data for Productline1?

A.spark.read.format(“delta”).load(“Files/ResearchProduct”)
B.spark.sql(“SELECT * FROM Lakehouse1.ResearchProduct ”)
C.spark.sql(“SELECT * FROM Lakehouse1.Tables.ResearchProduct ”)
D.external_table(ResearchProduct)

Answer: A

Explanation:

A.spark.read.format("delta").load("Files/ResearchProduct")

The key change is understanding that shortcuts do not behave like managed tables in Spark SQL. To read
data from a shortcut you need to use spark.read.format and use Files/ path, since it's a shortcut it behaves like
a file in the filesystem.

Question: 112
HOTSPOT
For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Answer:

Explanation:

1. YES - The .select() function in the code specifies the exact columns ('SalesOrderNumber', 'OrderDate',
'CustomerName', and 'UnitPrice') to be selected.

Therefore, only these columns will be read from the CSV.

2. NO - The withColumn("Year", year("OrderDate")) function adds a new column called "Year" by extracting
the year from the "OrderDate" column.

However, it does not replace the "OrderDate" column—it only adds the new "Year" column.
3. YES - The inferSchema='true' tells Spark to infer the data types of each column in the CSV, which requires
an extra scan of the data to determine these types.

Question: 113
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.

You have a Fabric tenant that contains a new semantic model in OneLake.

You use a Fabric notebook to read the data into a Spark DataFrame.

You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
numeric columns.

Solution: You use the following PySpark expression:

df.describe().show()

Does this meet the goal?

A.Yes
B.No

Answer: A

Explanation:

DataFrame.Describe = Computes basic statistics for numeric and string columns, including count, mean,
stddev, min, and max. If no columns are given, this function computes statistics for all numerical or string
columns.

Question: 114
You have a Fabric tenant.

You have a stored procedure that returns the number of active customers and their average sales for the current

You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be

Which type of activity should you add?

C.Append variable
D.Lookup
Answer: D
Explanation:

Question: 115
HOTSPOT
-

You have a Fabric tenant that contains a semantic model named model1. The two largest columns in model1 are
shown in the following table.

You need to optimize model1. The solution must meet the following requirements:

• Reduce the model size.


• Increase refresh performance when using Import mode.
• Ensure that the datetime value for each sales transaction is available in the model.

What should you do on each column? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.


Answer:

Explanation:

TransactionKey --> Remove the column.

There is no need to keep 160 GB for a surrogate key, it can be removed and SaleDateTime can be moved into
the SalesTransaction table that will for sure "Increase refresh performance when using Import mode." and
"Reduce the model size." since there is no need for an heavy join like this one.

SaleDateTime --> Split the column.

Splitting the column would lead for sure to "Ensure that the datetime value for each sales transaction is
available in the model." and i think it would lead to a better segmentation of the data meaning less cardinality
that would lead to "Increase refresh performance when using Import mode." and "Reduce the model size."

Question: 116
DRAG DROP
-

You have a Fabric tenant that contains a data warehouse named DW1. DW1 contains a table named DimCustomer.
DimCustomer contains the fields shown in the following table.

You need to identify duplicate email addresses in DimCustomer. The solution must return a maximum of 1,000
records.

Which four T-SQL statements should you run in sequence? To answer, move the appropriate statements from the
list of statements to the answer area and arrange them in the correct order.

Answer:
Explanation:

SELECT TOP(1000) CustomerAltKey, COUNT(*)

This selects the CustomerAltKey column and counts the number of occurrences of each key.

The TOP(1000) ensures that only the first 1,000 results are returned.

FROM DimCustomer

Specifies the DimCustomer table as the source of data.

GROUP BY CustomerAltKey

Groups the data by CustomerAltKey, meaning that each unique CustomerAltKey will have its count
calculated.

HAVING COUNT(*) > 1

Filters the results to include only those CustomerAltKey values that appear more than once in the table,
identifying duplicates.

Question: 117
HOTSPOT
For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Answer:

Explanation:

1. Dimension.GetDirectReports is a scalar T-SQL function.

Answer: No

A scalar function returns a single value, but GetDirectReports likely returns multiple rows (i.e., a table)
containing details of employees reporting to a manager. This suggests it is a table-valued function (TVF)
rather than a scalar function.

2. The Dimension.GetDirectReports function will run only once when the query runs."

Answer: No

If GetDirectReports is a table-valued function, it executes multiple times—once for each relevant row—
rather than just once per query execution.

3. The output rows will include at least one row for each row in the Dimension.Employee table."

Answer:yes.

the output of the query using OUTER APPLY will include at least one row for each row in Dimension.employee,
even if the Dimension.getDirectReports function does not return any results for some rows.

Question: 118
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.

NOT ( CALCULATE ( COUNTROWS ( 'Order Item' ) ) < 0)

Does this meet the goal?

A.Yes
B.No

Answer: B
Explanation:

Question: 119
Case study -

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -

Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -

Contoso identifies the following data preparation requirements:

• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -


Contoso identifies the following requirements for implementing and managing semantic models:

• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -

Contoso identifies the following high-level requirements that must be considered for all solutions:

• Follow the principle of least privilege when applicable.


• Minimize implementation and maintenance effort when possible.

Which syntax should you use in a notebook to access the Research division data for Productline1?

A.spark.read.format(“delta”).load(“Tables/ResearchProduct”)
B.spark.read.format(“delta”).load(“Files/ResearchProduct”)
C.external_table(‘Tables/ResearchProduct)
D.external_table(ResearchProduct)

Answer: A
Explanation:

The syntax spark.read.format("delta").load is specifically designed for loading Delta tables in Apache Spark.
Delta Lake is an optimized storage layer that brings reliability to data lakes.

The path "Tables/ResearchProduct" indicates that the data is organized within a directory structure named
"Tables," which aligns with best practices for storing structured data in a data lake. This structure is ideal for
accessing data in a Delta format.

Question: 120
You have a Microsoft Power BI Premium Per User (PPU) workspace that contains a semantic model.

You have an Azure App Service app named App1 that modifies row-level security (RLS) for the model by using the
XMLA endpoint.

App1 requires users to sign in by using their Microsoft Entra credentials to access the XMLA endpoint.

You need to configure App1 to use a service account to access the model.

What should you do first?

A. Add a managed identity to the workspace.


B. Modify the XMLA Endpoint setting.
C. Upgrade the workspace to Premium capacity.
D. Add a managed identity to App1.

Answer: D

Explanation:

Adding a managed identity to App1 will allow it to authenticate and access resources securely without
needing to manage credentials explicitly. This is the first step in configuring the app to use a service account
for accessing the model via the XMLA endpoint.

Question: 121
HOTSPOT
-

You have a Fabric tenant that contains a warehouse named WH1.

You have source data in a CSV file that has the following fields:

• SalesTransactionID
• SaleDate
• CustomerCode
• CustomerName
• CustomerAddress
• ProductCode
• ProductName
• Quantity
• UnitPrice

You plan to implement a star schema for the tables in WH1. The dimension tables in WH1 will implement Type 2
slowly changing dimension (SCD) logic.

You need to design the tables that will be used for sales transaction analysis and load the source data.

Which type of target table should you specify for the CustomerName, CustomerCode, and SaleDate fields? To
answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.


Answer:
Explanation:

Box 1: Dimension .

CustomerCode is an identifier used to track a customer.

It is not a measurable value but a descriptive attribute.

It belongs in the Customer Dimension Table along with CustomerName, CustomerAddress, etc.

Box 2 : Dimension .

CustomerName describes a customer and is not measurable.

It is part of the Customer Dimension Table.

It provides context but does not belong in the Fact Table.

Box 3:Fact.

Question: 122
You have a Fabric tenant.
You have a stored procedure that returns the number of active customers and their average sales for the current
month.

You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
available to the downstream activities of the pipeline.

Which type of activity should you add?

D.Append variable

Answer: C

Explanation:

Question: 123
HOTSPOT
-

You have a Fabric tenant that contains a warehouse named Warehouse1. Warehouse1 contains a fact table named
FactSales that has one billion rows.

You run the following T-SQL statement.

CREATE TABLE test.FactSales AS CLONE OF dbo.FactSales;

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

Yes

No

No

https://learn.microsoft.com/en-us/sql/t-sql/statements/create-table-as-clone-of-transact-sql?
view=fabric&preserve-view=true

Creates a new table as a zero-copy clone of another table in Warehouse in Microsoft Fabric. Only the
metadata of the table is copied. The underlying data of the table, stored as parquet files, is not copied.

https://learn.microsoft.com/en-us/fabric/data-warehouse/clone-table

Separate and independent

Upon creation, a table clone is an independent and separate copy of the data from its source.

Any changes made through DML or DDL on the source of the clone table are not reflected in the clone
table.Similarly, any changes made through DDL or DML on the table clone are not reflected on the source of
the clone table.

Question: 124

You need to configure authentication for the connection.

Which two values should you provide? Each correct answer presents part of the solution.

the shared access signature (SAS) token


the secret access key

Answer: BD

Explanation:

Secret Access Key (B): The secret access key is a private key used in conjunction with the access key ID to

Access Key ID (D): The access key ID is a public identifier for your AWS account. It is required to identify the
Question: 125
You have a Fabric tenant that contains a data warehouse.

You need to load rows into a large Type 2 slowly changing dimension (SCD). The solution must minimize resource
usage.

Which T-SQL statement should you use?

A.UPDATE AND INSERT


B.MERGE
C.TRUNCATE TABLE and INSERT
D.CREATE TABLE AS SELECT

Answer: B

Explanation:

B: "MERGE."

MERGE Statement: The MERGE statement in T-SQL is specifically designed for handling situations where you
need to perform both insert and update operations based on certain conditions. It's particularly useful for
managing Type 2 Slowly Changing Dimensions (SCDs) in data warehouses, where historical data needs to be
preserved, and changes are tracked over time.

Resource Efficiency: Using the MERGE statement minimizes resource usage because it consolidates multiple
operations into a single statement. Instead of running separate UPDATE and INSERT commands, MERGE
allows you to update existing rows and insert new rows in one go. This reduces the complexity of your T-SQL
code and improves performance by reducing the number of operations required.

Question: 126
Case study -

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.

To start the case study -


To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.

Overview -

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.

Existing Environment -

Identity Environment -

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.

Data Environment -

Contoso has the following data environment:

• The Sales division uses a Microsoft Power BI Premium capacity.


• The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
• The Research department uses an on-premises, third-party data warehousing product.
• Fabric is enabled for contoso.com.
• An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
• A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.

Requirements -

Planned Changes -

Contoso plans to make the following changes:

• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

Data Analytics Requirements -

Contoso identifies the following data analytics requirements:

• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.

Data Preparation Requirements -

Contoso identifies the following data preparation requirements:

• The Research division data for Productline2 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.

Semantic Model Requirements -


Contoso identifies the following requirements for implementing and managing semantic models:

• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.

General Requirements -

Contoso identifies the following high-level requirements that must be considered for all solutions:

• Follow the principle of least privilege when applicable.


• Minimize implementation and maintenance effort when possible.

Which syntax should you use in a notebook to access the Research division data for Productline1?

A.spark.sql("SELECT * FROM Lakehouse1.Tables.ResearchProduct")


B.spark.read.format("delta").load("Tables/productline1/ResearchProduct")
C.external_table(ResearchProduct)
D.spark.read.format("delta").load("Tables/ResearchProduct")

Answer: D

Explanation:

spark.read.format("delta").load("Tables/ResearchProduct") is designed for loading Delta tables using Apache


Spark. Delta Lake is an optimized storage layer that ensures reliability and performance for data lakes.

This syntax correctly specifies the format as Delta and loads the data from the specified table in the
lakehouse.

Question: 127
You have a Fabric tenant that contains a lakehouse named LH1.

You create new tables in LH1.

You need to ensure that the tables are added automatically to the default semantic model.

What should you do?

A.Disable Query Caching for the default semantic model.


B.From the settings pane of LH1, enable Sync the default Power BI semantic model.
C.Enable Refresh for the default semantic model.
D.From the Endorsement and discovery settings of LH1, select Make discoverable.

Answer: B

Explanation:

B. From the settings pane of LH1, enable "Sync the default Power BI semantic model".

When you create new tables in a Microsoft Fabric Lakehouse (LH1), they are not automatically added to the
default Power BI semantic model unless synchronization is enabled.

To ensure automatic addition of new tables, you must:

Enable "Sync the default Power BI semantic model" from the settings pane of LH1.
This feature allows newly created tables in the Lakehouse to be automatically included in the Power BI
semantic model, ensuring seamless data integration for reporting and analysis

Question: 128
You have a Fabric tenant.

You are creating a Fabric Data Factory pipeline.

You have a stored procedure that returns the number of active customers and their average sales for the current
month.

You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
available to the downstream activities of the pipeline.

Which type of activity should you add?

A.Append variable
B.Lookup
C.Copy data
D.KQL

Answer: B
Explanation:

B: "Lookup."

Lookup Activity: In Azure Data Factory (and its Fabric Data Factory counterpart), the Lookup activity is
designed to retrieve data from a data source and make the results available for downstream activities in the
pipeline. It is particularly useful when you need to fetch data from a stored procedure, SQL query, or any data
source and use that data within your pipeline.

Executing Stored Procedures: When you use the Lookup activity to execute a stored procedure, it will capture
the output of the stored procedure (in this case, the number of active customers and their average sales) and
store the results in a manner that can be used by subsequent activities in the pipeline. This makes it ideal for
scenarios where you need to perform additional operations based on the retrieved data.

Question: 129
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More

After you answer a question in this section, you will NOT be able to return. As a result, these questions do not

You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.

You query a measure that includes the USERNAME() function, and the query returns a blank result.
Does this meet the goal?

A.Yes
B.No

Answer: B

Explanation:

user principal name (UPN) of a user is returned by the USERPRINCIPALNAME() function.

Question: 130
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More
than one solution in the set might solve the problem. It is also possible that none of the solutions in the set solve
the problem.

After you answer a question in this section, you will NOT be able to return. As a result, these questions do not
appear on the Review Screen.

Your network contains an on-premises Active Directory Domain Services (AD DS) domain named contoso.com that
syncs with a Microsoft Entra tenant by using Microsoft Entra Connect.

You have a Fabric tenant that contains a semantic model.

You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.

You query a measure that includes the USERNAME() function, and the query returns a blank result.

You need to ensure that the measure returns the user principal name (UPN) of a user.

Solution: You update the measure to use the USERPRINCIPALNAME() function.

Does this meet the goal?

A.Yes
B.No

Answer: A

Explanation:

user principal name (UPN) of a user is returned by the USERPRINCIPALNAME() function.

Question: 131
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More
After you answer a question in this section, you will NOT be able to return. As a result, these questions do not
appear on the Review Screen.

Your network contains an on-premises Active Directory Domain Services (AD DS) domain named contoso.com that
syncs with a Microsoft Entra tenant by using Microsoft Entra Connect.

You have a Fabric tenant that contains a semantic model.

You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.

You query a measure that includes the USERNAME() function, and the query returns a blank result.

You need to ensure that the measure returns the user principal name (UPN) of a user.

Solution: You add user objects to the list of synced objects in Microsoft Entra Connect.

Does this meet the goal?

A.Yes
B.No

Answer: A

Explanation:

A. Yes.

Synced Objects in Microsoft Entra Connect: Adding user objects to the list of synced objects in Microsoft
Entra Connect means that those users will be included in the synchronization process between your on -
premises directory and Microsoft Entra ID (formerly Azure Active Directory). This allows for the seamless
integration and management of user identities across environments.

Goal Achievement: If the goal is to ensure that specific user objects are included in the synchronization
process, then adding them to the list of synced objects will meet this requirement. This ensures that the user
accounts and their associated attributes are correctly synchronized and available in Microsoft Entra ID.

Question: 132
You have a Fabric tenant that contains a warehouse named DW1 and a lakehouse named LH1. DW1 contains a table
named Sales.Product. LH1 contains a table named Sales.Orders.

You plan to schedule an automated process that will create a new point-in-time (PIT) table named
Sales.ProductOrder in DW1. Sales.ProductOrder will be built by using the results of a query that will join
Sales.Product and Sales.Orders.

You need to ensure that the types of columns in Sales.ProductOrder match the column types in the source tables.
The solution must minimize the number of operations required to create the new table.

Which operation should you use?

A.INSERT INTO
B.CREATE TABLE AS SELECT (CTAS)
C.CREATE TABLE AS CLONE OF
D.CREATE MATERIALIZED VIEW AS SELECT

Answer: B

Explanation:
Question: 133
HOTSPOT
-

You have a Fabric tenant that contains the semantic model shown in the following exhibit.

Use the drop-down menus to select the answer choice that completes each statement based on the information
presented in the graphic.
NOTE: Each correct selection is worth one point.

Answer:

Explanation:

1. Querying SQL Views from the Warehouse Causes a Fallback to "DirectQuery" Mode

Correct Answer: DirectQuery.

In Microsoft Fabric, the default query mode for a semantic model in a Lakehouse is Direct Lake, which
provides fast, real-time querying on delta tables.

However, when querying SQL Views from the Warehouse, the system cannot use Direct Lake because SQL
Views are not stored as delta tables.

As a result, it falls back to DirectQuery mode, where queries are executed on demand against the source
database.

2. Row and Column Security is Undefined.

Correct Answer: Undefined.

In Direct Lake mode, row-level security (RLS) and object-level security (OLS) are not yet supported in
Microsoft Fabric.

Since DirectQuery is being used (due to the SQL View fallback), security settings are undefined because they
are not automatically applied.

To enforce security, you must configure RLS or OLS manually in Power BI or SQL Warehouse.

Question: 134
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More
than one solution in the set might solve the problem. It is also possible that none of the solutions in the set solve
the problem.

After you answer a question in this section, you will NOT be able to return. As a result, these questions do not
appear on the Review Screen.

Your network contains an on-premises Active Directory Domain Services (AD DS) domain named contoso.com that
syncs with a Microsoft Entra tenant by using Microsoft Entra Connect.

You have a Fabric tenant that contains a semantic model.

You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.

You query a measure that includes the USERNAME() function, and the query returns a blank result.

You need to ensure that the measure returns the user principal name (UPN) of a user.

Solution: You create a role in the model.

Does this meet the goal?

A. Yes
B.No

Answer: B
Explanation:

Creating role will not solve the purpose , will have to use Userprincipalname().

Question: 135

You need to ensure that User1 can perform the following tasks:

Create a new domain.


Create two subdomains named subdomain1 and subdomain2.
Assign Workspace1 to subdomain1.

The solution must follow the principle of least privilege.

domain admin
domain

Answer: C
Explanation:

C .Fabric admin.

Comprehensive Permissions: The Fabric admin role has the necessary permissions to create and manage
domains, subdomains, and assign workspaces to them. This role encompasses the administrative capabilities
required for the tasks you mentioned.

Least Privilege Principle: While the Fabric admin role provides broad permissions, it is necessary in this
context because the tasks involve high-level administrative actions that cannot be accomplished with more
restricted roles. Thus, it adheres to the principle of least privilege by granting just enough access to perform
the required tasks without excessive permissions.

Question: 136
HOTSPOT
-

You have a Fabric tenant that contains three users named User1, User2, and User3. The tenant contains a security
group named Group1. User1 and User3 are members of Group1.

The tenant contains the workspaces shown in the following table.

The tenant contains the domains shown in the following table.

User1 creates a new workspace named Workspace3.

You assign Domain1 as the default domain of Group1.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

No, Yes, No

User2 is assigned the Contributor role for Workspace3.

No. User2 is not a member of Group1 and has no direct association with Domain1 or Workspace3. There's no
indication User2 has inherited any role here.

User3 is assigned the Viewer role for Workspace3.

Yes. User3 is a member of Group1, which is associated with Domain1 because it was created by User 1. When
Workspace3 is created under Domain1, User3 likely inherits the Viewer role.

User3 is assigned the Contributor role for Workspace1.

No. User3 is a member of Group1, but Workspace1 is not linked to Domain1. It's directly tied to User1 as the
admin. There's no evidence User3 has any Contributor role in Workspace1.

Question: 137
You have a Fabric warehouse named Warehouse1 that contains a table named Table1. Table1 contains customer

Which two objects should you create? Each correct answer presents part of the solution.

A.DATABASE ROLE
B.STORED PROCEDURE
C.CONSTRAINT
D.FUNCTION
POLICY

Answer: DE

Explanation:
D. FUNCTION.

E. SECURITY POLICY.

FUNCTION (D): Used to define a security function that filters the rows of the table based on the user's identity
or specific criteria.

SECURITY POLICY (E): Applies the security function to the table to restrict access to rows according to the
defined criteria.

Question: 138
You have a Fabric workspace named Workspace1 that contains a lakehouse named Lakehouse1. Lakehouse1
contains a table named Table1. Table1 contains the following data.

You need to perform the following actions:

• Load the data from Table1 into a star schema.


• Create a product dimension table named DimProduct and a fact table named FactSales.

Which three columns should you include in DimProduct?

A.ProductColor, ProductID, and ProductName.


B.ProductName, SalesAmount, and TransactionlD.
C.Date, ProductID, and TransactionlD.
D.ProductID, ProductName, and SalesAmount

Answer: A

Explanation:

ProductColor: This is another attribute describing the product. It adds further context and can be useful in
analyzing product variations and performance by color.

ProductID: This is a unique identifier for each product. It serves as the primary key in the dimension table and
will be used to join with the fact table.

ProductName: This provides a descriptive name for each product, which is useful for reporting and analysis.

Question: 139
DRAG DROP
-

You have a Fabric workspace that contains a Dataflow Gen2 query. The query returns the following data.

You need to filter the results to ensure that only the latest version of each customer’s record is retained. The
solution must ensure that no new columns are loaded to the semantic model.

Which four actions should you perform in sequence in Power Query Editor? To answer, move the appropriate
actions from the list of actions to the answer area and arrange them in the correct order.

Answer:

Explanation:

1. Group by Customer ID and calculate the max version date per customer ID.

This is the first step to identify the most recent version date for each customer.

2. Filter the query where the version date value equals the max version date value.

This ensures that only the rows with the most recent version date are included.

3. Remove duplicates based on Customer ID.

This step ensures that each customer appears only once in the final dataset.

4. Remove the max version date column.


After filtering and removing duplicates, the max version date column is no longer needed and can be removed
to clean up the dataset.

Question: 140
HOTSPOT
-

You have a Microsoft Power Bl project that contains a file named definition.pbir. definition.pbir contains the
following JSON.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

definition.pbir is in the PBIR-Legacy format.

Answer: No

The .pbir file format is a new format used in Power BI for report definitions, but it is not part of the legacy PBIR
format. The legacy format refers to older Power BI report types.

The semantic model referenced by definition.pbir is located in the Power BI service.

Answer: No

The .pbir file format does not inherently mean that the semantic model is stored in the Power BI service. It can
reference models from different sources, including local files or cloud services.

When the related report is opened, Power BI Desktop will open the semantic model in full edit mode.

Answer: Yes

Power BI Desktop allows full editing of the report and its semantic model when opening a .pbir file, assuming
the appropriate permissions and connections are available.

Question: 141

You need to optimize the deployment of Model1. The solution must minimize how long it takes to deploy Model1.

Select Small semantic model storage format.

Set Enable Cache for Shortcuts to On.


Select Large semantic model storage format.

Answer: D

Explanation:

Large semantic model storage format allows for efficient handling of larger datasets and optimized

Question: 142
You have a Microsoft Power BI project that contains a semantic model.

You plan to use Azure DevOps for version control.

You need to modify the .gitignore file to prevent the data values from the data sources from being pushed to the

Which file should you reference?


D.model.bim

Answer: B
Explanation:

B. cache .abf Justification: This file contains the cached data for the semantic model. It is a binary file and can
be very large. Excluding this file from version control is necessary to prevent pushing large amounts of
unnecessary data, and will prevent data from data sources from being pushed to the repository.Why it's the
correct option: Contains the cached data from the data model.

Question: 143
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 uses Pro license mode and
contains a semantic model named Model1.

You need to ensure that Model1 supports XMLA connections.

Which setting should modify?

A. Users can edit data models in the Power BI service


B. Enforce strict access control for all data connection types
C. Enable Cache for Shortcuts
D.License mode

Answer: D

Explanation:

The license mode of the workspace impacts the availability of XMLA endpoints. By using a Pro license mode,
you might not have full support for XMLA connections1. Switching to a Premium license mode would enable
XMLA endpoints, allowing for more advanced connectivity options, including XMLA connections.

Question: 144

You need to visualize a diagram of the model. The diagram must contain only the Sales table and related tables.

Data view

D.DAX query view

Answer: C
Explanation:

Question: 145
HOTSPOT
-

You have a Fabric tenant that contains a lakehouse named LH1.

You need to deploy a new semantic model. The solution must meet the following requirements:
• Support complex calculated columns that include aggregate functions, calculated tables, and Multidimensional
Expressions (MDX) user hierarchies.
• Minimize page rendering times.

How should you configure the model? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

Mode: Import.

Query Caching: On.

Question: 146
HOTSPOT

For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:

Explanation:

The query excludes sales that have a Status of Cancelled. Yes

This means the query filters out sales records where the status is "Cancelled."

The query calculates the total sales of each product category for the last 30 days. Yes

The query is aggregating sales data based on product categories within the last 30 days.

The query includes product categories that have had zero sales during the last 30 days. No

This indicates that product categories with no sales in the last 30 days are not included in the results.

Question: 147

You need to calculate the proportion of the total quantity of each sales order represented by each product within

A.
B.

C.

D.

Answer: D

Explanation:

SUM(OrderQty) OVER (PARTITION BY SalesOrderID):

This calculates the total OrderQty per SalesOrderID.

OrderQty * 1.0 / SUM(OrderQty) OVER (PARTITION BY SalesOrderID):

This divides each row’s OrderQty by the total OrderQty for the SalesOrderID to calculate the proportion.

Question: 148
HOTSPOT
You need build a T-SQL statement that will return the total sales amount by OrderDate only for the days that are
holidays in Australia. The total sales amount must sum the quantity multiplied by the price on each row in the
dbo.sales table.

How should you complete the statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:

Explanation:

Sum(s.Quantity * s.UnitPrice), Inner.

For the type of Joins, they only want it if it is Australia. Inner will only bring over the matches, making it an
ideal Join Type. To be fair, I had to look it up, but once I learned that I wanted to share.
Question: 149
HOTSPOT
-

You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains a warehouse named
DW1. DW1 contains two tables named Employees and Sales. All users have read access to Dw1.

You need to implement access controls to meet the following requirements:

• For the Sales table, ensure that the users can see only the sales data from their respective region.
• For the Employees table, restrict access to all Personally Identifiable Information (PII).
• Maintain access to unrestricted data for all the users.

What should you use for each table? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:
Explanation:

Employees: Column-level security.

Column-Level-Security (CLS):

https://learn.microsoft.com/en-us/fabric/data-warehouse/column-level-security

"allowing you to restrict column access to protect sensitive data..."

Sales: Row-Level-Security (RLS)

https://learn.microsoft.com/en-us/fabric/data-warehouse/row-level-security

"control access to rows in a database table. For example, you can ensure that workers access only those data
rows that are pertinent to their department." (or Sales Region in this case)

Question: 150
You need to create a Microsoft Power BI file that will be used to create multiple reports. The solution must meet

The file must include predefined data source connections.


The file must include the report structure and formatting.
The file must NOT contain any data.

Which file format should you use?

A.PBIT
B.PBIDS
C.PBIX
D.PBIP

Answer: A

Explanation:

A. PBIT (Power BI Template file)

Includes predefined data source connections – PBIT files store connection details and queries.

Includes the report structure and formatting – PBIT files retain visuals, layouts, and report configurations.

Does NOT contain any data – When opened, a PBIT file prompts the user to refresh the data from the
connected sources.

Question: 151
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains a single semantic
model that has two Microsoft Power BI reports.

You have a Microsoft 365 subscription that contains a data loss prevention (DLP) policy named DLP1.

You need to apply DLP1 to the items in Workspace1.

What should you do?

A.Create a workspace identity.


B.Apply a certified endorsement to the semantic model.
C.Apply sensitivity labels to the semantic model and reports.
D.Apply a master data endorsement to the semantic model.

Answer: C
Explanation:

C. Apply sensitivity labels to the semantic model and reports .

Sensitivity labels classify Power BI assets (semantic models, reports) and apply DLP policies.

Microsoft 365 DLP policies work with sensitivity labels to enforce security.

A label like "Confidential – Internal Use" could trigger a DLP rule to prevent external sharing.

Question: 152
You have a Fabric tenant that contains a workspace named Workspace1 and a user named User1. User1 is assigned

You plan to configure Workspace1 to use an Azure DevOps repository for version control.
Which two settings should you enable for User1? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

Users can sync workspace items with GitHub repositories


Users can create and use Data workflows

Users can synchronize workspace items with their Git repositories

Answer: CD
Explanation:

C. Users can create Fabric items.

To commit items to a repository, User1 must be able to create Fabric items in the workspace.

Required for managing and pushing changes in the Fabric workspace.

Question: 153

You perform the following actions:

Move Lakehouse1 to Folder1.


Run DeployPipeline1.

\Lakehouse1

Answer: D

Explanation:

D. \Folder1\Lakehouse1.
Question: 154
Your company has a finance department.

You have a Fabric tenant, an Azure Storage account named storage1, and a Microsoft Entra group named Group1.
Group1 contains the users in the finance department.

You need to create a new workspace named Workspace1 in the tenant. The solution must meet the following
requirements:

• Ensure that the finance department users can create and edit items in Workspace1.
• Ensure that Workspace1 can securely access storage1 to read and write data.
• Ensure that you are the only admin of Workspace1.
• Minimize administrative effort.

You create Workspace1.

Which two actions should you perform next? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

A.Assign the Contributor role to Group1.


B.Create a workspace identity.
C.Assign the Admin role to yourself.
D.Assign the Contributor role to each finance department user.

Answer: AB

Explanation:

A. Assign the Contributor role to Group1:Since Group1 already contains all finance department users,
assigning the Contributor role to the group grants them the necessary permissions (create and edit items) in
Workspace1 with minimal administrative effort. This avoids having to assign roles to each user individually.B.
Create a workspace identity:A workspace identity (or managed identity) allows Workspace1 to securely
access external resources like storage1. By creating and configuring a workspace identity, you can grant it the
appropriate permissions on storage1 for secure read and write access.Additionally, because you created the
workspace, you are automatically the admin—meeting the requirement that you are the only admin of
Workspace1.

Question: 155
You have a deployment pipeline named Pipeline1 that deploys items from Workspace_DEV to Workspace_TEST. In
Pipeline1, all items that have matching names are paired.

You deploy the contents of Workspace_DEV to Workspace_TEST by using Pipeline1.

What will the contents of Workspace_TEST be once the deployment is complete?

A.Lakehouse1 -

Lakehouse2 -

Notebook1 -

Notebook2 -

Pipeline1 -
SemanticModel1
B.Lakehouse1 -

Notebook1 -

Pipeline1 -
SemanticModel1
C.Lakehouse2 -

Notebook2 -
SemanticModel1
D.Lakehouse2 -

Notebook2 -

Pipeline1 -
SemanticModel1

Answer: A
Explanation:

Lakehouse1 -

Lakehouse2 -

Notebook1 -

Notebook2 -
When deploying contents from Workspace_DEV to Workspace_TEST using Pipeline1 where matching names
are paired, the resulting contents in Workspace_TEST will include all the items from Workspace_DEV and any
additional items from Workspace_TEST that do not have a matching name.

Question: 156
HOTSPOT
-

You have a Fabric tenant that contains a workspace named Workspace_DEV. Workspace_DEV contains the
semantic models shown in the following table.

Workspace_DEV contains the dataflows shown in the following table.

You create a new workspace named Workspace_TEST.

You create a deployment pipeline named Pipeline1 to move items from Workspace_DEV to Workspace_TEST.

You run Pipeline1.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.


Answer:

Explanation:

DF1 will be deployed to Workspace_TEST :Yes.

This means that DF1 (likely a Dataflow) is included in the deployment.

Data from Model1 will be deployed to Workspace_TEST :No .

This suggests that only the structure or metadata of Model1 is deployed, not the actual data.

The scheduled refresh policy for Model1 will be deployed to Workspace_TEST :No

This indicates that the refresh policy of Model1 is not carried over to the new workspace.

Question: 157
HOTSPOT

DimDate contains the following fields.

Trips contains the following fields.

You need to compare the average miles per trip for statutory holidays versus non-statutory holidays.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Answer:
Question: 158
You have a Fabric tenant.

You are creating a Fabric Data Factory pipeline.

You have a stored procedure that returns the number of active customers and their average sales for the current
month.

You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
available to the downstream activities of the pipeline.

Which type of activity should you add?

A.Get metadata
B.Switch
C.Lookup
D.Append variable

Answer: C

Explanation:

C. Lookup.

The Lookup activity in Azure Data Factory (ADF) or Azure Synapse


Analytics allows you to execute a stored procedure and retrieve the
returned values. The retrieved values can then be made available to
Question: 159
HOTSPOT
-

You have the following T-SQL statement.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Answer:

Explanation:

The statement returns Region values when a Sales item has a RefundStatus of Refunded: No
This means that the query does not filter Region values based on RefundStatus.

The statement only returns TransactionDate values that occurred during the current year: Yes

Question: 160
HOTSPOT
-

You have a Fabric warehouse that contains the following data.

The data has the following characteristics:

• Each customer is assigned a unique CustomerID value.


• Each customer is associated to a single SalesRegion value.
• Each customer is associated to a single CustomerAddress value.
• The Customer table contains 5 million rows.
• All foreign key values are non-null.

You need to create a view to denormalize the data into a customer dimension that contains one row per distinct
CustomerID value. The solution must minimize query processing time and resources.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.


Answer:
Explanation:

inner join is used to filter only matching records across the tables.

A.AddressID = CA.AddressID) ensures address details are correctly retrieved.

Question: 161

A.DateTime.LocalNow
B.Duration.Days
C.Duration.From
D.Date.AddDays

Answer: B

Explanation:

B. Duration.Days.

This function will calculate the duration (difference in days) between two date values.
Duration.Days([Shipping_Date] - [Order_Date])

Question: 162
DRAG DROP
-

You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 uses the Pro license mode
and contains a semantic model named Model1.

You have an Azure DevOps organization.

You need to enable version control for Workspace1. The solution must ensure that Model1 is added to the
repository.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of
actions to the answer area and arrange them in the correct order.

Answer:

Explanation:

Connect Workspace1 to a Git provider.


Sync Workspace1 with the repository .

Configure branch policies in Azure DevOps .

After connecting and syncing, configuring branch policies helps maintain quality by enforcing rules, such as
requiring code reviews before merging changes.

Question: 163
HOTSPOT
-

You have a Fabric tenant that contains a workspace named Enterprise. Enterprise contains a semantic model
named Model1. Model1 contains a date parameter named Date1 that was created in Power Query.

You build a deployment pipeline named Enterprise Data that includes two stages named Development and Test.
You assign the Enterprise workspace to the Development stage.

You need to perform the following actions:

• Create a workspace named Enterprise [Test] and assign the workspace to the Test stage.
• Configure a rule that will modify the value of Date1 when changes are deployed to the Test stage.

Which two settings should you use? To answer, select the appropriate settings in the answer area.

NOTE: Each correct answer is worth one point.

Answer:
Question: 164
HOTSPOT
-

You have two Microsoft Power BI queries named Employee and Retired Roles.

You need to merge the Employee query with the Retired Roles query. The solution must ensure that duplicate rows
in each query are removed.

Which column and Join Kind should you use in Power Query Editor? To answer, select the appropriate options in the
answer area.

NOTE: Each correct answer is worth one point.


Answer:
Explanation:

The Employee table has multiple entries in the Division column with the value "Services".

The Retired Roles table also contains multiple entries under the Division column with "Services".

The Join Kind selection options are visible, and the user has highlighted "Full Outer (all rows from both)" and
"Inner (only matching rows)".

Inner Join (only matching rows)

Keeps only the rows that have a match in both tables based on the Division column.

Since "Services" exists in both tables, only those matching rows will be included in the final result.

Only for Renewal Exam [Questions of DP-600 exam]:


Question 1:

You have a Fabric tenant.

All users are currently permitted to create Fabric items.

You plan to block specific users from creating Fabric items.

What should you do first?

Select only one answer.

A. From the Fabric admin portal, configure Capacity settings.

B. From the Fabric admin portal, configure delegation.

C. From the Microsoft 365 admin center, create a security group.

D. From the Microsoft 365 admin center, modify the role assignment of each user.

Answer: C

"From the Microsoft 365 admin center, create a security group."

Explanation:

To block specific users from creating Fabric items, you first need to group those users together. The best way
to do this is by creating a security group in the Microsoft 365 admin center.

Once the security group is created, you can go to the Fabric admin portal and modify the creation
permissions by restricting Fabric item creation to specific users or excluding certain groups.

Why not the other options?

"From the Fabric admin portal, configure Capacity settings."


Capacity settings control resource usage and performance, not user permissions.
"From the Fabric admin portal, configure delegation."
Delegation settings determine who can manage administrative tasks, not item creation permissions.
"From the Microsoft 365 admin center, modify the role assignment of each user."
Role assignments in Microsoft 365 manage admin roles but do not directly control Fabric item creation
permissions.

Next Steps After Creating the Security Group:

1. Go to the Fabric admin portal → Tenant settings.


2. Find the setting for Fabric item creation.
3. Restrict creation to specific security groups, allowing or blocking users as needed.

Questions 2

You have a Fabric tenant that contains a workspace named Workspace1.

You need to allow many users to view multiple reports and dashboards developed in Workspace1. The solution
must follow the principle of least privilege and minimize administrative effort.

What should you use?


Select only one answer.

A. a Microsoft Power BI app

B. item-level permissions

C. the Member workspace role

D. the Viewer workspace role

The correct answer is: A

"a Microsoft Power BI app"

Explanation:

A Microsoft Power BI app is the best choice because it allows many users to access reports and dashboards
without needing direct workspace access.
Apps simplify distribution by providing a read-only experience, following the principle of least privilege.
It also minimizes administrative effort since you can share the app with many users at once rather than
managing individual permissions.

Why not the other options?

"item-level permissions"
Managing permissions at the item level (report or dashboard) for many users is time-consuming and complex.
"the Member workspace role"
The Member role grants more privileges than necessary, including the ability to edit content. This violates
the least privilege principle.
"the Viewer workspace role"
The Viewer role allows read-only access, but users must be assigned directly to the workspace. Managing
permissions for a large number of users this way is inefficient.

Question 3:

You have a Fabric tenant.

You need to control which users can create Fabric items in all the workspaces assigned to Fabric capacities.

What should you configure in the Fabric admin portal?

Select only one answer.

A. Capacity settings

B. Item-level settings

C. Lakehouse settings

D. Tenant settings

E. Workspace settings

The correct answer is: D

"Tenant settings"
Explanation:

Tenant settings in the Fabric admin portal allow you to control who can create Fabric items across all
workspaces assigned to Fabric capacities.
This is the correct choice because item creation is a global setting that applies across the tenant, rather than
at a workspace or capacity level.

Why not the other options?

"Capacity settings"
Capacity settings manage resource allocation and performance, not user permissions for item creation.
"Item-level settings"
These settings control permissions on individual Fabric items (like reports or dashboards), not overall
creation rights.
"Lakehouse settings"
These are specific to managing Lakehouses, not general Fabric item creation.
"Workspace settings"
Workspace settings control permissions within a specific workspace but do not govern item creation tenant-
wide.

Question 4:

You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity. Workspace 1 contains the following items:

· Report1

· SemanticModel1

· Lakehouse1

· Dashboard1

Which items can you endorse?

Select only one answer.

A. Dashboard1 only

B. SemanticModel1 only

C. Lakehouse1 only

D. Report 1, SemanticModel1, and Lakehouse1

E. Report1 and SemanticModel1 only

F. Report1 only

The correct answer is: E

"Report1 and SemanticModel1 only"

Explanation:
In Microsoft Fabric, you can endorse certain items to highlight their quality and reliability. The two types of
endorsements are Promoted and Certified.

Endorsable items in Fabric include:


Reports
Semantic models (formerly datasets)
Items that cannot be endorsed:
Lakehouses (No endorsement feature)
Dashboards (No endorsement feature)

Why not the other options?

"Dashboard1 only" → Dashboards cannot be endorsed.


"SemanticModel1 only" → Reports can also be endorsed.
"Lakehouse1 only" → Lakehouses do not support endorsement.
"Report1, SemanticModel1, and Lakehouse1" → Lakehouses cannot be endorsed.
"Report1 only" → Semantic models can also be endorsed.

Question 5:

You have a Fabric tenant.

A user creates and shares a report that has default permissions to all the users in an organization.

Which actions can all the users perform with the shared report?

Select only one answer.

A. view and modify only

B. view and share only

C. view, modify, and share

D. view only

The correct answer is: B

"view and share only"

Explanation:

When a report is shared in Microsoft Fabric (Power BI) with default permissions to all users in the
organization:

1. Users can view the report


2. Users can reshare the report (if reshare permissions are enabled)
3. Users cannot modify the report (only owners and those with edit permissions can modify it)

Why not the other options?

"view and modify only" → Users cannot modify the report unless they are explicitly given edit
permissions.
"view, modify, and share" → Users cannot modify the report by default.
"view only" → By default, users can share the report with others unless sharing is restricted.
Question 6:

You have a Fabric tenant that contains a lakehouse named Lakehouse1.

You have an on-premises Microsoft SQL server that contains a database named DB1.

You plan to extract data from DB1 by using Data Factory and store the data in Lakehouse1.

You need to set up a connection to DB1.

What should you use?

Select only one answer.

A. a personal gateway

B. a shortcut

C. an on-premises gateway

D. mirroring

The correct answer is: C

"an on-premises gateway"

Explanation:

Since DB1 is an on-premises SQL Server database, you need a secure way to connect Fabric’s Data Factory
to it. The best solution is an on-premises data gateway, which allows Fabric to securely access on-premises
data sources.

Why not the other options?

"a personal gateway"


A personal mode gateway is only for individual use and does not support Data Factory.
"a shortcut"
Shortcuts are used to create virtual links to external storage (like OneLake, ADLS, or AWS S3), not to
extract data from on-prem SQL Server.
"mirroring"
Mirroring is used for real-time replication of cloud databases (like Azure SQL), not for on-premises SQL
Server.

Question 7:

You have a Fabric tenant that contains a lakehouse named Lakehouse1.

You plan to use the Delete data activity to remove the existing data in Lakehouse1.

What should you create first?

Select only one answer.

A. a dataflow
B. a lookup

C. a pipeline

D. a shortcut

The correct answer is: C

"a pipeline"

Explanation:

In Microsoft Fabric, the Delete data activity is part of Data Factory and is used in pipelines to remove data
from a Lakehouse or other storage.

To use the Delete data activity, you must first create a pipeline and then configure the activity within it.

Why not the other options?

"a dataflow"
Dataflows are used for data transformation and ingestion, not for deleting data.
"a lookup"
Lookups retrieve data for use in other pipeline activities but do not delete data.
"a shortcut"
Shortcuts provide virtual access to external storage but do not support data deletion.

Question 8:

You have a Fabric tenant.

You plan to create a pipeline that will download a comma-separated dataset from a public GitHub repository.

You need to specify which request method to use when defining the connection to the data source.

Which request method should you specify?

Select only one answer.

CONNECT

GET

OPTIONS

PUT

The correct answer is: B

"GET"

Explanation:

The GET request method is used to retrieve data from a web source, such as downloading a CSV file from a
public GitHub repository.
Since you are downloading data (not modifying or sending data), GET is the appropriate choice.

Why not the other options?


"CONNECT"
Used for establishing a network tunnel, typically for proxy connections—not for retrieving data.
"OPTIONS"
Used to retrieve allowed methods from a server but does not fetch data.
"PUT"
Used to upload or update data on a server, which is not needed here.

Question 9:

You have a Fabric tenant that contains a warehouse named Warehouse1.

You plan to use a Data Factory pipeline to ingest data from multiple supported data sources, transform the
data, and then merge the data before saving the results to Warehouse1.

Which pipeline activity should you use?

Select only one answer.

A. Copy data activity

B. Dataflow activity

C. Notebook activity

D. Spark job definition activity

E. Stored procedure activity

The correct answer is: B

"Dataflow activity"

Explanation:

The Dataflow activity in Data Factory is designed for ingesting, transforming, and merging data from
multiple sources before loading it into Warehouse1.
It provides a visual, no-code interface to apply transformations like joins, aggregations, and mappings
before writing to the warehouse.

Why not the other options?

"Copy data activity"


This is mainly used for simple data movement (ETL), but it does not support complex transformations or
merging.
"Notebook activity"
Executes custom Python, SQL, or Spark notebooks, but it's not the best choice for a structured
transformation pipeline.
"Spark job definition activity"
Used for running big data processing tasks in Spark, but overkill for structured transformations needed
here.
"Stored procedure activity"
Used to execute SQL stored procedures inside Warehouse1, but does not handle ingestion and
transformation in Data Factory.
Question 10:

You have a Fabric tenant that contains a lakehouse named Lakehouse1.

You plan to use a Data Factory pipeline to create a repeatable process to ingest Parquet files stored on a
network drive and load them to Lakehouse1 as a Delta table without any transformations.

Which pipeline activity should you use?

Select only one answer.

A. Copy data activity

B. Get Metadata activity

C. Lookup activity

D. Notebook activity

E. Stored procedure activity

The correct answer is: A

"Copy data activity"

Explanation:

The Copy data activity in Data Factory is the best choice when you need to ingest Parquet files from a
network drive and load them into Lakehouse1 as a Delta table, without transformations.
This activity is designed for efficient ETL (Extract, Load) operations, supporting direct movement of data
from various sources to destinations like Lakehouses, Warehouses, and Delta tables.

Why not the other options?

"Get Metadata activity"


This retrieves metadata about files or tables but does not perform ingestion.
"Lookup activity"
This is used for querying a dataset (e.g., fetching a single value), not for data ingestion.
"Notebook activity"
This runs custom code in a notebook, which is unnecessary for a simple file transfer.
"Stored procedure activity"
This is used to execute SQL stored procedures, not for moving files from a network drive to a Lakehouse.

Question 11:

You have a Fabric dataflow named Dataflow1 that contains a column named LastName with the Text data
type.

You notice that the LastName column contains several NULL entries and several blank entries.

You need to ensure that Dataflow1 contains only rows where the LastName column contains text.

What should you use in Power Query?

Select only one answer.

A. Clear filter command


B. Remove empty command

C. Remove errors transform

D. Replace values transform

The correct answer is: B

"Remove empty command"

Explanation:

In Power Query, the Remove empty command is used to filter out rows where a column contains NULL or
blank values.
Since your LastName column has both NULL and blank entries, using Remove empty ensures that only rows
with actual text values remain.

Why not the other options?

"Clear filter command"


Clears existing filters but does not specifically remove NULL or blank values.
"Remove errors transform"
Used to remove rows containing errors, but NULL and blanks are not considered errors.
"Replace values transform"
Used to replace specific values but does not remove rows with blanks or NULLs.

Question 12:

You have a Fabric tenant.

You create a Dataflow Gen2 activity that contains several aggregation transformations.

You need to view the M code that represents the transformations by using the Power Query Online interface.

Which element of Power Query Online should you use?

Select only one answer.

A. Advanced editor window

B. Diagram View

C. Queries pane

D. Query Settings pane

The correct answer is: A

"Advanced editor window"

Explanation:

The Advanced editor window in Power Query Online allows you to view and edit the M code that represents
the transformations in your Dataflow Gen2 activity.
Since you have aggregation transformations, the M code captures these steps in a structured format.

Why not the other options?


"Diagram View"
Provides a visual representation of query steps but does not show M code.
"Queries pane"
Lists all queries in the dataflow but does not display M code.
"Query Settings pane"
Shows applied steps for a query but does not allow direct M code editing.

Question 13:

You have a Fabric tenant that contains a lakehouse named Lakehouse1.

You plan to use Dataflow Gen2 to merge several tables into a single table, and then save the table to
Lakehouse1 as a Delta table.

You need to visually check all the transformations applied.

What should you use in Dataflow Gen2?

Select only one answer.

A. Data Preview pane

B. Diagram View

C. Queries pane

D. Query Settings pane

The correct answer is: B

"Diagram View"

Explanation:

Diagram View in Dataflow Gen2 provides a visual representation of all transformations applied to your data.
Since you are merging several tables and performing transformations before saving the final table as a Delta
table in Lakehouse1, Diagram View allows you to see the dataflow structure and transformations step by
step.

Why not the other options?

"Data Preview pane"


Shows sample data but does not provide a visual overview of transformations.
"Queries pane"
Lists available queries but does not show transformation steps visually.
"Query Settings pane"
Displays applied steps for an individual query but does not provide a full workflow visualization.

Question 14:

You have a Fabric tenant.


You have a single table in Dataflow Gen2 that contains the following columns:

· Employee

· Region

· Amount Sold

Each employee has multiple regions and an associated value in the Amount Sold column.

You need to transform the table to meet the following requirements:

· Each row must have a single record per employee.

· Columns must represent the Region with aggregate Amount Sold for each Employee per Region.

Which Power Query function should you use?

Select only one answer.

A. Group by

B. Pivot column

C. Split column

D. Transpose

The correct answer is: B

"Pivot column"

Explanation:

The Pivot column function in Power Query converts row values into column headers, which is exactly what
you need:
Each row should represent a single Employee.
Each Region should become a column.
The values in the new columns should be the aggregated Amount Sold per Region.

Why not the other options?

"Group by"
Groups data but does not transform row values into columns. It would summarize data but not create
separate columns for each region.
"Split column"
Used to divide a column into multiple columns, typically by delimiter or position, but not for aggregation.
"Transpose"
Swaps rows and columns entirely, which is not suitable for this scenario.

Question 15:

You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity and contains a KQL database named KQL1. KQL1 contains a table named Sales that has the following
columns:
· SalesAmount: decimal data type.

· Region: string data type.

You plan to calculate the total SalesAmount by Region.

Which KQL statement should you use?

Select only one answer.

A. Sales | extend TotalSales= sum(SalesAmount) | extend Region

B. Sales | extend TotalSales= sum(SalesAmount) by Region

C. Sales | summarize by TotalSales= sum(SalesAmount), Region

D. Sales | summarize TotalSales= sum(SalesAmount) by Region

The correct answer is: D

Sales | summarize TotalSales = sum(SalesAmount) by Region

Explanation:

In Kusto Query Language (KQL), the summarize operator is used to aggregate data.
The correct syntax for calculating total SalesAmount grouped by Region is:

kql

CopyEdit

Sales

| summarize TotalSales = sum(SalesAmount) by Region

This statement:
Groups the data by Region
Calculates the total SalesAmount for each Region
Names the result as TotalSales

Why not the other options?

Sales | extend TotalSales= sum(SalesAmount) | extend Region


extend is used to add calculated columns, but sum() must be used within summarize, not extend.
Sales | extend TotalSales= sum(SalesAmount) by Region
extend does not support grouping with by, so this syntax is incorrect.
Sales | summarize by TotalSales= sum(SalesAmount), Region
Incorrect syntax: by should come after the aggregation.

Question 16:

You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity and contains a KQL database named KQL1. KQL1 contains a table named Sales.

You need to display the following columns from Sales:


· OrderKey

· OrderId

· OrderDate

Which KQL statement should you use?

Select only one answer.

A. Sales | extend OrderKey, OrderId, OrderDate

B. Sales | project OrderKey, OrderId, OrderDate

C. Sales | summarize OrderKey, OrderId, OrderDate

D. Sales | where OrderKey, OrderId, OrderDate

The correct answer is: B

Sales | project OrderKey, OrderId, OrderDate

Explanation:

The project operator in Kusto Query Language (KQL) is used to select specific columns from a table.
The correct syntax to display only OrderKey, OrderId, and OrderDate is:

kql

CopyEdit

Sales

| project OrderKey, OrderId, OrderDate

This ensures that only the specified columns are included in the output.

Why not the other options?

Sales | extend OrderKey, OrderId, OrderDate


extend is used to create new calculated columns, not to select existing ones.
Sales | summarize OrderKey, OrderId, OrderDate
summarize is used for aggregations, but this query does not require aggregation.
Sales | where OrderKey, OrderId, OrderDate
where is used for filtering rows based on conditions, but no conditions are specified.

Question 17:

You have a Fabric workspace named Workspace1 that contains a lakehouse named Lakehouse1.

For Lakehouse1, you plan to create a semantic model named Model1.

For Model1, you need to create a table named Table1 that will support incremental refresh and will be
partitioned to meet the following requirements:

· Multiple import partitions will store older data.


· A single DirectQuery partition will be used for real-time data.

The solution must minimize costs and maximize performance.

What should you do first?

Select only one answer.

A. Add a calculated column to Table1.

B. Assign the Power BI Premium capacity to Workspace1.

C. Assign the Power BI Pro capacity to Workspace1.

D. Disable incremental refresh for Table1.

The correct answer is: B

"Assign the Power BI Premium capacity to Workspace1."

Explanation:

Incremental refresh and hybrid tables (which combine import mode for historical data and DirectQuery for
real-time data) require Power BI Premium capacity.
Power BI Pro does NOT support incremental refresh with DirectQuery partitions, so assigning Premium
capacity is necessary to enable this functionality.

Why not the other options?

"Add a calculated column to Table1."


Not relevant to enabling incremental refresh or hybrid tables.
"Assign the Power BI Pro capacity to Workspace1."
Pro does NOT support incremental refresh with DirectQuery partitions; Premium is required.
"Disable incremental refresh for Table1."
This would prevent partitioning, which is the opposite of what you need.

Question 18:

You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity and contains a lakehouse named Lakehouse1.

You plan to use the SQL analytics endpoint of Lakehouse1 to develop a semantic model. The solution must
meet the following requirements:

· Refresh the model at 6 AM every day.

· Support all DAX and M functions.

· Support calculated tables.

· Ensure that more than one source group is NOT comprised.

· Use only one storage mode.

Which storage mode should you use?


Select only one answer.

A. Direct Lake

B. DirectQuery

C. Dual

D. Import

The correct answer is Import mode because it meets all the specified requirements.

1. Supports scheduled refresh at 6 AM – Import mode loads data into memory and relies on scheduled
refreshes to update the model.
2. Works with all DAX and M functions – Unlike Direct Lake and DirectQuery, Import mode fully
supports advanced DAX calculations and M functions.
3. Allows calculated tables – Calculated tables are only available in Import mode, making it the only
viable choice.
4. Prevents multiple source group issues – Import mode loads data from a single source, ensuring
consistency.
5. Uses only one storage mode – Import mode is a single-mode storage option, fulfilling the
requirement.

Direct Lake and DirectQuery do not support calculated tables or full DAX/M functionality. Dual mode allows
switching between Import and DirectQuery, which does not align with the "one storage mode" requirement.

Thus, Import mode is the best choice.

Question 19:

You have a Fabric tenant that contains a lakehouse named Lakehouse1.

You have an Azure SQL database named SQLDB1.

SQLDB1 stores product inventory that is updated in real time.

You plan to develop a custom semantic model using Lakehouse1 that will provide up-to-date inventory levels
by querying live transactional data from SQLDB1.

You need to ensure that all the queries that use the model target live transactional data.

Which storage mode should you use?

Select only one answer.

Direct Lake

DirectQuery

Dual

Import

The correct answer is: B

DirectQuery

Explanation:
Since you need to provide up-to-date inventory levels by querying live transactional data from SQLDB1, the
best storage mode is DirectQuery.

1. DirectQuery ensures real-time data access – Queries retrieve data directly from SQLDB1 instead of
relying on a cached dataset.
2. Best for live transactional data – Since inventory data updates in real time, DirectQuery ensures that
every query reflects the latest data.
3. Avoids scheduled refresh limitations – Unlike Import mode, which requires periodic refreshes,
DirectQuery always fetches data from the source.

Why not the other options?

Direct Lake – Works only with OneLake and Delta tables, but SQLDB1 is an Azure SQL database, not a
lakehouse.
Import – Stores data in memory and requires scheduled refreshes, meaning the data would be stale
between refreshes.
Dual – Allows both Import and DirectQuery but is used for optimized performance, not for ensuring 100%
real-time queries.

Question 20:

You have a Fabric tenant that contains a lakehouse named Lakehouse1.

You create several tables in Lakehouse1. You load data into the tables.

You need to define relationships for the default semantic model in Lakehouse1.

What should you do first?

Select only one answer.

A. From the Fabric lakehouse view, connect to SQL analytics endpoint.

B. From the Fabric lakehouse view, disable the Sync the default Power BI semantic model option.

C. From the Microsoft Fabric portal, switch to the Model view of the Data Warehouse experience.

D. From the Microsoft Fabric portal, switch to the Query view of the Data Warehouse experience.

The correct answer is: A

"From the Fabric lakehouse view, connect to SQL analytics endpoint."

Explanation:

To define relationships in the default semantic model of Lakehouse1, you need to use Power BI's model view,
which requires connecting to the SQL analytics endpoint.

1. SQL analytics endpoint enables querying and modeling – It provides a SQL-compatible interface to
work with Lakehouse data, allowing you to define relationships.
2. Power BI automatically syncs with the semantic model – Once connected, you can define
relationships in Power BI Model view.

Why not the other options?

"Disable the Sync the default Power BI semantic model option" – This would prevent Power BI from
automatically maintaining the semantic model, which is the opposite of what you need.
"Switch to the Model view of the Data Warehouse experience" – Lakehouse1 is a lakehouse, not a Data
Warehouse, so this option does not apply.
"Switch to the Query view of the Data Warehouse experience" – Query view is for writing SQL queries,
not for defining relationships in a semantic model.

Question 21:

You use Microsoft Power BI Desktop to generate and display visuals within a report.

You use Performance analyzer to record the performance of your activities.

You plan to verify your original results by rerecording the performance by using the same visuals.

You need to ensure that the results are comparable.

What should you do first in Power BI Desktop?

Select only one answer.

A. Close the report.

B. Customize optimization presets.

C. Pause the visuals.

D. Refresh the visuals.

The correct answer is: D

"Refresh the visuals."

Explanation:

When using Performance Analyzer in Power BI Desktop, you need to ensure that the second recording is
measuring the same operations as the first one.

1. Refreshing the visuals ensures a new performance measurement – Power BI does not re-execute
queries unless the visuals are refreshed.
2. Ensures comparable results – If you don’t refresh, the data might be cached, making the second
measurement faster and not a true comparison.
3. Prepares the visuals for a fresh performance capture – Refreshing forces Power BI to reprocess
queries and render visuals from scratch, ensuring consistency.

Why not the other options?

"Close the report." – This isn’t necessary and could lose context from the first test.
"Customize optimization presets." – While useful for tuning performance, it does not ensure that the
second recording is identical to the first.
"Pause the visuals." – This prevents visuals from updating, which is the opposite of what you need.

Question 22:
You use Microsoft Power BI Desktop to generate and display DAX query-based visuals in a report.

From Performance analyzer, you review the statistics of different DAX queries.

You need to determine the smallest threshold above which the DAX queries execution time warrants further
investigation.

What execution time threshold should you use?

Select only one answer.

A. 1.2 ms

B. 12 ms

C. 120 ms

D. 1200 ms

The correct answer is: C

"120 ms"

Explanation:

When analyzing DAX query performance in Power BI using Performance Analyzer, a good rule of thumb is:

1. Queries executing in under 100 ms are generally not a concern.


2. Queries above 120 ms warrant further investigation because:

They may indicate inefficient DAX formulas.


They could be affected by poorly structured data models or relationships.
They might benefit from optimizations like indexing, aggregations, or better filtering.

Why not the other options?

1.2 ms – Extremely low, even simple operations take more time than this.
12 ms – Insignificant; even well-optimized queries can take longer.
1200 ms (1.2 sec) – Too high; waiting until this threshold could result in severe performance issues before
taking action.

Question 23:

You use Microsoft Power BI Desktop to generate and display reports.

You plan to identify opportunities for performance improvements by using Performance analyzer.

Which two features in Performance analyzer can help you optimize performance? Each correct answer
presents part of the solution.

Select all answers that apply.

A. data model

B. data source

C. environment
D. report visuals

The correct answers are: AD

Data model

Report visuals

Explanation:

Performance Analyzer in Power BI helps identify bottlenecks and optimization opportunities in reports. The
two key areas it helps with are:

1. Data model – Performance Analyzer can highlight:

Slow DAX queries due to inefficient measures or relationships.


Large or unoptimized tables that impact performance.
Unnecessary calculated columns that could be optimized with measures.

1. Report visuals – Performance Analyzer breaks down:

Rendering time of visuals.


Query execution time to identify slow-performing visuals.
Visual interactions that slow down report responsiveness.

Why not the other options?

Data source – While an inefficient data source can impact performance, Performance Analyzer does not
directly analyze the data source itself. It focuses on queries, visuals, and model performance.
Environment – Performance Analyzer does not assess system performance, hardware, or network
conditions.

Question 24:

You use Microsoft Power BI Desktop to generate and display DAX query-based visuals.

You are analyzing the performance of the following measure:

Sales YoY Growth =

DIVIDE (

( [Sales] - CALCULATE ( [Sales], PARALLELPERIOD ( 'Date'[Date], -12, MONTH ) ) ),

CALCULATE ( [Sales], PARALLELPERIOD ( 'Date'[Date], -12, MONTH ) )

You need to improve the performance of the measure.

What should you include in the measure definition?


Select only one answer.

A. DATESINPERIOD

B. nested calculations

C. SUMX

D. variables

The correct answer is: D

Variables

Explanation:

Using variables (VAR) in DAX improves performance by:

1. Avoiding redundant calculations – The measure currently repeats CALCULATE([Sales],


PARALLELPERIOD('Date'[Date], -12, MONTH)) twice, which results in unnecessary recalculations.
2. Improving readability and maintainability – Storing values in variables makes DAX easier to
understand and debug.
3. Enhancing performance – Power BI evaluates variables once and reuses the result, reducing query
execution time.

Optimized Measure Using Variables:

Sales YoY Growth =

VAR PreviousSales = CALCULATE ( [Sales], PARALLELPERIOD ( 'Date'[Date], -12, MONTH ) )

RETURN

DIVIDE ( [Sales] - PreviousSales, PreviousSales )

This approach ensures that PreviousSales is calculated once instead of twice.

Why not the other options?

DATESINPERIOD – This function works similarly to PARALLELPERIOD, but it does not inherently improve
performance.
Nested calculations – Nesting calculations increases complexity and can make measures slower, rather
than optimizing them.
SUMX – SUMX is useful for row-by-row calculations but is not needed in this case because
PARALLELPERIOD returns an aggregated value.

Question 25:

You use Microsoft Power BI Desktop to generate and display DAX query-based visuals.

You plan to use Power BI Desktop Performance analyzer to determine how long it takes to load a visual
included in a report.

You need to eliminate any impact of the visual cache on your performance analysis.

What should you do first?


Select only one answer.

A. Add a blank page to the report.

B. Change the filter on the visual.

C. Refresh the data model.

D. Reopen the Performance Analyzer pane

The correct answer is: B

Change the filter on the visual.

Explanation:

When using Performance Analyzer in Power BI Desktop, cached results can distort performance
measurements. To ensure that each query runs fresh, you need to force Power BI to re-execute the DAX
query rather than retrieving results from cache.

Changing the filter on the visual forces Power BI to:

Recompute the DAX query.


Fetch new data instead of using cached results.
Provide an accurate measure of how long the query actually takes to execute.

Why not the other options?

Add a blank page to the report – This does not refresh or impact the query execution of the original visual.
Refresh the data model – Refreshing the data model reloads all data from the source, which is
unnecessary when only testing visual performance.
Reopen the Performance Analyzer pane – Closing and reopening the pane does not reset cached results
or force query execution.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy