DP-600
DP-600
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.
Requirements -
Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
You need to ensure that Contoso can use version control to meet the data analytics requirements and the general
requirements.
What should you do?
A.Store at the semantic models and reports in Data Lake Gen2 storage.
B.Modify the settings of the Research workspaces to use a GitHub repository.
C.Modify the settings of the Research division workspaces to use an Azure Repos repository.
D.Store all the semantic models and reports in Microsoft OneDrive.
Answer: C
Explanation:
C. Modify the settings of the Research division workspaces to use an Azure Repos repository.
Contoso needs version control for its data analytics and general requirements. In Microsoft Fabric and Power
BI, version control is best managed through Git integration using Azure Repos or GitHub.
Azure Repos is a Git-based repository within Azure DevOps, providing versioning, tracking, and collaboration
for code, reports, and semantic models.
This is the most appropriate choice for version-controlling Power BI artifacts like datasets, reports, and
semantic models in a workspace.
Reference:
https://learn.microsoft.com/en-us/fabric/cicd/git-integration/intro-to-git-integration
Question: 2
HOTSPOT -
Case study -
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.
Requirements -
Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
You need to recommend a solution to group the Research division workspaces.
What should you include in the recommendation? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
Domain: Group the Research division workspaces based on their departmental context.
Domain allows you to group workspaces based on their purpose or business context. In this case, grouping by
department (Research division) aligns with the requirement.
The Fabric Admin Portal is the primary management interface for Microsoft Fabric.
Reference:
https://learn.microsoft.com/en-us/fabric/governance/domains#configure-domain-settings
Question: 3
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.
Requirements -
Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
You need to refresh the Orders table of the Online Sales department. The solution must meet the semantic model
requirements.
What should you include in the solution?
A.an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the maximum value of
the OrderID column in the destination lakehouse
B.an Azure Data Factory pipeline that executes a Stored procedure activity to retrieve the minimum value of the
OrderID column in the destination lakehouse
C.an Azure Data Factory pipeline that executes a dataflow to retrieve the minimum value of the OrderID column
in the destination lakehouse
D.an Azure Data Factory pipeline that executes a dataflow to retrieve the maximum value of the OrderID
column in the destination lakehouse
Answer: D
Explanation:
D. an Azure Data Factory pipeline that executes a dataflow to retrieve the maximum value of the OrderID
column in the destination lakehouse. A dataflow can be used to retrieve the max OrderID number (stored in
the destination table - OrderID is a sequencial number). This number can be used to set from which row data
must be added to the destination table (implementing an incremental load).
Question: 4
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Contoso has the following data environment:
The Sales division uses a Microsoft Power BI Premium capacity.
The semantic model of the Online Sales department includes a fact table named Orders that uses Import made. In
the system of origin, the OrderID value represents the sequence in which orders are created.
The Research department uses an on-premises, third-party data warehousing product.
Fabric is enabled for contoso.com.
An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division data for a product
line named Productline1. The data is in the delta format.
A Data Lake Storage Gen2 storage account named storage2 contains Research division data for a product line
named Productline2. The data is in the CSV format.
Requirements -
Planned Changes -
Contoso plans to make the following changes:
Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
Make all the data for the Sales division and the Research division available in Fabric.
For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
In Productline1ws, create a lakehouse named Lakehouse1.
In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Follow the principle of least privilege when applicable.
Minimize implementation and maintenance effort when possible.
Which syntax should you use in a notebook to access the Research division data for Productline1?
A.spark.read.format(“delta”).load(“Tables/productline1/ResearchProduct”)
B.spark.sql(“SELECT * FROM Lakehouse1.ResearchProduct ”)
C.external_table(‘Tables/ResearchProduct)
D.external_table(ResearchProduct)
Answer: B
Explanation:
https://learn.microsoft.com/en-us/fabric/onelake/onelake-shortcuts
Question: 5
HOTSPOT -
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
The customer satisfaction data contains the following tables:
Survey -
Question -
Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)
Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
You need to assign permissions for the data store in the AnalyticsPOC workspace. The solution must meet the
security requirements.
Which additional permissions should you assign when you share the data store? To answer, select the appropriate
options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
Data Engineers: Read all Apache Spark - because they need to be able to work with Spark for Data curation.
Data Analysts: Build Reports on the default dataset - because they are report builders
Data Scientists: Read All SQL analytics Endpoints data_ They leverage curated data (by engineers) to do
predictive analytics.
Question: 6
HOTSPOT -
Case study -
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
Survey -
Question -
Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)
A custom semantic model -
Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
You need to create a DAX measure to calculate the average overall satisfaction score.
How should you complete the DAX code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
1. Average(Survey[Response Value]) as this one is the only value that can be aggregated.
Customer Key Does not make any sense , as it just a key to identify a customer, this can never provide us the
score.
2. Period : Variable is defined as to select 1 year of date range. Can directly be passed in the Filter context of
Calculate formula.
Question: 7
HOTSPOT -
Case study -
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
Survey -
Question -
Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)
A custom semantic model -
Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
You need to resolve the issue with the pricing group classification.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
* WHEN ListPrice BETWEEN 50 AND 1000 THEN 'medium': the other two options miss value 1000; on the other
hand, the BETWEEN includes both 50 and 1000.
Question: 8
Case study -
to this section.
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
Survey -
Question -
Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)
Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
What should you recommend using to ingest the customer data into the data store in the AnalyticsPOC
workspace?
Answer: D
Explanation:
D. a dataflow.
Even though the text reads "Data will be loaded without transformation in one area of the AnalyticsPOC data
store": in general, dataflows are used when data transformations are involved after ingestion. As suggested
by user BHARAT, the Copy Activity should be the optimal solution.
Question: 9
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
Survey -
Question -
Response -
For each survey submitted, the following occurs:
One row is added to the Survey table.
One row is added to the Response table for each question in the survey.
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
The following three workspaces will be created:
AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
A data store (type to be decided)
Interactive reports -
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
The data store must support the following:
Read access by using T-SQL or Python
Semi-structured and unstructured data
Row-level security (RLS) for users executing T-SQL queries
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
List prices that are less than or equal to 50 are in the low pricing group.
List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Fabric administrators will be the workspace administrators.
The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
The date dimension must be available to all users of the data store.
The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
FabricAdmins: Fabric administrators
AnalyticsTeam: All the members of the analytics team
DataAnalysts: The data analysts on the analytics team
DataScientists: The data scientists on the analytics team
DataEngineers: The data engineers on the analytics team
AnalyticsEngineers: The analytics engineers on the analytics team
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
Shows data as soon as the data is updated in the data store.
Ensures that the report and the semantic model only contain data from the current and previous year.
Ensures that the report respects any table-level security specified in the source data store.
Minimizes the execution time of report queries.
Which type of data store should you recommend in the AnalyticsPOC workspace?
Answer: C
Explanation:
A lakehouse The data store must handle semi-structured and unstructured data, therefore a Lakehouse
should be the optimal solution supporting read access with T-SQL and Python.
Question: 10
You have a Fabric warehouse that contains a table named Staging.Sales. Staging.Sales contains the following
columns.
You need to write a T-SQL query that will return data for the year 2023 that displays ProductID and ProductName
and has a summarized Amount that is higher than 10,000.
Which query should you use?
A.
B.
C.
D.
Answer: A
Explanation:
Answer B is excluded because the filter on year should be included in the WHERE-clause. It is the result of a
function, but it is not an aggregate.
Question: 11
HOTSPOT -
You have a data warehouse that contains a table named Stage.Customers. Stage.Customers contains all the
customer record updates from a customer relationship management (CRM) system. There can be multiple updates
per customer.
You need to write a T-SQL query that will return the customer ID, name. postal code, and the last updated time of
the most recent row for each customer ID.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
Box 1: ROW_NUMBER()
ROW_NUMBER() is a window function that assigns a unique sequential number to each row within a partition,
based on the ORDER BY clause.
In this case, it is used to assign a row number to each CustomerID, ordering by LastUpdated DESC.
This ensures that the most recent record for each CustomerID gets ROW_NUMBER() = 1.
Box 2: WHERE X = 1
Since we only want the most recent row per CustomerID, we filter for X = 1.
This approach ensures we get only the latest entry for each customer, based on LastUpdated.
Question: 12
HOTSPOT -
You have a Fabric tenant.
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
Answer: No
Embedding a Power BI report typically involves using the Power BI Embedded API, which is not indicated here.
Answer: Yes
If the code generates a new Power BI report (e.g., using Python, Power BI REST API, or Power BI Desktop
automation), then this is correct.
Answer: No
Displaying a summary of a DataFrame would involve functions like df.describe(), df.info(), or df.head() in
Pandas. Since this is about Power BI, this function is not relevant.
Question: 13
You are the administrator of a Fabric workspace that contains a lakehouse named Lakehouse1. Lakehouse1
contains the following tables:
Table1: A Delta table created by using a shortcut
Table2: An external table created by using Spark
A.Read Table3.
B.Update the data Table3.
C.Read Table2.
D.Update the data in Table1.
Answer: A
Explanation:
The right answer is A. A managed table, it is stored within the Fabric storage and becomes immediately
accessible through the SQL endpoint upon connection.D is not right becouse tables created using shortcuts
might not be immediately accessible through the SQL endpoint for updates.
Question: 14
You have a Fabric tenant that contains a warehouse.
You use a dataflow to load a new dataset from OneLake to the warehouse.
You need to add a PowerQuery step to identify the maximum values for the numeric columns.
Which function should you include in the step?
A.Table.MaxN
B.Table.Max
C.Table.Range
D.Table.Profile
Answer: D
Explanation:
The Table.Profile function in PowerQuery is specifically designed to provide statistical information about a
table's columns, including the maximum values for numeric columns. When you apply Table.Profile to your
dataset, it generates a summary of each column, which includes the minimum, maximum, average, standard
deviation, count of distinct values, and other useful statistics.
Reference:
https://learn.microsoft.com/en-us/powerquery-m/table-profile
Question: 15
You need to use the model to generate predictions by using the PREDICT function in a Fabric notebook.
NOTE: Each correct answer is worth one point.
A.T-SQL
B.DAX
C.Spark SQL
D.PySpark
Answer: CD
Explanation:
Spark SQL: This language leverages the Spark framework, which is commonly used for big data processing
and machine learning. Spark SQL allows you to run SQL queries over data and includes support for machine
learning model scoring with the PREDICT function.
PySpark: This is the Python API for Spark. PySpark allows you to interact with Spark using Python, which is
one of the most popular programming languages for data analysis and machine learning. PySpark includes
capabilities for integrating with machine learning models and running predictions using the PREDICT function.
T-SQL: This is Transact-SQL, primarily used for managing and querying data in SQL Server and Azure SQL
Database. It doesn't support the PREDICT function for machine learning model scoring in Fabric notebooks.
DAX: This is Data Analysis Expressions, a language used in Power BI, Power Pivot, and Analysis Services for
data modeling and calculations. It doesn't support the PREDICT function for machine learning model scoring
in Fabric notebooks.
Question: 16
A.displayHTML
B.show
Answer: D
Explanation:
Display: This function is specifically designed for rendering visual representations of data within interactive
notebooks. When you use the display function on a Spark DataFrame, it enables the data to be shown in a
Question: 17
You need all rows to appear in the visual.
What should you do?
Answer: C
Explanation:
Question: 18
DRAG DROP -
You have a Fabric tenant that contains a semantic model. The model contains data about retail stores.
You need to write a DAX query that will be executed by using the XMLA endpoint. The query must return a table of
stores that have opened since December 1, 2023.
How should you complete the DAX expression? To answer, drag the appropriate values to the correct targets.
Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or
scroll to view content.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
DEFINE
This section defines a variable (VAR _SalesSince) that stores a specific date (2023-12-01).
EVALUATE
This keyword is required for executing and returning a table result in DAX queries.
SUMMARIZE
Here, it summarizes data from the Store table, including Store[Name] and Store[OpenDate].
Question: 19
What can you identify about the pickupLongitude column?
Answer: A
Explanation:
Answer A
Distinct Value: This refers to all different values present in a dataset. When you retrieve distinct values from a
column, you eliminate duplicate values so that each value is shown once. For example, if a column contains
the values [1, 2, 2, 3, 3, 3], the distinct values would be [1, 2, 3].
Unique Value: This usually refers to values that appear only once in the dataset. Unlike distinct values, a
unique value will only be considered if it has no duplicates at all. For example, if a column contains the values
[1, 2, 2, 3, 3, 3], the unique values would be [1], since only 1 appears without repetition.
Question: 20
You have a Fabric tenant named Tenant1 that contains a workspace named WS1. WS1 uses a capacity named C1
and contains a dataset named DS1.
Answer: C
Explanation:
As XMLA is set to Read-Only first, you must go to the capacity settings to enable read-write.
https://learn.microsoft.com/en-us/power-bi/enterprise/service-premium-connect-tools#enable-xmla-read-
Question: 21
You need to recommend a solution to provide users with the ability to create and publish custom Direct Lake
Answer: ADE
Explanation:
From the Tenant settings, set Allow XMLA Endpoints and Analyze in Excel with on-premises datasets to
Enabled. This allows users to interact with the dataset via XMLA endpoints and analyze data using Excel.
In the Capacity settings, configure the XMLA Endpoint to Read Write. This ensures that users have the
necessary permissions to create and modify semantic models through external tools.
From the Tenant settings, set Users can create Fabric items to Enabled. This grants users the ability to create
custom semantic models within the Fabric workspace.
Question: 22
You are creating a semantic model in Microsoft Power BI Desktop.
You plan to make bulk changes to the model by using the Tabular Model Definition Language (TMDL) extension for
Microsoft Visual Studio Code.
You need to save the semantic model to a file.
Which file format should you use?
A.PBIP
B.PBIX
C.PBIT
D.PBIDS
Answer: A
Explanation:
A. PBIP.
PBIP (Power BI Project) is a file format that supports the open-source TMDL format and is designed for
integrating Power BI with external development environments like Visual Studio Code.
PBIX is the common Power BI report file format but is not intended for bulk edits through TMDL.
PBIT is a Power BI template file, used for creating new reports based on an existing structure but not for bulk
editing in Visual Studio Code.
PBIDS is for creating Power BI dataset connections and is unrelated to TMDL editing.
Question: 23
HOTSPOT -
You have a Fabric tenant that contains a warehouse named Warehouse1. Warehouse1 contains three schemas
named schemaA, schemaB, and schemaC.
You need to ensure that a user named User1 can truncate tables in schemaA only.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
Question: 24
Ensure that the developers can deploy items to the workspaces for Development and Test.
Answer: BDE
Explanation:
Admin access to the deployment pipeline : Admin access to the deployment pipeline is necessary to manage
the pipeline stages and deployment processes. This allows developers to configure deployments between
stages but does not grant direct access to the workspaces themselves.
Viewer access to the Production workspace : Providing Viewer access to the Production workspace ensures
that developers can see the items in the Production stage but cannot make changes or deploy items to the
Production workspace. This restriction aligns with the requirement to prevent developers from deploying to
the Production workspace.
Contributor access to the Development and Test workspaces : Contributor access to the Development and
Test workspaces allows developers to deploy and modify items within these stages. This access is necessary
to enable developers to work on and test their items before they move to production.
A. Build permission to the production semantic models: This does not relate directly to the deployment
pipeline access requirements and does not provide the needed permissions for deployment.
C. Viewer access to the Development and Test workspaces: Viewer access would only allow developers to
see the items without making any changes or deployments, which does not meet the requirement to deploy
items to these workspaces.
F. Contributor access to the Production workspace: This would grant developers the ability to deploy and
modify items in the Production workspace, which contradicts the requirement to prevent developers from
deploying to the Production stage.
Question: 25
You have a Fabric workspace that contains a DirectQuery semantic model. The model queries a data source that
has 500 million rows.
You have a Microsoft Power Bi report named Report1 that uses the model. Report1 contains visuals on multiple
pages.
You need to reduce the query execution time for the visuals on all the pages.
What are two features that you can use? Each correct answer presents a complete solution,
NOTE: Each correct answer is worth one point.
A.user-defined aggregations
B.automatic aggregation
C.query caching
D.OneLake integration
Answer: AB
Explanation:
A. User-defined aggregations.
Users define aggregations based on expected queries, reducing the need to scan massive datasets.
Automatic aggregations optimize query performance by learning from query patterns and storing frequently
used aggregate results in memory.
Question: 26
You have a Fabric tenant that contains 30 CSV files in OneLake. The files are updated daily.
You create a Microsoft Power BI semantic model named Model1 that uses the CSV files as a data source. You
configure incremental refresh for Model1 and publish the model to a Premium capacity in the Fabric tenant.
When you initiate a refresh of Model1, the refresh fails after running out of resources.
What is a possible cause of the failure?
Answer: D
Explanation:
As described in Incremental refresh and real-time data for models - Requirements, incremental refresh is
designed for data sources that support query folding. Make sure data source queries are being folded in
Power BI Desktop before publishing to the service, where query folding issues can be significantly
compounded.
So having said this, csv files is not a data source that support query folding.
Reference:
https://learn.microsoft.com/en-us/power-bi/connect-data/incremental-refresh-troubleshoot#problem-
loading-data-takes-too-long
Question: 27
You have a Fabric tenant that uses a Microsoft Power BI Premium capacity.
At the semantic model level, set Large dataset storage format to Off.
At the tenant level, set Create and use Metrics to Enabled.
At the semantic model level, set Large dataset storage format to On.
At the tenant level, set Data Activator to Enabled.
Answer: C
Explanation:
At the semantic model level, set Large dataset storage format to On.
Reference:
https://learn.microsoft.com/en-us/power-bi/enterprise/service-premium-scale-out-configure
Question: 28
You have a Fabric tenant that contains a warehouse. The warehouse uses row-level security (RLS).
You create a Direct Lake semantic model that uses the Delta tables and RLS of the warehouse.
When users interact with a report built from the model, which mode will be used by the DAX queries?
A.DirectQuery
B.Dual
C.Direct Lake
D.Import
Answer: A
Explanation:
A. Direct Query "Row-level security only applies to queries on a Warehouse or SQL analytics endpoint in
Fabric. Power BI queries on a warehouse in Direct Lake mode will fall back to Direct Query mode to abide by
row-level security."
When users interact with a report built from a Direct Lake semantic model, which leverages row-level security
(RLS) and Delta tables from a warehouse, the DAX queries will operate in DirectQuery mode.
Reference:
https://learn.microsoft.com/en-us/fabric/data-warehouse/row-level-security
Question: 29
You need to create a diagram of the model. The diagram must contain only the Sales table and related tables.
Data view
Answer: C
Explanation:
C. Model view.
Model view in Microsoft Power BI Desktop allows you to visualize and manage relationships between tables in
a semantic model.
Since the requirement is to create a diagram that includes only the Sales table and its related tables, Model
view is the best choice.
In Model view, you can filter and arrange tables to focus only on relevant parts of the data model, making it
easier to analyze relationships.
Question: 30
You have a Fabric tenant that contains a semantic model. The model uses Direct Lake mode.
You suspect that some DAX queries load unnecessary columns into memory.
You need to identify the frequently used columns that are loaded into memory.
What are two ways to achieve the goal? Each correct answer presents a complete solution.
NOTE: Each correct answer is worth one point.
Answer: BC
Explanation:
Vertipaq Analyzer: This tool helps analyze the internal structure of your Power BI model. It provides detailed
information about the storage and memory usage of your model, including which columns are frequently
accessed and loaded into memory. This can help you identify unnecessary columns that are consuming
resources.
Steps:
Use the Vertipaq Analyzer tool to analyze the model and review the column usage statistics.
DMVs: Dynamic Management Views (DMVs) provide detailed information about the operations of your Power
BI models. Specifically, the $System.DISCOVER_STORAGE_TABLE_COLUMN_SEGMENTS DMV can give you
insights into the storage and usage patterns of individual columns within your model.
Question: 31
HOTSPOT -
You have the source data model shown in the following exhibit.
The primary keys of the tables are indicated by a key symbol beside the columns involved in each key.
You need to create a dimensional data model that will enable the analysis of order items by date, product, and
customer.
What should you include in the solution? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
1. "The relationship between OrderItem and Product must be based on: The ProductID column.
This selection indicates that the relationship between OrderItem and Product is based only on ProductID.
This implies that a ProductID is unique across the system or that OrderItem only needs to reference a product
without considering other attributes like CompanyID.
If products can exist independently and are not company-specific, then this choice is valid.
However, if multiple companies sell the same ProductID, it may be necessary to include CompanyID to avoid
ambiguity.
2. "The Company entity must be: Denormalized into the Customer and Product entities.
Denormalization means storing redundant data in multiple places to improve performance, at the cost of
increased storage and potential update anomalies.
Here, the Company entity is not kept as a separate table but instead integrated into both the Customer and
Product tables.
This can reduce the need for joins when querying customer and product data but might introduce redundancy.
Question: 32
You have a Fabric tenant that contains a semantic model named Model1. Model1 uses Import mode. Model1
contains a table named Orders. Orders has 100 million rows and the following fields.
You need to reduce the memory used by Model1 and the time it takes to refresh the model.
Which two actions should you perform? Each correct answer presents part of the solution.
Answer: AD
Explanation:
In Import mode, Power BI compresses data more efficiently when Date and Time are stored separately instead
of a single DateTime column.
DateTime values have higher cardinality (more unique values), which increases memory usage.
Splitting into Date and Time reduces the number of unique values, improving compression and reducing
memory usage.
Measures in Power BI are calculated at query time and do not consume additional memory, unlike calculated
columns, which take up storage.
TotalSalesAmount is likely derived from Quantity * UnitPrice, so instead of storing it as a column, it should be
computed dynamically using a DAX measure.
Question: 33
You have a Fabric tenant that contains a semantic model.
You need to prevent report creators from populating visuals by using implicit measures.
What are two tools that you can use to achieve the goal? Each correct answer presents a complete solution.
NOTE: Each correct answer is worth one point.
Answer: AB
Explanation:
A. Microsoft Power BI Desktop: Power BI Desktop allows you to control and manage how measures are used
within your reports. By carefully defining and using explicit measures within your data model, you can ensure
that report creators use only these predefined measures instead of creating implicit measures automatically.
B. Tabular Editor: Tabular Editor is a powerful tool for managing and editing Power BI and Analysis Services
tabular models. It allows you to enforce best practices, such as disabling implicit measures, by modifying the
model's properties and ensuring that only explicit measures are available for use in reports.
Question: 34
HOTSPOT -
You have a Fabric tenant that contains two lakehouses.
You are building a dataflow that will combine data from the lakehouses. The applied steps from one of the queries
in the dataflow is shown in the following exhibit.
Use the drop-down menus to select the answer choice that completes each statement based on the information
presented in the graphic.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
Query folding refers to the ability of Power Query to push transformations back to the data source rather than
processing them in Power Query itself.
Some transformations are not foldable and will be executed locally in Power Query.
This is common because basic transformations (filtering, joins, aggregations, etc.) often fold, but more
complex operations (custom columns, advanced functions, etc.) may not.
2. The Added custom step will be performed in the Microsoft Power Query engine.
When a transformation cannot be folded, it is processed in the Power Query engine instead of being pushed to
the source.
Custom calculations
Such transformations are executed locally within Power Query rather than at the source (lakehouse query
engine).
Question: 35
You have a Fabric tenant that contains a lakehouse named Lakehouse’. Lakehouse1 contains a table named Tablet.
You are creating a new data pipeline.
You plan to copy external data to Table’. The schema of the external data changes regularly.
You need the copy operation to meet the following requirements:
Replace Table1 with the schema of the external data.
Replace all the data in Table1 with the rows in the external data.
You add a Copy data activity to the pipeline.
What should you do for the Copy data activity?
Answer: B
Explanation:
The best way to achieve this is by setting Table action to "Overwrite" in the Destination tab of the Copy Data
activity.
"Overwrite" ensures that the table is completely replaced, including the schema and data, which meets both
requirements.
This means that every time the pipeline runs, Table1 will have the latest schema and data from the external
source.
Adding columns does not replace the schema; it only appends extra fields.
This does not ensure that the table structure matches the external data.
Staging is useful for performance improvements in large data transfers, but it does not affect schema
replacement.
Partition discovery helps when reading partitioned data, but it does not affect replacing the table schema or
data.
E. From the Source tab, select Recursively
Recursion is used for reading multiple files in hierarchical folders but does not replace the schema or data in
the destination table.
Question: 36
You have a Fabric tenant that contains a lakehouse.
You plan to query sales data files by using the SQL endpoint. The files will be in an Amazon Simple Storage Service
(Amazon S3) storage bucket.
You need to recommend which file format to use and where to create a shortcut.
Which two actions should you include in the recommendation? Each correct answer presents part of the solution.
NOTE: Each correct answer is worth one point.
Answer: BD
Explanation:
Creating a shortcut in the Files section would treat the data more like raw files, which might not leverage the
full potential of SQL querying capabilities. Creating the shortcut in the Tables section aligns more closely with
the structured query requirements and provides a better-integrated experience.
Question: 37
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a subfolder named
Subfolder1 that contains CSV files.
You need to convert the CSV files into the delta format that has V-Order optimization enabled.
What should you do from Lakehouse explorer?
Answer: A
Explanation:
To convert CSV files into Delta format with V-Order optimization, you need to load the files into tables in the
Lakehouse.
"Load to Tables" is a feature in Microsoft Fabric's Lakehouse Explorer that allows you to convert files (such as
CSV) into managed Delta tables.
During this process, the data is converted into the Delta format, which supports V-Order optimization for
better performance.
V-Order is an advanced columnar storage format that enhances query performance and compression for
Delta tables.
https://learn.microsoft.com/en-us/fabric/data-engineering/load-to-tables#load-to-table-capabilities-overview
Question: 38
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains an unpartitioned table
named Table1.
You plan to copy data to Table1 and partition the table based on a date column in the source data.
You create a Copy activity to copy the data to Table1.
You need to specify the partition column in the Destination settings of the Copy activity.
What should you do first?
Answer: D
Explanation:
Expand Advanced, in Table action, select Overwrite, and then select Enable partition, under Partition columns,
select Add column, and choose the column you want to use as the partition column. You can choose to use a
single column or multiple columns as the partition column.
Reference:
https://learn.microsoft.com/en-us/fabric/data-factory/tutorial-lakehouse-partition#load-data-to-lakehouse-
using-partition-columns
Question: 39
HOTSPOT -
Explanation:
1. Yes
A replica of dbo.Sales is created in the test schema by copying the metadata only.
If metadata-only copying is used, it means the structure (schema) is copied but not the data.
This is possible with views, synonyms, or schema-only copies (e.g., CREATE TABLE ... AS NO DATA).
If the intent is to create an empty table with the same structure, this would be true.
2. No
If test.FactSales was independently created (i.e., copied separately from dbo.FactSales), then any schema
changes to dbo.FactSales will NOT automatically reflect in test.FactSales.
Schema changes would only propagate if test.FactSales was a synonym, view, or linked table.
Since it's likely an independent table, changes to dbo.FactSales do not apply automatically.
3.No
Changes to dbo.FactSales do not automatically apply to test.FactSales unless explicitly linked (e.g., via a view
or a synonym.
Question: 40
You have source data in a folder on a local computer.
You need to create a solution that will use Fabric to populate a data store. The solution must meet the following
requirements:
Support the use of dataflows to load and append data to the data store.
Ensure that Delta tables are V-Order optimized and compacted automatically.
Which type of data store should you use?
A.a lakehouse
B.an Azure SQL database
C.a warehouse
D.a KQL database
Answer: A
Explanation:
The Lakehouse and the Delta Lake table format are central to Microsoft Fabric, assuring that tables are
optimized for analytics is a key requirement.
To meet the requirements of supporting dataflows to load and append data to the data store while ensuring
that Delta tables are V-Order optimized and compacted automatically, you should use a lakehouse in Fabric as
your solution.
Reference:
https://learn.microsoft.com/en-us/fabric/data-engineering/delta-optimization-and-v-order?tabs=sparksql
Question: 41
HOTSPOT -
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
1. "The results will form a hierarchy of folders for each partition key."
Yes
Partitioning creates a folder structure where each folder represents a partition key value.
bash
Copy
Edit
/sales/year=2023/month=01/
/sales/year=2023/month=02/
2. "The resulting file partitions can be read in parallel across multiple nodes."
Yes
In distributed environments like Spark, Synapse, or Fabric Lakehouse, each partition can be processed
independently across multiple compute nodes, improving performance.
This is especially true for columnar formats like Parquet and Delta tables, which support distributed query
execution.
Yes
File compression is commonly used in partitioned storage to reduce storage costs and improve read
performance.
Formats like Parquet, ORC, and Delta support built-in compression (e.g., Snappy, Gzip, or ZSTD).
Compressed partitions reduce I/O costs and improve query performance, making them a best practice.
Question: 42
You have a Fabric workspace named Workspace1 that contains a data flow named Dataflow1 contains a query that
returns the data shown in the following exhibit.
You need to transform the data columns into attribute-value pairs, where columns become rows.
You select the VendorID column.
Which transformation should you select from the context menu of the VendorID column?
A. Group by
B. Unpivot columns
C. Unpivot other columns
D. Split column
E. Remove other columns
Answer: C
Explanation:
You need to transform data columns into attribute-value pairs, meaning that columns will be converted into
rows. This is achieved using Unpivoting in Power Query.
"Unpivot Other Columns" (Option C) → Converts all columns except the selected column into rows.
Since you are selecting the VendorID column and want to transform the other columns into attribute-value
pairs, you should choose "Unpivot Other Columns."
Question: 43
You need to ensure that the pipeline runs every four hours on Mondays and Fridays.
By the minute
Answer: C
Explanation:
Answer C : Weekly.
The only way to do this is to set the schedule to ''Weekly'', set the days on Monday and Friday and add
manually 6 Time of 4 hour intervals.
You need the pipeline to run every four hours on Mondays and Fridays.
Question: 44
You have a Fabric tenant that contains a warehouse.
Several times a day, the performance of all warehouse queries degrades. You suspect that Fabric is throttling the
compute used by the warehouse.
What should you use to identify whether throttling is occurring?
Answer: D
Explanation:
The Microsoft Capacity Metrics app, also known as the metrics app, serves as a monitoring tool within the
Microsoft Power BI service. It offers functionalities to track and analyze the resource utilization.
Reference:
https://learn.microsoft.com/en-us/fabric/data-warehouse/compute-capacity-smoothing-throttling
Question: 45
HOTSPOT -
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
No: The Spark engine will initially read all columns from the CSV file because the .select() transformation is
applied after the data has been read into memory. Therefore, all 12 columns from Sales_raw.csv are read
before the selection of specific columns is applied.
Yes: Adding inferSchema = 'true' will increase the execution time of the query because Spark will need to
read through the entire dataset to determine the data types of each column. This extra pass over the data
adds to the initial read time.
Question: 46
You have a Fabric tenant that contains a warehouse.
A user discovers that a report that usually takes two minutes to render has been running for 45 minutes and has
still not rendered.
You need to identify what is preventing the report query from completing.
Which dynamic management view (DMV) should you use?
A.sys.dm_exec_requests
B.sys.dm_exec_sessions
C.sys.dm_exec_connections
D.sys.dm_pdw_exec_requests
Answer: A
Explanation:
The sys.dm_exec_requests dynamic management view provides real-time information about all requests
currently executing on the SQL Server. This view helps you understand the status of each request, including
whether it is blocked, waiting, or running.
Reference:
https://learn.microsoft.com/en-us/fabric/data-warehouse/monitor-using-dmv
Question: 47
DRAG DROP -
How should you complete the code? To answer, drag the appropriate values to the correct targets. Each value may
be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view
Answer:
Explanation:
Value.
NativeQuery.
NativeQuery allows executing a direct SQL query within Power Query M code.
EnableFolding .
This enables query folding, ensuring that transformations are executed in SQL Server rather than in Power
Query.
Question: 48
DRAG DROP -
Readings from 100 IoT devices are appended to a Delta table in Lakehouse1. Each set of readings is approximately
All the table and SparkSession settings are set to the default.
You discover that queries are slow to execute. In addition, the lakehouse storage contains data and log files that
are no longer used.
You need to remove the files that are no longer used and combine small files into larger files with a target size of 1
GB per file.
What should you do? To answer, drag the appropriate actions to the correct requirements. Each action may be
used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view
content.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
Run the VACUUM command on a schedule: The VACUUM command cleans up old files and log files that are
no longer needed by the Delta table, helping to free up storage and potentially improve performance by
reducing the number of files the query engine needs to consider.
Run the OPTIMIZE command on a schedule: The OPTIMIZE command compacts small files into larger ones,
improving read performance by reducing the overhead associated with opening many small files. This can be
particularly useful when you have a large number of small files due to frequent appends of small data sets.
Question: 49
You need to create a data loading pattern for a Type 1 slowly changing dimension (SCD).
Which two actions should you include in the process? Each correct answer presents part of the solution.
NOTE: Each correct answer is worth one point.
Answer: AD
Explanation:
A. Update rows when the non-key attributes have changed: In a Type 1 SCD, when a change is detected in any
of the non-key attributes of an existing row, the current row is updated with the new values. This type of SCD
does not keep any historical data; it simply overwrites the old data with the new data.
D. Insert new records when the natural key is a new value in the table: When a new record (with a new natural
key) is encountered that does not already exist in the dimension table, it is inserted as a new row. This is
necessary to ensure that all new entities are captured in the dimension.
Question: 50
HOTSPOT -
You have a Fabric workspace named Workspace1 and an Azure Data Lake Storage Gen2 account named storage1.
Workspace1 contains a lakehouse named Lakehouse1.
You need to create a shortcut to storage1 in Lakehouse1.
Which connection and endpoint should you specify? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
The selection of dfs (Distributed File System) means that the connection is using Azure Data Lake Storage
Gen2.
dfs endpoints enable Hierarchical Namespace (HNS), which allows for file system-like directory structures
and optimized big data analytics.
Question: 51
You are analyzing customer purchases in a Fabric notebook by using PySpark.
You have the following DataFrames:
transactions: Contains five columns named transaction_id, customer_id, product_id, amount, and date and has 10
million rows, with each row representing a transaction. customers: Contains customer details in 1,000 rows and
three columns named customer_id, name, and country.
You need to join the DataFrames on the customer_id column. The solution must minimize data shuffling.
You write the following code.
from pyspark.sql import functions as F
results =
Which code should you run to populate the results DataFrame?
Answer: A
Explanation:
In Apache Spark, broadcasting refers to an optimization technique for join operations. When you join two
DataFrames or RDDs and one of them is significantly smaller than the other, Spark can "broadcast" the
smaller table to all nodes in the cluster. This approach avoids the need for network shuffles for each row of
the larger table, significantly reducing the execution time of the join operation.
https://sparkbyexamples.com/spark/broadcast-join-in-spark/"
Question: 52
HOTSPOT -
Use the drop-down menus to select the answer choice that completes each statement based on the information
presented in the graphic.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
Direct Lake Mode enables real-time querying of Delta tables in a Fabric lakehouse without requiring data
import or DirectQuery.
When fallback behavior is set to Automatic, Power BI will automatically switch between Direct Lake Mode and
DirectQuery when necessary (e.g., if a feature isn't supported in Direct Lake).
Direct Lake Mode allows Power BI to load data directly from Delta tables in a Fabric lakehouse without
preloading or sending queries to an external database.
It provides better performance than DirectQuery by eliminating query overhead and improving response times.
Question: 53
HOTSPOT -
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a table named
Nyctaxi_raw. Nyctaxi_row contains the following table:
1. df.withColumn.
It is appropriate here because we are creating a new column (pickupDate) based on an existing column
(tpepPickupDateTime).
Alternative options like df.select or df.withColumnsRenamed would not modify a column directly.
2 .cast('date')
Since tpepPickupDateTime is likely a timestamp, converting it to a date ensures it only contains the date
portion.
Other options like .alias('date') only rename the column without changing its type.
The .filter() function in PySpark is used to apply SQL-like filtering conditions to DataFrames.
The selected condition ensures that only rows where fareAmount is greater than 0 and less than 100 are
retained.
Other options like .where(df.fareAmount.isin([1,100])) filter for specific values rather than a range.
Question: 54
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a new semantic model in OneLake.
You use a Fabric notebook to read the data into a Spark DataFrame.
You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
numeric columns.
Solution: You use the following PySpark expression:
df.explain()
Does this meet the goal?
A.Yes
B.No
Answer: B
Explanation:
The df.explain() method in PySpark is used to print the logical and physical plans of a DataFrame, which helps
in understanding how Spark plans to execute the query. It does not compute any statistical values like min,
max, mean, or standard deviation.
A.Yes
B.No
Answer: B
Explanation:
Correct methods: Use df.describe().show() for basic statistics and df.agg() with appropriate functions (min,
max, mean, stddev) for detailed statistics.
Question: 56
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a new semantic model in OneLake.
You use a Fabric notebook to read the data into a Spark DataFrame.
You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
numeric columns.
Solution: You use the following PySpark expression:
df.summary()
Does this meet the goal?
A.Yes
B.No
Answer: A
Explanation:
A. Yes
The df.summary() function in PySpark provides summary statistics for both numeric and string columns in a
DataFrame, including:
A.Yes
B.No
Answer: A
Explanation:
Yes, running DESCRIBE HISTORY customer meets the goal of identifying whether maintenance tasks were
performed on the Delta table.
Question: 58
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table named
Customer.
When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT
performed on the table.
You need to identify whether maintenance tasks were performed on Customer.
Solution: You run the following Spark SQL statement:
A.Yes
B.No
Answer: B
Explanation:
No, running REFRESH TABLE customer does not meet the goal of identifying whether maintenance tasks
were performed on the Delta table. This Spark SQL command is used to refresh the metadata of a table. It
ensures that the latest schema and data are available for queries but does not give any historical information
Question: 59
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table named
Customer.
When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT
performed on the table.
You need to identify whether maintenance tasks were performed on Customer.
Solution: You run the following Spark SQL statement:
A.Yes
B.No
Answer: B
Explanation:
The EXPLAIN statement in Spark SQL is used to display the execution plan of a query. This plan shows how
Spark will execute the query, including details about the operations and stages involved. While it is useful for
understanding and optimizing query performance, it does not provide historical information about
maintenance tasks like optimization, compaction, or vacuuming performed on the table.
Question: 60
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
• Survey
• Question
• Response
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
• Fabric administrators will be the workspace administrators.
• The data engineers must be able to read from and write to the data store. No access must be granted to datasets
or reports.
• The analytics engineers must be able to read from, write to, and create schemas in the data store. They also must
be able to create and share semantic models with the data analysts and view and modify all reports in the
workspace.
• The data scientists must be able to read from the data store, but not write to it. They will access the data by using
a Spark notebook
• The data analysts must have read access to only the dimensional model objects in the data store. They also must
have access to create Power BI reports by using the semantic models created by the analytics engineers.
• The date dimension must be available to all users of the data store.
• The principle of least privilege must be followed.
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.
You need to recommend a solution to prepare the tenant for the PoC.
Which two actions should you recommend performing from the Fabric Admin portal? Each correct answer presents
part of the solution.
A. Enable the Users can try Microsoft Fabric paid features option for the entire organization.
B. Enable the Users can try Microsoft Fabric paid features option for specific security groups.
C. Enable the Allow Azure Active Directory guest users to access Microsoft Fabric option for specific security
groups.
D. Enable the Users can create Fabric items option and exclude specific security groups.
E. Enable the Users can create Fabric items option for specific security groups.
Answer: BE
Explanation:
Enable the Users can try Microsoft Fabric paid features option for specific security groups: This will allow
specific security groups (like the AnalyticsTeam, DataAnalysts, DataScientists, DataEngineers, and
AnalyticsEngineers) to access and use the paid features of Microsoft Fabric necessary for the PoC. This is
important to ensure that only the relevant team members can utilize these advanced features while
preventing unnecessary access for other users.
Enable the Users can create Fabric items option for specific security groups: This will allow only specific
security groups to create Fabric items, ensuring that the creation of these items is controlled and managed by
the appropriate team members. This helps maintain the principle of least privilege and ensures that only
authorized personnel can create and manage Fabric items during the PoC.
Question: 61
HOTSPOT
-
Case study
-
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview
-
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment
-
Fabric Environment
-
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data
-
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
• Survey
• Question
• Response
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems
-
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements
-
Planned Changes
-
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The following will be created in the AnalyticsPOC workspace:
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements
-
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.
Security Requirements
-
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
Report Requirements
-
The data analysts must create a customer satisfaction report that meets the following requirements:
• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.
You need to design a semantic model for the customer satisfaction report.
Which data source authentication method and mode should you use? To answer, select the appropriate options in
the answer area.
Answer:
Explanation:
SSO (Single Sign-On) allows users to authenticate once and gain access to multiple applications without
needing to log in separately.
This is a secure and seamless authentication method commonly used in enterprise environments for Microsoft
Fabric, Power BI, and other services.
Direct Lake mode is a new hybrid query mode that combines the performance of Import mode with the real-
time access of DirectQuery.
Unlike Import mode, which loads data into memory, or DirectQuery, which queries the source live, Direct Lake
queries Delta tables directly from OneLake without the need for pre-loading.
This mode is optimized for high performance and scalability, making it an ideal choice for Power BI and Fabric
scenarios.
Question: 62
Case study -
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
To start the case study -
To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore
the content of the case study before you answer the questions. Clicking these buttons displays information such
as business requirements, existing environment, and problem statements. If the case study has an All Information
tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you
are ready to answer a question, click the Question button to return to the question.
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
• Survey
• Question
• Response
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.
You need to implement the date dimension in the data store. The solution must meet the technical requirements.
What are two ways to achieve the goal? Each correct answer presents a complete solution.
Answer: AD
Explanation:
Dataflows in Microsoft Fabric (or Power BI) allow ETL (Extract, Transform, Load) operations to create and
populate tables.
Stored procedures allow for predefined logic and transformations to be applied while inserting or updating
records in the date dimension table.
Question: 63
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team at Litware
contains data engineers, analytics engineers, data analysts, and data scientists.
Existing Environment -
Fabric Environment -
Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabric capacities
and features.
Available Data -
Litware has data that must be analyzed as shown in the following table.
The Product data contains a single table and the following columns.
• Survey
• Question
• Response
The Question table contains the text of each survey question. The third question in each survey response is an
overall satisfaction score. Customers can submit a survey after each purchase.
User Problems -
The analytics team has large volumes of data, some of which is semi-structured. The team wants to use Fabric to
create a new data store.
Product data is often classified into three pricing groups: high, medium, and low. This logic is implemented in
several databases and semantic models, but the logic does NOT always match across implementations.
Requirements -
Planned Changes -
Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new data store as a
proof of concept (PoC). The remaining Liware users will only get access to the Fabric features once the PoC is
complete. The PoC will be completed by using a Fabric trial capacity
• AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooks used to
populate the data store
• DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake
• DataSciPOC: Will contain all the notebooks and reports created by the data scientists
The data engineers will create data pipelines to load data to OneLake either hourly or daily depending on the data
source. The analytics engineers will create processes to ingest, transform, and load the data to the data store in
the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data
ingestion. The choice of which data cleansing and transformation tools to use will be at the data engineers’
discretion.
All the semantic models and reports in the Analytics POC workspace will use the data store as the sole data
source.
Technical Requirements -
Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet Delta Lake
specifications.
Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will then be
cleansed, merged, and transformed into a dimensional model
The data load process must ensure that the raw and cleansed data is updated completely before populating the
dimensional model
The dimensional model must contain a date dimension. There is no existing data source for the date dimension. The
Litware fiscal year matches the calendar year. The date dimension must always contain dates from 2010 through
the end of the current year.
The product pricing group logic must be maintained by the analytics engineers in a single location. The pricing
group data must be made available in the data store for T-SOL. queries and in the default semantic model. The
following logic must be used:
• List prices that are less than or equal to 50 are in the low pricing group.
• List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.
• List prices that are greater than 1,000 are in the high pricing group.
Security Requirements -
Only Fabric administrators and the analytics team must be able to see the Fabric items created as part of the PoC.
Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:
Both the default and custom semantic models must include only tables or views from the dimensional model in the
data store. Litware already has the following Microsoft Entra security groups:
• FabricAdmins: Fabric administrators
• AnalyticsTeam: All the members of the analytics team
• DataAnalysts: The data analysts on the analytics team
• DataScientists: The data scientists on the analytics team
• DataEngineers: The data engineers on the analytics team
• AnalyticsEngineers: The analytics engineers on the analytics team
Report Requirements -
The data analysts must create a customer satisfaction report that meets the following requirements:
• Enables a user to select a product to filter customer survey responses to only those who have purchased that
product.
• Displays the average overall satisfaction score of all the surveys submitted during the last 12 months up to a
selected dat.
• Shows data as soon as the data is updated in the data store.
• Ensures that the report and the semantic model only contain data from the current and previous year.
• Ensures that the report respects any table-level security specified in the source data store.
• Minimizes the execution time of report queries.
You need to ensure the data loading activities in the AnalyticsPOC workspace are executed in the appropriate
sequence. The solution must meet the technical requirements.
A.Create a dataflow that has multiple steps and schedule the dataflow.
B.Create and schedule a Spark notebook.
C.Create and schedule a Spark job definition.
D.Create a pipeline that has dependencies between activities and schedule the pipeline.
Answer: D
Explanation:
D. Create a pipeline that has dependencies between activities and schedule the pipeline.
To ensure that data loading activities in the AnalyticsPOC workspace are executed in the appropriate
sequence while meeting technical requirements, the best approach is to use a data pipeline with
dependencies between activities.
Pipelines in Microsoft Fabric (or Azure Data Factory) allow you to orchestrate data movement, transformation,
and processing across multiple steps.
You can define activity dependencies, ensuring that each step executes only after the previous step
completes successfully.
You can schedule the pipeline to run at specific times, making it automated and reliable.
Question: 64
Case study -
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Requirements -
Planned Changes -
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.
Contoso identifies the following requirements for implementing and managing semantic models:
• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
You need to recommend which type of Fabric capacity SKU meets the data analytics requirements for the
Research division.
A.A
B.EM
C.P
D.F
Answer: D
Explanation:
Fabric capacity SKUs all start with F, they allows for Pay as you go by the minute. Reserved capacity is also
available.
Reference:
https://azure.microsoft.com/en-us/pricing/details/microsoft-fabric/
Question: 65
HOTSPOT
Case
-
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview
-
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment
-
Identity Environment
-
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment
-
Requirements
-
Planned Changes
-
Contoso plans to make the following changes:
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.
Contoso identifies the following requirements for implementing and managing semantic models:
• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements
-
Contoso identifies the following high-level requirements that must be considered for all solutions:
You need to migrate the Research division data for Productline1. The solution must meet the data preparation
requirements.
How should you complete the code? To answer, select the appropriate options in the answer area.
Explanation:
"Tables/productline1"
If you use saveAsTable() you don't need to specify the path "Table/"
Question: 66
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Requirements -
Planned Changes -
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.
Contoso identifies the following requirements for implementing and managing semantic models:
• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
• Follow the principle of least privilege when applicable.
• Minimize implementation and maintenance effort when possible.
What should you use to implement calculation groups for the Research division semantic models?
Answer: D
Explanation:
D .Tabular Editor
The semantic models in the Research division workspaces must use Direct Lake mode.
Question: 67
HOTSPOT
-
Case study
-
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview
-
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment
-
Identity Environment
-
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment
-
Requirements
-
Planned Changes
-
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productine2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
Contoso identifies the following requirements for implementing and managing semantic models:
The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements
-
Contoso identifies the following high-level requirements that must be considered for all solutions:
Which workspace role assignments should you recommend for ResearchReviewersGroup1 and
ResearchReviewersGroup2? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
1. ResearchReviewersGroup1: Viewer.
Members of this group can only view resources but cannot edit, add, or delete them.
Suitable for users who need read-only access to reports, datasets, or workspaces.
2.ResearchReviewersGroup2: Contributor.
Members can edit and add content but cannot manage access or settings.
Suitable for users who need to modify or update reports and datasets but should not have full administrative
privileges.
Question: 68
HOTSPOT
You need to configure OneLake security for users shown in the following table.
Which permission should you assign to each user? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
User1: "ReadAll"
This permission allows access to all data and metadata within the system.
Suitable for users who need full read access across multiple datasets, reports, or tables.
User2: "ReadData"
This permission allows access to the data only, but not the metadata or system-level configurations.
Suitable for users who need to query and analyze specific datasets but should not see system-wide metadata.
Reference:
https://learn.microsoft.com/en-us/fabric/data-engineering/lakehouse-sharing#sharing-and-permissions
Question: 69
You have an Azure Repos repository named Repo1 and a Fabric-enabled Microsoft Power BI Premium capacity. The
capacity contains two workspaces named Workspace1 and Workspace2. Git integration is enabled at the
workspace level.
You plan to use Microsoft Power BI Desktop and Workspace1 to make version-controlled changes to a semantic
model stored in Repo1. The changes will be built and deployed to Workspace2 by using Azure Pipelines.
You need to ensure that report and semantic model definitions are saved as individual text files in a folder
hierarchy. The solution must minimize development and maintenance effort.
A.PBIP
B.PBIDS
C.PBIT
D.PBIX
Answer: A
Explanation:
PBIP format is designed to work with version control systems like Azure Repos. It breaks down Power BI
artifacts into individual files that can be managed and versioned separately, facilitating better collaboration
and change tracking.
Folder Hierarchy: It saves the project structure in a folder hierarchy, where each component of the Power BI
project (like datasets, reports, data sources) is stored in separate files.
Text-Based: Being a text-based format, it integrates well with Git repositories and supports diff and merge
operations.
Question: 70
DRAG DROP
You have a lakehouse that contains the Bronze and Silver layers and a warehouse that contains the Gold layer.
You need to ensure that the layers are populated daily in sequential order such that Silver is populated only after
Bronze is complete, and Gold is populated only after Silver is complete. The solution must minimize development
effort and complexity.
What should you use to execute each set of items? To answer, drag the appropriate options to the correct items.
Each option may be used once, more than once, or not at all. You may need to drag the split bar between panes or
scroll to view content.
Answer:
Explanation:
The orchestration pipeline ensures that the entire data pipeline runs on a set schedule.
The Bronze layer typically ingests raw data from various sources.
The Gold layer is responsible for aggregating and modeling data for analytics.
A Stored Procedure Activity is used to execute complex SQL logic on the transformed data.
Question: 71
DRAG DROP
-
You have a Spark DataFrame assigned to a variable named df. The DataFrame returns four columns.
You need to change the data type of a string column named Age to integer. The solution must return a DataFrame
that includes all the columns.
How should you complete the code? To answer, drag the appropriate values to the correct targets. Each value may
be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view
content.
Answer:
Explanation:
The cast function is applied to convert the "age" column to an integer type.
Question: 72
HOTSPOT
Using a notebook in Workspace1, you need to load the content of the file to the default lakehouse. The solution
must ensure that the content will display automatically as a table named Sales in Lakehouse explorer.
How should you complete the code? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
df = spark.read.parquet("abfss://fs1@storage1.dfs.core.windows.net/files/sales.parquet")
df.write.mode("overwrite").format("delta").saveAsTable("sales")
https://learn.microsoft.com/en-us/fabric/data-engineering/lakehouse-notebook-load-data
Question: 73
You have a Fabric workspace named Workspace1 that contains a lakehouse named Lakehouse1.
D.Copy data
Answer: A
Explanation:
A. Dataflow.
To copy data from CSV files in Azure Storage to Lakehouse1 while supporting Power Query M formula
Question: 74
HOTSPOT
You have a Fabric tenant that contains lakehouse named Lakehouse1. Lakehouse1 contains a Delta table with eight
columns.
You receive new data that contains the same eight columns and two additional columns.
You create a Spark DataFrame and assign the DataFrame to a variable named df. The DataFrame contains the new
data.
You need to add the new data to the Delta table to meet the following requirements:
How should you complete the code? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
1. Mode: "append"
The "append" mode adds new data to the existing Delta table without overwriting it.
This ensures that previously stored data is retained, and only new data is added.
mergeSchema=true allows automatic schema evolution when new columns are introduced in the DataFrame
but are not present in the existing Delta table.
This prevents schema mismatch errors and enables seamless schema updates.
Question: 75
HOTSPOT
You need to write a T-SQL query that will return the following columns.
How should you complete the code? To answer, select the appropriate options in the answer area.
Explanation:
The COALESCE function returns the first non-null value from the provided arguments.
In this case, it checks Weight first. If it is NULL, it checks Quantity. If both are NULL, it defaults to 1.
The LEAST function returns the smallest value among the provided arguments.
It selects the lowest price between ListPrice and SalePrice, ensuring that the minimum price is used for
Question: 76
You need to ensure that the query returns all the rows in both tables.
Answer: B
Explanation:
If there is no match, it still includes the row, filling missing values with NULLs.
Question: 77
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table that has
one million Parquet files.
You need to remove files that were NOT referenced by the table during the past 30 days. The solution must ensure
that the transaction log remains consistent, and the ACID properties of the table are maintained.
Answer: D
Explanation:
The VACUUM command is used in Delta tables to remove old, unreferenced Parquet files while ensuring that:
The transaction log remains consistent
By default, Delta Lake retains all historical data files for seven days to allow for time travel and rollback. The
VACUUM command permanently deletes these unreferenced files, but you can specify a retention period
(e.g., 30 days) to control the cleanup.
Question: 78
DRAG DROP
-
You are implementing two dimension tables named Customers and Products in a Fabric warehouse.
You need to use slowly changing dimension (SCD) to manage the versioning of data. The solution must meet the
requirements shown in the following table.
Which type of SCD should you use for each table? To answer, drag the appropriate SCD types to the correct tables.
Each SCD type may be used once, more than once, or not at all. You may need to drag the split bar between panes
or scroll to view content.
Answer:
Explanation:
Customers :Type 2
Customer records often require history tracking (e.g., change of address, email, or preferences).
Question: 79
You have a Fabric workspace named Workspace1 and an Azure SQL database.
You plan to create a dataflow that will read data from the database, and then transform the data by performing an
inner join.
You need to ignore spaces in the values when performing the inner join. The solution must minimize development
effort.
Answer: B
Explanation:
You need to perform an inner join while ignoring spaces in the values. The best way to achieve this with
minimal development effort in a Fabric dataflow is by using fuzzy matching in a merge query.
Fuzzy Matching → Allows approximate matching, which can ignore leading, trailing, or extra spaces.
https://learn.microsoft.com/en-us/power-query/merge-queries-fuzzy-match
Question: 80
You have a Fabric tenant that contains a warehouse named Warehouse1. Warehouse1 contains two schemas name
schema1 and schema2 and a table named schema1.city.
You need to make a copy of schema1.city in schema2. The solution must minimize the copying of data.
Answer: C
Explanation:
This statement creates a new table named city in schema2 that has the same structure as the city table in
schema1 without copying any data. It essentially creates a metadata reference to the original table, which
minimizes the data copying.
Reference:
https://learn.microsoft.com/en-us/fabric/data-warehouse/clone-table
Question: 81
You need to prevent new tables added to Lakehouse1 from being added automatically to the default semantic
Answer: A
Explanation:
analytics endpoint settings is indeed the correct approach to prevent new tables added to Lakehouse1 from
Question: 82
HOTSPOT
-
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains a lakehouse named
Lakehouse1 and a warehouse named Warehouse1.
You need to create a new table in Warehouse1 named POSCustomers by querying the customer table in
Lakehouse1.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.
FROM lakehouse1.dbo.customer.
Question: 83
You have a Fabric tenant.
You have a stored procedure that returns the number of active customers and their average sales for the current
You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
Copy data
Answer: D
Explanation:
executing a query or stored procedure. The output can be a singleton value or an array of attributes, which
can be consumed in a subsequent copy, transformation, or control flow activities like ForEach activity.
https://learn.microsoft.com/en-us/azure/data-factory/control-flow-lookup-activity
Question: 84
A.a shortcut
B.a dataflow
C.a view
D.a managed table
Answer: A
Explanation:
Shortcuts are specifically designed for cross-lakehouse referencing and provide a seamless and efficient way
Views typically reference tables within the same database or lakehouse. While you can create views that
reference tables in different schemas or databases, they generally do not support cross-workspace
references directly. Creating views for cross-workspace data might require additional configuration or data
duplication.
Question: 85
You have a Fabric tenant that contains a warehouse.
You are designing a star schema model that will contain a customer dimension. The customer dimension table will
be a Type 2 slowly changing dimension (SCD).
You need to recommend which columns to add to the table. The columns must NOT already exist in the source.
Which three types of columns should you recommend? Each correct answer presents part of the solution.
Answer: CDE
Explanation:
Surrogate keys are typically used in dimension tables rather than fact tables. In a data warehouse, a
surrogate key is a unique identifier assigned to each record in a dimension table, usually for internal
processing and joining purposes. It provides a stable reference to the dimension record, regardless of any
changes in the natural key or other attributes.
Question: 86
You have a Fabric tenant.
You plan to create a data pipeline named Pipeline1. Pipeline1 will include two activities that will execute in
sequence.
You need to ensure that a failure of the first activity will NOT block the second activity.
Which conditional path should you configure between the first activity and the second activity?
Upon Failure
Upon Completion
Upon Skip
Upon Skip
Answer: B
Explanation:
Upon Completion ensures that the second activity executes regardless of whether the first activity succeeds
or fails.
Question: 87
You have a Microsoft Power BI semantic model.
You need to identify any surrogate key columns in the model that have the Summarize By property set to a value
other than to None. The solution must minimize effort.
Answer: D
Explanation:
The Best Practice Analyzer (BPA) in Tabular Editor can be configured to check for specific properties and
configurations in your model. This includes identifying columns with certain "Summarize By" settings.
You can create or use existing BPA rules to quickly identify any surrogate key columns that do not have the
"Summarize By" property set to "None," thus minimizing the manual effort required.
Question: 88
DRAG DROP
Which five actions should you perform in sequence? To answer, move the appropriate actions from the list of
actions to the answer area and arrange them in the correct order.
Answer:
Explanation:
Before analyzing query performance, you must first capture a recording in the Performance Analyzer.
Sort the Duration (ms) Column in Descending Order by DAX Query Time.
Sorting the DAX query time in descending order helps in identifying the slowest queries.
After identifying the slowest query, copy it into DAX Studio for further analysis and debugging.
Enable Query Timings and Server Timings, Then Run the Query.
In DAX Studio, enabling Query Timings and Server Timings helps to distinguish between:
The final step is to analyze the query's execution at the server level to identify bottlenecks and optimize
performance.
Question: 89
You need to programmatically change each column that ends in the word Key to meet the following requirements:
Set Nullable to False
Answer: C
Explanation:
C. Tabular Editor.
Tabular Editor is a powerful tool for modifying and managing Power BI semantic models (Tabular models)
Question: 90
HOTSPOT
You need to create a calculation item that will change the context from the selected date to month-to-date (MTD).
How should you complete the DAX expression? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
CALCULATE:
SELECTEDMEASURE:
Question: 91
You have a Microsoft Power BI report named Report1 that uses a Fabric semantic model.
You open Performance analyzer and identify that a visual named Orders By Date is the slowest to render. The
duration breakdown for Orders By Date is shown in the following table.
What will provide the greatest reduction in the rendering duration of Report1?
Answer: D
Explanation:
Rendering duration in Power BI is significantly impacted by the number of visuals on a report page.
Each visual requires query execution, data processing, and rendering time.
Reducing the number of visuals will lower the computational load, leading to faster rendering.
Question: 92
You have a custom Direct Lake semantic model named Model1 that has one billion rows of data.
You use Tabular Editor to connect to Model1 by using the XMLA endpoint.
You need to ensure that when users interact with reports based on Model1, their queries always use Direct Lake
Explanation:
https://learn.microsoft.com/en-us/power-bi/enterprise/directlake-overview#fallback-behavior
The DirectLakeBehavior property can be configured by using Tabular Object Model (TOM) or Tabular Model
Scripting Language (TMSL).
The following example specifies all queries use Direct Lake mode only:
C#
Copy
//
database.Model.DirectLakeBehavior = DirectLakeBehavior.DirectLakeOnly = 1;
database.Model.SaveChanges();
Additional Question
Answer:
3 Set the Object Level Security to None for the sales manager role.
Question: 93
DRAG DROP
-
You create a semantic model by using Microsoft Power BI Desktop. The model contains one security role named
SalesRegionManager and the following tables:
• Sales
• SalesRegion
• SalesAddress
You need to modify the model to ensure that users assigned the SalesRegionManager role cannot see a column
named Address in SalesAddress.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of
actions to the answer area and arrange them in the correct order.
Answer:
Explanation:
Since Tabular Editor is used for advanced metadata management, including setting OLS, this is the correct
tool.
To modify OLS settings, the relevant column (Address) must be selected in SalesAddress.
This ensures that SalesRegionManager has no restrictions on accessing the Address column.
If security were set to Default, the column could be hidden or restricted based on predefined roles.
Question: 94
functions and a FILTER function.
In which use case will replacing the FILTER function with the KEEPFILTERS function reduce execution time?
Answer: D
Explanation:
D. When the FILTER function references a column from a single table that uses Import mode.
The FILTER function creates a new row context, which can be slow when iterating over large tables.
KEEPFILTERS applies additional filters without overriding the existing ones, improving query performance by
avoiding unnecessary row context transitions.
When FILTER references a column from a single Import mode table, replacing it with KEEPFILTERS reduces
the overhead of row context evaluation and improves execution speed.
Question: 95
You have a semantic model named Model1. Model1 contains five tables that all use Import mode. Model1 contains a
dynamic row-level security (RLS) role named HR. The HR role filters employee data so that HR managers only see
the data of the department to which they are assigned.
You publish Model1 to a Fabric tenant and configure RLS role membership. You share the model and related
reports to users.
Answer: C
Explanation:
Option C (Select "Test as role" to view the report as the HR manager) is the best approach as it directly
validates what the specific HR manager sees under the dynamic RLS conditions, ensuring the completeness
and accuracy of the data.
A. Select "Test as role" to view the data as the HR role: This option is useful, but it doesn't specify viewing the
report as the specific HR manager, which is crucial to identify user-specific issues.
Question: 96
You have a Microsoft Fabric tenant that contains a dataflow.
From Power Query, you need to view column information as shown in the following exhibit.
Which three Data view options should you select? Each correct answer presents part of the solution.
Answer: ACD
Explanation:
Show column value distribution: This option provides a visual representation of the distribution of values in
each column, which is visible in the exhibit.
Enable column profile: This option displays statistics and other detailed information about each column,
including value distribution, which aligns with the data shown in the exhibit.
Show column quality details: This option shows the quality of the data in each column, indicating valid, error,
and empty values, as displayed in the exhibit.
Question: 97
HOTSPOT
You need to write a T-SQL query that will return the following columns.
How should you complete the code? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
The GREATEST function returns the highest value from the given columns.
This is useful for finding the maximum price among different price categories.
If any value is NULL, it does not affect the result unless all values are NULL.
The COALESCE function returns the first non-null value from the list of columns.
This ensures that TradePrice takes the first available price instead of leaving it NULL.
Question: 98
You have a Fabric notebook that has the Python code and output shown in the following exhibit.
Which type of analytics are you performing?
A.descriptive
B.diagnostic
C.prescriptive
D.predictive
Answer: A
Explanation:
Descriptive Analytics: This type of analytics focuses on summarizing historical data to understand what has
happened. For example, generating reports and dashboards to show past sales performance.
Reference:
https://azure.microsoft.com/es-es/blog/answering-whats-happening-whys-happening-and-what-will-happen-
with-iot-analytics/
Question: 99
HOTSPOT
-
You have a Fabric warehouse that contains a table named Sales.Orders. Sales.Orders contains the following
columns.
You need to write a T-SQL query that will return the following columns.
How should you complete the code? To answer, select the appropriate options in the answer area.
Explanation:
DATETRUNC
Weekday
Reference:
https://learn.microsoft.com/en-us/sql/t-sql/functions/datetrunc-transact-sql?view=sql-server-ver16
Question: 100
You have a Fabric tenant that contains JSON files in OneLake. The files have one billion items.
You need to transform the data, visualize the data to find insights, perform anomaly detection, and share the
insights with other business users. The solution must meet the following requirements:
Answer: A
Explanation:
PySpark provides robust capabilities for data transformation and manipulation, making it well-suited for
preparing data for time series analysis. You can then use libraries like pandas for further data manipulation if
needed and leverage Spark's machine learning capabilities for anomaly detection.
Question: 101
You have a Fabric tenant that contains customer churn data stored as Parquet files in OneLake. The data contains
details about customer demographics and product usage.
You create a Fabric notebook to read the data into a Spark DataFrame. You then create column charts in the
notebook that show the distribution of retained customers as compared to lost customers based on geography, the
number of products purchased, age, and customer tenure.
A.diagnostic
B.descriptive
C.prescriptive
D.predictive
Answer: B
Explanation:
Descriptive - it's just describing the customers, not saying why they stayed or left
Descriptive analytics tells what happened in the past, presenting it as numbers and visuals in reports and
dashboards.Diagnostic analytics gives the reason why something happened.Predictive analytics determines
the potential outcomes of present and past actions and trends.Prescriptive analytics offers decision support
for the best course of action.Given the scenario in the question where data is read into a Spark DataFrame and
column charts are created to show the distribution of retained customers compared to lost customers based
on various factors, this falls under the definition of descriptive analytics. No future predictions or
prescriptions are made, nor are reasons for the past events provided.Final Answer:The type of analytics being
performed in the scenario described is descriptive analytics.
https://www.selecthub.com/business-intelligence/predictive-descriptive-prescriptive-analytics/
Question: 102
HOTSPOT
-
You have a Fabric tenant that contains a semantic model. The model contains data about retail stores.
You need to write a DAX query that will be executed by using the XMLA endpoint. The query must return the total
amount of sales from the same period last year.
How should you complete the DAX expression? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
CALCULATE.
-LY Sales.
The entire expression calculates the sales amount for the same period last year based on the current date
context.
The SUMMARIZE function creates a summary table based on grouping and aggregation.
RETURN
_ LYSales
Question: 103
You need to identify the min and max values for each column in the query results.
Which three Data view options should you select? Each correct answer presents part of the solution.
Answer: BCE
Explanation:
Required to view the detailed column profile statistics.mn profile in details pane E - Enable details pane; this
needs to be enabled to be able to see the output of (C).
Question: 104
You have a Fabric tenant that contains a Microsoft Power BI report.
A.Table.schema
B.Table.view
C.Table.FuzzyGroup
D.Table.Profile
Answer: D
Explanation:
https://learn.microsoft.com/en-us/powerquery-m/table-profile
Question: 105
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
You have a Fabric tenant that contains a lakehouse named Lakehouse1. Lakehouse1 contains a Delta table named
DESCRIBE DETAIL customer -
Yes
Answer: B
Explanation:
Question: 106
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
A.Yes
B.No
Answer: B
Explanation:
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a semantic model named Model1.
You discover that the following query performs slowly against Model1.
A.Yes
B.No
Answer: B
Explanation:
B is correct (NO) because the suggested change does not meet the goal because it reverses the logic of the
query.
Question: 108
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
You have a Fabric tenant that contains a semantic model named Model1.
You need to reduce the execution time of the query.
A.Yes
B.No
Answer: A
Explanation:
A — The proposed solution improves efficiency by reducing the number of calculations required. Instead of
counting all the rows for each customer and then checking if the count is greater than zero, it simply checks if
there are any rows at all, which requires fewer computational resources and execution time
Yes, replacing CALCULATE ( COUNTROWS( 'Order Item' ) ) > 0 with NOT ISEMPTY ( CALCULATETABLE (
'Order Item ' ) ) should reduce the execution time of the query. It is a simpler, more meaningful, and faster way
to check if a table is empty.
https://www.sqlbi.com/articles/check-empty-table-condition-with-dax/
Question: 109
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
You have a Fabric tenant that contains a semantic model named Model1.
A.Yes
B.No
Answer: B
Explanation:
Question: 110
HOTSPOT
-
Case study
-
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview
-
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment
-
Identity Environment
-
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment
-
Requirements
-
Planned Changes
-
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.
Contoso identifies the following requirements for implementing and managing semantic models:
• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements
-
Contoso identifies the following high-level requirements that must be considered for all solutions:
You need to migrate the Research division data for Productline2. The solution must meet the data preparation
requirements.
How should you complete the code? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
format("delta") → Saves the file in Delta Lake format, which supports ACID transactions and versioning.
save("Tables/productline2") → Saves the Delta table in the default location for managed tables.
Case study -
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Requirements -
Planned Changes -
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
Data Preparation Requirements -
• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.
Contoso identifies the following requirements for implementing and managing semantic models:
• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Which syntax should you use in a notebook to access the Research division data for Productline1?
A.spark.read.format(“delta”).load(“Files/ResearchProduct”)
B.spark.sql(“SELECT * FROM Lakehouse1.ResearchProduct ”)
C.spark.sql(“SELECT * FROM Lakehouse1.Tables.ResearchProduct ”)
D.external_table(ResearchProduct)
Answer: A
Explanation:
A.spark.read.format("delta").load("Files/ResearchProduct")
The key change is understanding that shortcuts do not behave like managed tables in Spark SQL. To read
data from a shortcut you need to use spark.read.format and use Files/ path, since it's a shortcut it behaves like
a file in the filesystem.
Question: 112
HOTSPOT
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
1. YES - The .select() function in the code specifies the exact columns ('SalesOrderNumber', 'OrderDate',
'CustomerName', and 'UnitPrice') to be selected.
2. NO - The withColumn("Year", year("OrderDate")) function adds a new column called "Year" by extracting
the year from the "OrderDate" column.
However, it does not replace the "OrderDate" column—it only adds the new "Year" column.
3. YES - The inferSchema='true' tells Spark to infer the data types of each column in the CSV, which requires
an extra scan of the data to determine these types.
Question: 113
Note: This question is part of a series of questions that present the same scenario. Each question in the series
contains a unique solution that might meet the stated goals. Some question sets might have more than one correct
solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
You have a Fabric tenant that contains a new semantic model in OneLake.
You use a Fabric notebook to read the data into a Spark DataFrame.
You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and
numeric columns.
df.describe().show()
A.Yes
B.No
Answer: A
Explanation:
DataFrame.Describe = Computes basic statistics for numeric and string columns, including count, mean,
stddev, min, and max. If no columns are given, this function computes statistics for all numerical or string
columns.
Question: 114
You have a Fabric tenant.
You have a stored procedure that returns the number of active customers and their average sales for the current
You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
C.Append variable
D.Lookup
Answer: D
Explanation:
Question: 115
HOTSPOT
-
You have a Fabric tenant that contains a semantic model named model1. The two largest columns in model1 are
shown in the following table.
You need to optimize model1. The solution must meet the following requirements:
What should you do on each column? To answer, select the appropriate options in the answer area.
Explanation:
There is no need to keep 160 GB for a surrogate key, it can be removed and SaleDateTime can be moved into
the SalesTransaction table that will for sure "Increase refresh performance when using Import mode." and
"Reduce the model size." since there is no need for an heavy join like this one.
Splitting the column would lead for sure to "Ensure that the datetime value for each sales transaction is
available in the model." and i think it would lead to a better segmentation of the data meaning less cardinality
that would lead to "Increase refresh performance when using Import mode." and "Reduce the model size."
Question: 116
DRAG DROP
-
You have a Fabric tenant that contains a data warehouse named DW1. DW1 contains a table named DimCustomer.
DimCustomer contains the fields shown in the following table.
You need to identify duplicate email addresses in DimCustomer. The solution must return a maximum of 1,000
records.
Which four T-SQL statements should you run in sequence? To answer, move the appropriate statements from the
list of statements to the answer area and arrange them in the correct order.
Answer:
Explanation:
This selects the CustomerAltKey column and counts the number of occurrences of each key.
The TOP(1000) ensures that only the first 1,000 results are returned.
FROM DimCustomer
GROUP BY CustomerAltKey
Groups the data by CustomerAltKey, meaning that each unique CustomerAltKey will have its count
calculated.
Filters the results to include only those CustomerAltKey values that appear more than once in the table,
identifying duplicates.
Question: 117
HOTSPOT
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
Answer: No
A scalar function returns a single value, but GetDirectReports likely returns multiple rows (i.e., a table)
containing details of employees reporting to a manager. This suggests it is a table-valued function (TVF)
rather than a scalar function.
2. The Dimension.GetDirectReports function will run only once when the query runs."
Answer: No
If GetDirectReports is a table-valued function, it executes multiple times—once for each relevant row—
rather than just once per query execution.
3. The output rows will include at least one row for each row in the Dimension.Employee table."
Answer:yes.
the output of the query using OUTER APPLY will include at least one row for each row in Dimension.employee,
even if the Dimension.getDirectReports function does not return any results for some rows.
Question: 118
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not
appear in the review screen.
A.Yes
B.No
Answer: B
Explanation:
Question: 119
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Requirements -
Planned Changes -
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
• The Research division data for Productline1 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.
• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Which syntax should you use in a notebook to access the Research division data for Productline1?
A.spark.read.format(“delta”).load(“Tables/ResearchProduct”)
B.spark.read.format(“delta”).load(“Files/ResearchProduct”)
C.external_table(‘Tables/ResearchProduct)
D.external_table(ResearchProduct)
Answer: A
Explanation:
The syntax spark.read.format("delta").load is specifically designed for loading Delta tables in Apache Spark.
Delta Lake is an optimized storage layer that brings reliability to data lakes.
The path "Tables/ResearchProduct" indicates that the data is organized within a directory structure named
"Tables," which aligns with best practices for storing structured data in a data lake. This structure is ideal for
accessing data in a Delta format.
Question: 120
You have a Microsoft Power BI Premium Per User (PPU) workspace that contains a semantic model.
You have an Azure App Service app named App1 that modifies row-level security (RLS) for the model by using the
XMLA endpoint.
App1 requires users to sign in by using their Microsoft Entra credentials to access the XMLA endpoint.
You need to configure App1 to use a service account to access the model.
Answer: D
Explanation:
Adding a managed identity to App1 will allow it to authenticate and access resources securely without
needing to manage credentials explicitly. This is the first step in configuring the app to use a service account
for accessing the model via the XMLA endpoint.
Question: 121
HOTSPOT
-
You have source data in a CSV file that has the following fields:
• SalesTransactionID
• SaleDate
• CustomerCode
• CustomerName
• CustomerAddress
• ProductCode
• ProductName
• Quantity
• UnitPrice
You plan to implement a star schema for the tables in WH1. The dimension tables in WH1 will implement Type 2
slowly changing dimension (SCD) logic.
You need to design the tables that will be used for sales transaction analysis and load the source data.
Which type of target table should you specify for the CustomerName, CustomerCode, and SaleDate fields? To
answer, select the appropriate options in the answer area.
Box 1: Dimension .
It belongs in the Customer Dimension Table along with CustomerName, CustomerAddress, etc.
Box 2 : Dimension .
Box 3:Fact.
Question: 122
You have a Fabric tenant.
You have a stored procedure that returns the number of active customers and their average sales for the current
month.
You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
available to the downstream activities of the pipeline.
D.Append variable
Answer: C
Explanation:
Question: 123
HOTSPOT
-
You have a Fabric tenant that contains a warehouse named Warehouse1. Warehouse1 contains a fact table named
FactSales that has one billion rows.
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
Yes
No
No
https://learn.microsoft.com/en-us/sql/t-sql/statements/create-table-as-clone-of-transact-sql?
view=fabric&preserve-view=true
Creates a new table as a zero-copy clone of another table in Warehouse in Microsoft Fabric. Only the
metadata of the table is copied. The underlying data of the table, stored as parquet files, is not copied.
https://learn.microsoft.com/en-us/fabric/data-warehouse/clone-table
Upon creation, a table clone is an independent and separate copy of the data from its source.
Any changes made through DML or DDL on the source of the clone table are not reflected in the clone
table.Similarly, any changes made through DDL or DML on the table clone are not reflected on the source of
the clone table.
Question: 124
Which two values should you provide? Each correct answer presents part of the solution.
Answer: BD
Explanation:
Secret Access Key (B): The secret access key is a private key used in conjunction with the access key ID to
Access Key ID (D): The access key ID is a public identifier for your AWS account. It is required to identify the
Question: 125
You have a Fabric tenant that contains a data warehouse.
You need to load rows into a large Type 2 slowly changing dimension (SCD). The solution must minimize resource
usage.
Answer: B
Explanation:
B: "MERGE."
MERGE Statement: The MERGE statement in T-SQL is specifically designed for handling situations where you
need to perform both insert and update operations based on certain conditions. It's particularly useful for
managing Type 2 Slowly Changing Dimensions (SCDs) in data warehouses, where historical data needs to be
preserved, and changes are tracked over time.
Resource Efficiency: Using the MERGE statement minimizes resource usage because it consolidates multiple
operations into a single statement. Instead of running separate UPDATE and INSERT commands, MERGE
allows you to update existing rows and insert new rows in one go. This reduces the complexity of your T-SQL
code and improves performance by reducing the number of operations required.
Question: 126
Case study -
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to
complete each case. However, there may be additional case studies and sections on this exam. You must manage
your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the
case study. Case studies might contain exhibits and other resources that provide more information about the
scenario that is described in the case study. Each question is independent of the other questions in this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to
make changes before you move to the next section of the exam. After you begin a new section, you cannot return
to this section.
Overview -
Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales and Research.
The Sales division contains two departments named Online Sales and Retail Sales. The Research division assigns
internally developed product lines to individual teams of researchers and analysts.
Existing Environment -
Identity Environment -
Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups named
ResearchReviewersGroup1 and ResearchReviewersGroup2.
Data Environment -
Requirements -
Planned Changes -
• Enable support for Fabric in the Power BI Premium capacity used by the Sales division.
• Make all the data for the Sales division and the Research division available in Fabric.
• For the Research division, create two Fabric workspaces named Productline1ws and Productline2ws.
• In Productline1ws, create a lakehouse named Lakehouse1.
• In Lakehouse1, create a shortcut to storage1 named ResearchProduct.
• All the workspaces for the Sales division and the Research division must support all Fabric experiences.
• The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.
• The Research division workspaces must be grouped together logically to support OneLake data hub filtering
based on the department name.
• For the Research division workspaces, the members of ResearchReviewersGroup1 must be able to read lakehouse
and warehouse data and shortcuts by using SQL endpoints.
• For the Research division workspaces, the members of ResearchReviewersGroup2 must be able to read lakehouse
data by using Lakehouse explorer.
• All the semantic models and reports for the Research division must use version control that supports branching.
• The Research division data for Productline2 must be retrieved from Lakehouse1 by using Fabric notebooks.
• All the Research division data in the lakehouses must be presented as managed tables in Lakehouse explorer.
• The number of rows added to the Orders table during refreshes must be minimized.
• The semantic models in the Research division workspaces must use Direct Lake mode.
General Requirements -
Contoso identifies the following high-level requirements that must be considered for all solutions:
Which syntax should you use in a notebook to access the Research division data for Productline1?
Answer: D
Explanation:
This syntax correctly specifies the format as Delta and loads the data from the specified table in the
lakehouse.
Question: 127
You have a Fabric tenant that contains a lakehouse named LH1.
You need to ensure that the tables are added automatically to the default semantic model.
Answer: B
Explanation:
B. From the settings pane of LH1, enable "Sync the default Power BI semantic model".
When you create new tables in a Microsoft Fabric Lakehouse (LH1), they are not automatically added to the
default Power BI semantic model unless synchronization is enabled.
Enable "Sync the default Power BI semantic model" from the settings pane of LH1.
This feature allows newly created tables in the Lakehouse to be automatically included in the Power BI
semantic model, ensuring seamless data integration for reporting and analysis
Question: 128
You have a Fabric tenant.
You have a stored procedure that returns the number of active customers and their average sales for the current
month.
You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
available to the downstream activities of the pipeline.
A.Append variable
B.Lookup
C.Copy data
D.KQL
Answer: B
Explanation:
B: "Lookup."
Lookup Activity: In Azure Data Factory (and its Fabric Data Factory counterpart), the Lookup activity is
designed to retrieve data from a data source and make the results available for downstream activities in the
pipeline. It is particularly useful when you need to fetch data from a stored procedure, SQL query, or any data
source and use that data within your pipeline.
Executing Stored Procedures: When you use the Lookup activity to execute a stored procedure, it will capture
the output of the stored procedure (in this case, the number of active customers and their average sales) and
store the results in a manner that can be used by subsequent activities in the pipeline. This makes it ideal for
scenarios where you need to perform additional operations based on the retrieved data.
Question: 129
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More
After you answer a question in this section, you will NOT be able to return. As a result, these questions do not
You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.
You query a measure that includes the USERNAME() function, and the query returns a blank result.
Does this meet the goal?
A.Yes
B.No
Answer: B
Explanation:
Question: 130
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More
than one solution in the set might solve the problem. It is also possible that none of the solutions in the set solve
the problem.
After you answer a question in this section, you will NOT be able to return. As a result, these questions do not
appear on the Review Screen.
Your network contains an on-premises Active Directory Domain Services (AD DS) domain named contoso.com that
syncs with a Microsoft Entra tenant by using Microsoft Entra Connect.
You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.
You query a measure that includes the USERNAME() function, and the query returns a blank result.
You need to ensure that the measure returns the user principal name (UPN) of a user.
A.Yes
B.No
Answer: A
Explanation:
Question: 131
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More
After you answer a question in this section, you will NOT be able to return. As a result, these questions do not
appear on the Review Screen.
Your network contains an on-premises Active Directory Domain Services (AD DS) domain named contoso.com that
syncs with a Microsoft Entra tenant by using Microsoft Entra Connect.
You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.
You query a measure that includes the USERNAME() function, and the query returns a blank result.
You need to ensure that the measure returns the user principal name (UPN) of a user.
Solution: You add user objects to the list of synced objects in Microsoft Entra Connect.
A.Yes
B.No
Answer: A
Explanation:
A. Yes.
Synced Objects in Microsoft Entra Connect: Adding user objects to the list of synced objects in Microsoft
Entra Connect means that those users will be included in the synchronization process between your on -
premises directory and Microsoft Entra ID (formerly Azure Active Directory). This allows for the seamless
integration and management of user identities across environments.
Goal Achievement: If the goal is to ensure that specific user objects are included in the synchronization
process, then adding them to the list of synced objects will meet this requirement. This ensures that the user
accounts and their associated attributes are correctly synchronized and available in Microsoft Entra ID.
Question: 132
You have a Fabric tenant that contains a warehouse named DW1 and a lakehouse named LH1. DW1 contains a table
named Sales.Product. LH1 contains a table named Sales.Orders.
You plan to schedule an automated process that will create a new point-in-time (PIT) table named
Sales.ProductOrder in DW1. Sales.ProductOrder will be built by using the results of a query that will join
Sales.Product and Sales.Orders.
You need to ensure that the types of columns in Sales.ProductOrder match the column types in the source tables.
The solution must minimize the number of operations required to create the new table.
A.INSERT INTO
B.CREATE TABLE AS SELECT (CTAS)
C.CREATE TABLE AS CLONE OF
D.CREATE MATERIALIZED VIEW AS SELECT
Answer: B
Explanation:
Question: 133
HOTSPOT
-
You have a Fabric tenant that contains the semantic model shown in the following exhibit.
Use the drop-down menus to select the answer choice that completes each statement based on the information
presented in the graphic.
NOTE: Each correct selection is worth one point.
Answer:
Explanation:
1. Querying SQL Views from the Warehouse Causes a Fallback to "DirectQuery" Mode
In Microsoft Fabric, the default query mode for a semantic model in a Lakehouse is Direct Lake, which
provides fast, real-time querying on delta tables.
However, when querying SQL Views from the Warehouse, the system cannot use Direct Lake because SQL
Views are not stored as delta tables.
As a result, it falls back to DirectQuery mode, where queries are executed on demand against the source
database.
In Direct Lake mode, row-level security (RLS) and object-level security (OLS) are not yet supported in
Microsoft Fabric.
Since DirectQuery is being used (due to the SQL View fallback), security settings are undefined because they
are not automatically applied.
To enforce security, you must configure RLS or OLS manually in Power BI or SQL Warehouse.
Question: 134
Note: This section contains one or more sets of questions with the same scenario and problem. Each question
presents a unique solution to the problem. You must determine whether the solution meets the stated goals. More
than one solution in the set might solve the problem. It is also possible that none of the solutions in the set solve
the problem.
After you answer a question in this section, you will NOT be able to return. As a result, these questions do not
appear on the Review Screen.
Your network contains an on-premises Active Directory Domain Services (AD DS) domain named contoso.com that
syncs with a Microsoft Entra tenant by using Microsoft Entra Connect.
You enable dynamic row-level security (RLS) for the model and deploy the model to the Fabric service.
You query a measure that includes the USERNAME() function, and the query returns a blank result.
You need to ensure that the measure returns the user principal name (UPN) of a user.
A. Yes
B.No
Answer: B
Explanation:
Creating role will not solve the purpose , will have to use Userprincipalname().
Question: 135
You need to ensure that User1 can perform the following tasks:
domain admin
domain
Answer: C
Explanation:
C .Fabric admin.
Comprehensive Permissions: The Fabric admin role has the necessary permissions to create and manage
domains, subdomains, and assign workspaces to them. This role encompasses the administrative capabilities
required for the tasks you mentioned.
Least Privilege Principle: While the Fabric admin role provides broad permissions, it is necessary in this
context because the tasks involve high-level administrative actions that cannot be accomplished with more
restricted roles. Thus, it adheres to the principle of least privilege by granting just enough access to perform
the required tasks without excessive permissions.
Question: 136
HOTSPOT
-
You have a Fabric tenant that contains three users named User1, User2, and User3. The tenant contains a security
group named Group1. User1 and User3 are members of Group1.
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
No, Yes, No
No. User2 is not a member of Group1 and has no direct association with Domain1 or Workspace3. There's no
indication User2 has inherited any role here.
Yes. User3 is a member of Group1, which is associated with Domain1 because it was created by User 1. When
Workspace3 is created under Domain1, User3 likely inherits the Viewer role.
No. User3 is a member of Group1, but Workspace1 is not linked to Domain1. It's directly tied to User1 as the
admin. There's no evidence User3 has any Contributor role in Workspace1.
Question: 137
You have a Fabric warehouse named Warehouse1 that contains a table named Table1. Table1 contains customer
Which two objects should you create? Each correct answer presents part of the solution.
A.DATABASE ROLE
B.STORED PROCEDURE
C.CONSTRAINT
D.FUNCTION
POLICY
Answer: DE
Explanation:
D. FUNCTION.
E. SECURITY POLICY.
FUNCTION (D): Used to define a security function that filters the rows of the table based on the user's identity
or specific criteria.
SECURITY POLICY (E): Applies the security function to the table to restrict access to rows according to the
defined criteria.
Question: 138
You have a Fabric workspace named Workspace1 that contains a lakehouse named Lakehouse1. Lakehouse1
contains a table named Table1. Table1 contains the following data.
Answer: A
Explanation:
ProductColor: This is another attribute describing the product. It adds further context and can be useful in
analyzing product variations and performance by color.
ProductID: This is a unique identifier for each product. It serves as the primary key in the dimension table and
will be used to join with the fact table.
ProductName: This provides a descriptive name for each product, which is useful for reporting and analysis.
Question: 139
DRAG DROP
-
You have a Fabric workspace that contains a Dataflow Gen2 query. The query returns the following data.
You need to filter the results to ensure that only the latest version of each customer’s record is retained. The
solution must ensure that no new columns are loaded to the semantic model.
Which four actions should you perform in sequence in Power Query Editor? To answer, move the appropriate
actions from the list of actions to the answer area and arrange them in the correct order.
Answer:
Explanation:
1. Group by Customer ID and calculate the max version date per customer ID.
This is the first step to identify the most recent version date for each customer.
2. Filter the query where the version date value equals the max version date value.
This ensures that only the rows with the most recent version date are included.
This step ensures that each customer appears only once in the final dataset.
Question: 140
HOTSPOT
-
You have a Microsoft Power Bl project that contains a file named definition.pbir. definition.pbir contains the
following JSON.
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
Answer: No
The .pbir file format is a new format used in Power BI for report definitions, but it is not part of the legacy PBIR
format. The legacy format refers to older Power BI report types.
Answer: No
The .pbir file format does not inherently mean that the semantic model is stored in the Power BI service. It can
reference models from different sources, including local files or cloud services.
When the related report is opened, Power BI Desktop will open the semantic model in full edit mode.
Answer: Yes
Power BI Desktop allows full editing of the report and its semantic model when opening a .pbir file, assuming
the appropriate permissions and connections are available.
Question: 141
You need to optimize the deployment of Model1. The solution must minimize how long it takes to deploy Model1.
Answer: D
Explanation:
Large semantic model storage format allows for efficient handling of larger datasets and optimized
Question: 142
You have a Microsoft Power BI project that contains a semantic model.
You need to modify the .gitignore file to prevent the data values from the data sources from being pushed to the
Answer: B
Explanation:
B. cache .abf Justification: This file contains the cached data for the semantic model. It is a binary file and can
be very large. Excluding this file from version control is necessary to prevent pushing large amounts of
unnecessary data, and will prevent data from data sources from being pushed to the repository.Why it's the
correct option: Contains the cached data from the data model.
Question: 143
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 uses Pro license mode and
contains a semantic model named Model1.
Answer: D
Explanation:
The license mode of the workspace impacts the availability of XMLA endpoints. By using a Pro license mode,
you might not have full support for XMLA connections1. Switching to a Premium license mode would enable
XMLA endpoints, allowing for more advanced connectivity options, including XMLA connections.
Question: 144
You need to visualize a diagram of the model. The diagram must contain only the Sales table and related tables.
Data view
Answer: C
Explanation:
Question: 145
HOTSPOT
-
You need to deploy a new semantic model. The solution must meet the following requirements:
• Support complex calculated columns that include aggregate functions, calculated tables, and Multidimensional
Expressions (MDX) user hierarchies.
• Minimize page rendering times.
How should you configure the model? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
Mode: Import.
Question: 146
HOTSPOT
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
This means the query filters out sales records where the status is "Cancelled."
The query calculates the total sales of each product category for the last 30 days. Yes
The query is aggregating sales data based on product categories within the last 30 days.
The query includes product categories that have had zero sales during the last 30 days. No
This indicates that product categories with no sales in the last 30 days are not included in the results.
Question: 147
You need to calculate the proportion of the total quantity of each sales order represented by each product within
A.
B.
C.
D.
Answer: D
Explanation:
This divides each row’s OrderQty by the total OrderQty for the SalesOrderID to calculate the proportion.
Question: 148
HOTSPOT
You need build a T-SQL statement that will return the total sales amount by OrderDate only for the days that are
holidays in Australia. The total sales amount must sum the quantity multiplied by the price on each row in the
dbo.sales table.
How should you complete the statement? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
For the type of Joins, they only want it if it is Australia. Inner will only bring over the matches, making it an
ideal Join Type. To be fair, I had to look it up, but once I learned that I wanted to share.
Question: 149
HOTSPOT
-
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains a warehouse named
DW1. DW1 contains two tables named Employees and Sales. All users have read access to Dw1.
• For the Sales table, ensure that the users can see only the sales data from their respective region.
• For the Employees table, restrict access to all Personally Identifiable Information (PII).
• Maintain access to unrestricted data for all the users.
What should you use for each table? To answer, select the appropriate options in the answer area.
Answer:
Explanation:
Column-Level-Security (CLS):
https://learn.microsoft.com/en-us/fabric/data-warehouse/column-level-security
https://learn.microsoft.com/en-us/fabric/data-warehouse/row-level-security
"control access to rows in a database table. For example, you can ensure that workers access only those data
rows that are pertinent to their department." (or Sales Region in this case)
Question: 150
You need to create a Microsoft Power BI file that will be used to create multiple reports. The solution must meet
A.PBIT
B.PBIDS
C.PBIX
D.PBIP
Answer: A
Explanation:
Includes predefined data source connections – PBIT files store connection details and queries.
Includes the report structure and formatting – PBIT files retain visuals, layouts, and report configurations.
Does NOT contain any data – When opened, a PBIT file prompts the user to refresh the data from the
connected sources.
Question: 151
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains a single semantic
model that has two Microsoft Power BI reports.
You have a Microsoft 365 subscription that contains a data loss prevention (DLP) policy named DLP1.
Answer: C
Explanation:
Sensitivity labels classify Power BI assets (semantic models, reports) and apply DLP policies.
Microsoft 365 DLP policies work with sensitivity labels to enforce security.
A label like "Confidential – Internal Use" could trigger a DLP rule to prevent external sharing.
Question: 152
You have a Fabric tenant that contains a workspace named Workspace1 and a user named User1. User1 is assigned
You plan to configure Workspace1 to use an Azure DevOps repository for version control.
Which two settings should you enable for User1? Each correct answer presents part of the solution.
Answer: CD
Explanation:
To commit items to a repository, User1 must be able to create Fabric items in the workspace.
Question: 153
\Lakehouse1
Answer: D
Explanation:
D. \Folder1\Lakehouse1.
Question: 154
Your company has a finance department.
You have a Fabric tenant, an Azure Storage account named storage1, and a Microsoft Entra group named Group1.
Group1 contains the users in the finance department.
You need to create a new workspace named Workspace1 in the tenant. The solution must meet the following
requirements:
• Ensure that the finance department users can create and edit items in Workspace1.
• Ensure that Workspace1 can securely access storage1 to read and write data.
• Ensure that you are the only admin of Workspace1.
• Minimize administrative effort.
Which two actions should you perform next? Each correct answer presents part of the solution.
Answer: AB
Explanation:
A. Assign the Contributor role to Group1:Since Group1 already contains all finance department users,
assigning the Contributor role to the group grants them the necessary permissions (create and edit items) in
Workspace1 with minimal administrative effort. This avoids having to assign roles to each user individually.B.
Create a workspace identity:A workspace identity (or managed identity) allows Workspace1 to securely
access external resources like storage1. By creating and configuring a workspace identity, you can grant it the
appropriate permissions on storage1 for secure read and write access.Additionally, because you created the
workspace, you are automatically the admin—meeting the requirement that you are the only admin of
Workspace1.
Question: 155
You have a deployment pipeline named Pipeline1 that deploys items from Workspace_DEV to Workspace_TEST. In
Pipeline1, all items that have matching names are paired.
A.Lakehouse1 -
Lakehouse2 -
Notebook1 -
Notebook2 -
Pipeline1 -
SemanticModel1
B.Lakehouse1 -
Notebook1 -
Pipeline1 -
SemanticModel1
C.Lakehouse2 -
Notebook2 -
SemanticModel1
D.Lakehouse2 -
Notebook2 -
Pipeline1 -
SemanticModel1
Answer: A
Explanation:
Lakehouse1 -
Lakehouse2 -
Notebook1 -
Notebook2 -
When deploying contents from Workspace_DEV to Workspace_TEST using Pipeline1 where matching names
are paired, the resulting contents in Workspace_TEST will include all the items from Workspace_DEV and any
additional items from Workspace_TEST that do not have a matching name.
Question: 156
HOTSPOT
-
You have a Fabric tenant that contains a workspace named Workspace_DEV. Workspace_DEV contains the
semantic models shown in the following table.
You create a deployment pipeline named Pipeline1 to move items from Workspace_DEV to Workspace_TEST.
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Explanation:
This suggests that only the structure or metadata of Model1 is deployed, not the actual data.
The scheduled refresh policy for Model1 will be deployed to Workspace_TEST :No
This indicates that the refresh policy of Model1 is not carried over to the new workspace.
Question: 157
HOTSPOT
You need to compare the average miles per trip for statutory holidays versus non-statutory holidays.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.
Answer:
Question: 158
You have a Fabric tenant.
You have a stored procedure that returns the number of active customers and their average sales for the current
month.
You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be
available to the downstream activities of the pipeline.
A.Get metadata
B.Switch
C.Lookup
D.Append variable
Answer: C
Explanation:
C. Lookup.
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
Answer:
Explanation:
The statement returns Region values when a Sales item has a RefundStatus of Refunded: No
This means that the query does not filter Region values based on RefundStatus.
The statement only returns TransactionDate values that occurred during the current year: Yes
Question: 160
HOTSPOT
-
You need to create a view to denormalize the data into a customer dimension that contains one row per distinct
CustomerID value. The solution must minimize query processing time and resources.
How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.
inner join is used to filter only matching records across the tables.
Question: 161
A.DateTime.LocalNow
B.Duration.Days
C.Duration.From
D.Date.AddDays
Answer: B
Explanation:
B. Duration.Days.
This function will calculate the duration (difference in days) between two date values.
Duration.Days([Shipping_Date] - [Order_Date])
Question: 162
DRAG DROP
-
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 uses the Pro license mode
and contains a semantic model named Model1.
You need to enable version control for Workspace1. The solution must ensure that Model1 is added to the
repository.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of
actions to the answer area and arrange them in the correct order.
Answer:
Explanation:
After connecting and syncing, configuring branch policies helps maintain quality by enforcing rules, such as
requiring code reviews before merging changes.
Question: 163
HOTSPOT
-
You have a Fabric tenant that contains a workspace named Enterprise. Enterprise contains a semantic model
named Model1. Model1 contains a date parameter named Date1 that was created in Power Query.
You build a deployment pipeline named Enterprise Data that includes two stages named Development and Test.
You assign the Enterprise workspace to the Development stage.
• Create a workspace named Enterprise [Test] and assign the workspace to the Test stage.
• Configure a rule that will modify the value of Date1 when changes are deployed to the Test stage.
Which two settings should you use? To answer, select the appropriate settings in the answer area.
Answer:
Question: 164
HOTSPOT
-
You have two Microsoft Power BI queries named Employee and Retired Roles.
You need to merge the Employee query with the Retired Roles query. The solution must ensure that duplicate rows
in each query are removed.
Which column and Join Kind should you use in Power Query Editor? To answer, select the appropriate options in the
answer area.
The Employee table has multiple entries in the Division column with the value "Services".
The Retired Roles table also contains multiple entries under the Division column with "Services".
The Join Kind selection options are visible, and the user has highlighted "Full Outer (all rows from both)" and
"Inner (only matching rows)".
Keeps only the rows that have a match in both tables based on the Division column.
Since "Services" exists in both tables, only those matching rows will be included in the final result.
D. From the Microsoft 365 admin center, modify the role assignment of each user.
Answer: C
Explanation:
To block specific users from creating Fabric items, you first need to group those users together. The best way
to do this is by creating a security group in the Microsoft 365 admin center.
Once the security group is created, you can go to the Fabric admin portal and modify the creation
permissions by restricting Fabric item creation to specific users or excluding certain groups.
Questions 2
You need to allow many users to view multiple reports and dashboards developed in Workspace1. The solution
must follow the principle of least privilege and minimize administrative effort.
B. item-level permissions
Explanation:
A Microsoft Power BI app is the best choice because it allows many users to access reports and dashboards
without needing direct workspace access.
Apps simplify distribution by providing a read-only experience, following the principle of least privilege.
It also minimizes administrative effort since you can share the app with many users at once rather than
managing individual permissions.
"item-level permissions"
Managing permissions at the item level (report or dashboard) for many users is time-consuming and complex.
"the Member workspace role"
The Member role grants more privileges than necessary, including the ability to edit content. This violates
the least privilege principle.
"the Viewer workspace role"
The Viewer role allows read-only access, but users must be assigned directly to the workspace. Managing
permissions for a large number of users this way is inefficient.
Question 3:
You need to control which users can create Fabric items in all the workspaces assigned to Fabric capacities.
A. Capacity settings
B. Item-level settings
C. Lakehouse settings
D. Tenant settings
E. Workspace settings
"Tenant settings"
Explanation:
Tenant settings in the Fabric admin portal allow you to control who can create Fabric items across all
workspaces assigned to Fabric capacities.
This is the correct choice because item creation is a global setting that applies across the tenant, rather than
at a workspace or capacity level.
"Capacity settings"
Capacity settings manage resource allocation and performance, not user permissions for item creation.
"Item-level settings"
These settings control permissions on individual Fabric items (like reports or dashboards), not overall
creation rights.
"Lakehouse settings"
These are specific to managing Lakehouses, not general Fabric item creation.
"Workspace settings"
Workspace settings control permissions within a specific workspace but do not govern item creation tenant-
wide.
Question 4:
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity. Workspace 1 contains the following items:
· Report1
· SemanticModel1
· Lakehouse1
· Dashboard1
A. Dashboard1 only
B. SemanticModel1 only
C. Lakehouse1 only
F. Report1 only
Explanation:
In Microsoft Fabric, you can endorse certain items to highlight their quality and reliability. The two types of
endorsements are Promoted and Certified.
Question 5:
A user creates and shares a report that has default permissions to all the users in an organization.
Which actions can all the users perform with the shared report?
D. view only
Explanation:
When a report is shared in Microsoft Fabric (Power BI) with default permissions to all users in the
organization:
"view and modify only" → Users cannot modify the report unless they are explicitly given edit
permissions.
"view, modify, and share" → Users cannot modify the report by default.
"view only" → By default, users can share the report with others unless sharing is restricted.
Question 6:
You have an on-premises Microsoft SQL server that contains a database named DB1.
You plan to extract data from DB1 by using Data Factory and store the data in Lakehouse1.
A. a personal gateway
B. a shortcut
C. an on-premises gateway
D. mirroring
Explanation:
Since DB1 is an on-premises SQL Server database, you need a secure way to connect Fabric’s Data Factory
to it. The best solution is an on-premises data gateway, which allows Fabric to securely access on-premises
data sources.
Question 7:
You plan to use the Delete data activity to remove the existing data in Lakehouse1.
A. a dataflow
B. a lookup
C. a pipeline
D. a shortcut
"a pipeline"
Explanation:
In Microsoft Fabric, the Delete data activity is part of Data Factory and is used in pipelines to remove data
from a Lakehouse or other storage.
To use the Delete data activity, you must first create a pipeline and then configure the activity within it.
"a dataflow"
Dataflows are used for data transformation and ingestion, not for deleting data.
"a lookup"
Lookups retrieve data for use in other pipeline activities but do not delete data.
"a shortcut"
Shortcuts provide virtual access to external storage but do not support data deletion.
Question 8:
You plan to create a pipeline that will download a comma-separated dataset from a public GitHub repository.
You need to specify which request method to use when defining the connection to the data source.
CONNECT
GET
OPTIONS
PUT
"GET"
Explanation:
The GET request method is used to retrieve data from a web source, such as downloading a CSV file from a
public GitHub repository.
Since you are downloading data (not modifying or sending data), GET is the appropriate choice.
Question 9:
You plan to use a Data Factory pipeline to ingest data from multiple supported data sources, transform the
data, and then merge the data before saving the results to Warehouse1.
B. Dataflow activity
C. Notebook activity
"Dataflow activity"
Explanation:
The Dataflow activity in Data Factory is designed for ingesting, transforming, and merging data from
multiple sources before loading it into Warehouse1.
It provides a visual, no-code interface to apply transformations like joins, aggregations, and mappings
before writing to the warehouse.
You plan to use a Data Factory pipeline to create a repeatable process to ingest Parquet files stored on a
network drive and load them to Lakehouse1 as a Delta table without any transformations.
C. Lookup activity
D. Notebook activity
Explanation:
The Copy data activity in Data Factory is the best choice when you need to ingest Parquet files from a
network drive and load them into Lakehouse1 as a Delta table, without transformations.
This activity is designed for efficient ETL (Extract, Load) operations, supporting direct movement of data
from various sources to destinations like Lakehouses, Warehouses, and Delta tables.
Question 11:
You have a Fabric dataflow named Dataflow1 that contains a column named LastName with the Text data
type.
You notice that the LastName column contains several NULL entries and several blank entries.
You need to ensure that Dataflow1 contains only rows where the LastName column contains text.
Explanation:
In Power Query, the Remove empty command is used to filter out rows where a column contains NULL or
blank values.
Since your LastName column has both NULL and blank entries, using Remove empty ensures that only rows
with actual text values remain.
Question 12:
You create a Dataflow Gen2 activity that contains several aggregation transformations.
You need to view the M code that represents the transformations by using the Power Query Online interface.
B. Diagram View
C. Queries pane
Explanation:
The Advanced editor window in Power Query Online allows you to view and edit the M code that represents
the transformations in your Dataflow Gen2 activity.
Since you have aggregation transformations, the M code captures these steps in a structured format.
Question 13:
You plan to use Dataflow Gen2 to merge several tables into a single table, and then save the table to
Lakehouse1 as a Delta table.
B. Diagram View
C. Queries pane
"Diagram View"
Explanation:
Diagram View in Dataflow Gen2 provides a visual representation of all transformations applied to your data.
Since you are merging several tables and performing transformations before saving the final table as a Delta
table in Lakehouse1, Diagram View allows you to see the dataflow structure and transformations step by
step.
Question 14:
· Employee
· Region
· Amount Sold
Each employee has multiple regions and an associated value in the Amount Sold column.
· Columns must represent the Region with aggregate Amount Sold for each Employee per Region.
A. Group by
B. Pivot column
C. Split column
D. Transpose
"Pivot column"
Explanation:
The Pivot column function in Power Query converts row values into column headers, which is exactly what
you need:
Each row should represent a single Employee.
Each Region should become a column.
The values in the new columns should be the aggregated Amount Sold per Region.
"Group by"
Groups data but does not transform row values into columns. It would summarize data but not create
separate columns for each region.
"Split column"
Used to divide a column into multiple columns, typically by delimiter or position, but not for aggregation.
"Transpose"
Swaps rows and columns entirely, which is not suitable for this scenario.
Question 15:
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity and contains a KQL database named KQL1. KQL1 contains a table named Sales that has the following
columns:
· SalesAmount: decimal data type.
Explanation:
In Kusto Query Language (KQL), the summarize operator is used to aggregate data.
The correct syntax for calculating total SalesAmount grouped by Region is:
kql
CopyEdit
Sales
This statement:
Groups the data by Region
Calculates the total SalesAmount for each Region
Names the result as TotalSales
Question 16:
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity and contains a KQL database named KQL1. KQL1 contains a table named Sales.
· OrderId
· OrderDate
Explanation:
The project operator in Kusto Query Language (KQL) is used to select specific columns from a table.
The correct syntax to display only OrderKey, OrderId, and OrderDate is:
kql
CopyEdit
Sales
This ensures that only the specified columns are included in the output.
Question 17:
You have a Fabric workspace named Workspace1 that contains a lakehouse named Lakehouse1.
For Model1, you need to create a table named Table1 that will support incremental refresh and will be
partitioned to meet the following requirements:
Explanation:
Incremental refresh and hybrid tables (which combine import mode for historical data and DirectQuery for
real-time data) require Power BI Premium capacity.
Power BI Pro does NOT support incremental refresh with DirectQuery partitions, so assigning Premium
capacity is necessary to enable this functionality.
Question 18:
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 is assigned to a Fabric
capacity and contains a lakehouse named Lakehouse1.
You plan to use the SQL analytics endpoint of Lakehouse1 to develop a semantic model. The solution must
meet the following requirements:
A. Direct Lake
B. DirectQuery
C. Dual
D. Import
The correct answer is Import mode because it meets all the specified requirements.
1. Supports scheduled refresh at 6 AM – Import mode loads data into memory and relies on scheduled
refreshes to update the model.
2. Works with all DAX and M functions – Unlike Direct Lake and DirectQuery, Import mode fully
supports advanced DAX calculations and M functions.
3. Allows calculated tables – Calculated tables are only available in Import mode, making it the only
viable choice.
4. Prevents multiple source group issues – Import mode loads data from a single source, ensuring
consistency.
5. Uses only one storage mode – Import mode is a single-mode storage option, fulfilling the
requirement.
Direct Lake and DirectQuery do not support calculated tables or full DAX/M functionality. Dual mode allows
switching between Import and DirectQuery, which does not align with the "one storage mode" requirement.
Question 19:
You plan to develop a custom semantic model using Lakehouse1 that will provide up-to-date inventory levels
by querying live transactional data from SQLDB1.
You need to ensure that all the queries that use the model target live transactional data.
Direct Lake
DirectQuery
Dual
Import
DirectQuery
Explanation:
Since you need to provide up-to-date inventory levels by querying live transactional data from SQLDB1, the
best storage mode is DirectQuery.
1. DirectQuery ensures real-time data access – Queries retrieve data directly from SQLDB1 instead of
relying on a cached dataset.
2. Best for live transactional data – Since inventory data updates in real time, DirectQuery ensures that
every query reflects the latest data.
3. Avoids scheduled refresh limitations – Unlike Import mode, which requires periodic refreshes,
DirectQuery always fetches data from the source.
Direct Lake – Works only with OneLake and Delta tables, but SQLDB1 is an Azure SQL database, not a
lakehouse.
Import – Stores data in memory and requires scheduled refreshes, meaning the data would be stale
between refreshes.
Dual – Allows both Import and DirectQuery but is used for optimized performance, not for ensuring 100%
real-time queries.
Question 20:
You create several tables in Lakehouse1. You load data into the tables.
You need to define relationships for the default semantic model in Lakehouse1.
B. From the Fabric lakehouse view, disable the Sync the default Power BI semantic model option.
C. From the Microsoft Fabric portal, switch to the Model view of the Data Warehouse experience.
D. From the Microsoft Fabric portal, switch to the Query view of the Data Warehouse experience.
Explanation:
To define relationships in the default semantic model of Lakehouse1, you need to use Power BI's model view,
which requires connecting to the SQL analytics endpoint.
1. SQL analytics endpoint enables querying and modeling – It provides a SQL-compatible interface to
work with Lakehouse data, allowing you to define relationships.
2. Power BI automatically syncs with the semantic model – Once connected, you can define
relationships in Power BI Model view.
"Disable the Sync the default Power BI semantic model option" – This would prevent Power BI from
automatically maintaining the semantic model, which is the opposite of what you need.
"Switch to the Model view of the Data Warehouse experience" – Lakehouse1 is a lakehouse, not a Data
Warehouse, so this option does not apply.
"Switch to the Query view of the Data Warehouse experience" – Query view is for writing SQL queries,
not for defining relationships in a semantic model.
Question 21:
You use Microsoft Power BI Desktop to generate and display visuals within a report.
You plan to verify your original results by rerecording the performance by using the same visuals.
Explanation:
When using Performance Analyzer in Power BI Desktop, you need to ensure that the second recording is
measuring the same operations as the first one.
1. Refreshing the visuals ensures a new performance measurement – Power BI does not re-execute
queries unless the visuals are refreshed.
2. Ensures comparable results – If you don’t refresh, the data might be cached, making the second
measurement faster and not a true comparison.
3. Prepares the visuals for a fresh performance capture – Refreshing forces Power BI to reprocess
queries and render visuals from scratch, ensuring consistency.
"Close the report." – This isn’t necessary and could lose context from the first test.
"Customize optimization presets." – While useful for tuning performance, it does not ensure that the
second recording is identical to the first.
"Pause the visuals." – This prevents visuals from updating, which is the opposite of what you need.
Question 22:
You use Microsoft Power BI Desktop to generate and display DAX query-based visuals in a report.
From Performance analyzer, you review the statistics of different DAX queries.
You need to determine the smallest threshold above which the DAX queries execution time warrants further
investigation.
A. 1.2 ms
B. 12 ms
C. 120 ms
D. 1200 ms
"120 ms"
Explanation:
When analyzing DAX query performance in Power BI using Performance Analyzer, a good rule of thumb is:
1.2 ms – Extremely low, even simple operations take more time than this.
12 ms – Insignificant; even well-optimized queries can take longer.
1200 ms (1.2 sec) – Too high; waiting until this threshold could result in severe performance issues before
taking action.
Question 23:
You plan to identify opportunities for performance improvements by using Performance analyzer.
Which two features in Performance analyzer can help you optimize performance? Each correct answer
presents part of the solution.
A. data model
B. data source
C. environment
D. report visuals
Data model
Report visuals
Explanation:
Performance Analyzer in Power BI helps identify bottlenecks and optimization opportunities in reports. The
two key areas it helps with are:
Data source – While an inefficient data source can impact performance, Performance Analyzer does not
directly analyze the data source itself. It focuses on queries, visuals, and model performance.
Environment – Performance Analyzer does not assess system performance, hardware, or network
conditions.
Question 24:
You use Microsoft Power BI Desktop to generate and display DAX query-based visuals.
DIVIDE (
A. DATESINPERIOD
B. nested calculations
C. SUMX
D. variables
Variables
Explanation:
RETURN
DATESINPERIOD – This function works similarly to PARALLELPERIOD, but it does not inherently improve
performance.
Nested calculations – Nesting calculations increases complexity and can make measures slower, rather
than optimizing them.
SUMX – SUMX is useful for row-by-row calculations but is not needed in this case because
PARALLELPERIOD returns an aggregated value.
Question 25:
You use Microsoft Power BI Desktop to generate and display DAX query-based visuals.
You plan to use Power BI Desktop Performance analyzer to determine how long it takes to load a visual
included in a report.
You need to eliminate any impact of the visual cache on your performance analysis.
Explanation:
When using Performance Analyzer in Power BI Desktop, cached results can distort performance
measurements. To ensure that each query runs fresh, you need to force Power BI to re-execute the DAX
query rather than retrieving results from cache.
Add a blank page to the report – This does not refresh or impact the query execution of the original visual.
Refresh the data model – Refreshing the data model reloads all data from the source, which is
unnecessary when only testing visual performance.
Reopen the Performance Analyzer pane – Closing and reopening the pane does not reset cached results
or force query execution.