Az-204 0
Az-204 0
https://www.2passeasy.com/dumps/AZ-204/
NEW QUESTION 1
- (Topic 8)
The solution must receive and store messages until they can be processed. You create an Azure Service Bus instance by providing a name, pricing tier,
subscription, resource group, and location.
You need to complete the configuration.
Which Azure CLI or PowerShell command should you run?
A)
B)
C)
D)
A. Option A
B. Option B
C. Option C
D. Option D
Answer: C
NEW QUESTION 2
- (Topic 8)
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the
stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You develop an HTTP triggered Azure Function app to process Azure Storage blob data. The app is triggered using an output binding on the blob.
The app continues to time out after four minutes. The app must process the blob data. You need to ensure the app does not time out and processes the blob data.
Solution: Update the functionTimeout property of the host.json project file to 10 minutes. Does the solution meet the goal?
A. Yes
B. No
Answer: B
Explanation:
Instead pass the HTTP trigger payload into an Azure Service Bus queue to be processed by a queue trigger function and return an immediate HTTP success
response.
Note: Large, long-running functions can cause unexpected timeout issues. General best practices include:
Whenever possible, refactor large functions into smaller function sets that work together and return responses fast. For example, a webhook or HTTP trigger
function might require an acknowledgment response within a certain time limit; it's common for webhooks to require an immediate response. You can pass the
HTTP trigger payload into a queue to be
processed by a queue trigger function. This approach lets you defer the actual work and return an immediate response.
Reference:
https://docs.microsoft.com/en-us/azure/azure-functions/functions-best-practices
NEW QUESTION 3
- (Topic 8)
You are building a web application that performs image analysis on user photos and returns metadata containing objects identified. The image is very costly in
terms of time and compute resources. You are planning to use Azure Redis Cache so duplicate uploads do not need to be reprocessed.
In case of an Azure data center outage, metadata loss must be kept to a minimum. You need to configure the Azure Redis cache instance.
Which two actions should you perform?
Answer: BC
NEW QUESTION 4
DRAG DROP - (Topic 8)
You manage several existing Logic Apps.
You need to change definitions, add new logic, and optimize these apps on a regular basis. What should you use? To answer, drag the appropriate tools to the
correct functionalities.
Each tool may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
NOTE:Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Enterprise Integration Pack
After you create an integration account that has partners and agreements, you are ready to create a business to business (B2B) workflow for your logic app with
the Enterprise Integration Pack.
Box 2: Code View Editor
To work with logic app definitions in JSON, open the Code View editor when working in the Azure portal or in Visual Studio, or copy the definition into any editor
that you want.
Box 3: Logical Apps Designer
You can build your logic apps visually with the Logic Apps Designer, which is available in the Azure portal through your browser and in Visual Studio.
References:
https://docs.microsoft.com/en-us/azure/logic-apps/logic-apps-enterprise-integration-b2b https://docs.microsoft.com/en-us/azure/logic-apps/logic-apps-author-
definitions https://docs.microsoft.com/en-us/azure/logic-apps/logic-apps-overview
NEW QUESTION 5
HOTSPOT - (Topic 8)
You develop and deploy the following staticwebapp.config.json file to the app_location value specified in the workflow file of an Azure Static Web app.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
NEW QUESTION 6
HOTSPOT - (Topic 8)
You are developing a .NET application that communicates with Azure Storage. A message must be stored when the application initializes.
You need to implement the message.
How should you complete the code segment? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
NEW QUESTION 7
HOTSPOT - (Topic 8)
You have an Azure Web app that uses Cosmos DB as a data store. You create a CosmosDB container by running the following PowerShell script:
$resourceGroupName = "testResourceGroup"
$accountName = "testCosmosAccount"
$databaseName = "testDatabase"
$containerName = "testContainer"
$partitionKeyPath = "/EmployeeId"
$autoscaleMaxThroughput = 5000 New-AzCosmosDBSqlContainer
-ResourceGroupName $resourceGroupName
-AccountName $accountName
-DatabaseName $databaseName
-Name $containerName
-PartitionKeyKind Hash
-PartitionKeyPath $partitionKeyPath
-AutoscaleMaxThroughput $autoscaleMaxThroughput You create the following queries that target the container:
SELECT * FROM c WHERE c.EmployeeId > '12345' SELECT * FROM c WHERE c.UserID = '12345'
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE:Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: No
You set the highest, or maximum RU/s Tmax you don't want the system to exceed. The system automatically scales the throughput T such that 0.1* Tmax <= T <=
Tmax.
In this example we have autoscaleMaxThroughput = 5000, so the minimum throughput for the container is 500 R/Us.
Box 2: No
First query:SELECT * FROM c WHERE c.EmployeeId > '12345'
Here's a query that has a range filter on the partition key and won't be scoped to a single physical partition. In order to be an in-partition query, the query must
have an equality filter that includes the partition key:
SELECT * FROM c WHERE c.DeviceId > 'XMS-0001'
Box 3: Yes
Example of In-partition query:
Consider the below query with an equality filter on DeviceId. If we run this query on a container partitioned on DeviceId, this query will filter to a single physical
partition.
SELECT * FROM c WHERE c.DeviceId = 'XMS-0001'
NEW QUESTION 8
- (Topic 8)
You develop a REST API. You implement a user delegation SAS token to communicate
with Azure Blob storage.
The token is compromised. You need to revoke the token.
What are two possible ways to achieve this goal? Each correct answer presents a complete solution.
NOTE: Each correct selection is worth one point.
Answer: AB
Explanation:
A: Revoke a user delegation SAS
To revoke a user delegation SAS from the Azure CLI, call the az storage account revoke- delegation-keys command. This command revokes all of the user
delegation keys associated with the specified storage account. Any shared access signatures associated with those keys are invalidated.
B: To revoke a stored access policy, you can either delete it, or rename it by changing the signed identifier.
Changing the signed identifier breaks the associations between any existing signatures and the stored access policy. Deleting or renaming the stored access
policy immediately effects all of the shared access signatures associated with it. D18912E1457D5D1DDCBD40AB3BF70D5D
Reference:
https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/storage/blobs/storage-blob-user-delegationsas-create-cli.md
https://docs.microsoft.com/en-us/rest/api/storageservices/define-stored-access-policy#modifying-or-revoking-astored-access-policy
NEW QUESTION 9
- (Topic 8)
You are developing a web application that runs as an Azure Web App. The web application stores data in Azure SQL Database and stores files in an Azure
Storage account. The web application makes HTTP requests to external services as part of normal operations.
The web application is instrumented with Application Insights. The external services are OpenTelemetry compliant.
You need to ensure that the customer ID of the signed in user is associated with all operations throughout the overall system.
What should you do?
A. Create a new SpanContext with the TraceRags value set to the customer ID for the signed in user.
B. On the current SpanContext, set the Traceld to the customer ID for the signed in user.
C. Add the customer ID for the signed in user to the CorrelationContext in the web application.
D. Set the header Ocp-Apim-Trace to the customer ID for the signed in user.
Answer: C
Explanation:
Reference:
https://docs.microsoft.com/en-us/azure/azure-monitor/app/correlation
NEW QUESTION 10
- (Topic 8)
You are developing an application that uses Azure Blob storage.
The application must read the transaction logs of all the changes that occur to the blobs and the blob metadata in the storage account for auditing purposes. The
changes must be in the order in which they occurred, include only create, update, delete, and copy operations and be retained for compliance reasons.
You need to process the transaction logs asynchronously. What should you do?
A. Process all Azure Blob storage events by using Azure Event Grid with a subscriber Azure Function app.
B. Enable the change feed on the storage account and process all changes for available events.
C. Process all Azure Storage Analytics logs for successful blob events.
D. Use the Azure Monitor HTTP Data Collector API and scan the request body for successful blob events.
Answer: B
Explanation:
Change feed support in Azure Blob Storage
The purpose of the change feed is to provide transaction logs of all the changes that occur to the blobs and the blob metadata in your storage account. The
change feed provides ordered, guaranteed, durable, immutable, read-only log of these changes. Client applications can read these logs at any time, either in
streaming or in batch mode. The change feed enables you to build efficient and scalable solutions that process change events that occur in your Blob Storage
account at a low cost.
Reference:
https://docs.microsoft.com/en-us/azure/storage/blobs/storage-blob-change-feed
NEW QUESTION 10
- (Topic 8)
You are developing a solution that will use a multi-partitioned Azure Cosmos DB database. You plan to use the latest Azure Cosmos DB SDK for development.
The solution must meet the following requirements:
? Send insert and update operations to an Azure Blob storage account.
? Process changes to all partitions immediately.
? Allow parallelization of change processing.
You need to process the Azure Cosmos DB operations.
What are two possible ways to achieve this goal? Each correct answer presents a complete solution.
NOTE:Each correct selection is worth one point.
A. Create an Azure App Service API and implement the change feed estimator of the SD
B. Scale the API by using multiple Azure App Service instances.
C. Create a background job in an Azure Kubernetes Service and implement the change feed feature of the SDK.
D. Create an Azure Function to use a trigger for Azure Cosmos D
E. Configure the trigger toconnect to the container.
F. Create an Azure Function that uses a Feedlterator object that processes the change feed by using the pull model on the containe
G. Use a FeedRange objext to parallelize the processing of the change feed across multiple functions.
Answer: CD
Explanation:
Azure Functions is the simplest option if you are just getting started using the change feed. Due to its simplicity, it is also the recommended option for most change
feed use cases. When you create an Azure Functions trigger for Azure Cosmos DB, you select the container to connect, and the Azure Function gets triggered
whenever there is a change in the container. Because Azure Functions uses the change feed processor behind the scenes, it automatically parallelizes change
processing across your container's partitions.
Note: You can work with change feed using the following options:
? Using change feed with Azure Functions
? Using change feed with change feed processor
Reference:
https://docs.microsoft.com/en-us/azure/cosmos-db/read-change-feed
https://docs.microsoft.com/en-us/azure/cosmos-db/change-feed-pull-model https://docs.microsoft.com/en-us/azure/cosmos-db/read-change-feed#azure-functions
https://docs.microsoft.com/en-us/azure/cosmos-db/change-feed-pull-model#using-feedrange-for-parallelization
NEW QUESTION 12
- (Topic 8)
You a web application that provides access to legal documents that are stored on Azure Blob Storage with version level immutability policies. Documents are
protected with both time-based policies legal hold policies. All time—based retention policies have AllowProtectedAppendWrites property enabled.
You have a requirement to prevent the user from attempting to perform operations that would fail only a legal is in effect and when all other are expired
You reed to meet the requirement.
Which two operations you prevent?
A. overwriting existing
B. adding data to documents
C. deleting documents
D. creating document
Answer: AC
NEW QUESTION 13
HOTSPOT - (Topic 8)
You are building a website to access project data related to terms within your organization. The website does not allow anonymous access. Authentication
performed using an Azure Active Directory (Azure AD) app named internal.
The website has the following authentication requirements:
•Azure AD users must be able to login to the website.
•Personalization of the website must be based on membership in Active Directory groups. You need to configure the application’s manifest to meet the
authentication requirements.
How should you configure the manifest? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: groupMembershipClaims
Personalization of the website must be based on membership in Active Directory groups. Group claims can also be configured in the Optional Claims section of the
Application Manifest. Enable group membership claims by changing the groupMembershipClaim The valid values are:
- "All"
- "SecurityGroup"
- "DistributionList"
- "DirectoryRole"
Here we need to mention that we want to get the groups for the users. Hence we need to mention to set the groupMembershipClaims property to All.
Box 2: oauth2AllowImplicitFlow
Azure AD users must be able to login to the website.
auth2Permissions can only accept collections value like an array, not a boolean. oauth2AllowImplicitFlow accepts boolean value.
Here from the list of options given, if we want the application to fetch the required tokens , we would need to allow Implicit Flow.
NEW QUESTION 14
- (Topic 8)
You are a developing a SaaS application that stores data as key value pairs.
You must make multiple editions of the application available. In the lowest cost edition, the performance must be best-effort, and there is no regional failover.
In higher cos! editions customers must be able to select guaranteed performance and support for multiple regions. Azure costs must be minimized.
Which Azure Cosmos OB API should you use for the application?
A. Core
B. MongoDB
C. Cassandra
D. Table API
Answer: D
NEW QUESTION 17
- (Topic 8)
You ate developing an application that allows users to find musicians that ate looking for work. The application must store information about musicians, the
instruments that they play, and other related data.
The application must also allow users to determine which musicians have played together, including groups of three or more musicians that have performed
together at a specific location.
Which Azure Cosmos D6 API should you use for the application?
A. Core
B. MongoDB
C. Cassandra
D. Gremlin
Answer: B
NEW QUESTION 19
HOTSPOT - (Topic 8)
You are implementing a software as a service (SaaS) ASP.NET Core web service that will run as an Azure Web App. The web service will use an on-premises
SQL Server database for storage. The web service also includes a WebJob that processes data updates. Four customers will use the web service.
•Each instance of the WebJob processes data for a single customer and must run as a singleton instance.
•Each deployment must be tested by using deployment slots prior to serving production data.
•Azure costs must be minimized.
•Azure resources must be located in an isolated network. You need to configure the App Service plan for the Web App.
How should you configure the App Service plan? To answer, select the appropriate settings in the answer area.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Number of VM instances: 4
You are not charged extra for deployment slots.
Pricing tier: Isolated
The App Service Environment (ASE) is a powerful feature offering of the Azure App Service that gives network isolation and improved scale capabilities. It is
essentially a deployment of the Azure App Service into a subnet of a customer’s Azure Virtual Network (VNet).
References:
https://azure.microsoft.com/sv-se/blog/announcing-app-service-isolated-more-power-scale-and-ease-of-use/
NEW QUESTION 21
- (Topic 8)
You must implement Application Insights instrumentation capabilities utilizing the Azure Mobile Apps SDK to provide meaningful analysis of user interactions with a
mobile app.
You need to capture the data required to implement the Usage Analytics feature of Application Insights. Which three data values should you capture? Each correct
answer presents part of the solution
NOTE: Each correct selection is worth one point.
A. Trace
B. Session Id
C. Exception
D. User Id
E. Events
Answer: ADE
Explanation:
Application Insights is a service for monitoring the performance and usage of your apps. This module allows you to send telemetry of various kinds (events, traces,
etc.) to the Application Insights service where your data can be visualized in the Azure Portal.
Application Insights manages the ID of a session for you. References:
https://github.com/microsoft/ApplicationInsights-Android
NEW QUESTION 23
- (Topic 8)
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the
stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this question, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
Margie’s Travel is an international travel and bookings management service. The company is expanding into restaurant bookings. You are tasked with
implementing Azure Search for the restaurants listed in their solution.
You create the index in Azure Search.
You need to import the restaurant data into the Azure Search service by using the Azure Search .NET SDK.
Solution:
* 1. Create a SearchIndexClient o bject to connect to the search index.
A. Yes
B. No
Answer: B
Explanation:
Use the following method:
* 1.- Create a SearchIndexClient object to connect to the search index
* 2.- Create an IndexBatch that contains the documents which must be added.
* 3.- Call the Documents.Index method of the SearchIndexClient and pass the IndexBatch.
References:
https://docs.microsoft.com/en-us/azure/search/search-howto-dotnet-sdk
NEW QUESTION 28
- (Topic 8)
You use Azure Table storage to store customer information for an application. The data contains customer details and is partitioned by last name. You need to
create a query that returns all customers with the last name Smith. Which code segment should you use?
Answer: C
Explanation:
Retrieve all entities in a partition. The following code example specifies a filter for entities where 'Smith' is the partition key. This example prints the fields of each
entity in the query results to the console.
Construct the query operation for all customer entities where PartitionKey="Smith".
TableQuery<CustomerEntity> query = new TableQuery<CustomerEntity>().Where(TableQuery.GenerateFilterCondition("PartitionKey", QueryComparisons.Equal,
"Smith"));
References:
https://docs.microsoft.com/en-us/azure/cosmos-db/table-storage-how-to-use-dotnet
NEW QUESTION 30
- (Topic 8)
You are developing an Azure-based web application. The application goes offline periodically to perform offline data processing. While the application is offline,
numerous Azure Monitor alerts fire which result in the on-call developer being paged.
The application must always log when the application is offline for any reason.
You need to ensure that the on-call developer is not paged during offline processing. What should you do?
Answer: C
NEW QUESTION 35
DRAG DROP - (Topic 8)
You are authoring a set of nested Azure Resource Manager templates to deploy multiple Azure resources.
The templates must be tested before deployment and must follow recommended practices. You need to validate and test the templates before deployment.
Which tools should you use? To answer, drag the appropriate tools to the correct requirements. Each tool may be used once, more than once, or not at all. You
may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
https://docs.microsoft.com/en-us/azure/azure-resource-manager/templates/test-toolkit
https://docs.microsoft.com/en-us/azure/azure-resource-manager/templates/deploy-what-if?tabs=azure-powershell
NEW QUESTION 40
- (Topic 8)
You are developing several Azure API Management (APIM) hosted APIs.
You must transform the APIs to hide private backend information and obscure the technology stack used to implement the backend processing.
You need to protect all APIs. What should you do?
Answer: A
NEW QUESTION 43
HOTSPOT - (Topic 8)
You are developing a solution to store documents in Azure Blob storage. Customers upload documents to multiple containers. Documents consist of PDF, CSV,
Microsoft Office format, and plain text files.
The solution must process millions of documents across hundreds of containers. The solution must meet the following requirements:
* Document must the categorized by a customer identifier as they are uploaded to the storage account.
* Allow filtering by the customer identifier.
* Allow searching of information contained within a document.
* Minimize costs.
You created and configure a standard general-purpose v2 storage account to support the solution.
You need to implement the solution.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Azure Blob Index tags: https://docs.microsoft.com/en-us/azure/storage/blobs/storage-blob- index-how-to?tabs=azure-portal
Azure Cognitive Search: Search inside documents
NEW QUESTION 45
HOTSPOT - (Topic 8)
You are developing an application to store and retrieve data in Azure Blob storage. The application will be hosted in an on-premises virtual machine (VM). The VM
is connected to Azure by using a Site-to-Site VPN gateway connection. The application is secured by using Azure Active Directory (Azure AD) credentials.
The application must be granted access to the Azure Blob storage account with a start time, expiry time, and read permissions. The Azure Blob storage account
access must use the Azure AD credentials of the application to secure data access. Data access must be able to be revoked if the client application security is
breached.
You need to secure the application access to Azure Blob storage.
Which security features should you use? To answer select the appropriate options in the answer area.
NOTE:Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Shared access signature (SAS) token
When your application design requires shared access signatures for access to Blob storage, use Azure AD credentials to create a user delegation SAS when
possible for superior security.
Box 2: Stored access policy
Stored access policies give you the option to revoke permissions for a service SAS without
having to regenerate the storage account keys.
A shared access signature can take one of the following two forms:
? Service SAS with stored access policy. A stored access policy is defined on a resource container, which can be a blob container, table, queue, or file share. The
stored access policy can be used to manage constraints for one or more service shared access signatures. When you associate a service SAS with a stored
access policy, the SAS inherits the constraints – the start time, expiry time, and permissions – defined for the stored access policy.
? Ad hoc SAS.
NEW QUESTION 48
HOTSPOT - (Topic 8)
You develop several Azure Grid to include hundreds of event types, such as billing, inventory, and shipping updates.
Events must be sent to a single endpoint for the Azure Functions app to process. The events must be filtered by event type before processing. You must have
authorization and authentication control to partition your tenants to receive the event data.
You need to configure Azure Event Grid.
Which configuration should you use? To answer, select the appropriate values in the answer area.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
NEW QUESTION 53
HOTSPOT - (Topic 8)
You are developing an application that uses a premium block blob storage account. You are optimizing costs by automating Azure Blob Storage access tiers.
You apply the following policy rules to the storage account. You must determine the implications of applying the rules to the data. (Line numbers are included for
reference only.)
A. Mastered
B. Not Mastered
Answer: A
Explanation:
* 1. Yes
* 2. Yes
* 3. Yes
* 4. No
https://docs.microsoft.com/en-us/azure/storage/blobs/lifecycle-management-overview?tabs=azure-portal#move-aging-data-to-a-cooler-tier
NEW QUESTION 56
HOTSPOT - (Topic 8)
You are configuring a development environment for your team. You deploy the latest Visual Studio image from the Azure Marketplace to your Azure subscription.
The development environment requires several software development kits (SDKs) and third-party components to support application development across the
organization. You install and customize the deployed virtual machine (VM) for your development team. The customized VM must be saved to allow provisioning of
a new team member development environment.
You need to save the customized VM for future provisioning.
Which tools or services should you use? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Azure Powershell
Creating an image directly from the VM ensures that the image includes all of the disks associated with the VM, including the OS disk and any data disks.
Before you begin, make sure that you have the latest version of the Azure PowerShell module.
You use Sysprep to generalize the virtual machine, then use Azure PowerShell to create the image.
Box 2: Azure Blob Storage References:
https://docs.microsoft.com/en-us/azure/virtual-machines/windows/capture-image-resource#create-an-image-of-a-vm-using-powershell
NEW QUESTION 61
- (Topic 8)
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution. Determine whether the
solution meets the stated goals.
You are developing and deploying several ASP.Net web applications to Azure App Service. You plan to save session state information and HTML output. You
must use a storage mechanism with the following requirements:
•Share session state across all ASP.NET web applications
•Support controlled, concurrent access to the same session state data for multiple readers
and a single writer
•Save full HTTP responses for concurrent requests You need to store the information.
Proposed Solution: Add the web applications to Docker containers. Deploy the containers. Deploy the containers to Azure Kubernetes Service (AKS).
Does the solution meet the goal?
A. Yes
B. No
Answer: B
Explanation:
Instead use Azure Cache for Redis.
Note: Azure Cache for Redis provides a session state provider that you can use to store your session state in-memory with Azure Cache for Redis instead of a
SQL Server database. To use the caching session state provider, first configure your cache, and then configure your ASP.NET application for cache using the
Azure Cache for Redis Session State NuGet package.
References:
https://docs.microsoft.com/en-us/azure/azure-cache-for-redis/cache-aspnet-session-state- provider
NEW QUESTION 62
HOTSPOT - (Topic 8)
You are developing an application that includes two Docker containers. The application must meet the following requirements
? The containers must not run as root.
? The containers must be deployed to Azure Container Instances by using a YAML
file.
? The containers must share a lifecycle, resources, local network and storage volume.
? The storage volume must persist through container crashes.
? The storage volume must be destroyed on stop or restart of the containers.
You need to configure Azure Container Instances for the application.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
NEW QUESTION 67
HOTSPOT - (Topic 8)
You are a developer building a web site using a web app. The web site stores configuration data in Azure App Configuration. Access to Azure App Configuration
has been configured to use the identity of the web app for authentication. Security requirements specify that no other authentication systems must be used.
You need to load configuration data from Azure App Configuration.
How should you complete the code? To answer, select the appropriate options in the answer area.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
NEW QUESTION 69
HOTSPOT - (Topic 8)
You are developing a solution that uses the Azure Storage Client library for .NET. You have the following code: (Line numbers are included for reference only.)
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE:Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Yes
AcquireLeaseAsync does not specify leaseTime.
leaseTime is a TimeSpan representing the span of time for which to acquire the lease, which will be rounded down to seconds. If null, an infinite lease will be
acquired. If not null, this must be 15 to 60 seconds.
Box 2: No
The GetBlockBlobReference method just gets a reference to a block blob in this container.
Box 3: Yes
The BreakLeaseAsync method initiates an asynchronous operation that breaks the current lease on this container.
NEW QUESTION 73
- (Topic 8)
You are developing an Azure Function App that runs in an App Service Plan. The Azure Function is triggered by a Timer object. You observe that the Azure
Function does not reliably trigger when scheduled. Which two actions should you perform?
Answer: AC
NEW QUESTION 75
- (Topic 8)
You develop and deploy an Azure App Service web app to a production environment. You enable the Always On setting and the Application Insights site
extensions. You deploy a code update and receive multiple failed requests and exceptions in the web app. You need to validate the performance and failure
counts of the web app in near real time. Which Application Insights tool should you use?
A. Snapshot Debugger
B. Profiler
C. Smart Detection
D. Live Metrics Stream
E. Application Map
Answer: D
NEW QUESTION 79
- (Topic 8)
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution. Determine whether the
solution meets the stated goals.
You are developing and deploying several ASP.Net web applications to Azure App Service. You plan to save session state information and HTML output. You
must use a storage mechanism with the following requirements:
•Share session state across all ASP.NET web applications
•Support controlled, concurrent access to the same session state data for multiple readers and a single writer
•Save full HTTP responses for concurrent requests You need to store the information.
Proposed Solution: Deploy and configure an Azure Database for PostgreSQL. Update the web applications.
Does the solution meet the goal?
A. Yes
B. No
Answer: B
Explanation:
Instead deploy and configure Azure Cache for Redis. Update the web applications. Reference:
https://docs.microsoft.com/en-us/azure/architecture/best-practices/caching#managing-concurrency-in-a-cache
NEW QUESTION 83
HOTSPOT - (Topic 8)
You are working for Contoso, Ltd.
You define an API Policy object by using the following XML markup:
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE:Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Yes
Use the set-backend-service policy to redirect an incoming request to a different backend than the one specified in the API settings for that operation. Syntax: <set-
backend-service base-url="base URL of the backend service" />
Box 2: No
The condition is on 512k, not on 256k.
Box 3: No
The set-backend-service policy changes the backend service base URL of the incoming
request to the one specified in the policy.
NEW QUESTION 84
DRAG DROP - (Topic 8)
You are developing an application. You have an Azure user account that has access to two subscriptions.
You need to retrieve a storage account key secret from Azure Key Vault.
In which order should you arrange the PowerShell commands to develop the solution? To answer, move all commands from the list of commands to the answer
area and arrange them in the correct order.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Step 1: Get-AzSubscription
If you have multiple subscriptions, you might have to specify the one that was used to create your key vault. Enter the following to see the subscriptions for your
account: Get-AzSubscription
Step 2: Set-AzContext -SubscriptionId
To specify the subscription that's associated with the key vault you'll be logging, enter: Set-AzContext -SubscriptionId <subscriptionID>
Step 3: Get-AzStorageAccountKey You must get that storage account key.
Step 4: $secretvalue = ConvertTo-SecureString <storageAccountKey> -AsPlainText -Force
Set-AzKeyVaultSecret -VaultName <vaultName> -Name <secretName> -SecretValue
$secretvalue
After retrieving your secret (in this case, your storage account key), you must convert that key to a secure string, and then create a secret with that value in your
key vault.
Step 5: Get-AzKeyVaultSecret
Next, get the URI for the secret you created. You'll need this URI in a later step to call the key vault and retrieve your secret. Run the following PowerShell
command and make note of the ID value, which is the secret's URI:
Get-AzKeyVaultSecret –VaultName <vaultName>
NEW QUESTION 88
- (Topic 8)
You are building a B2B web application that uses Azure B2B collaboration for authentication Paying customers authenticate to Azure B2B using federation
The application allows users to sign up for trial accounts using any email address
When a user converts to a paying customer, the data associated with the teal should be kept, but the user must authenticate using federation
You need to update the user in Azure Active Directory (Azure AD) when they convert to a paying customer
Which Graph API parameter is used to change authentication from one-time passé odes to federation?
A. uscrFlowType
B. Status
C. invittdUstr
D. resetRedemption
Answer: B
NEW QUESTION 91
DRAG DROP - (Topic 8)
You are implementing an order processing system. A point of sale application publishes orders to topics in an Azure Service Bus queue. The label property for the
topic includes the following data:
You need to implement filtering and maximize throughput while evaluating filters.
Which filter types should you implement? To answer, drag the appropriate filter types to the correct subscriptions. Each filter type may be used once,more than
once, or not at all. You may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
FutureOrders: SQLFilter HighPriortyOrders: CorrelationFilter
CorrelationID only
InternationalOrders: SQLFilter
Country NOT USA requires an SQL Filter
HighQuantityOrders: SQLFilter
Need to use relational operators so an SQL Filter is needed. AllOrders: No Filter
SQL Filter: SQL Filters - A SqlFilter holds a SQL-like conditional expression that is evaluated in the broker against the arriving messages' user-defined properties
and system properties. All system properties must be prefixed with sys. in the conditional expression. The SQL-language subset for filter conditions tests for the
existence of properties (EXISTS), as well as for null-values (IS NULL), logical NOT/AND/OR, relational operators, simple numeric arithmetic, and simple text
pattern matching with LIKE.
Correlation Filters - A CorrelationFilter holds a set of conditions that are matched against one or more of an arriving message's user and system properties. A
common use is to match against the CorrelationId property, but the application can also choose to match against ContentType, Label, MessageId, ReplyTo,
ReplyToSessionId, SessionId, To, and any user-defined properties. A match exists when an arriving message's value for a property is equal to the value specified
in the correlation filter. For string expressions, the comparison is case-sensitive. When specifying multiple match properties, the filter combines them as a logical
AND condition, meaning for the filter to match, all conditions must match.
Boolean filters - The TrueFilter and FalseFilter either cause all arriving messages (true) or none of the arriving messages (false) to be selected for the subscription.
References:
https://docs.microsoft.com/en-us/azure/service-bus-messaging/topic-filters
NEW QUESTION 92
DRAG DROP - (Topic 8)
You are developing an ASP.NET Core Web API web service that uses Azure Application Insights to monitor performance and track events.
You need to enable logging and ensure that log messages can be correlated to events tracked by Application Insights.
How should you complete the code? To answer, drag the appropriate code segments to the correct locations. Each code segment may be used once, more than
once, or not at all. You may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: ApplicationInsightsLoggerOptions
If you want to include the EventId and EventName properties, then add the following to the ConfigureServices method:
services AddOptions<ApplicationInsightsLoggerOptions>() Configure(o => o.IncludeEventId = true);
Box 2: IncludeEventID
Box 3: ApplicationServices
In Asp.Net core apps it turns out that trace logs do not show up in Application Insights out of the box. We need to add the following code snippet to our Configure
method in Startup.cs:
loggerFactory.AddApplicationInsights(app.ApplicationServices, logLevel); References:
https://blog.computedcloud.com/enabling-application-insights-trace-logging-in-asp-net-
core/
NEW QUESTION 93
HOTSPOT - (Topic 8)
You develop a news and blog content app for Windows devices.
A notification must arrive on a user’s device when there is a new article available for them to view.
You need to implement push notifications.
How should you complete the code segment? To answer, select the appropriate options in the answer area.
NOTE:Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: NotificationHubClient
Box 2: NotificationHubClient
Box 3: CreateClientFromConnectionString
// Initialize the Notification Hub NotificationHubClient hub =
NotificationHubClient.CreateClientFromConnectionString(listenConnString, hubName);
Box 4: SendWindowsNativeNotificationAsync Send the push notification.
var result = await hub.SendWindowsNativeNotificationAsync(windowsToastPayload);
NEW QUESTION 95
HOTSPOT - (Topic 8)
You are using Azure Front Door Service.
You are expecting inbound files to be compressed by using Brotli compression. You discover that inbound XML files are not compressed. The files are 9
megabytes (MB) in size.
You need to determine the root cause for the issue.
To answer, select the appropriate options in the answer area.
NOTE:Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: No
Front Door can dynamically compress content on the edge, resulting in a smaller and faster response to your clients. All files are eligible for compression.
However, a file must be of a MIME type that is eligible for compression list.
Box 2: No
Sometimes you may wish to purge cached content from all edge nodes and force them all to retrieve new updated assets. This might be due to updates to your
web application, or to quickly update assets that contain incorrect information.
Box 3: Yes
These profiles support the following compression encodings: Gzip (GNU zip), Brotli
Answer: DF
A. Azure portal
B. Azure PowerShell
C. Azure Functions Core Tools (Azure CLI)
D. Visual Studio
Answer: A
Explanation:
The function editor built into the Azure portal lets you update the function.json file and the code file for a function. The host.json file, which contains some runtime-
specific configurations, is in the root folder of the function app.
References:
https://docs.microsoft.com/en-us/azure/azure-functions/functions-reference#fileupdate
A. Yes
B. No
Answer: A
Explanation:
The session state provider for Azure Cache for Redis enables you to share session information between different instances of an ASP.NET web application.
The same connection can be used by multiple concurrent threads. Redis supports both read and write operations.
The output cache provider for Azure Cache for Redis enables you to save the HTTP responses generated by an ASP.NET web application.
Note: Using the Azure portal, you can also configure the eviction policy of the cache, and control access to the cache by adding users to the roles provided. These
roles, which define the operations that members can perform, include Owner, Contributor, and Reader. For example, members of the Owner role have complete
control over the cache (including security) and its contents, members of the Contributor role can read and write information
in the cache, and members of the Reader role can only retrieve data from the cache.
Reference:
https://docs.microsoft.com/en-us/azure/architecture/best-practices/caching
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Service principal
Applications and container orchestrators can perform unattended, or "headless," authentication by using an Azure Active Directory (Azure AD) service principal.
Box 2: AcrPush
AcrPush provides pull/push permissions only and meets the principle of least privilege.
If the encryption key has been corrupted it must not be used for encryption. You need to ensure that the blob is encrypted.
How should you complete the code segment? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
A. Yes
B. No
Answer: B
Explanation:
Instead pass the HTTP trigger payload into an Azure Service Bus queue to be processed by a queue trigger function and return an immediate HTTP success
response.
Note: Large, long-running functions can cause unexpected timeout issues. General best practices include:
Whenever possible, refactor large functions into smaller function sets that work together and return responses fast. For example, a webhook or HTTP trigger
function might require an acknowledgment response within a certain time limit; it's common for webhooks to require an immediate response. You can pass the
HTTP trigger payload into a queue to be processed by a queue trigger function. This approach lets you defer the actual work and return an immediate response.
Reference:
https://docs.microsoft.com/en-us/azure/azure-functions/functions-best-practices
A. Yes
B. No
Answer: B
Explanation:
Instead update the web.config file to include the applicationInitialization configuration element. Specify custom initialization actions to run the scripts.
Note: Some apps might require custom warm-up actions before the swap. The applicationInitialization configuration element in web.config lets you specify custom
initialization actions. The swap operation waits for this custom warm-up to finish before swapping with the target slot. Here's a sample web.config fragment.
<system.webServer>
<applicationInitialization>
<add initializationPage="/" hostName="[app hostname]" />
<add initializationPage="/Home/About" hostName="[app hostname]" />
</applicationInitialization>
</system.webServer>
Reference:
https://docs.microsoft.com/en-us/azure/app-service/deploy-staging-slots#troubleshoot- swaps
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: SecretClient
Box 2: DefaultAzureCredential
In below example, the name of your key vault is expanded to the key vault URI, in the format "https://<your-key-vault-name>.vault.azure.net". This example is
using 'DefaultAzureCredential()' class from Azure Identity Library, which allows to use the same code across different environments with different options to
provide identity.
string keyVaultName = Environment.GetEnvironmentVariable("KEY_VAULT_NAME"); var kvUri = "https://" + keyVaultName + ".vault.azure.net";
var client = new SecretClient(new Uri(kvUri), new DefaultAzureCredential());
A. Use PowerShell on a remote machine to make a request to the local managed identity for Azure resources endpoint.
B. Use PowerShell on the VM to make a request to the local managed identity for Azure resources endpoint.
C. From the code on the V
D. call Azure Resource Manager using an access token.
E. From the code on the V
F. call Azure Resource Manager using a SAS token.
G. From the code on the V
H. generate a user delegation SAS token.
Answer: BC
A. Mastered
B. Not Mastered
Answer: A
Explanation:
A. Mastered
B. Not Mastered
Answer: A
Explanation:
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: config
To Configure logging for a web app use the command: az webapp log config
Box 2: --docker-container-logging Syntax include:
az webapp log config [--docker-container-logging {filesystem, off}]
Box 3: webapp
To download a web app's log history as a zip file use the command: az webapp log download
Box 4: download References:
https://docs.microsoft.com/en-us/cli/azure/webapp/log
You need to write an Azure CLI script that will create the jobs, tasks, and the pool.
In which order should you arrange the commands to develop the solution? To answer, move the appropriate commands from the list of command segments to the
answer area and arrange them in the correct order.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Step 1: az batch pool create
# Create a new Linux pool with a virtual machine configuration. az batch pool create \
--id mypool \
--vm-size Standard_A1 \
--target-dedicated 2 \
--image canonical:ubuntuserver:16.04-LTS \
--node-agent-sku-id "batch.node.ubuntu 16.04"
Step 2: az batch job create
# Create a new job to encapsulate the tasks that are added. az batch job create \
--id myjob \
--pool-id mypool
Step 3: az batch task create
# Add tasks to the job. Here the task is a basic shell command. az batch task create \
--job-id myjob \
--task-id task1 \
--command-line "/bin/bash -c 'printenv AZ_BATCH_TASK_WORKING_DIR'" Step 4: for i in {1..$numberOfJobs} do
References:
https://docs.microsoft.com/bs-latn-ba/azure/batch/scripts/batch-cli-sample-run-job
A. Mastered
B. Not Mastered
Answer: A
Explanation:
A. Column
B. Table
C. Trigger
D. Index
E. Schema
Answer: ABE
Explanation:
In the Dynamic Data Masking configuration page, you may see some database columns that the recommendations engine has flagged for masking. In order to
accept the recommendations, just click Add Mask for one or more columns and a mask is created based on the default type for this column. You can change the
masking function by clicking on the masking rule and editing the masking field format to a different format of your choice.
References:
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-dynamic-data-masking-get-started-portal
A. Yes
B. No
Answer: B
Explanation:
Instead create an AKS cluster that supports network policy. Create and apply a network to allow traffic only from within a defined namespace
References:
https://docs.microsoft.com/en-us/azure/aks/use-network-policies
A. Mastered
B. Not Mastered
Answer: A
Explanation:
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Answer: A
Explanation:
With the Premium plan the max outbound connections per instance is unbounded compared to the 600 active (1200 total) in a Consumption plan.
Note: The number of available connections is limited partly because a function app runs in
a sandbox environment. One of the restrictions that the sandbox imposes on your code is a limit on the number of outbound connections, which is currently 600
active (1,200 total) connections per instance. When you reach this limit, the functions runtime writes the following message to the logs: Host thresholds exceeded:
Connections.
Reference:
https://docs.microsoft.com/en-us/azure/azure-functions/manage-connections https://docs.microsoft.com/en-us/azure/azure-functions/functions-scale#service-limits
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Service bus queue
You are developing a back-end Azure App Service that scales based on the number of messages contained in a Service Bus queue.
Box 2: ActiveMessage Count
ActiveMessageCount: Messages in the queue or subscription that are in the active state and ready for delivery.
Box 3: Count
Box 4: Less than or equal to
You need to add a new rule that will continuously scale down the App Service as long as the scale up condition is not met.
Box 5: Decrease count by
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: The monitored container
The monitored container has the data from which the change feed is generated. Any inserts and updates to the monitored container are reflected in the change
feed of the container.
Box 2: The lease container
The lease container acts as a state storage and coordinates processing the change feed across multiple workers. The lease container can be stored in the same
account as the monitored container or in a separate account.
Box 3: The host: A host is an application instance that uses the change feed processor to listen for changes. Multiple instances with the same lease configuration
can run in parallel, but each instance should have a different instance name.
Box 4: The delegate
The delegate is the code that defines what you, the developer, want to do with each batch of changes that the change feed processor reads.
A. integration
B. multi-step web
C. URL ping
D. unit
E. load
Answer: BC
Explanation:
There are three types of availability tests:
? URL ping test: a simple test that you can create in the Azure portal.
? Multi-step web test: A recording of a sequence of web requests, which can be played back to test more complex scenarios. Multi-step web tests are created in
Visual Studio Enterprise and uploaded to the portal for execution.
? Custom Track Availability Tests: If you decide to create a custom application to run availability tests, the TrackAvailability() method can be used to send the
results to Application Insights.
Reference:
https://docs.microsoft.com/en-us/azure/azure-monitor/app/monitor-web-app-availability
A. Yes
B. No
Answer: A
Explanation:
These are valid warm-up behavior options, but are not helpful in fixing swap problems. Instead update the web.config file to include the applicationInitialization
configuration
element. Specify custom initialization actions to run the scripts.
Note: Some apps might require custom warm-up actions before the swap. The
applicationInitialization configuration element in web.config lets you specify custom initialization actions. The swap operation waits for this custom warm-up to finish
before swapping with the target slot. Here's a sample web.config fragment.
<system.webServer>
<applicationInitialization>
<add initializationPage="/" hostName="[app hostname]" />
<add initializationPage="/Home/About" hostName="[app hostname]" />
</applicationInitialization>
</system.webServer>
Reference:
https://docs.microsoft.com/en-us/azure/app-service/deploy-staging-slots#troubleshoot- swaps
Answer: C
Explanation:
Azure Cosmos DB now provides a new RBAC role, Cosmos DB Operator. This new role lets you provision Azure Cosmos accounts, databases, and containers,
but can’t access the keys that are required to access the data. This role is intended for use in scenarios
where the ability to grant access to Azure Active Directory service principals to manage deployment operations for Cosmos DB is needed, including the account,
database, and containers.
Reference:
https://azure.microsoft.com/en-us/updates/azure-cosmos-db-operator-role-for-role-based-access-control-rbac-is-now-available/
A. Create a customer-managed key (CMK) and store the key in a new Azure Key Vault instance.
B. Create an Azure AD managed identity and assign the identity to a new Azure Key Vault instance.
C. Create a data encryption key (DEK) by using the Azure Cosmos DB SDK and store the key in Azure Cosmos DB.
D. Create a new container to include an encryption policy with the JSON properties to be encrypted.
Answer: A
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Azure Instance Metadata Service endpoints "/oauth2/token" Box 1: http://169.254.169.254/metadata/identity/oauth2/token
Sample request using the Azure Instance Metadata Service (IMDS) endpoint (recommended):
GET 'http://169.254.169.254/metadata/identity/oauth2/token?api-version=2018-02- 01&resource=https://management.azure.com/' HTTP/1.1 Metadata: true
Box 2: JsonConvert.DeserializeObject<Dictionary<string,string>>(payload); Deserialized token response; returning access code.
A. Core
B. Mongo DB
C. Gremlin
D. Table API
Answer: D
In higher cos! editions customers must be able to select guaranteed performance and support for multiple regions. Azure costs must be minimized.
Which Azure Cosmos OB API should you use for the application?
A. Core
B. MongoDB
C. Cassandra
D. Table API
Answer: C
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Accessing the client certificate from App Service.
If you are using ASP.NET and configure your app to use client certificate authentication, the certificate will be available through the HttpRequest.ClientCertificate
property. For other application stacks, the client cert will be available in your app through a base64 encodedvalue in the "X-ARR-ClientCert" request header. Your
application can create a certificate from this value and then use it for authentication and authorization purposes in your application.
References:
https://docs.microsoft.com/en-us/azure/app-service/app-service-web-configure-tls-mutual- auth
Answer: AC
Explanation:
Reference:
https://azure.microsoft.com/en-us/blog/creating-a-web-test-alert-programmatically-with-application-insights/
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Remove the update that causes the problem
? Take a snapshot of the OS disk of the affected VM as a backup.
? Attach the OS disk to a recovery VM.
? Once the OS disk is attached on the recovery VM, run diskmgmt.msc to open Disk Management, and ensure the attached disk is ONLINE.
? (Step 1) Open an elevated command prompt instance (Run as administrator). Run the following command to get the list of the update packages that are on the
attached OS disk:
dism /image:<Attached OS disk>:\ /get-packages > c:\temp\Patch_level
? (Step 2) Open the C:\temp\Patch_level.txt file, and then read it from the bottom up.
Locate the update that's in Install Pending or Uninstall Pending state.
? Remove the update that caused the problem:
dism /Image:<Attached OS disk>:\ /Remove-Package /PackageName:<PACK
? (Step 4) Detach the OS disk and recreate the VM. Then check whether the issue is resolved.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Step 1: A user requests the image..
A user requests a file (also called an asset) by using a URL with a special domain name, such as <endpoint name>.azureedge.net. This name can be an endpoint
hostname or a custom domain. The DNS routes the request to the best performing POP location, which is usually the POP that is geographically closest to the
user.
Step 2: If no edge servers in the POP have the..
If no edge servers in the POP have the file in their cache, the POP requests the file from the origin server. The origin server can be an Azure Web App, Azure
Cloud Service, Azure Storage account, or any publicly accessible web server.
Step 3: The origin server returns the..
The origin server returns the file to an edge server in the POP.
An edge server in the POP caches the file and returns the file to the original requestor (Alice). The file remains cached on the edge server in the POP until the time-
to-live (TTL) specified by its HTTP headers expires. If the origin server didn't specify a TTL, the default TTL is seven days.
Step 4: Subsequent requests for..
Additional users can then request the same file by using the same URL that the original user used, and can also be directed to the same POP.
If the TTL for the file hasn't expired, the POP edge server returns the file directly from the cache. This process results in a faster, more responsive user experience.
References:
https://docs.microsoft.com/en-us/azure/cdn/cdn-overview
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Core (SQL)
Core(SQL) API stores data in document format. It offers the best end-to-end experience as
we have full control over the interface, service, and the SDK client libraries. SQL API supports analytics and offers performance isolation between operational and
analytical workloads.
Box 2: item id
item id is a unique identifier and is suitable for the partition key.
You must ensure that converted files are placed in the container referenced by the outputContainerSasUrl parameter. Files which fail to convert are places in the
container referenced by the failedContainerSasUrl parameter.
You need to ensure the files are correctly processed.
How should you complete the code segment? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: CreateJob
Box 2: TaskSuccess
TaskSuccess: Upload the file(s) only after the task process exits with an exit code of 0.
Incorrect: TaskCompletion: Upload the file(s) after the task process exits, no matter what the exit code was.
Box 3: TaskFailure
TaskFailure:Upload the file(s) only after the task process exits with a nonzero exit code.
Box 4: OutputFiles
To specify output files for a task, create a collection of OutputFile objects and assign it to
the CloudTask.OutputFiles property when you create the task.
References: https://docs.microsoft.com/en-
us/dotnet/api/microsoft.azure.batch.protocol.models.outputfileuploadcondition https://docs.microsoft.com/en-us/azure/batch/batch-task-output-files
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Strong
Strong: Strong consistency offers a linearizability guarantee. The reads are guaranteed to return the most recent committed version of an item. A client never sees
an uncommitted or partial write. Users are always guaranteed to read the latest committed write.
Box 2: Bounded staleness
Bounded staleness: The reads are guaranteed to honor the consistent-prefix guarantee. The reads might lag behind writes by at most "K" versions (that is
"updates") of an item or by "t" time interval. When you choose bounded staleness, the "staleness" can be configured in two ways:
The number of versions (K) of the item
The time interval (t) by which the reads might lag behind the writes
Box 3: Eventual
Eventual: There's no ordering guarantee for reads. In the absence of any further writes, the replicas eventually converge.
A. Create an environment variable Set the secureValue property to the secret value.
B. Add the secret value to the container imag
C. Use a managed identity.
D. Add the secret value to the application code Set the container startup command.
E. Add the secret value to an Azure Blob storage accoun
F. Generate a SAS token.
G. Mount a secret volume containing the secret value in a secrets file.
Answer: AE
Explanation:
Objects with secure values are intended to hold sensitive information like passwords or keys for your application. Using secure values for environment variables is
both safer and more flexible than including it in your container's image. Another option is to use secret volumes, described in Mount a secret volume in Azure
Container Instances.....https://docs.microsoft.com/en-us/azure/container-instances/container- instances-environment-variables
A. JobOperations.CreateJobO
B. CloudJob.Enable(IEnumerable<BatchClientBehavior>)
C. CloudJob.CommitAsync(IEnumerable<BatchClientBehavior>, CancellationToken)
D. JobOperations.EnableJob(String, IEnumerable<BatchClientBehavior>)
E. JobOperations.EnableJobAsync(Strin
F. IEnumerable<BatchClientBehavior>. CancellationToken)
Answer: C
Explanation:
A Batch job is a logical grouping of one or more tasks. A job includes settings common to the tasks, such as priority and the pool to run tasks on. The app uses the
BatchClient.JobOperations.CreateJob method to create a job on your pool.
The Commit method submits the job to the Batch service. Initially the job has no tasks.
{
CloudJob job = batchClient.JobOperations.CreateJob(); job.Id = JobId;
job.PoolInformation = new PoolInformation { PoolId = PoolId };
job.Commit();
}
References:
https://docs.microsoft.com/en-us/azure/batch/quick-run-dotnet
- (Topic 8)
You are developing an application to store information about the organizational structure for a company.
Users must be able to determine which people report to a particular manager, the office where employees work, and the projects that are assigned to an
employee.
Which Azure Cosmos DB API should you use for the application?
A. Core
B. Cassandra
C. Table API
D. Gremlin
E. MongoDB
Answer: E
A. Assign the value of the hazard message SessionID property to the ReplyToSessionId property.
B. Assign the value of the hazard message MessageId property to the DevileryCount property.
C. Assign the value of the hazard message SessionID property to the SequenceNumber property.
D. Assign the value of the hazard message MessageId property to the CorrelationIdproperty.
E. Assign the value of the hazard message SequenceNumber property to the DeliveryCount property.
F. Assign the value of the hazard message MessageId property to the SequenceNumber property.
Answer: AC
Explanation:
Reference:
https://docs.microsoft.com/en-us/azure/service-bus-messaging/service-bus-messages- payloads
A. await subscriptionClient.CloseAsync();
B. await subscriptionClient.AddRuleAsync(new RuleDescription(RuleDescription.DefaultRuleName, new TrueFilter()));
C. subscriptionClient.RegisterMessageHandler(ProcessMessagesAsync, messageHandlerOptions);
D. subscriptionClient = new SubscriptionClient(ServiceBusConnectionString, TopicName, SubscriptionName);
Answer: C
Explanation:
Using topic client, call RegisterMessageHandler which is used to receive messages continuously from the entity. It registers a message handler and begins a new
thread to receive messages. This handler is waited on every time a new message is received by the receiver.
subscriptionClient.RegisterMessageHandler(ReceiveMessagesAsync, messageHandlerOptions);
References:
https://www.c-sharpcorner.com/article/azure-service-bus-topic-and-subscription-pub-sub/
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: Yes
You can configure Always Encrypted for individual database columns containing your sensitive data. When setting up encryption for a column, you specify the
information about the encryption algorithm and cryptographic keys used to protect the data in the column.
Box 2: No
Box 3: Yes
In SQL Database, the VIEW permissions are not granted by default to the public fixed database role. This enables certain existing, legacy tools (using older
versions of DacFx) to work properly. Consequently, to work withencrypted columns (even if not decrypting them) a database administrator must explicitly grant the
two VIEW permissions.
Box 4: No
All cryptographic keys are stored in an Azure Key Vault.
References:
https://docs.microsoft.com/en-us/sql/relational-databases/security/encryption/always-encrypted-database-engine
A.
B.
C.
D.
Answer: D
For each of the following statements, select Yes if the statement is true. Otherwise, select No.
NOTE: Each correct selection is worth one point.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Many-to-many relationships without an entity class to represent the join table are not yet supported. However, you can represent a many-to-many relationship by
including an entity class for the join table and mapping two separate one-to-many relationships.
protected override void OnModelCreating(ModelBuilder modelBuilder)
{
modelBuilder.Entity<PostTag>()
HasKey(t => new { t.PostId, t.TagId });
modelBuilder.Entity<PostTag>() HasOne(pt => pt.Post) WithMany(p => p.PostTags) HasForeignKey(pt => pt.PostId);
modelBuilder.Entity<PostTag>() HasOne(pt => pt.Tag) WithMany(t => t.PostTags) HasForeignKey(pt => pt.TagId);
}
}
HOTSPOT - (Topic 7)
You need to implement the corporate website. How should you configure the solution?
A. Mastered
B. Not Mastered
Answer: A
Explanation:
A. Mastered
B. Not Mastered
Answer: A
Explanation:
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Step 1: Build a new application image by using dockerfile
Step 2: Create an alias if the image with the fully qualified path to the registry
Before you can push the image to a private registry, you’ve to ensure a proper image name. This can be achieved using the docker tag command. For
demonstration purpose, we’ll use Docker’s hello world image, rename it and push it to ACR.
# pulls hello-world from the public docker hub
$ docker pull hello-world
# tag the image in order to be able to push it to a private registry
$ docker tag hello-word <REGISTRY_NAME>/hello-world
# push the image
$ docker push <REGISTRY_NAME>/hello-world
Step 3: Log in to the registry and push image
In order to push images to the newly created ACR instance, you need to login to ACR form the Docker CLI. Once logged in, you can push any existing docker
image to your ACR instance.
Scenario:
Coho Winery plans to move the application to Azure and continue to support label creation. LabelMaker app
Azure Monitor Container Health must be used to monitor the performance of workloads that are deployed to Kubernetes environments and hosted on Azure
Kubernetes Service (AKS).
You must use Azure Container Registry to publish images that support the AKS deployment.
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: orders o
Scenario: Order data is stored as nonrelational JSON and must be queried using SQL. Box 2:li
Box 3: o.line_items
Box 4: o.city
The city field is in Order, not in the 2s.
Answer: D
Explanation:
You can create a snapshot of a blob. A snapshot is a read-only version of a blob that's taken at a point in time. Once a snapshot has been created, it can be read,
copied, or deleted, but not modified. Snapshots provide a way to back up a blob as it appears at a moment in time.
Scenario: Processing is performed by an Azure Function that uses version 2 of the Azure Function runtime. Once processing is completed, results are stored in
Azure Blob Storage and an Azure SQL database. Then, an email summary is sent to the user with a link to the processing report. The link to the report must
remain valid if the email is forwarded to another user.
Reference:
https://docs.microsoft.com/en-us/rest/api/storageservices/creating-a-snapshot-of-a-blob
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: id
id is a unique identifier for the event.
Box 2: eventType
eventType is one of the registered event types for this event source.
Box 3: dataVersion
dataVersion is the schema version of the data object. The publisher defines the schema version.
Scenario: Authentication events are used to monitor users signing in and signing out. All authentication events must be processed by Policy service. Sign outs
must be processed as quickly as possible.
The following example shows the properties that are used by all event publishers: [
{
"topic": string, "subject": string, "id": string, "eventType": string, "eventTime": string, "data":{
object-unique-to-each-publisher
},
"dataVersion": string, "metadataVersion": string
}
]
A. Connect Live Metrics Stream from Application Insights to the Azure Function app and filter the metrics.
B. Create a new Azure Log Analytics workspace and instrument the Azure Function app with Application Insights.
C. Update the Azure Function app with extension methods from Microsoft.Extensions.Logging to log events by using the log instance.
D. Add a new diagnostic setting to the Azure Function app to send logs to Log Analytics.
Answer: A
Explanation:
Azure Functions offers built-in integration with Azure Application Insights to monitor functions.
The following areas of Application Insights can be helpful when evaluating the behavior, performance, and errors in your functions:
Live Metrics: View metrics data as it's created in near real-time. Failures
Performance Metrics
Reference:
https://docs.microsoft.com/en-us/azure/azure-functions/functions-monitoring
Answer: C
Explanation:
Scenario: All Internal services must only be accessible from Internal Virtual Networks (VNets)
There are three Network Location types – Private, Public and Domain Reference:
https://devblogs.microsoft.com/powershell/setting-network-location-to-private/
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: volumeMounts Example: volumeMounts:
- mountPath: /mnt/secrets name: secretvolume1 volumes:
- name: secretvolume1 secret:
mysecret1: TXkgZmlyc3Qgc2VjcmV0IEZPTwo= Box 2: volumes
Box 3: secret
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: [BlobTrigger(..)]
Box 2: [Blob(..)]
Azure Blob storage output binding for Azure Functions. The output binding allows you to modify and delete blob storage data in an Azure Function.
The attribute's constructor takes the path to the blob and a FileAccess parameter indicating read or write, as shown in the following example:
[FunctionName("ResizeImage")] public static void Run(
[BlobTrigger("sample-images/{name}")] Stream image,
[Blob("sample-images-md/{name}", FileAccess.Write)] Stream imageSmall)
{
}
Scenario: You must create an Azure Function named CheckUserContent to perform the content checks.
The company’s data science group built ContentAnalysisService which accepts user generated content as a string and returns a probable value for inappropriate
content. Any values over a specific threshold must be reviewed by an employee of Contoso, Ltd.
Answer: B
Explanation:
Azure Event Hub is used for telemetry and distributed data streaming.
This service provides a single solution that enables rapid data retrieval for real-time processing as well as repeated replay of stored raw data. It can capture the
streaming data into a file for processing and analysis.
It has the following characteristics:
? low latency
? capable of receiving and processing millions of events per second
? at least once delivery
Reference:
https://docs.microsoft.com/en-us/azure/event-grid/compare-messaging-services
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Box 1: AllowedOrigins
A CORS request will fail if Access-Control-Allow-Origin is missing.
Scenario:
The following error message displays while you are testing the website:
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Scenario: Shipping Function app: Implement secure function endpoints by using app-level security and include Azure Active Directory (Azure AD).
Box 1: Function
A. Mastered
B. Not Mastered
Answer: A
Explanation:
Scenario: Shipping website
Use Azure Content Delivery Network (CDN) and ensure maximum performance for dynamic content while minimizing latency and costs.
Tier: Standard Profile: Akamai
Optimization: Dynamic site acceleration
Dynamic site acceleration (DSA) is available for Azure CDN Standard from Akamai, Azure CDN Standard from Verizon, and Azure CDN Premium from Verizon
profiles.
DSA includes various techniques that benefit the latency and performance of dynamic content. Techniques include route and network optimization, TCP
optimization, and more.
You can use this optimization to accelerate a web app that includes numerous responses that aren't cacheable. Examples are search results, checkout
transactions, or real-time data. You can continue to use core Azure CDN caching capabilities for static data.
Visit Our Site to Purchase the Full Set of Actual AZ-204 Exam Questions With Answers.
We Also Provide Practice Exam Software That Simulates Real Exam Environment And Has Many Self-Assessment Features. Order the
AZ-204 Product From:
https://www.2passeasy.com/dumps/AZ-204/
* AZ-204 Most Realistic Questions that Guarantee you a Pass on Your FirstTry
* AZ-204 Practice Test Questions in Multiple Choice Formats and Updatesfor 1 Year