DP-203: Data Engineering on Microsoft Azure

34%

Question 151

Applications that publish messages to Azure Event Hub very frequently will get the best performance using Advanced Message Queuing Protocol (AMQP) because it establishes a persistent socket. True or False?
True
False




Answer is True

Publishers can use either HTTPS or AMQP. AMQP opens a socket and can send multiple messages over that socket.

Question 152

By default, how many partitions will a new Event Hub have?
1
2
3
4
8




Answer is 4

Event Hubs default to 4 partitions. Partitions are the buckets within an Event Hub. Each publication will go into only one partition. Each consumer group may read from one or more than one partition.

Question 153

If an Event Hub goes offline before a consumer group can process the events it holds, those events will be lost. True or False?
True
False




Answer is False

Events are persistent. Each consumer group has its own cursor maintaining its position within the partition. The consumer groups can resume processing at their cursor position when the Event Hub is again available.

Question 154

Which job input consumes data streams from applications at low latencies and high throughput?
Azure Blob
Event Hubs
IoT hubs




Answer is Event Hubs

Event hubs consumes data streams from applications at low latencies and high throughput. Azure Blob stores static data that can be rerun through a streaming job. Azure IoT Hub is a data stream ingestion service that consumes events from IoT devices and also provides bi-directional communication between Azure and IoT devices.

Question 155

Streaming Analytics Query Language is a subset of which query language
WQL
TSQL
JSON




Answer is TSQL

Streaming Analytics Query Language is a subset of Transact-SQL. WQL is WIndows Management Instrumentation query language. JSON is an open standard file format

Question 156

You are a Data Engineer for Contoso. You want to view key health metrics of your Stream Analytics jobs. Which tool in Streaming Analytics should you use?
Diagnostics
Alerts
Dashboards




Answer is Dashboards

Dashboard are used to view the key health metrics of your Stream Analytics jobs. Diagnostic logging is turned off by default and can help with root-cause analysis in production deployments. Alerts enable proactive detection of issues in Stream Analytics.

Question 157

You are developing a solution that will stream to Azure Stream Analytics. The solution will have both streaming data and reference data.

Which input type should you use for the reference data?
Azure Cosmos DB
Azure Event Hubs
Azure Blob storage
Azure IoT Hub




Answer is Azure Blob storage

Stream Analytics supports Azure Blob storage and Azure SQL Database as the storage layer for Reference Data.

References:
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-use-reference-data

Question 158

You develop data engineering solutions for a company.
You need to ingest and visualize real-time Twitter data by using Microsoft Azure.

Which three technologies should you use? Each correct answer presents part of the solution.
Event Grid topic
Azure Stream Analytics Job that queries Twitter data from an Event Hub
Azure Stream Analytics Job that queries Twitter data from an Event Grid
Logic App that sends Twitter posts which have target keywords to Azure
Event Grid subscription
Event Hub instance




Answers are;
Azure Stream Analytics Job that queries Twitter data from an Event Hub
Logic App that sends Twitter posts which have target keywords to Azure
Event Hub instance


You can use Azure Logic apps to send tweets to an event hub and then use a Stream Analytics job to read from event hub and send them to PowerBI.

References:
https://community.powerbi.com/t5/Integrations-with-Files-and/Twitter-streaming-analytics-step-by-step/td-p/9594

Question 159

You have an Azure Stream Analytics query. The query returns a result set that contains 10,000 distinct values for a column named clusterID.
You monitor the Stream Analytics job and discover high latency.
You need to reduce the latency.

Which two actions should you perform?
Add a pass-through query.
Add a temporal analytic function.
Scale out the query by using PARTITION BY.
Convert the query to a reference query.
Increase the number of streaming units.




Answers are;
Scale out the query by using PARTITION BY.
Increase the number of streaming units.


Scaling a Stream Analytics job takes advantage of partitions in the input or output. Partitioning lets you divide data into subsets based on a partition key. A process that consumes the data (such as a Streaming Analytics job) can consume and write different partitions in parallel, which increases throughput.

Streaming Units (SUs) represents the computing resources that are allocated to execute a Stream Analytics job. The higher the number of SUs, the more CPU and memory resources are allocated for your job. This capacity lets you focus on the query logic and abstracts the need to manage the hardware to run your Stream Analytics job in a timely manner.

References:
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-parallelization
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-streaming-unit-consumption

Question 160

You are monitoring an Azure Stream Analytics job.
You discover that the Backlogged Input Events metric is increasing slowly and is consistently non-zero.
You need to ensure that the job can handle all the events.

What should you do?
Change the compatibility level of the Stream Analytics job.
Increase the number of streaming units (SUs).
Create an additional output stream for the existing input stream.
Remove any named consumer groups from the connection and use $default.




Answer is Increase the number of streaming units (SUs).

Backlogged Input Events: Number of input events that are backlogged. A non-zero value for this metric implies that your job isn't able to keep up with the number of incoming events. If this value is slowly increasing or consistently non-zero, you should scale out your job. You should increase the Streaming Units.

Note: Streaming Units (SUs) represents the computing resources that are allocated to execute a Stream Analytics job. The higher the number of SUs, the more CPU and memory resources are allocated for your job.

Reference:
https://docs.microsoft.com/bs-cyrl-ba/azure/stream-analytics/stream-analytics-monitoring

< Previous PageNext Page >

Quick access to all questions in this exam