How should you stru...
 
Notifications
Clear all

How should you structure the output of the Stream Analytics job? To answer, select the appropriate options in the answer area. NOTE:

1 Posts
1 Users
0 Likes
140 Views
(@cornadante)
Noble Member
Joined: 2 years ago
Posts: 716
Topic starter  

HOTSPOT

You have an Azure event hub named retailhub that has 16 partitions. Transactions are posted to retailhub. Each transaction includes the transaction ID, the individual line items, and the payment details. The transaction ID is used as the partition key.

You are designing an Azure Stream Analytics job to identify potentially fraudulent transactions at a retail store. The job will use retailhub as the input. The job will output the transaction ID, the individual line items, the payment details, a fraud score, and a fraud indicator.

You plan to send the output to an Azure event hub named fraudhub.

You need to ensure that the fraud detection solution is highly scalable and processes transactions as quickly as possible.

How should you structure the output of the Stream Analytics job? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

Show Answer Hide Answer

Suggested Answer:

Explanation:

Box 1: 16

For Event Hubs you need to set the partition key explicitly.

An embarrassingly parallel job is the most scalable scenario in Azure Stream Analytics. It connects one partition of the input to one instance of the query to one partition of the output.

Box 2: Transaction ID


   
Quote

Latest Microsoft DP-203 Dumps Valid Version

Latest And Valid Q&A | Instant Download | Once Fail, Full Refund
Share: