Topic 5: Misc. Questions
You configure version control for an Azure Data Factory instance as shown in the following exhibit.
A company plans to use Apache Spark analytics to analyze intrusion detection data.
You need to recommend a solution to analyze network and system activity data for
malicious activities and policy violations. The solution must minimize administrative efforts.
What should you recommend?
A.
Azure Data Lake Storage
B.
Azure Databricks
C.
Azure HDInsight
D.
Azure Data Factory
Azure HDInsight
Explanation:
Azure HDInsight offers pre-made, monitoring dashboards in the form of solutions that can
be used to monitor the workloads running on your clusters. There are solutions for Apache
Spark, Hadoop, Apache Kafka, live long and process (LLAP), Apache HBase, and Apache
Storm available in the Azure Marketplace.
Note: With Azure HDInsight you can set up Azure Monitor alerts that will trigger when the
value of a metric or the results of a query meet certain conditions. You can condition on a
query returning a record with a value that is greater than or less than a certain threshold, or
even on the number of results returned by a query. For example, you could create an alert
to send an email if a Spark job fails or if a Kafka disk usage becomes over 90 percent full.
Reference:
https://azure.microsoft.com/en-us/blog/monitoring-on-azure-hdinsight-part-4-workloadmetrics-
and-logs/
From a website analytics system, you receive data extracts about user interactions such as
downloads, link clicks, form submissions, and video plays.
The data contains the following columns:
You have an Azure Active Directory (Azure AD) tenant named contoso.com that contains a
user named user1@contoso.com and an Azure SQL managed instance named SQLMI1.
You need to ensure that user1@contoso.com can create logins in SQLMI1 that map to
Azure AD service principals.
Which three actions should you perform in sequence? To answer, move the appropriate
actions from the list of actions to the answer area and arrange them in the correct order.
You have an Azure SQL database named db1 on a server named server1.
You need to modify the MAXDOP settings for db1. What should you do?
A.
Connect to db1 and run the sp_configure command.
B.
Connect to the master database of server1 and run the sp_configure command.
C.
Configure the extended properties of db1.
D.
Modify the database scoped configuration of db1.
Modify the database scoped configuration of db1.
Reference:
https://docs.microsoft.com/en-us/azure/azure-sql/database/configure-max-degree-ofparallelism
You have an Azure SQL database named db1 on a server named server1.
You use Query Performance Insight to monitor db1.
You need to modify the Query Store configuration to ensure that performance monitoring
data is available as soon as possible.
Which configuration setting should you modify and which value should you configure? To
answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
You have an Azure SQL database.
You are reviewing a slow performing query as shown in the following exhibit
You are performing exploratory analysis of bus fare data in an Azure Data Lake Storage Gen2 account by using an Azure Synapse Analytics serverless SQL pool.
You execute the Transact-SQL query shown in the following exhibit
You plan to perform batch processing in Azure Databricks once daily.
Which type of Databricks cluster should you use?
A.
automated
B.
interactive
C.
High Concurrency
automated
Explanation:
Azure Databricks makes a distinction between all-purpose clusters and job clusters. You
use all-purpose clusters to analyze data collaboratively using interactive notebooks. You
use job clusters to run fast and robust automated jobs.
The Azure Databricks job scheduler creates a job cluster when you run a job on a new job
cluster and
terminates the cluster when the job is complete.
Reference:
https://docs.microsoft.com/en-us/azure/databricks/clusters
You have an Azure subscription that contains a server named Server1. Server1 hosts two
Azure SQL databases named DB1 and DB2.
You plan to deploy a Windows app named App1 that will authenticate to DB2 by using SQL
authentication.
You need to ensure that App1 can access DB2. The solution must meet the following
requirements: What should you create?
A.
a contained database user for App1 on DB2
B.
a login for App1 on Server1
C.
a contained database user from an external provider for App1 on DB2
D.
a contained database user from a Windows login for App1 on DB2
a contained database user for App1 on DB2
Reference:
https://docs.microsoft.com/en-us/sql/relational-databases/security/contained-databaseusers-
making-your-database-portable?view=sql-server-ver15
You have an Azure Stream Analytics job.
You need to ensure that the job has enough streaming units provisioned.
You configure monitoring of the SU % Utilization metric.
Which two additional metrics should you monitor? Each correct answer presents part of the
solution.
NOTE: Each correct selection is worth one point.
A.
Late Input Events
B.
Out of order Events
C.
Backlogged Input Events
D.
Watermark Delay
E.
Function Events
Backlogged Input Events
Watermark Delay
Explanation:
To react to increased workloads and increase streaming units, consider setting an alert of
80% on the SU Utilization metric. Also, you can use watermark delay and backlogged
events metrics to see if there is an impact.
Note: Backlogged Input Events: Number of input events that are backlogged. A non-zero
value for this metric implies that your job isn't able to keep up with the number of incoming
events. If this value is slowly increasing or consistently non-zero, you should scale out your
job, by increasing the SUs.
Reference:
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-monitoring
You are designing a star schema for a dataset that contains records of online orders. Each
record includes an order date, an order due date, and an order ship date.
You need to ensure that the design provides the fastest query times of the records when
querying for arbitrary date ranges and aggregating by fiscal calendar attributes.
Which two actions should you perform? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
A.
Create a date dimension table that has a DateTime key.
B.
Create a date dimension table that has an integer key in the format of YYYYMMDD.
C.
Use built-in SQL functions to extract date attributes.
D.
Use integer columns for the date fields.
E.
Use DateTime columns for the date fields.
Create a date dimension table that has an integer key in the format of YYYYMMDD.
Use integer columns for the date fields.
https://community.idera.com/database-tools/blog/b/community_blog/posts/why-use-a-datedimension-
table-ina-data-warehouse
Page 8 out of 18 Pages |
Previous |