Azure Data Engineer Certified DP-201 Exam Dumps Questions

You need to get correct DP-201 study materials to prepare for your Microsoft Certified: Azure Data Engineer Associate certification. Azure Data Engineer certification requires to pass DP-200 and DP-201 exams to design and implement the management, monitoring, security, and privacy of data using the full stack of Azure data services to satisfy business needs. We updated Azure Data Engineer Certified DP-201 Exam Dumps Questions with 141 real exam questions and accurate answers to ensure that you can pass DP-201 exam successfully. More, based on the latest reviews, the current DP-200 exam questions are also great for 100% passing.

Here we have free DP-201 exam dumps online for checking:

1. Testlet 1

Case study

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next section of the exam. After you begin a new section, you cannot return to this section.

To start the case study

To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question.

Background

Trey Research is a technology innovator. The company partners with regional transportation department office to build solutions that improve traffic flow and safety.

The company is developing the following solutions:

Regional transportation departments installed traffic sensor systems on major highways across North America.

Sensors record the following information each time a vehicle passes in front of a sensor:

– Time

– Location in latitude and longitude

– Speed in kilometers per second (kmps)

– License plate number

– Length of vehicle in meters

Sensors provide data by using the following structure:

Traffic sensors will occasionally capture an image of a vehicle for debugging purposes.

You must optimize performance of saving/storing vehicle images.

Traffic sensor data

– Sensors must have permission only to add items to the SensorData collection.

– Traffic data insertion rate must be maximized.

– Once every three months all traffic sensor data must be analyzed to look for data patterns that indicate sensor malfunctions.

– Sensor data must be stored in a Cosmos DB named treydata in a collection named SensorData

– The impact of vehicle images on sensor data throughout must be minimized.

Backtrack

This solution reports on all data related to a specific vehicle license plate. The report must use data from the SensorData collection.

Users must be able to filter vehicle data in the following ways:

– vehicles on a specific road

– vehicles driving above the speed limit

Planning Assistance

Data used for Planning Assistance must be stored in a sharded Azure SQL Database.

Data from the Sensor Data collection will automatically be loaded into the Planning Assistance database once a week by using Azure Data Factory. You must be able to manually trigger the data load process.

Privacy and security policy

– Azure Active Directory must be used for all services where it is available.

– For privacy reasons, license plate number information must not be accessible in Planning Assistance.

– Unauthorized usage of the Planning Assistance data must be detected as quickly as possible. Unauthorized usage is determined by looking for an unusual pattern of usage.

– Data must only be stored for seven years.

Performance and availability

– The report for Backtrack must execute as quickly as possible.

– The SLA for Planning Assistance is 70 percent, and multiday outages are permitted.

– All data must be replicated to multiple geographic regions to prevent data loss.

– You must maximize the performance of the Real Time Response system.

Financial requirements

Azure resource costs must be minimized where possible.

You need to design the vehicle images storage solution.

What should you recommend?

 
 
 
 

2. You need to design a sharding strategy for the Planning Assistance database.

What should you recommend?

 
 
 
 

3. HOTSPOT

You need to design the SensorData collection.

What should you recommend? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

4. You need to recommend an Azure SQL Database pricing tier for Planning Assistance.

Which pricing tier should you recommend?

 
 
 
 

5. HOTSPOT

You need to design the Planning Assistance database.

For each of the following statements, select Yes if the statement is true. Otherwise, select No. NOTE: Each correct selection is worth one point.

6. Testlet 2

Case study

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next section of the exam. After you begin a new section, you cannot return to this section.

To start the case study

To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question.

Overview

You develop data engineering solutions for Graphics Design Institute, a global media company with offices in New York City, Manchester, Singapore, and Melbourne.

The New York office hosts SQL Server databases that stores massive amounts of customer data. The company also stores millions of images on a physical server located in the New York office. More than 2 TB of image data is added each day. The images are transferred from customer devices to the server in New York.

Many images have been placed on this server in an unorganized manner, making it difficult for editors to search images. Images should automatically have object and color tags generated. The tags must be stored in a document database, and be queried by SQL

You are hired to design a solution that can store, transform, and visualize customer data.

Requirements

Business

The company identifies the following business requirements:

– You must transfer all images and customer data to cloud storage and remove on-premises servers.

– You must develop an analytical processing solution for transforming customer data.

– You must develop an image object and color tagging solution.

– Capital expenditures must be minimized.

– Cloud resource costs must be minimized.

Technical

The solution has the following technical requirements:

– Tagging data must be uploaded to the cloud from the New York office location.

– Tagging data must be replicated to regions that are geographically close to company office locations.

– Image data must be stored in a single data store at minimum cost.

– Customer data must be analyzed using managed Spark clusters.

– Power BI must be used to visualize transformed customer data.

– All data must be backed up in case disaster recovery is required.

Security and optimization

All cloud data must be encrypted at rest and in transit.

The solution must support:

– parallel processing of customer data

– hyper-scale storage of images

– global region data replication of processed image data

You need to recommend a solution for storing the image tagging data.

What should you recommend?

 
 
 
 
 

7. You need to design the solution for analyzing customer data.

What should you recommend?

 
 
 
 
 

8. You need to recommend a solution for storing customer data.

What should you recommend?

 
 
 
 

9. HOTSPOT

You need to design storage for the solution.

Which storage services should you recommend? To answer, select the appropriate configuration in the answer area. NOTE: Each correct selection is worth one point.

10. Testlet 3

Case study

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next section of the exam. After you begin a new section, you cannot return to this section.

To start the case study

To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question.

Background

Current environment

The company has the following virtual machines (VMs):

Requirements

Storage and processing

You must be able to use a file system view of data stored in a blob.

You must build an architecture that will allow Contoso to use the DB FS filesystem layer over a blob store. The architecture will need to support data files, libraries, and images. Additionally, it must provide a web-based interface to documents that contain runnable command, visualizations, and narrative text such as a notebook.

CONT_SQL3 requires an initial scale of 35000 IOPS.

CONT_SQL1 and CONT_SQL2 must use the vCore model and should include replicas. The solution must support 8000 IOPS.

The storage should be configured to optimized storage for database OLTP workloads.

Migration

– You must be able to independently scale compute and storage resources.

– You must migrate all SQL Server workloads to Azure. You must identify related machines in the on-premises environment, get disk size data usage information.

– Data from SQL Server must include zone redundant storage.

– You need to ensure that app components can reside on-premises while interacting with components that run in the Azure public cloud.

– SAP data must remain on-premises.

– The Azure Site Recovery (ASR) results should contain per-machine data.

Business requirements

– You must design a regional disaster recovery topology.

– The database backups have regulatory purposes and must be retained for seven years.

– CONT_SQL1 stores customers sales data that requires ETL operations for data analysis. A solution is required that reads data from SQL, performs ETL, and outputs to Power BI. The solution should use managed clusters to minimize costs. To optimize logistics, Contoso needs to analyze customer sales data to see if certain products are tied to specific times in the year.

– The analytics solution for customer sales data must be available during a regional outage.

Security and auditing

– Contoso requires all corporate computers to enable Windows Firewall.

– Azure servers should be able to ping other Contoso Azure servers.

– Employee PII must be encrypted in memory, in motion, and at rest. Any data encrypted by SQL Server must support equality searches, grouping, indexing, and joining on the encrypted data.

– Keys must be secured by using hardware security modules (HSMs).

– CONT_SQL3 must not communicate over the default ports

Cost

– All solutions must minimize cost and resources.

– The organization does not want any unexpected charges.

– The data engineers must set the SQL Data Warehouse compute resources to consume 300 DWUs.

– CONT_SQL2 is not fully utilized during non-peak hours. You must minimize resource costs for during non-peak hours.

You need to design a solution to meet the SQL Server storage requirements for CONT_SQL3.

Which type of disk should you recommend?

 
 
 

11. You need to recommend an Azure SQL Database service tier.

What should you recommend?

 
 
 
 
 

12. You need to recommend the appropriate storage and processing solution?

What should you recommend?

 
 
 
 
 

13. Testlet 4

Case study

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next section of the exam. After you begin a new section, you cannot return to this section.

To start the case study

To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question.

Overview

General Overview

ADatum Corporation is a medical company that has 5,000 physicians located in more than 300 hospitals across the US. The company has a medical department, a sales department, a marketing department, a medical research department, and a human resources department.

You are redesigning the application environment of ADatum.

Physical Locations

ADatum has three main offices in New York, Dallas, and Los Angeles. The offices connect to each other by using a WAN link. Each office connects directly to the Internet. The Los Angeles office also has a datacenter that hosts all the company’s applications.

Existing Environment

Health Review

ADatum has a critical OLTP web application named Health Review that physicians use to track billing, patient care, and overall physician best practices.

Health Interface

ADatum has a critical application named Health Interface that receives hospital messages related to patient care and status updates. The messages are sent in batches by each hospital’s enterprise relationship management (ERM) system by using a VPN. The data sent from each hospital can have varying columns and formats.

Currently, a custom C# application is used to send the data to Health Interface. The application uses deprecated libraries and a new solution must be designed for this functionality.

Health Insights

ADatum has a web-based reporting system named Health Insights that shows hospital and patient insights to physicians and business users. The data is created from the data in Health Review and Health Interface, as well as manual entries.

Database Platform

Currently, the databases for all three applications are hosted on an out-of-date VMware cluster that has a single instance of Microsoft SQL Server 2012.

Problem Statements

ADatum identifies the following issues in its current environment:

– Over time, the data received by Health Interface from the hospitals has slowed, and the number of messages has increased.

– When a new hospital joins ADatum, Health Interface requires a schema modification due to the lack of data standardization.

– The speed of batch data processing is inconsistent.

Business Requirements

Business Goals

ADatum identifies the following business goals:

– Migrate the applications to Azure whenever possible.

– Minimize the development effort required to perform data movement.

– Provide continuous integration and deployment for development, test, and production environments.

– Provide faster access to the applications and the data and provide more consistent application performance.

– Minimize the number of services required to perform data processing, development, scheduling, monitoring, and the operationalizing of pipelines.

Health Review Requirements

ADatum identifies the following requirements for the Health Review application:

– Ensure that sensitive health data is encrypted at rest and in transit.

– Tag all the sensitive health data in Health Review. The data will be used for auditing.

Health Interface Requirements

ADatum identifies the following requirements for the Health Interface application:

– Upgrade to a data storage solution that will provide flexible schemas and increased throughput for writing data. Data must be regionally located close to each hospital, and reads must display be the most recent committed version of an item.

– Reduce the amount of time it takes to add data from new hospitals to Health Interface.

– Support a more scalable batch processing solution in Azure.

– Reduce the amount of development effort to rewrite existing SQL queries.

Health Insights Requirements

ADatum identifies the following requirements for the Health Insights application:

– The analysis of events must be performed over time by using an organizational date dimension table.

– The data from Health Interface and Health Review must be available in Health Insights within 15 minutes of being committed.

– The new Health Insights application must be built on a massively parallel processing (MPP) architecture that will support the high performance of joins on large fact tables.

You need to design a solution that meets the business requirements of Health Insights.

What should you include in the recommendation?

 
 
 
 

14. HOTSPOT

Which Azure data storage solution should you recommend for each application? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

15. You need to recommend a solution that meets the data platform requirements of Health Interface. The solution must minimize redevelopment efforts for the application.

What should you include in the recommendation?

 
 
 
 

16. Which consistency level should you use for Health Interface?

 
 
 
 

17. HOTSPOT

You need to design the storage for the Health Insights data platform.

Which types of tables should you include in the design? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

18. Testlet 5

Overview

You are a data engineer for Trey Research. The company is close to completing a joint project with the government to build smart highways infrastructure across North America. This involves the placement of sensors and cameras to measure traffic flow, car speed, and vehicle details.

You have been asked to design a cloud solution that will meet the business and technical requirements of the smart highway.

Solution components

Telemetry Capture

The telemetry capture system records each time a vehicle passes in front of a sensor.

The sensors run on a custom embedded operating system and record the following telemetry data:

– Time

– Location in latitude and longitude

– Speed in kilometers per hour (kmph)

– Length of vehicle in meters

Visual Monitoring

The visual monitoring system is a network of approximately 1,000 cameras placed near highways that capture images of vehicle traffic every 2 seconds. The cameras record high resolution images. Each image is approximately 3 MB in size.

Requirements: Business

The company identifies the following business requirements:

– External vendors must be able to perform custom analysis of data using machine learning technologies.

– You must display a dashboard on the operations status page that displays the following metrics: telemetry, volume, and processing latency.

– Traffic data must be made available to the Government Planning Department for the purpose of modeling changes to the highway system. The traffic data will be used in conjunction with other data such as information about events such as sporting events, weather conditions, and population statistics. External data used during the modeling is stored in on-premises SQL Server 2016 databases and CSV files stored in an Azure Data Lake Storage Gen2 storage account.

– Information about vehicles that have been detected as going over the speed limit during the last 30 minutes must be available to law enforcement officers. Several law enforcement organizations may respond to speeding vehicles.

– The solution must allow for searches of vehicle images by license plate to support law enforcement investigations. Searches must be able to be performed using a query language and must support fuzzy searches to compensate for license plate detection errors.

Requirements: Security

The solution must meet the following security requirements:

– External vendors must not have direct access to sensor data or images.

– Images produced by the vehicle monitoring solution must be deleted after one month. You must minimize costs associated with deleting images from the data store.

– Unauthorized usage of data must be detected in real time. Unauthorized usage is determined by looking for unusual usage patterns.

– All changes to Azure resources used by the solution must be recorded and stored. Data must be provided to the security team for incident response purposes.

Requirements: Sensor data

You must write all telemetry data to the closest Azure region. The sensors used for the telemetry capture system have a small amount of memory available and so must write data as quickly as possible to avoid losing telemetry data.

You need to design the storage for the visual monitoring system.

Which storage solution should you recommend?

 
 
 
 

19. You need to design the storage for the telemetry capture system.

What storage solution should you use in the design?

 
 
 

20. You need to design the solution for the government planning department.

Which services should you include in the design?

 
 
 
 

21. Testlet 6

Case study

This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided.

To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other questions in this case study.

At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next section of the exam. After you begin a new section, you cannot return to this section.

To start the case study

To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question.

Overview

Litware, Inc. owns and operates 300 convenience stores across the US. The company sells a variety of packaged foods and drinks, as well as a variety of prepared foods, such as sandwiches and pizzas.

Litware has a loyalty club whereby members can get daily discounts on specific items by providing their membership number at checkout.

Litware employs business analysts who prefer to analyze data by using Microsoft Power BI, and data scientists who prefer analyzing data in Azure Databricks notebooks.

Requirements. Business Goals

Litware wants to create a new analytics environment in Azure to meet the following requirements:

– See inventory levels across the stores. Data must be updated as close to real time as possible.

– Execute ad hoc analytical queries on historical data to identify whether the loyalty club discounts increase sales of the discounted products.

– Every four hours, notify store employees about how many prepared food items to produce based on historical demand from the sales data.

Requirements. Technical Requirements

Litware identifies the following technical requirements:

– Minimize the number of different Azure services needed to achieve the business goals

– Use platform as a service (PaaS) offerings whenever possible and avoid having to provision virtual machines that must be managed by Litware.

– Ensure that the analytical data store is accessible only to the company’s on-premises network and Azure services.

– Use Azure Active Directory (Azure AD) authentication whenever possible.

– Use the principle of least privilege when designing security.

– Stage inventory data in Azure Data Lake Storage Gen2 before loading the data into the analytical data store. Litware wants to remove transient data from Data Lake Storage once the data is no longer in use. Files that have a modified date that is older than 14 days must be removed.

– Limit the business analysts’ access to customer contact information, such as phone numbers, because this type of data is not analytically relevant.

– Ensure that you can quickly restore a copy of the analytical data store within one hour in the event of corruption or accidental deletion.

Requirements. Planned Environment

Litware plans to implement the following environment:

– The application development team will create an Azure event hub to receive real-time sales data, including store number, date, time, product ID, customer loyalty number, price, and discount amount, from the point of sale (POS) system and output the data to data storage in Azure.

– Customer data, including name, contact information, and loyalty number, comes from Salesforce and can be imported into Azure once every eight hours. Row modified dates are not trusted in the source table.

– Product data, including product ID, name, and category, comes from Salesforce and can be imported into Azure once every eight hours. Row modified dates are not trusted in the source table.

– Daily inventory data comes from a Microsoft SQL server located on a private network.

– Litware currently has 5 TB of historical sales data and 100 GB of customer data. The company expects approximately 100 GB of new data per month for the next year.

– Litware will build a custom application named FoodPrep to provide store employees with the calculation results of how many prepared food items to produce every four hours.

– Litware does not plan to implement Azure ExpressRoute or a VPN between the on-premises network and Azure.

Which Azure service should you recommend for the analytical data store so that the business analysts and data scientists can execute ad hoc queries as quickly as possible?

 
 
 
 

22. Question Set 7

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are designing an HDInsight/Hadoop cluster solution that uses Azure Data Lake Gen1 Storage.

The solution requires POSIX permissions and enables diagnostics logging for auditing.

You need to recommend solutions that optimize storage.

Proposed

Solution: Ensure that files stored are larger than 250MB.

Does the solution meet the goal?

 
 

23. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are designing an HDInsight/Hadoop cluster solution that uses Azure Data Lake Gen1 Storage.

The solution requires POSIX permissions and enables diagnostics logging for auditing.

You need to recommend solutions that optimize storage.

Proposed

Solution: Implement compaction jobs to combine small files into larger files.

Does the solution meet the goal?

 
 

24. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are designing an HDInsight/Hadoop cluster solution that uses Azure Data Lake Gen1 Storage.

The solution requires POSIX permissions and enables diagnostics logging for auditing.

You need to recommend solutions that optimize storage.

Proposed

Solution: Ensure that files stored are smaller than 250MB.

Does the solution meet the goal?

 
 

25. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are designing an Azure SQL Database that will use elastic pools. You plan to store data about customers in a table. Each record uses a value for CustomerID. You need to recommend a strategy to partition data based on values in CustomerID. Proposed

Solution: Separate data into customer regions by using vertical partitioning.

Does the solution meet the goal?

 
 

26. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are designing an Azure SQL Database that will use elastic pools. You plan to store data about customers in a table. Each record uses a value for CustomerID. You need to recommend a strategy to partition data based on values in CustomerID. Proposed

Solution: Separate data into customer regions by using horizontal partitioning.

Does the solution meet the goal?

 
 

27. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You are designing an Azure SQL Database that will use elastic pools. You plan to store data about customers in a table. Each record uses a value for CustomerID.

You need to recommend a strategy to partition data based on values in CustomerID.

Proposed

Solution: Separate data into shards by using horizontal partitioning.

Does the solution meet the goal?

 
 

28. HOTSPOT

You are designing a data processing solution that will run as a Spark job on an HDInsight cluster. The solution will be used to provide near real-time information about online ordering for a retailer.

The solution must include a page on the company intranet that displays summary information.

The summary information page must meet the following requirements:

– Display a summary of sales to date grouped by product categories, price range, and review scope.

– Display sales summary information including total sales, sales as compared to one day ago and sales as compared to one year ago.

– Reflect information for new orders as quickly as possible.

You need to recommend a design for the solution.

What should you recommend? To answer, select the appropriate configuration in the answer area.

29. You are evaluating data storage solutions to support a new application.

You need to recommend a data storage solution that represents data by using nodes and relationships in graph structures.

Which data storage solution should you recommend?

 
 
 
 

30. HOTSPOT

You have an on-premises data warehouse that includes the following fact tables. Both tables have the following columns: DataKey, ProductKey, RegionKey.

There are 120 unique product keys and 65 unique region keys.

Queries that use the data warehouse take a long time to complete.

You plan to migrate the solution to use Azure SQL Data Warehouse. You need to ensure that the Azure-based solution optimizes query performance and minimizes processing skew.

What should you recommend? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

31. You are designing a data processing solution that will implement the lambda architecture pattern. The solution will use Spark running on HDInsight for data processing.

You need to recommend a data storage technology for the solution.

Which two technologies should you recommend? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.

 
 
 
 
 

32. A company manufactures automobile parts. The company installs IoT sensors on manufacturing machinery.

You must design a solution that analyzes data from the sensors.

You need to recommend a solution that meets the following requirements:

– Data must be analyzed in real-time.

– Data queries must be deployed using continuous integration.

– Data must be visualized by using charts and graphs.

– Data must be available for ETL operations in the future.

– The solution must support high-volume data ingestion.

Which three actions should you recommend? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.

 
 
 
 
 
 
 
 

33. You are designing an Azure Databricks interactive cluster.

You need to ensure that the cluster meets the following requirements:

– Enable auto-termination

– Retain cluster configuration indefinitely after cluster termination.

What should you recommend?

 
 
 
 

34. You are designing a solution for a company. The solution will use model training for objective classification.

You need to design the solution.

What should you recommend?

 
 
 
 
 

35. A company stores data in multiple types of cloud-based databases.

You need to design a solution to consolidate data into a single relational database. Ingestion of data will occur at set times each day.

What should you recommend?

 
 
 
 
 

36. HOTSPOT

You manage an on-premises server named Server1 that has a database named Database1. The company purchases a new application that can access data from Azure SQL Database. You recommend a solution to migrate Database1 to an Azure SQL Database instance.

What should you recommend? To answer, select the appropriate configuration in the answer area. NOTE: Each correct selection is worth one point.

37. You are designing an application. You plan to use Azure SQL Database to support the application.

The application will extract data from the Azure SQL Database and create text documents. The text documents will be placed into a cloud-based storage solution. The text storage solution must be accessible from an SMB network share.

You need to recommend a data storage solution for the text documents.

Which Azure data storage type should you recommend?

 
 
 
 

38. You are designing an application that will have an Azure virtual machine. The virtual machine will access an Azure SQL database. The database will not be accessible from the Internet You need to recommend a solution to provide the required level of access to the database.

What should you include in the recommendation?

 
 
 
 

39. HOTSPOT

You are designing an application that will store petabytes of medical imaging data When the data is first created, the data will be accessed frequently during the first week. After one month, the data must be accessible within 30 seconds, but files will be accessed infrequently. After one year, the data will be accessed infrequently but must be accessible within five minutes.

You need to select a storage strategy for the data. The solution must minimize costs.

Which storage tier should you use for each time frame? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

40. You are designing a data store that will store organizational information for a company. The data will be used to identify the relationships between users. The data will be stored in an Azure Cosmos DB database and will contain several million objects.

You need to recommend which API to use for the database. The API must minimize the complexity to query the user relationships. The solution must support fast traversals.

Which API should you recommend?

 
 
 
 

41. HOTSPOT

You are designing a new application that uses Azure Cosmos DB. The application will support a variety of data patterns including log records and social media mentions. You need to recommend which Cosmos DB API to use for each data pattern. The solution must minimize resource utilization.

Which API should you recommend for each data pattern? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

42. You need to recommend a storage solution to store flat files and columnar optimized files.

The solution must meet the following requirements:

– Store standardized data that data scientists will explore in a curated folder.

– Ensure that applications cannot access the curated folder.

– Store staged data for import to applications in a raw folder.

– Provide data scientists with access to specific folders in the raw folder and all the content the curated folder.

Which storage solution should you recommend?

 
 
 
 

43. Your company is an online retailer that can have more than 100 million orders during a 24-hour period, 95 percent of which are placed between 16:30 and 17:00. All the orders are in US dollars.

The current product line contains the following three item categories:

– Games with 15,123 items

– Books with 35,312 items

– Pens with 6,234 items

You are designing an Azure Cosmos DB data solution for a collection named Orders Collection. The following documents is a typical order in Orders Collection.

Order Collection is expected to have a balanced read/write-intensive workload.

Which partition key provides the most efficient throughput?

 
 
 
 

44. You have a MongoDB database that you plan to migrate to an Azure Cosmos DB account that uses the MongoDB API.

During testing, you discover that the migration takes longer than expected.

You need to recommend a solution that will reduce the amount of time it takes to migrate the data.

What are two possible recommendations to achieve this goal? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.

 
 
 
 
 

45. You need to recommend a storage solution for a sales system that will receive thousands of small files per minute. The files will be in JSON, text, and CSV formats. The files will be processed and transformed before they are loaded into an Azure data warehouse. The files must be stored and secured in folders.

Which storage solution should you recommend?

 
 
 
 

46. You are designing an Azure Cosmos DB database that will support vertices and edges.

Which Cosmos DB API should you include in the design?

 
 
 
 

47. You are designing a big data storage solution.

The solution must meet the following requirements:

– Provide unlimited account sizes.

– Support a hierarchical file system.

– Be optimized for parallel analytics workloads.

Which storage solution should you use?

 
 
 
 

48. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You plan to store delimited text files in an Azure Data Lake Storage account that will be organized into department folders.

You need to configure data access so that users see only the files in their respective department folder.

Solution: From the storage account, you enable a hierarchical namespace, and you use RBAC.

Does this meet the goal?

 
 

49. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You plan to store delimited text files in an Azure Data Lake Storage account that will be organized into department folders. You need to configure data access so that users see only the files in their respective department folder.

Solution: From the storage account, you disable a hierarchical namespace, and you use RBAC.

Does this meet the goal?

 
 

50. Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You plan to store delimited text files in an Azure Data Lake Storage account that will be organized into department folders.

You need to configure data access so that users see only the files in their respective department folder.

Solution: From the storage account, you disable a hierarchical namespace, and you use access control lists (ACLs).

Does this meet the goal?

 
 

51. You plan to store 100 GB of data used by a line-of-business (LOB) app. You need to recommend a data storage solution for the data.

The solution must meet the following requirements:

– Minimize storage costs.

– Natively support relational queries.

– Provide a recovery time objective (RTO) of less than one minute.

What should you include in the recommendation?

 
 
 
 

52. HOTSPOT

You have a data model that you plan to implement in an Azure SQL data warehouse as shown in the following exhibit.

All the dimension tables will be less than 5 GB after compression, and the fact table will be approximately 6 TB.

Which type of table should you use for each table? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

53. You are designing a data storage solution for a database that is expected to grow to 50 TB. The usage pattern is singleton inserts, singleton updates, and reporting.

Which storage solution should you use?

 
 
 
 
 

54. HOTSPOT

You are designing a solution that will use Azure Table storage.

The solution will log records in the following entity.

You are evaluating which partition key to use based on the following two scenarios:

– Scenario1: Minimize hotspots under heavy write workloads.

– Scenario2: Ensure that date lookups are as efficient as possible for read workloads.

Which partition key should you use for each scenario? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

55. DRAG DROP You have data on the 75,000 employees of your company.

The data contains the properties shown in the following table.

You need to store the employee data in an Azure Cosmos DB container. Most queries on the data will filter by the Current Department and the Employee Surname properties.

Which partition key and item ID should you use for the container? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

56. DRAG DROP

You need to design a data architecture to bring together all your data at any scale and provide insights into all your users through the use of analytical dashboards, operational reports, and advanced analytics.

How should you complete the architecture? To answer, drag the appropriate Azure services to the correct

locations in the architecture. Each service may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content. NOTE: Each correct selection is worth one point.


Administration of Veritas Backup Exec 20.1 VCS-325 Dumps
Great Materials for Passing 70-742 Windows Server 2016 Exam

Add a Comment

Your email address will not be published. Required fields are marked *