A company uses an on-premises Microsoft SQL Server database to store financial transaction data. The company migrates the transaction data from the on-premises database to AWS at the end of each month. The company has noticed that the cost to migrate data from the on-premises database to an Amazon RDS for SQL Server database has increased recently.
The company requires a cost-effective solution to migrate the data to AWS. The solution must cause minimal downtown for the applications that access the database.
Which AWS service should the company use to meet these requirements?
A. AWS Lambda
B. AWS Database Migration Service (AWS DMS)
C. AWS Direct Connect
D. AWS DataSync
正解:B
質問 2:
A retail company uses AWS Glue for extract, transform, and load (ETL) operations on a dataset that contains information about customer orders. The company wants to implement specific validation rules to ensure data accuracy and consistency.
Which solution will meet these requirements?
A. Create custom AWS Glue Data Quality rulesets to define specific data quality checks.
B. Use AWS Glue job bookmarks to track the data for accuracy and consistency.
C. Use AWS Glue Data Catalog to maintain a centralized data schema and metadata repository.
D. Use the built-in AWS Glue Data Quality transforms for standard data quality validations.
正解:A
解説: (Pass4Test メンバーにのみ表示されます)
質問 3:
A company uses Amazon RDS for MySQL as the database for a critical application. The database workload is mostly writes, with a small number of reads.
A data engineer notices that the CPU utilization of the DB instance is very high. The high CPU utilization is slowing down the application. The data engineer must reduce the CPU utilization of the DB Instance.
Which actions should the data engineer take to meet this requirement? (Choose two.)
A. Upgrade to a larger instance size.
B. Use the Performance Insights feature of Amazon RDS to identify queries that have high CPU utilization. Optimize the problematic queries.
C. Reboot the RDS DB instance once each week.
D. Modify the database schema to include additional tables and indexes.
E. Implement caching to reduce the database query load.
正解:A,B
質問 4:
An airline company is collecting metrics about flight activities for analytics. The company is conducting a proof of concept (POC) test to show how analytics can provide insights that the company can use to increase on-time departures.
The POC test uses objects in Amazon S3 that contain the metrics in .csv format. The POC test uses Amazon Athena to query the data. The data is partitioned in the S3 bucket by date.
As the amount of data increases, the company wants to optimize the storage solution to improve query performance.
Which combination of solutions will meet these requirements? (Choose two.)
A. Add a randomized string to the beginning of the keys in Amazon S3 to get more throughput across partitions.
B. Preprocess the .csv data to JSON format by fetching only the document keys that the query requires.
C. Use an S3 bucket that is in the same AWS Region where the company runs Athena queries.
D. Preprocess the .csv data to Apache Parquet format by fetching only the data blocks that are needed for predicates.
E. Use an S3 bucket that is in the same account that uses Athena to query the data.
正解:C,D
解説: (Pass4Test メンバーにのみ表示されます)
質問 5:
A gaming company uses a NoSQL database to store customer information. The company is planning to migrate to AWS.
The company needs a fully managed AWS solution that will handle high online transaction processing (OLTP) workload, provide single-digit millisecond performance, and provide high availability around the world.
Which solution will meet these requirements with the LEAST operational overhead?
A. Amazon DynamoDB
B. Amazon DocumentDB (with MongoDB compatibility)
C. Amazon Timestream
D. Amazon Keyspaces (for Apache Cassandra)
正解:A
質問 6:
A manufacturing company wants to collect data from sensors. A data engineer needs to implement a solution that ingests sensor data in near real time.
The solution must store the data to a persistent data store. The solution must store the data in nested JSON format. The company must have the ability to query from the data store with a latency of less than 10 milliseconds.
Which solution will meet these requirements with the LEAST operational overhead?
A. Use Amazon Simple Queue Service (Amazon SQS) to buffer incoming sensor data. Use AWS Glue to store the data in Amazon RDS for querying.
B. Use a self-hosted Apache Kafka cluster to capture the sensor data. Store the data in Amazon S3 for querying.
C. Use Amazon Kinesis Data Streams to capture the sensor data. Store the data in Amazon DynamoDB for querying.
D. Use AWS Lambda to process the sensor data. Store the data in Amazon S3 for querying.
正解:A
質問 7:
A company is planning to upgrade its Amazon Elastic Block Store (Amazon EBS) General Purpose SSD storage from gp2 to gp3. The company wants to prevent any interruptions in its Amazon EC2 instances that will cause data loss during the migration to the upgraded storage.
Which solution will meet these requirements with the LEAST operational overhead?
A. Create new gp3 volumes. Gradually transfer the data to the new gp3 volumes. When the transfer is complete, mount the new gp3 volumes to the EC2 instances to replace the gp2 volumes.
B. Create snapshots of the gp2 volumes. Create new gp3 volumes from the snapshots. Attach the new gp3 volumes to the EC2 instances.
C. Use AWS DataSync to create new gp3 volumes. Transfer the data from the original gp2 volumes to the new gp3 volumes.
D. Change the volume type of the existing gp2 volumes to gp3. Enter new values for volume size, IOPS, and throughput.
正解:D
質問 8:
A banking company uses an application to collect large volumes of transactional data. The company uses Amazon Kinesis Data Streams for real-time analytics. The company's application uses the PutRecord action to send data to Kinesis Data Streams.
A data engineer has observed network outages during certain times of day. The data engineer wants to configure exactly-once delivery for the entire processing pipeline.
Which solution will meet this requirement?
A. Update the checkpoint configuration of the Amazon Managed Service for Apache Flink (previously known as Amazon Kinesis Data Analytics) data collection application to avoid duplicate processing of events.
B. Stop using Kinesis Data Streams. Use Amazon EMR instead. Use Apache Flink and Apache Spark Streaming in Amazon EMR.
C. Design the data source so events are not ingested into Kinesis Data Streams multiple times.
D. Design the application so it can remove duplicates during processing by embedding a unique ID in each record at the source.
正解:D
Yamaguchi -
Pass4TestのDEA-C01問題集の模擬試験の機能にとても気に入って、模擬試験を繰り返し行うことで、試験形式に慣れることができるので、本試験で緊張したり焦ったりせず、冷静に落ち着いて試験に臨むことができました。合格だ!