You have a Dataflow pipeline that processes website traffic logs stored in Cloud Storage and writes the processed data to BigQuery. You noticed that the pipeline is failing intermittently. You need to troubleshoot the issue. What should you do?
A. Use the Dataflow job monitoring interface to check the pipeline's status every hour. Use Cloud Profiler to analyze the pipeline's metrics, such as CPU utilization and memory usage.
B. Use Cloud Logging to create a chart displaying the pipeline's error logs. Use Metrics Explorer to validate the findings from the chart.
C. Use Cloud Logging to identify error groups in the pipeline's logs. Use Cloud Monitoring to create a dashboard that tracks the number of errors in each group.
D. Use Cloud Logging to view error messages in the pipeline's logs. Use Cloud Monitoring to analyze the pipeline's metrics, such as CPU utilization and memory usage.
正解:D
解説: (Pass4Test メンバーにのみ表示されます)
質問 2:
You are responsible for managing Cloud Storage buckets for a research company. Your company has well- defined data tiering and retention rules. You need to optimize storage costs while achieving your data retention needs. What should you do?
A. Configure the buckets to use the Standard storage class and enable Object Versioning.
B. Configure the buckets to use the Autoclass feature.
C. Configure the buckets to use the Archive storage class.
D. Configure a lifecycle management policy on each bucket to downgrade the storage class and remove objects based on age.
正解:D
解説: (Pass4Test メンバーにのみ表示されます)
質問 3:
You need to create a weekly aggregated sales report based on a large volume of data. You want to use Python to design an efficient process for generating this report. What should you do?
A. Create a Cloud Data Fusion and Wrangler flow. Schedule the flow to run once a week.
B. Create a Colab Enterprise notebook and use the bigframes.pandas library. Schedule the notebook to execute once a week.
C. Create a Cloud Run function that uses NumPy. Use Cloud Scheduler to schedule the function to run once a week.
D. Create a Dataflow directed acyclic graph (DAG) coded in Python. Use Cloud Scheduler to schedule the code to run once a week.
正解:D
解説: (Pass4Test メンバーにのみ表示されます)
質問 4:
Your organization has a petabyte of application logs stored as Parquet files in Cloud Storage. You need to quickly perform a one-time SQL-based analysis of the files and join them to data that already resides in BigQuery. What should you do?
A. Create a Dataproc cluster, and write a PySpark job to join the data from BigQuery to the files in Cloud Storage.
B. Use the bq load command to load the Parquet files into BigQuery, and perform SQL joins to analyze the data.
C. Launch a Cloud Data Fusion environment, use plugins to connect to BigQuery and Cloud Storage, and use the SQL join operation to analyze the data.
D. Create external tables over the files in Cloud Storage, and perform SQL joins to tables in BigQuery to analyze the data.
正解:D
解説: (Pass4Test メンバーにのみ表示されます)
質問 5:
Your company is building a near real-time streaming pipeline to process JSON telemetry data from small appliances. You need to process messages arriving at a Pub/Sub topic, capitalize letters in the serial number field, and write results to BigQuery. You want to use a managed service and write a minimal amount of code for underlying transformations. What should you do?
A. Use a Pub/Sub push subscription, write a Cloud Run service that accepts the messages, performs the transformations, and writes the results to BigQuery.
B. Use a Pub/Sub to Cloud Storage subscription, write a Cloud Run service that is triggered when objects arrive in the bucket, performs the transformations, and writes the results to BigQuery.
C. Use a Pub/Sub to BigQuery subscription, write results directly to BigQuery, and schedule a transformation query to run every five minutes.
D. Use the "Pub/Sub to BigQuery" Dataflow template with a UDF, and write the results to BigQuery.
正解:D
解説: (Pass4Test メンバーにのみ表示されます)
木村** -
前回の試験では及ばず落ちましが4月の試験でPass4Testのこの問題集を購入して今回合格出来ました。
説明が非常に分かりやすく