Winter Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

Professional-Data-Engineer Exam Dumps - Google Professional Data Engineer Exam

Searching for workable clues to ace the Google Professional-Data-Engineer Exam? You’re on the right place! ExamCert has realistic, trusted and authentic exam prep tools to help you achieve your desired credential. ExamCert’s Professional-Data-Engineer PDF Study Guide, Testing Engine and Exam Dumps follow a reliable exam preparation strategy, providing you the most relevant and updated study material that is crafted in an easy to learn format of questions and answers. ExamCert’s study tools aim at simplifying all complex and confusing concepts of the exam and introduce you to the real exam scenario and practice it with the help of its testing engine and real exam dumps

Go to page:
Question # 49

You have a data analyst team member who needs to analyze data by using BigQuery. The data analyst wants to create a data pipeline that would load 200 CSV files with an average size of 15MB from a Cloud Storage bucket into BigQuery daily. The data needs to be ingested and transformed before being accessed in BigQuery for analysis. You need to recommend a fully managed, no-code solution for the data analyst. What should you do?

A.

Create a Cloud Run function and schedule it to run daily using Cloud Scheduler to load the data into BigQuery.

B.

Use the BigQuery Data Transfer Service to load files from Cloud Storage to BigQuery, create a BigQuery job which transforms the data using BigQuery SQL and schedule it to run daily.

C.

Build a custom Apache Beam pipeline and run it on Dataflow to load the file from Cloud Storage to BigQuery and schedule it to run daily using Cloud Composer.

D.

Create a pipeline by using BigQuery pipelines and schedule it to load the data into BigQuery daily.

Full Access
Question # 50

You have several different unstructured data sources, within your on-premises data center as well as in the cloud. The data is in various formats, such as Apache Parquet and CSV. You want to centralize this data in Cloud Storage. You need to set up an object sink for your data that allows you to use your own encryption keys. You want to use a GUI-based solution. What should you do?

A.

Use Cloud Data Fusion to move files into Cloud Storage.

B.

Use Storage Transfer Service to move files into Cloud Storage.

C.

Use Dataflow to move files into Cloud Storage.

D.

Use BigQuery Data Transfer Service to move files into BigQuery.

Full Access
Question # 51

Your new customer has requested daily reports that show their net consumption of Google Cloud compute resources and who used the resources. You need to quickly and efficiently generate these daily reports. What should you do?

A.

Do daily exports of Cloud Logging data to BigQuery. Create views filtering by project, log type, resource, and user.

B.

Filter data in Cloud Logging by project, resource, and user; then export the data in CSV format.

C.

Filter data in Cloud Logging by project, log type, resource, and user, then import the data into BigQuery.

D.

Export Cloud Logging data to Cloud Storage in CSV format. Cleanse the data using Dataprep, filtering by project, resource, and user.

Full Access
Question # 52

You are collecting loT sensor data from millions of devices across the world and storing the data in BigQuery. Your access pattern is based on recent data tittered by location_id and device_version with the following query:

You want to optimize your queries for cost and performance. How should you structure your data?

A.

Partition table data by create_date, location_id and device_version

B.

Partition table data by create_date cluster table data by tocation_id and device_version

C.

Cluster table data by create_date location_id and device_version

D.

Cluster table data by create_date, partition by location and device_version

Full Access
Question # 53

You need to deploy additional dependencies to all of a Cloud Dataproc cluster at startup using an existing initialization action. Company security policies require that Cloud Dataproc nodes do not have access to the Internet so public initialization actions cannot fetch resources. What should you do?

A.

Deploy the Cloud SQL Proxy on the Cloud Dataproc master

B.

Use an SSH tunnel to give the Cloud Dataproc cluster access to the Internet

C.

Copy all dependencies to a Cloud Storage bucket within your VPC security perimeter

D.

Use Resource Manager to add the service account used by the Cloud Dataproc cluster to the Network User role

Full Access
Question # 54

You work for a mid-sized enterprise that needs to move its operational system transaction data from an on-premises database to GCP. The database is about 20 TB in size. Which database should you choose?

A.

Cloud SQL

B.

Cloud Bigtable

C.

Cloud Spanner

D.

Cloud Datastore

Full Access
Question # 55

You’ve migrated a Hadoop job from an on-prem cluster to dataproc and GCS. Your Spark job is a complicated analytical workload that consists of many shuffing operations and initial data are parquet files (on average 200-400 MB size each). You see some degradation in performance after the migration to Dataproc, so you’d like to optimize for it. You need to keep in mind that your organization is very cost-sensitive, so you’d like to continue using Dataproc on preemptibles (with 2 non-preemptible workers only) for this workload.

What should you do?

A.

Increase the size of your parquet files to ensure them to be 1 GB minimum.

B.

Switch to TFRecords formats (appr. 200MB per file) instead of parquet files.

C.

Switch from HDDs to SSDs, copy initial data from GCS to HDFS, run the Spark job and copy results back to GCS.

D.

Switch from HDDs to SSDs, override the preemptible VMs configuration to increase the boot disk size.

Full Access
Question # 56

Your team runs a complex analytical query daily that processes terabytes of data. Recently, after running for 20 minutes, the query fails with a "Resources exceeded" error. You need to resolve this issue. What should you do?

A.

Increase your project's BigQuery API request quota.

B.

Analyze the SQL syntax for errors.

C.

Increase the maximum table size limit.

D.

Move from BigQuery on-demand to slot reservations.

Full Access
Go to page: