Summer Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

Professional-Data-Engineer Exam Dumps - Google Professional Data Engineer Exam

Go to page:
Question # 9

You are administering shared BigQuery datasets that contain views used by multiple teams in your organization. The marketing team is concerned about the variability of their monthly BigQuery analytics spend using the on-demand billing model. You need to help the marketing team establish a consistent BigQuery analytics spend each month. What should you do?

A.

Create a BigQuery Standard pay-as-you go reservation with a baseline of 0 slots and autoscaling set to 500 for the marketing team, and bill them back accordingly.

B.

Create a BigQuery reservation with a baseline of 500 slots with no autoscaling for the marketing team, and bill them back accordingly.

C.

Establish a BigQuery quota for the marketing team, and limit the maximum number of bytes scanned each day.

D.

Create a BigQuery Enterprise reservation with a baseline of 250 slots and autoscaling set to 500 for the marketing team, and bill them back accordingly.

Full Access
Question # 10

You are designing a data warehouse in BigQuery to analyze sales data for a telecommunication service provider. You need to create a data model for customers, products, and subscriptions All customers, products, and subscriptions can be updated monthly, but you must maintain a historical record of all data. You plan to use the visualization layer for current and historical reporting. You need to ensure that the data model is simple, easy-to-use. and cost-effective. What should you do?

A.

Create a normalized model with tables for each entity. Use snapshots before updates to track historical data

B.

Create a normalized model with tables for each entity. Keep all input files in a Cloud Storage bucket to track historical data

C.

Create a denormalized model with nested and repeated fields Update the table and use snapshots to track historical data

D.

Create a denormalized, append-only model with nested and repeated fields Use the ingestion timestamp to track historical data.

Full Access
Question # 11

You’ve migrated a Hadoop job from an on-prem cluster to dataproc and GCS. Your Spark job is a complicated analytical workload that consists of many shuffing operations and initial data are parquet files (on average 200-400 MB size each). You see some degradation in performance after the migration to Dataproc, so you’d like to optimize for it. You need to keep in mind that your organization is very cost-sensitive, so you’d like to continue using Dataproc on preemptibles (with 2 non-preemptible workers only) for this workload.

What should you do?

A.

Increase the size of your parquet files to ensure them to be 1 GB minimum.

B.

Switch to TFRecords formats (appr. 200MB per file) instead of parquet files.

C.

Switch from HDDs to SSDs, copy initial data from GCS to HDFS, run the Spark job and copy results back to GCS.

D.

Switch from HDDs to SSDs, override the preemptible VMs configuration to increase the boot disk size.

Full Access
Question # 12

Which of these rules apply when you add preemptible workers to a Dataproc cluster (select 2 answers)?

A.

Preemptible workers cannot use persistent disk.

B.

Preemptible workers cannot store data.

C.

If a preemptible worker is reclaimed, then a replacement worker must be added manually.

D.

A Dataproc cluster cannot have only preemptible workers.

Full Access
Question # 13

Which of the following statements is NOT true regarding Bigtable access roles?

A.

Using IAM roles, you cannot give a user access to only one table in a project, rather than all tables in a project.

B.

To give a user access to only one table in a project, grant the user the Bigtable Editor role for

that table.

C.

You can configure access control only at the project level.

D.

To give a user access to only one table in a project, you must configure access through your application.

Full Access
Question # 14

What is the HBase Shell for Cloud Bigtable?

A.

The HBase shell is a GUI based interface that performs administrative tasks, such as creating and deleting tables.

B.

The HBase shell is a command-line tool that performs administrative tasks, such as creating and deleting tables.

C.

The HBase shell is a hypervisor based shell that performs administrative tasks, such as creating and deleting new virtualized instances.

D.

The HBase shell is a command-line tool that performs only user account management functions to grant access to Cloud Bigtable instances.

Full Access
Question # 15

Which action can a Cloud Dataproc Viewer perform?

A.

Submit a job.

B.

Create a cluster.

C.

Delete a cluster.

D.

List the jobs.

Full Access
Question # 16

If a dataset contains rows with individual people and columns for year of birth, country, and income, how many of the columns are continuous and how many are categorical?

A.

1 continuous and 2 categorical

B.

3 categorical

C.

3 continuous

D.

2 continuous and 1 categorical

Full Access
Go to page: