11.11 Sale Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: buysanta

Exact2Pass Menu

Google Professional Data Engineer Exam

Last Update 5 hours ago Total Questions : 387

The Google Professional Data Engineer Exam content is now fully updated, with all current exam questions added 5 hours ago. Deciding to include Professional-Data-Engineer practice exam questions in your study plan goes far beyond basic test preparation.

You'll find that our Professional-Data-Engineer exam questions frequently feature detailed scenarios and practical problem-solving exercises that directly mirror industry challenges. Engaging with these Professional-Data-Engineer sample sets allows you to effectively manage your time and pace yourself, giving you the ability to finish any Google Professional Data Engineer Exam practice test comfortably within the allotted time.

Question # 4

You have two projects where you run BigQuery jobs:

• One project runs production jobs that have strict completion time SLAs. These are high priority jobs that must have the required compute resources available when needed. These jobs generally never go below a 300 slot utilization, but occasionally spike up an additional 500 slots.

• The other project is for users to run ad-hoc analytical queries. This project generally never uses more than 200 slots at a time. You want these ad-hoc queries to be billed based on how much data users scan rather than by slot capacity.

You need to ensure that both projects have the appropriate compute resources available. What should you do?

A.

Create a single Enterprise Edition reservation for both projects. Set a baseline of 300 slots. Enable autoscaling up to 700 slots.

B.

Create two reservations, one for each of the projects. For the SLA project, use an Enterprise Edition with a baseline of 300 slots and enable autoscaling up to 500 slots. For the ad-hoc project, configure on-demand billing.

C.

Create two Enterprise Edition reservations, one for each of the projects. For the SLA project, set a baseline of 300 slots and enableautoscaling up to 500 slots. For the ad-hoc project, set a reservation baseline of 0 slots and set the ignore_idle_slot3 flag to False.

D.

Create two Enterprise Edition reservations, one for each of the projects. For the SLA project, set a baseline of 800 slots. For the ad-hocproject, enable autoscaling up to 200 slots.

Question # 5

You have one BigQuery dataset which includes customers' street addresses. You want to retrieve all occurrences of street addresses from the dataset. What should you do?

A.

Create a deep inspection job on each table in your dataset with Cloud Data Loss Prevention and create an inspection template that includes the STREET_ADDRESS infoType.

B.

Create a de-identification job in Cloud Data Loss Prevention and use the masking transformation.

C.

Write a SQL query in BigQuery by using REGEXP_CONTAINS on all tables in your dataset to find rows where the word "street" appears.

D.

Create a discovery scan configuration on your organization with Cloud Data Loss Prevention and create an inspection template thatincludes the STREET_ADDRESS infoType.

Question # 6

You have historical data covering the last three years in BigQuery and a data pipeline that delivers new data to BigQuery daily. You have noticed that when the Data Science team runs a query filtered on a date column and limited to 30–90 days of data, the query scans the entire table. You also noticed that your bill is increasing more quickly than you expected. You want to resolve the issue as cost-effectively as possible while maintaining the ability to conduct SQL queries. What should you do?

A.

Re-create the tables using DDL. Partition the tables by a column containing a TIMESTAMP or DATE Type.

B.

Recommend that the Data Science team export the table to a CSV file on Cloud Storage and use Cloud Datalab to explore the data by reading the files directly.

C.

Modify your pipeline to maintain the last 30–90 days of data in one table and the longer history in a different table to minimize full table scans over the entire history.

D.

Write an Apache Beam pipeline that creates a BigQuery table per day. Recommend that the Data Science team use wildcards on the table name suffixes to select the data they need.

Question # 7

You are building an ELT solution in BigQuery by using Dataform. You need to perform uniqueness and null value checks on your final tables. What should you do to efficiently integrate these checks into your pipeline?

A.

Build Dataform assertions into your code

B.

Write a Spark-based stored procedure.

C.

Build BigQuery user-defined functions (UDFs).

D.

Create Dataplex data quality tasks.

Question # 8

You work for a shipping company that has distribution centers where packages move on delivery lines to route them properly. The company wants to add cameras to the delivery lines to detect and track any visual damage to the packages in transit. You need to create a way to automate the detection of damaged packages and flag them for human review in real time while the packages are in transit. Which solution should you choose?

A.

Use BigQuery machine learning to be able to train the model at scale, so you can analyze the packages in batches.

B.

Train an AutoML model on your corpus of images, and build an API around that model to integrate with the package tracking applications.

C.

Use the Cloud Vision API to detect for damage, and raise an alert through Cloud Functions. Integrate the package tracking applications with this function.

D.

Use TensorFlow to create a model that is trained on your corpus of images. Create a Python notebook in Cloud Datalab that uses this model so you can analyze for damaged packages.

Question # 9

Cloud Bigtable is Google's ______ Big Data database service.

A.

Relational

B.

mySQL

C.

NoSQL

D.

SQL Server

Question # 10

What is the HBase Shell for Cloud Bigtable?

A.

The HBase shell is a GUI based interface that performs administrative tasks, such as creating and deleting tables.

B.

The HBase shell is a command-line tool that performs administrative tasks, such as creating and deleting tables.

C.

The HBase shell is a hypervisor based shell that performs administrative tasks, such as creating and deleting new virtualized instances.

D.

The HBase shell is a command-line tool that performs only user account management functions to grant access to Cloud Bigtable instances.

Go to page: