Summer Sale Special 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: ex2p65

Exact2Pass Menu

Google Cloud Certified - Professional Cloud Database Engineer

Last Update 19 hours ago Total Questions : 141

The Google Cloud Certified - Professional Cloud Database Engineer content is now fully updated, with all current exam questions added 19 hours ago. Deciding to include Professional-Cloud-Database-Engineer practice exam questions in your study plan goes far beyond basic test preparation.

You'll find that our Professional-Cloud-Database-Engineer exam questions frequently feature detailed scenarios and practical problem-solving exercises that directly mirror industry challenges. Engaging with these Professional-Cloud-Database-Engineer sample sets allows you to effectively manage your time and pace yourself, giving you the ability to finish any Google Cloud Certified - Professional Cloud Database Engineer practice test comfortably within the allotted time.

Question # 4

You are configuring a brand new PostgreSQL database instance in Cloud SQL. Your application team wants to have an optimal and highly available environment with automatic failover to avoid any unplanned outage. What should you do?

A.

Create one regional Cloud SQL instance with a read replica in another region.

B.

Create one regional Cloud SQL instance in one zone with a standby instance in another zone in the same region.

C.

Create two read-write Cloud SQL instances in two different zones with a standby instance in another region.

D.

Create two read-write Cloud SQL instances in two different regions with a standby instance in another zone.

Question # 5

You are evaluating Cloud SQL for PostgreSQL as a possible destination for your on-premises PostgreSQL instances. Geography is becoming increasingly relevant to customer privacy worldwide. Your solution must support data residency requirements and include a strategy to:

configure where data is stored

control where the encryption keys are stored

govern the access to data

What should you do?

A.

Replicate Cloud SQL databases across different zones.

B.

Create a Cloud SQL for PostgreSQL instance on Google Cloud for the data that does not need to adhere to data residency requirements. Keep the data that must adhere to data residency requirements on-premises. Make application changes to support both databases.

C.

Allow application access to data only if the users are in the same region as the Google Cloud region for the Cloud SQL for PostgreSQL database.

D.

Use features like customer-managed encryption keys (CMEK), VPC Service Controls, and Identity and Access Management (IAM) policies.

Question # 6

Your company uses Cloud Spanner for a mission-critical inventory management system that is globally available. You recently loaded stock keeping unit (SKU) and product catalog data from a company acquisition and observed hot-spots in the Cloud Spanner database. You want to follow Google-recommended schema design practices to avoid performance degradation. What should you do? (Choose two.)

A.

Use an auto-incrementing value as the primary key.

B.

Normalize the data model.

C.

Promote low-cardinality attributes in multi-attribute primary keys.

D.

Promote high-cardinality attributes in multi-attribute primary keys.

E.

Use bit-reverse sequential value as the primary key.

Question # 7

Your team is building an application that stores and analyzes streaming time series financial data. You need a database solution that can perform time series-based scans with sub-second latency. The solution must scale into the hundreds of terabytes and be able to write up to 10k records per second and read up to 200 MB per second. What should you do?

A.

Use Firestore.

B.

Use Bigtable

C.

Use BigQuery.

D.

Use Cloud Spanner.

Question # 8

You host an application in Google Cloud. The application is located in a single region and uses Cloud SQL for transactional data. Most of your users are located in the same time zone and expect the application to be available 7 days a week, from 6 AM to 10PM. You want to ensure regular maintenance updates to your Cloud SQL instance without creating downtime for your users. What should you do?

A.

Configure a maintenance window during a period when no users will be on the system. Control the order of update by setting non-production instances to earlier and production instances to later.

B.

Create your database with one primary node and one read replica in the region.

C.

Enable maintenance notifications for users, and reschedule maintenance activities to a specific time after notifications have been sent.

D.

Configure your Cloud SQL instance with high availability enabled.

Question # 9

Your organization stores marketing data such as customer preferences and purchase history on Bigtable. The consumers of this database are predominantly data analysts and operations users. You receive a service ticket from the database operations department citing poor database performance between 9 AM-10 AM every day. Theapplication team has confirmed no latency from their logs. A new cohort of pilot users that is testing a dataset loaded from a third-party data provider is experiencing poor database performance. Other users are not affected. You need to troubleshoot the issue. What should you do?

A.

Isolate the data analysts and operations user groups to use different Bigtable instances.

B.

Check the Cloud Monitoring table/bytes_used metric from Bigtable.

C.

Use Key Visualizer for Bigtable.

D.

Add more nodes to the Bigtable cluster.

Question # 10

You are choosing a new database backend for an existing application. The current database is running PostgreSQL on an on-premises VM and is managed by a database administrator and operations team. The application data is relational and has light traffic. You want to minimize costs and the migration effort for this application. What should you do?

A.

Migrate the existing database to Firestore.

B.

Migrate the existing database to Cloud SQL for PostgreSQL.

C.

Migrate the existing database to Cloud Spanner.

D.

Migrate the existing database to PostgreSQL running on Compute Engine.

Go to page: