Spring Sale Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: buysanta

Exact2Pass Menu

Google Cloud Certified - Professional Cloud Database Engineer

Last Update 10 hours ago Total Questions : 141

The Google Cloud Certified - Professional Cloud Database Engineer content is now fully updated, with all current exam questions added 10 hours ago. Deciding to include Professional-Cloud-Database-Engineer practice exam questions in your study plan goes far beyond basic test preparation.

You'll find that our Professional-Cloud-Database-Engineer exam questions frequently feature detailed scenarios and practical problem-solving exercises that directly mirror industry challenges. Engaging with these Professional-Cloud-Database-Engineer sample sets allows you to effectively manage your time and pace yourself, giving you the ability to finish any Google Cloud Certified - Professional Cloud Database Engineer practice test comfortably within the allotted time.

Question # 4

You are managing a Cloud SQL for MySQL environment in Google Cloud. You have deployed a primary instance in Zone A and a read replica instance in Zone B, both in the same region. You are notified that the replica instance in Zone B was unavailable for 10 minutes. You need to ensure that the read replica instance is still working. What should you do?

A.

Use the Google Cloud Console or gcloud CLI to manually create a new clone database.

B.

Use the Google Cloud Console or gcloud CLI to manually create a new failover replica from backup.

C.

Verify that the new replica is created automatically.

D.

Start the original primary instance and resume replication.

Question # 5

You are starting a large CSV import into a Cloud SQL for MySQL instance that has many open connections. You checked memory and CPU usage, and sufficient resources are available. You want to follow Google-recommended practices to ensure that the import will not time out. What should you do?

A.

Close idle connections or restart the instance before beginning the import operation.

B.

Increase the amount of memory allocated to your instance.

C.

Ensure that the service account has the Storage Admin role.

D.

Increase the number of CPUs for the instance to ensure that it can handle the additional import operation.

Question # 6

Your organization has a production Cloud SQL for MySQL instance. Your instance is configured with 16 vCPUs and 104 GB of RAM that is running between 90% and 100% CPU utilization for most of the day. You need to scale up the database and add vCPUs with minimal interruption and effort. What should you do?

A.

Issue a gcloud sql instances patch command to increase the number of vCPUs.

B.

Update a MySQL database flag to increase the number of vCPUs.

C.

Issue a gcloud compute instances update command to increase the number of vCPUs.

D.

Back up the database, create an instance with additional vCPUs, and restore the database.

Question # 7

An analytics team needs to read data out of Cloud SQL for SQL Server and update a table in Cloud Spanner. You need to create a service account and grant least privilege access using predefined roles. What roles should you assign to the service account?

A.

roles/cloudsql.viewer and roles/spanner.databaseUser

B.

roles/cloudsql.editor and roles/spanner.admin

C.

roles/cloudsql.client and roles/spanner.databaseReader

D.

roles/cloudsql.instanceUser and roles/spanner.databaseUser

Question # 8

Your team is building an application that stores and analyzes streaming time series financial data. You need a database solution that can perform time series-based scans with sub-second latency. The solution must scale into the hundreds of terabytes and be able to write up to 10k records per second and read up to 200 MB per second. What should you do?

A.

Use Firestore.

B.

Use Bigtable

C.

Use BigQuery.

D.

Use Cloud Spanner.

Question # 9

Your ecommerce application connecting to your Cloud SQL for SQL Server is expected to have additional traffic due to the holiday weekend. You want to follow Google-recommended practices to set up alerts for CPU and memory metrics so you can be notified by text message at the first sign of potential issues. What should you do?

A.

Use a Cloud Function to pull CPU and memory metrics from your Cloud SQL instance and to call a custom service to send alerts.

B.

Use Error Reporting to monitor CPU and memory metrics and to configure SMS notification channels.

C.

Use Cloud Logging to set up a log sink for CPU and memory metrics and to configure a sink destination to send a message to Pub/Sub.

D.

Use Cloud Monitoring to set up an alerting policy for CPU and memory metrics and to configure SMS notification channels.

Question # 10

You currently have a MySQL database running on Cloud SQL with a read replica in a different zone for non-mission critical analytics workloads. You want to enable high availability (HA) for the analytic workloads while keeping costs low. What should you do?

A.

Increase the size of the read replica Instance and enable MA.

B.

Enable HA on the current read replica.

C.

Create a new HA instance in the same zone db lie primary.

D.

Create a new MA Instance in a different region than the primary.

Go to page: