Summer Sale Special 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: ex2p65

Exact2Pass Menu

Question # 4

You are configuring a brand new PostgreSQL database instance in Cloud SQL. Your application team wants to have an optimal and highly available environment with automatic failover to avoid any unplanned outage. What should you do?

A.

Create one regional Cloud SQL instance with a read replica in another region.

B.

Create one regional Cloud SQL instance in one zone with a standby instance in another zone in the same region.

C.

Create two read-write Cloud SQL instances in two different zones with a standby instance in another region.

D.

Create two read-write Cloud SQL instances in two different regions with a standby instance in another zone.

Full Access
Question # 5

You are evaluating Cloud SQL for PostgreSQL as a possible destination for your on-premises PostgreSQL instances. Geography is becoming increasingly relevant to customer privacy worldwide. Your solution must support data residency requirements and include a strategy to:

configure where data is stored

control where the encryption keys are stored

govern the access to data

What should you do?

A.

Replicate Cloud SQL databases across different zones.

B.

Create a Cloud SQL for PostgreSQL instance on Google Cloud for the data that does not need to adhere to data residency requirements. Keep the data that must adhere to data residency requirements on-premises. Make application changes to support both databases.

C.

Allow application access to data only if the users are in the same region as the Google Cloud region for the Cloud SQL for PostgreSQL database.

D.

Use features like customer-managed encryption keys (CMEK), VPC Service Controls, and Identity and Access Management (IAM) policies.

Full Access
Question # 6

Your company uses Cloud Spanner for a mission-critical inventory management system that is globally available. You recently loaded stock keeping unit (SKU) and product catalog data from a company acquisition and observed hot-spots in the Cloud Spanner database. You want to follow Google-recommended schema design practices to avoid performance degradation. What should you do? (Choose two.)

A.

Use an auto-incrementing value as the primary key.

B.

Normalize the data model.

C.

Promote low-cardinality attributes in multi-attribute primary keys.

D.

Promote high-cardinality attributes in multi-attribute primary keys.

E.

Use bit-reverse sequential value as the primary key.

Full Access
Question # 7

Your team is building an application that stores and analyzes streaming time series financial data. You need a database solution that can perform time series-based scans with sub-second latency. The solution must scale into the hundreds of terabytes and be able to write up to 10k records per second and read up to 200 MB per second. What should you do?

A.

Use Firestore.

B.

Use Bigtable

C.

Use BigQuery.

D.

Use Cloud Spanner.

Full Access
Question # 8

You host an application in Google Cloud. The application is located in a single region and uses Cloud SQL for transactional data. Most of your users are located in the same time zone and expect the application to be available 7 days a week, from 6 AM to 10PM. You want to ensure regular maintenance updates to your Cloud SQL instance without creating downtime for your users. What should you do?

A.

Configure a maintenance window during a period when no users will be on the system. Control the order of update by setting non-production instances to earlier and production instances to later.

B.

Create your database with one primary node and one read replica in the region.

C.

Enable maintenance notifications for users, and reschedule maintenance activities to a specific time after notifications have been sent.

D.

Configure your Cloud SQL instance with high availability enabled.

Full Access
Question # 9

Your organization stores marketing data such as customer preferences and purchase history on Bigtable. The consumers of this database are predominantly data analysts and operations users. You receive a service ticket from the database operations department citing poor database performance between 9 AM-10 AM every day. Theapplication team has confirmed no latency from their logs. A new cohort of pilot users that is testing a dataset loaded from a third-party data provider is experiencing poor database performance. Other users are not affected. You need to troubleshoot the issue. What should you do?

A.

Isolate the data analysts and operations user groups to use different Bigtable instances.

B.

Check the Cloud Monitoring table/bytes_used metric from Bigtable.

C.

Use Key Visualizer for Bigtable.

D.

Add more nodes to the Bigtable cluster.

Full Access
Question # 10

You are choosing a new database backend for an existing application. The current database is running PostgreSQL on an on-premises VM and is managed by a database administrator and operations team. The application data is relational and has light traffic. You want to minimize costs and the migration effort for this application. What should you do?

A.

Migrate the existing database to Firestore.

B.

Migrate the existing database to Cloud SQL for PostgreSQL.

C.

Migrate the existing database to Cloud Spanner.

D.

Migrate the existing database to PostgreSQL running on Compute Engine.

Full Access
Question # 11

Your team is running a Cloud SQL for MySQL instance with a 5 TB database that must be available 24/7. You need to save database backups on object storage with minimal operational overhead or risk to your production workloads. What should you do?

A.

Use Cloud SQL serverless exports.

B.

Create a read replica, and then use the mysqldump utility to export each table.

C.

Clone the Cloud SQL instance, and then use the mysqldump utlity to export the data.

D.

Use the mysqldump utility on the primary database instance to export the backup.

Full Access
Question # 12

You are responsible for designing a new database for an airline ticketing application in Google Cloud. This application must be able to:

Work with transactions and offer strong consistency.

Work with structured and semi-structured (JSON) data.

Scale transparently to multiple regions globally as the operation grows.

You need a Google Cloud database that meets all the requirements of the application. What should you do?

A.

Use Cloud SQL for PostgreSQL with both cross-region read replicas.

B.

Use Cloud Spanner in a multi-region configuration.

C.

Use Firestore in Datastore mode.

D.

Use a Bigtable instance with clusters in multiple regions.

Full Access
Question # 13

You are migrating your data center to Google Cloud. You plan to migrate your applications to Compute Engine and your Oracle databases to Bare Metal Solution for Oracle. You must ensure that the applications in different projects can communicate securely and efficiently with the Oracle databases. What should you do?

A.

Set up a Shared VPC, configure multiple service projects, and create firewall rules.

B.

Set up Serverless VPC Access.

C.

Set up Private Service Connect.

D.

Set up Traffic Director.

Full Access
Question # 14

Your organization is currently updating an existing corporate application that is running in another public cloud to access managed database services in Google Cloud. The application will remain in the other public cloud while the database is migrated to Google Cloud. You want to follow Google-recommended practices for authentication. You need to minimize user disruption during the migration. What should you do?

A.

Use workload identity federation to impersonate a service account.

B.

Ask existing users to set their Google password to match their corporate password.

C.

Migrate the application to Google Cloud, and use Identity and Access Management (IAM).

D.

Use Google Workspace Password Sync to replicate passwords into Google Cloud.

Full Access
Question # 15

You are migrating critical production database from Amazon RDS for MySQL to Cloud SQL for MYSQL by using Google Cloud’s Migration Service.

You want to keep disruption to your production database to minimum and, at the same time, optimize migration performance. What should you do?

A.

Create and start multiple Database Migration Service jobs to migrate your database to the target Cloud SQL for MySQL instance.

B.

Upgrade the Amazon RDS for MySQL primary instance to an instance with more vCPUs and memory, and then run Google Cloud's Database Migration Service.

C.

Create a single Database Migration Service migration job with initial load parallelism configured to maximum on the source Amazon RDS for MySQL read replica.

D.

Create a single Database Migration Service migration job with initial Load Parallelism configured to Maximum on the Amazon RDS for MySQL primary instance.

Full Access
Question # 16

You are choosing a database backend for a new application. The application will ingest data points from IoT sensors. You need to ensure that the application can scale up to millions of requests per second with sub-10ms latency and store up to 100 TB of history. What should you do?

A.

Use Cloud SQL with read replicas for throughput.

B.

Use Firestore, and rely on automatic serverless scaling.

C.

Use Memorystore for Memcached, and add nodes as necessary to achieve the required throughput.

D.

Use Bigtable, and add nodes as necessary to achieve the required throughput.

Full Access
Question # 17

You are managing a mission-critical Cloud SQL for PostgreSQL instance. Your application team is running important transactions on the database when another DBA starts an on-demand backup. You want to verify the status of the backup. What should you do?

A.

Check the cloudsql.googleapis.com/postgres.log instance log.

B.

Perform the gcloud sql operations list command.

C.

Use Cloud Audit Logs to verify the status.

D.

Use the Google Cloud Console.

Full Access
Question # 18

Your company wants to migrate its MySQL, PostgreSQL, and Microsoft SQL Server on-premises databases to Google Cloud. You need a solution that provides near-zero downtime, requires no application changes, and supports change data capture (CDC). What should you do?

A.

Use the native export and import functionality of the source database.

B.

Create a database on Google Cloud, and use database links to perform the migration.

C.

Create a database on Google Cloud, and use Dataflow for database migration.

D.

Use Database Migration Service.

Full Access
Question # 19

You are starting a large CSV import into a Cloud SQL for MySQL instance that has many open connections. You checked memory and CPU usage, and sufficient resources are available. You want to follow Google-recommended practices to ensure that the import will not time out. What should you do?

A.

Close idle connections or restart the instance before beginning the import operation.

B.

Increase the amount of memory allocated to your instance.

C.

Ensure that the service account has the Storage Admin role.

D.

Increase the number of CPUs for the instance to ensure that it can handle the additional import operation.

Full Access
Question # 20

Your company is migrating their MySQL database to Cloud SQL and cannot afford any planned downtime during the month of December. The company is also concerned with cost, so you need the most cost-effective solution. What should you do?

A.

Open a support ticket in Google Cloud to prevent any maintenance in that MySQL instance during the month of December.

B.

Use Cloud SQL maintenance settings to prevent any maintenance during the month of December.

C.

Create MySQL read replicas in different zones so that, if any downtime occurs, the read replicas will act as the primary instance during the month of December.

D.

Create a MySQL regional instance so that, if any downtime occurs, the standby instance will act as the primary instance during the month of December.

Full Access
Question # 21

You are writing an application that will run on Cloud Run and require a database running in the Cloud SQL managed service. You want to secure this instance so that it only receives connections from applications running in your VPC environment in Google Cloud. What should you do?

A.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Use Cloud SQL Auth proxy to connect to the instance.

B.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Connect to the instance using a connection pool to best manage connections to the instance.

C.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Use Cloud SQL Auth proxy to connect to the instance.

D.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Connect to the instance using a connection pool to best manage connections to the instance.

Full Access
Question # 22

You manage a production MySQL database running on Cloud SQL at a retail company. You perform routine maintenance on Sunday at midnight when traffic is slow, but you want to skip routine maintenance during the year-end holiday shopping season. You need to ensure that your production system is available 24/7 during the holidays. What should you do?

A.

Define a maintenance window on Sundays between 12 AM and 1 AM, and deny maintenance periods between November 1 and January 15.

B.

Define a maintenance window on Sundays between 12 AM and 5 AM, and deny maintenance periods between November 1 and February 15.

C.

Build a Cloud Composer job to start a maintenance window on Sundays between 12 AM and 1AM, and deny maintenance periods between November 1 and January 15.

D.

Create a Cloud Scheduler job to start maintenance at 12 AM on Sundays. Pause the Cloud Scheduler job between November 1 and January 15.

Full Access
Question # 23

You support a consumer inventory application that runs on a multi-region instance of Cloud Spanner. A customer opened a support ticket to complain about slow response times. You notice a Cloud Monitoring alert about high CPU utilization. You want to follow Google-recommended practices to address the CPU performance issue. What should you do first?

A.

Increase the number of processing units.

B.

Modify the database schema, and add additional indexes.

C.

Shard data required by the application into multiple instances.

D.

Decrease the number of processing units.

Full Access
Question # 24

Your organization has an existing app that just went viral. The app uses a Cloud SQL for MySQL backend database that is experiencing slow disk performance while using hard disk drives (HDDs). You need to improve performance and reduce disk I/O wait times. What should you do?

A.

Export the data from the existing instance, and import the data into a new instance with solid-state drives (SSDs).

B.

Edit the instance to change the storage type from HDD to SSD.

C.

Create a high availability (HA) failover instance with SSDs, and perform a failover to the new instance.

D.

Create a read replica of the instance with SSDs, and perform a failover to the new instance

Full Access
Question # 25

You need to redesign the architecture of an application that currently uses Cloud SQL for PostgreSQL. The users of the application complain about slow query response times. You want to enhance your application architecture to offer sub-millisecond query latency. What should you do?

A.

Configure Firestore, and modify your application to offload queries.

B.

Configure Bigtable, and modify your application to offload queries.

C.

Configure Cloud SQL for PostgreSQL read replicas to offload queries.

D.

Configure Memorystore, and modify your application to offload queries.

Full Access
Question # 26

You work for a large retail and ecommerce company that is starting to extend their business globally. Your company plans to migrate to Google Cloud. You want to use platforms that will scale easily, handle transactions with the least amount of latency, and provide a reliable customer experience. You need a storage layer for sales transactions and current inventory levels. You want to retain the same relational schema that your existing platform uses. What should you do?

A.

Store your data in Firestore in a multi-region location, and place your compute resources in one of the constituent regions.

B.

Deploy Cloud Spanner using a multi-region instance, and place your compute resources close to the default leader region.

C.

Build an in-memory cache in Memorystore, and deploy to the specific geographic regions where your application resides.

D.

Deploy a Bigtable instance with a cluster in one region and a replica cluster in another geographic region.

Full Access
Question # 27

You are managing a Cloud SQL for PostgreSQL instance in Google Cloud. You need to test the high availability of your Cloud SQL instance by performing a failover. You want to use the cloud command.

What should you do?

A.

Use gcloud sql instances failover .

B.

Use gcloud sql instances failover .

C.

Use gcloud sql instances promote-replica .

D.

Use gcloud sql instances promote-replica .

Full Access
Question # 28

You are managing a Cloud SQL for PostgreSQL instance in Google Cloud. You have a primary instance in region 1 and a read replica in region 2. After a failure of region 1, you need to make the Cloud SQL instance available again. You want to minimize data loss and follow Google-recommended practices. What should you do?

A.

Restore the Cloud SQL instance from the automatic backups in region 3.

B.

Restore the Cloud SQL instance from the automatic backups in another zone in region 1.

C.

Check "Lag Bytes" in the monitoring dashboard for the primary instance in the read replica instance. Check the replication status usingpg_catalog.pg_last_wal_receive_lsn(). Then, fail over to region 2 by promoting the read replica instance.

D.

Check your instance operational log for the automatic failover status. Look for time, type, and status of the operations. If the failover operation is successful, no action is necessary. Otherwise, manually perform gcloud sql instances failover .

Full Access
Question # 29

Your organization has a ticketing system that needs an online marketing analytics and reporting application. You need to select a relational database that can manage hundreds of terabytes of data to support this new application. Which database should you use?

A.

Cloud SQL

B.

BigQuery

C.

Cloud Spanner

D.

Bigtable

Full Access
Question # 30

Your online delivery business that primarily serves retail customers uses Cloud SQL for MySQL for its inventory and scheduling application. The required recovery time objective (RTO) and recovery point objective (RPO) must be in minutes rather than hours as a part of your high availability and disaster recovery design. You need a high availability configuration that can recover without data loss during a zonal or a regional failure. What should you do?

A.

Set up all read replicas in a different region using asynchronous replication.

B.

Set up all read replicas in the same region as the primary instance with synchronous replication.

C.

Set up read replicas in different zones of the same region as the primary instance with synchronous replication, and set up read replicas in different regions with asynchronous replication.

D.

Set up read replicas in different zones of the same region as the primary instance with asynchronous replication, and set up read replicas in different regions with synchronous replication.

Full Access
Question # 31

Your company is migrating all legacy applications to Google Cloud. All on-premises applications are using legacy Oracle 12c databases with Oracle Real Application Cluster (RAC) for high availability (HA) and Oracle Data Guard for disaster recovery. You need a solution that requires minimal code changes, provides the same high availability you have today on-premises, and supports a low latency network for migrated legacy applications. What should you do?

A.

Migrate the databases to Cloud Spanner.

B.

Migrate the databases to Cloud SQL, and enable a standby database.

C.

Migrate the databases to Compute Engine using regional persistent disks.

D.

Migrate the databases to Bare Metal Solution for Oracle.

Full Access
Question # 32

You have a large Cloud SQL for PostgreSQL instance. The database instance is not mission-critical, and you want to minimize operational costs. What should you do to lower the cost of backups in this environment?

A.

Set the automated backups to occur every other day to lower the frequency of backups.

B.

Change the storage tier of the automated backups from solid-state drive (SSD) to hard disk drive (HDD).

C.

Select a different region to store your backups.

D.

Reduce the number of automated backups that are retained to two (2).

Full Access
Question # 33

You are designing an augmented reality game for iOS and Android devices. You plan to use Cloud Spanner as the primary backend database for game state storage and player authentication. You want to track in-game rewards that players unlock at every stage of the game. During the testing phase, you discovered that costs are much higher than anticipated, but the query response times are within the SLA. You want to follow Google-recommended practices. You need the database to be performant and highly available while you keep costs low. What should you do?

A.

Manually scale down the number of nodes after the peak period has passed.

B.

Use interleaving to co-locate parent and child rows.

C.

Use the Cloud Spanner query optimizer to determine the most efficient way to execute the SQL query.

D.

Use granular instance sizing in Cloud Spanner and Autoscaler.

Full Access
Question # 34

You currently have a MySQL database running on Cloud SQL with a read replica in a different zone for non-mission critical analytics workloads. You want to enable high availability (HA) for the analytic workloads while keeping costs low. What should you do?

A.

Increase the size of the read replica Instance and enable MA.

B.

Enable HA on the current read replica.

C.

Create a new HA instance in the same zone db lie primary.

D.

Create a new MA Instance in a different region than the primary.

Full Access
Question # 35

You are designing a highly available (HA) Cloud SQL for PostgreSQL instance that will be used by 100 databases. Each database contains 80 tables that were migrated from your on-premises environment to Google Cloud. The applications that use these databases are located in multiple regions in the US, and you need to ensure that read and write operations have low latency. What should you do?

A.

Deploy 2 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-east1 and us-west1.

B.

Deploy 2 Cloud SQL instances in the us-central1 region, and create read replicas in us-east1 and us-west1.

C.

Deploy 4 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-central1, us-east1, and us-west1.

D.

Deploy 4 Cloud SQL instances in the us-central1 region, and create read replicas in us-central1, us-east1 and us-west1.

Full Access
Question # 36

You need to provision several hundred Cloud SQL for MySQL instances for multiple project teams over a one-week period. You must ensure that all instances adhere to company standards such as instance naming conventions, database flags, and tags. What should you do?

A.

Automate instance creation by writing a Dataflow job.

B.

Automate instance creation by setting up Terraform scripts.

C.

Create the instances using the Google Cloud Console UI.

D.

Create clones from a template Cloud SQL instance.

Full Access
Question # 37

Your organization deployed a new version of a critical application that uses Cloud SQL for MySQL with high availability (HA) and binary logging enabled to store transactional information. The latest release of the application had an error that caused massive data corruption in your Cloud SQL for MySQL database. You need to minimize data loss. What should you do?

A.

Open the Google Cloud Console, navigate to SQL > Backups, and select the last version of the automated backup before the corruption.

B.

Reload the Cloud SQL for MySQL database using the LOAD DATA command to load data from CSV files that were used to initialize the instance.

C.

Perform a point-in-time recovery of your Cloud SQL for MySQL database, selecting a date and time before the data was corrupted.

D.

Fail over to the Cloud SQL for MySQL HA instance. Use that instance to recover the transactions that occurred before the corruption.

Full Access
Question # 38

You are running a mission-critical application on a Cloud SQL for PostgreSQL database with a multi-zonal setup. The primary and read replica instances are in the same regionbut in different zones. You need to ensure that you split the application load between both instances. What should you do?

A.

Use Cloud Load Balancing for load balancing between the Cloud SQL primary and read replica instances.

B.

Use PgBouncer to set up database connection pooling between the Cloud SQL primary and read replica instances.

C.

Use HTTP(S) Load Balancing for database connection pooling between the Cloud SQL primary and read replica instances.

D.

Use the Cloud SQL Auth proxy for database connection pooling between the Cloud SQL primary and read replica instances.

Full Access
Question # 39

You are a DBA of Cloud SQL for PostgreSQL. You want the applications to have password-less authentication for read and write access to the database. Which authentication mechanism should you use?

A.

Use Identity and Access Management (IAM) authentication.

B.

Use Managed Active Directory authentication.

C.

Use Cloud SQL federated queries.

D.

Use PostgreSQL database's built-in authentication.

Full Access
Question # 40

You are managing a Cloud SQL for MySQL environment in Google Cloud. You have deployed a primary instance in Zone A and a read replica instance in Zone B, both in the same region. You are notified that the replica instance in Zone B was unavailable for 10 minutes. You need to ensure that the read replica instance is still working. What should you do?

A.

Use the Google Cloud Console or gcloud CLI to manually create a new clone database.

B.

Use the Google Cloud Console or gcloud CLI to manually create a new failover replica from backup.

C.

Verify that the new replica is created automatically.

D.

Start the original primary instance and resume replication.

Full Access
Question # 41

Your company's mission-critical, globally available application is supported by a Cloud Spanner database. Experienced users of the application have read and write access to the database, but new users are assigned read-only access to the database. You need to assign the appropriate Cloud Spanner Identity and Access Management (IAM) role to new users being onboarded soon. What roles should you set up?

A.

roles/spanner.databaseReader

B.

roles/spanner.databaseUser

C.

roles/spanner.viewer

D.

roles/spanner.backupWriter

Full Access