Black Friday Sale - Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 65percent

Welcome To DumpsPedia

Professional-Cloud-Database-Engineer Sample Questions Answers

Questions 4

Your digital-native business runs its database workloads on Cloud SQL. Your website must be globally accessible 24/7. You need to prepare your Cloud SQL instance for high availability (HA). You want to follow Google-recommended practices. What should you do? (Choose two.)

Options:

A.

Set up manual backups.

B.

Create a PostgreSQL database on-premises as the HA option.

C.

Configure single zone availability for automated backups.

D.

Enable point-in-time recovery.

E.

Schedule automated backups.

Buy Now
Questions 5

Your organization is running a low-latency reporting application on Microsoft SQL Server. In addition to the database engine, you are using SQL Server Analysis Services (SSAS), SQL Server Reporting Services (SSRS), and SQL Server Integration Services (SSIS) in your on-premises environment. You want to migrate your Microsoft SQL Server database instances to Google Cloud. You need to ensure minimal disruption to the existing architecture during migration. What should you do?

Options:

A.

Migrate to Cloud SQL for SQL Server.

B.

Migrate to Cloud SQL for PostgreSQL.

C.

Migrate to Compute Engine.

D.

Migrate to Google Kubernetes Engine (GKE).

Questions 6

You are setting up a new AlloyDB instance and want users to be able to use their existing identity and Access Managemen (IAM) identities to connect to AlloyDB. You have performed the following steps:

Manually enabled IAM authentication on the AlloyDB instance

Granted the allowdb-databaseUser and serviceusage.serviceusageConsumer IAM roles to the users

Created new AllowDB database users based on corresponding IAM identities

Users are able to connect but are reporting that they are not able to SELECT from application tables. What should you do?

Options:

A.

Grant the alloydb.client IAM role to each user.

B.

Grant the alloydb.viewer IAM role to each user.

C.

Grant the new database users access privileges to the appropriate tables.

D.

Grant the alloydb.alloydbreplica IAM role to each user.

Buy Now
Questions 7

Your organization has a production Cloud SQL for MySQL instance. Your instance is configured with 16 vCPUs and 104 GB of RAM that is running between 90% and 100% CPU utilization for most of the day. You need to scale up the database and add vCPUs with minimal interruption and effort. What should you do?

Options:

A.

Issue a gcloud sql instances patch command to increase the number of vCPUs.

B.

Update a MySQL database flag to increase the number of vCPUs.

C.

Issue a gcloud compute instances update command to increase the number of vCPUs.

D.

Back up the database, create an instance with additional vCPUs, and restore the database.

Questions 8

You want to migrate your on-premises PostgreSQL database to Compute Engine. You need to migrate this database with the minimum downtime possible. What should you do?

Options:

A.

Perform a full backup of your on-premises PostgreSQL, and then, in the migration window, perform an incremental backup.

B.

Create a read replica on Cloud SQL, and then promote it to a read/write standalone instance.

C.

Use Database Migration Service to migrate your database.

D.

Create a hot standby on Compute Engine, and use PgBouncer to switch over the connections.

Buy Now
Questions 9

You are managing a mission-critical Cloud SQL for PostgreSQL instance. Your application team is running important transactions on the database when another DBA starts an on-demand backup. You want to verify the status of the backup. What should you do?

Options:

A.

Check the cloudsql.googleapis.com/postgres.log instance log.

B.

Perform the gcloud sql operations list command.

C.

Use Cloud Audit Logs to verify the status.

D.

Use the Google Cloud Console.

Buy Now
Questions 10

Your online delivery business that primarily serves retail customers uses Cloud SQL for MySQL for its inventory and scheduling application. The required recovery time objective (RTO) and recovery point objective (RPO) must be in minutes rather than hours as a part of your high availability and disaster recovery design. You need a high availability configuration that can recover without data loss during a zonal or a regional failure. What should you do?

Options:

A.

Set up all read replicas in a different region using asynchronous replication.

B.

Set up all read replicas in the same region as the primary instance with synchronous replication.

C.

Set up read replicas in different zones of the same region as the primary instance with synchronous replication, and set up read replicas in different regions with asynchronous replication.

D.

Set up read replicas in different zones of the same region as the primary instance with asynchronous replication, and set up read replicas in different regions with synchronous replication.

Buy Now
Questions 11

Your company is migrating their MySQL database to Cloud SQL and cannot afford any planned downtime during the month of December. The company is also concerned with cost, so you need the most cost-effective solution. What should you do?

Options:

A.

Open a support ticket in Google Cloud to prevent any maintenance in that MySQL instance during the month of December.

B.

Use Cloud SQL maintenance settings to prevent any maintenance during the month of December.

C.

Create MySQL read replicas in different zones so that, if any downtime occurs, the read replicas will act as the primary instance during the month of December.

D.

Create a MySQL regional instance so that, if any downtime occurs, the standby instance will act as the primary instance during the month of December.

Questions 12

You are managing multiple applications connecting to a database on Cloud SQL for PostgreSQL. You need to be able to monitor database performance to easily identify applications with long-running and resource-intensive queries. What should you do?

Options:

A.

Use log messages produced by Cloud SQL.

B.

Use Query Insights for Cloud SQL.

C.

Use the Cloud Monitoring dashboard with available metrics from Cloud SQL.

D.

Use Cloud SQL instance monitoring in the Google Cloud Console.

Questions 13

You are building an Android game that needs to store data on a Google Cloud serverless database. The database will log user activity, store user preferences, and receive in-game updates. The target audience resides in developing countries that have intermittent internet connectivity. You need to ensure that the game can synchronize game data to the backend database whenever an internet network is available. What should you do?

Options:

A.

Use Firestore.

B.

Use Cloud SQL with an external (public) IP address.

C.

Use an in-app embedded database.

D.

Use Cloud Spanner.

Buy Now
Questions 14

Your company is developing a global ecommerce website on Google Cloud. Your development team is working on a shopping cart service that is durable and elastically scalable with live traffic. Business disruptions from unplanned downtime are expected to be less than 5 minutes per month. In addition, the application needs to have very low latency writes. You need a data storage solution that has high write throughput and provides 99.99% uptime. What should you do?

Options:

A.

Use Cloud SQL for data storage.

B.

Use Cloud Spanner for data storage.

C.

Use Memorystore for data storage.

D.

Use Bigtable for data storage.

Buy Now
Questions 15

You host an application in Google Cloud. The application is located in a single region and uses Cloud SQL for transactional data. Most of your users are located in the same time zone and expect the application to be available 7 days a week, from 6 AM to 10PM. You want to ensure regular maintenance updates to your Cloud SQL instance without creating downtime for your users. What should you do?

Options:

A.

Configure a maintenance window during a period when no users will be on the system. Control the order of update by setting non-production instances to earlier and production instances to later.

B.

Create your database with one primary node and one read replica in the region.

C.

Enable maintenance notifications for users, and reschedule maintenance activities to a specific time after notifications have been sent.

D.

Configure your Cloud SQL instance with high availability enabled.

Buy Now
Questions 16

You are running a mission-critical application on a Cloud SQL for PostgreSQL database with a multi-zonal setup. The primary and read replica instances are in the same regionbut in different zones. You need to ensure that you split the application load between both instances. What should you do?

Options:

A.

Use Cloud Load Balancing for load balancing between the Cloud SQL primary and read replica instances.

B.

Use PgBouncer to set up database connection pooling between the Cloud SQL primary and read replica instances.

C.

Use HTTP(S) Load Balancing for database connection pooling between the Cloud SQL primary and read replica instances.

D.

Use the Cloud SQL Auth proxy for database connection pooling between the Cloud SQL primary and read replica instances.

Buy Now
Questions 17

You are running an instance of Cloud Spanner as the backend of your ecommerce website. You learn that the quality assurance (QA) team has doubled the number of their test cases. You need to create a copy of your Cloud Spanner database in a new test environment to accommodate the additional test cases. You want to follow Google-recommended practices. What should you do?

Options:

A.

Use Cloud Functions to run the export in Avro format.

B.

Use Cloud Functions to run the export in text format.

C.

Use Dataflow to run the export in Avro format.

D.

Use Dataflow to run the export in text format.

Questions 18

You finished migrating an on-premises MySQL database to Cloud SQL. You want to ensure that the daily export of a table, which was previously a cron job running on the database server, continues. You want the solution to minimize cost and operations overhead. What should you do?

Options:

A.

Use Cloud Scheduler and Cloud Functions to run the daily export.

B.

Create a streaming Datatlow job to export the table.

C.

Set up Cloud Composer, and create a task to export the table daily.

D.

Run the cron job on a Compute Engine instance to continue the export.

Questions 19

Your company wants to migrate an Oracle-based application to Google Cloud. The application team currently uses Oracle Recovery Manager (RMAN) to back up the database to tape for long-term retention (LTR). You need a cost-effective backup and restore solution that meets a 2-hour recovery time objective (RTO) and a 15-minute recovery point objective (RPO). What should you do?

Options:

A.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and store backups on tapes on-premises.

B.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and use Actifio to store backup files on Cloud Storage using the Nearline Storage class.

C.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and back up the Oracle databases to Cloud Storage using the Standard Storage class.

D.

Migrate the Oracle databases to Compute Engine, and store backups on tapes on-premises.

Buy Now
Questions 20

Your team is building a new inventory management application that will require read and write database instances in multiple Google Cloud regions around the globe. Your database solution requires 99.99% availability and global transactional consistency. You need a fully managed backend relational database to store inventory changes. What should you do?

Options:

A.

Use Bigtable.

B.

Use Firestore.

C.

Use Cloud SQL for MySQL

D.

Use Cloud Spanner.

Buy Now
Questions 21

You are running a transactional application on Cloud SQL for PostgreSQL in Google Cloud. The database is running in a high availability configuration within one region. You have encountered issues with data and want to restore to the last known pristine version of the database. What should you do?

Options:

A.

Create a clone database from a read replica database, and restore the clone in the same region.

B.

Create a clone database from a read replica database, and restore the clone into a different zone.

C.

Use the Cloud SQL point-in-time recovery (PITR) feature. Restore the copy from two hours ago to a new database instance.

D.

Use the Cloud SQL database import feature. Import last week's dump file from Cloud Storage.

Buy Now
Questions 22

Your organization stores marketing data such as customer preferences and purchase history on Bigtable. The consumers of this database are predominantly data analysts and operations users. You receive a service ticket from the database operations department citing poor database performance between 9 AM-10 AM every day. Theapplication team has confirmed no latency from their logs. A new cohort of pilot users that is testing a dataset loaded from a third-party data provider is experiencing poor database performance. Other users are not affected. You need to troubleshoot the issue. What should you do?

Options:

A.

Isolate the data analysts and operations user groups to use different Bigtable instances.

B.

Check the Cloud Monitoring table/bytes_used metric from Bigtable.

C.

Use Key Visualizer for Bigtable.

D.

Add more nodes to the Bigtable cluster.

Questions 23

You have a non-critical business application running on Google Kubernetes Engine (GKE) in the app-dev VPC. You have created an AllayDB cluster with private Service Access (PSA) and no public IP address in the db-dev VPC. You want your application to securely connect to AllowDB in a cost-effective way. What should you do?

Options:

A.

Set up a high availability VPN between the app-dev and db-dev VPCs. Connect the application directly to AlloyDB.

B.

Connect by using the private IP address of the AlloyDB cluster directly from the application.

C.

Connect by using AlloyDB Auth Proxy installed in the GKE cluster.

D.

Install a SOCKS proxy in a VM in the db-dev VPC. Install AlloyDB Auth Proxy in your GKE cluster, and connect to the AlloyDB cluster through the SOCKS server and port.

Buy Now
Questions 24

You are migrating an on-premises application to Google Cloud. The application requires a high availability (HA) PostgreSQL database to support business-critical functions. Your company's disaster recovery strategy requires a recovery time objective (RTO) and recovery point objective (RPO) within 30 minutes of failure. You plan to use a Google Cloud managed service. What should you do to maximize uptime for your application?

Options:

A.

Deploy Cloud SQL for PostgreSQL in a regional configuration. Create a read replica in a different zone in the same region and a read replica in another region for disaster recovery.

B.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Take periodic backups, and use this backup to restore to a new Cloud SQL for PostgreSQL instance in another region during a disaster recovery event.

C.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Create a cross-region read replica, and promote the read replica as the primary node for disaster recovery.

D.

Migrate the PostgreSQL database to multi-regional Cloud Spanner so that a single region outage will not affect your application. Update the schema to support Cloud Spanner data types, and refactor the application.

Buy Now
Questions 25

You recently launched a new product to the US market. You currently have two Bigtable clusters in one US region to serve all the traffic. Your marketing team is planning an immediate expansion to APAC. You need to roll out the regional expansion while implementing high availability according to Google-recommended practices. What should you do?

Options:

A.

Maintain a target of 23% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone europe-west1-d

cluster-c in zone asia-east1-b

B.

Maintain a target of 23% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone us-central1-b

cluster-c in zone us-east1-a

C.

Maintain a target of 35% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone australia-southeast1-a

cluster-c in zone europe-west1-d

cluster-d in zone asia-east1-b

D.

Maintain a target of 35% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone us-central2-a

cluster-c in zone asia-northeast1-b

cluster-d in zone asia-east1-b

Questions 26

Your company uses Bigtable for a user-facing application that displays a low-latency real-time dashboard. You need to recommend the optimal storage type for this read-intensive database. What should you do?

Options:

A.

Recommend solid-state drives (SSD).

B.

Recommend splitting the Bigtable instance into two instances in order to load balance the concurrent reads.

C.

Recommend hard disk drives (HDD).

D.

Recommend mixed storage types.

Buy Now
Questions 27

Your organization is migrating 50 TB Oracle databases to Bare Metal Solution for Oracle. Database backups must be available for quick restore. You also need to have backups available for 5 years. You need to design a cost-effective architecture that meets a recovery time objective (RTO) of 2 hours and recovery point objective (RPO) of 15 minutes. What should you do?

Options:

A.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on all flash storage.

Keep backups older than one day stored in Actifio OnVault storage.

B.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on standard storage.

Keep backups older than one day stored in Actifio OnVault storage.

C.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on standard storage.

Use the Oracle Recovery Manager (RMAN) backup utility to move backups older than one day to a Coldline Storage bucket.

D.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on all flash storage.

Use the Oracle Recovery Manager (RMAN) backup utility to move backups older than one day to an Archive Storage bucket.

Buy Now
Questions 28

You are using Compute Engine on Google Cloud and your data center to manage a set of MySQL databases in a hybrid configuration. You need to create replicas to scale reads and to offload part of the management operation. What should you do?

Options:

A.

Use external server replication.

B.

Use Data Migration Service.

C.

Use Cloud SQL for MySQL external replica.

D.

Use the mysqldump utility and binary logs.

Buy Now
Questions 29

You are building an application that allows users to customize their website and mobile experiences. The application will capture user information and preferences. User profiles have a dynamic schema, and users can add or delete information from their profile. You need to ensure that user changes automatically trigger updates to your downstream BigQuery data warehouse. What should you do?

Options:

A.

Store your data in Bigtable, and use the user identifier as the key. Use one column family to store user profile data, and use another column family to store user preferences.

B.

Use Cloud SQL, and create different tables for user profile data and user preferences from your recommendations model. Use SQL to join the user profile data and preferences

C.

Use Firestore in Native mode, and store user profile data as a document. Update the user profile with preferences specific to that user and use the user identifier to query.

D.

Use Firestore in Datastore mode, and store user profile data as a document. Update the user profile with preferences specific to that user and use the user identifier to query.

Buy Now
Questions 30

You want to migrate an on-premises 100 TB Microsoft SQL Server database to Google Cloud over a 1 Gbps network link. You have 48 hours allowed downtime to migrate this database. What should you do? (Choose two.)

Options:

A.

Use a change data capture (CDC) migration strategy.

B.

Move the physical database servers from on-premises to Google Cloud.

C.

Keep the network bandwidth at 1 Gbps, and then perform an offline data migration.

D.

Increase the network bandwidth to 2 Gbps, and then perform an offline data migration.

E.

Increase the network bandwidth to 10 Gbps, and then perform an offline data migration.

Questions 31

Your organization is currently updating an existing corporate application that is running in another public cloud to access managed database services in Google Cloud. The application will remain in the other public cloud while the database is migrated to Google Cloud. You want to follow Google-recommended practices for authentication. You need to minimize user disruption during the migration. What should you do?

Options:

A.

Use workload identity federation to impersonate a service account.

B.

Ask existing users to set their Google password to match their corporate password.

C.

Migrate the application to Google Cloud, and use Identity and Access Management (IAM).

D.

Use Google Workspace Password Sync to replicate passwords into Google Cloud.

Buy Now
Questions 32

Your company is shutting down their data center and migrating several MySQL and PostgreSQL databases to Google Cloud. Your database operations team is severely constrained by ongoing production releases and the lack of capacity for additional on-premises backups. You want to ensure that the scheduled migrations happen with minimal downtime and that the Google Cloud databases stay in sync with the on-premises data changes until the applications can cut over.

What should you do? (Choose two.)

Options:

A.

Use an external read replica to migrate the databases to Cloud SQL.

B.

Use a read replica to migrate the databases to Cloud SQL.

C.

Use Database Migration Service to migrate the databases to Cloud SQL.

D.

Use a cross-region read replica to migrate the databases to Cloud SQL.

E.

Use replication from an external server to migrate the databases to Cloud SQL.

Buy Now
Questions 33

Your company is developing a 24/7. global, real-lime analytics platform that needs to store and process large amounts of versioned time-series data. You need to

design a platform that Is highly scalable to accommodate traffic spikes and ensure high availability for mission-critical operations. What should you do?

Options:

A.

Implement a single-cluster Bigtable Instance with autoscaling enabled and row key design

B.

Implement a multi-cluster Bigtable Instance with autoscaling enabled and optimal schema design.

C.

Implement a multi-cluster Billable instance across multiple regions with replication.

D.

implement AlloyDB for PostgreSQL to handle the analytical workload using read replica

Buy Now
Questions 34

You are setting up a Bare Metal Solution environment. You need to update the operating system to the latest version. You need to connect the Bare Metal Solution environment to the internet so you can receive software updates. What should you do?

Options:

A.

Setup a static external IP address in your VPC network.

B.

Set up bring your own IP (BYOIP) in your VPC.

C.

Set up a Cloud NAT gateway on the Compute Engine VM.

D.

Set up Cloud NAT service.

Buy Now
Questions 35

You are designing a new gaming application that uses a highly transactional relational database to store player authentication and inventory data in Google Cloud. You want to launch the game in multiple regions. What should you do?

Options:

A.

Use Cloud Spanner to deploy the database.

B.

Use Bigtable with clusters in multiple regions to deploy the database

C.

Use BigQuery to deploy the database

D.

Use Cloud SQL with a regional read replica to deploy the database.

Buy Now
Questions 36

Your company uses the Cloud SQL out-of-disk recommender to analyze the storage utilization trends of production databases over the last 30 days. Your database operations team uses these recommendations to proactively monitor storage utilization and implement corrective actions. You receive a recommendation that the instance is likely to run out of disk space. What should you do to address this storage alert?

Options:

A.

Normalize the database to the third normal form.

B.

Compress the data using a different compression algorithm.

C.

Manually or automatically increase the storage capacity.

D.

Create another schema to load older data.

Questions 37

Your organization is running a MySQL workload in Cloud SQL. Suddenly you see a degradation in database performance. You need to identify the root cause of the performance degradation. What should you do?

Options:

A.

Use Logs Explorer to analyze log data.

B.

Use Cloud Monitoring to monitor CPU, memory, and storage utilization metrics.

C.

Use Error Reporting to count, analyze, and aggregate the data.

D.

Use Cloud Debugger to inspect the state of an application.

Questions 38

You use Python scripts to generate weekly SQL reports to assess the state of your databases and determine whether you need to reorganize tables or run statistics. You want to automate this report but need to minimize operational costs and overhead. What should you do?

Options:

A.

Create a VM in Compute Engine, and run a cron job.

B.

Create a Cloud Composer instance, and create a directed acyclic graph (DAG).

C.

Create a Cloud Function, and call the Cloud Function using Cloud Scheduler.

D.

Create a Cloud Function, and call the Cloud Function from a Cloud Tasks queue.

Buy Now
Questions 39

Your team uses thousands of connected IoT devices to collect device maintenance data for your oil and gas customers in real time. You want to design inspection routines, device repair, and replacement schedules based on insights gathered from the data produced by these devices. You need a managed solution that is highly scalable, supports a multi-cloud strategy, and offers low latency for these IoT devices. What should you do?

Options:

A.

Use Firestore with Looker.

B.

Use Cloud Spanner with Data Studio.

C.

Use MongoD8 Atlas with Charts.

D.

Use Bigtable with Looker.

Buy Now
Questions 40

You are starting a large CSV import into a Cloud SQL for MySQL instance that has many open connections. You checked memory and CPU usage, and sufficient resources are available. You want to follow Google-recommended practices to ensure that the import will not time out. What should you do?

Options:

A.

Close idle connections or restart the instance before beginning the import operation.

B.

Increase the amount of memory allocated to your instance.

C.

Ensure that the service account has the Storage Admin role.

D.

Increase the number of CPUs for the instance to ensure that it can handle the additional import operation.

Questions 41

Your organization has strict policies on tracking rollouts to production and periodically shares this information with external auditors to meet compliance requirements. You need to enable auditing on several Cloud Spanner databases. What should you do?

Options:

A.

Use replication to roll out changes to higher environments.

B.

Use backup and restore to roll out changes to higher environments.

C.

Use Liquibase to roll out changes to higher environments.

D.

Manually capture detailed DBA audit logs when changes are rolled out to higher environments.

Buy Now
Questions 42

You are developing a new application on a VM that is on your corporate network. The application will use Java Database Connectivity (JDBC) to connect to Cloud SQL for PostgreSQL. Your Cloud SQL instance is configured with IP address 192.168.3.48, and SSL is disabled. You want to ensure that your application can access your database instance without requiring configuration changes to your database. What should you do?

Options:

A.

Define a connection string using your Google username and password to point to the external (public) IP address of your Cloud SQL instance.

B.

Define a connection string using a database username and password to point to the internal (private) IP address of your Cloud SQL instance.

C.

Define a connection string using Cloud SQL Auth proxy configured with a service account to point to the internal (private) IP address of your Cloud SQL instance.

D.

Define a connection string using Cloud SQL Auth proxy configured with a service account to point to the external (public) IP address of your Cloud SQL instance.

Buy Now
Exam Code: Professional-Cloud-Database-Engineer
Exam Name: Google Cloud Certified - Professional Cloud Database Engineer
Last Update: Nov 22, 2025
Questions: 141
$57.75  $164.99
$43.75  $124.99
$36.75  $104.99
buy now Professional-Cloud-Database-Engineer