Weekend Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: cramtreat

Professional-Cloud-Database-Engineer Google Cloud Certified - Professional Cloud Database Engineer Questions and Answers

Questions 4

You want to migrate an on-premises mission-critical PostgreSQL database to Cloud SQL. The database must be able to withstand a zonal failure with less than five minutes of downtime and still not lose any transactions. You want to follow Google-recommended practices for the migration. What should you do?

Options:

A.

Take nightly snapshots of the primary database instance, and restore them in a secondary zone.

B.

Build a change data capture (CDC) pipeline to read transactions from the primary instance, and replicate them to a secondary instance.

C.

Create a read replica in another region, and promote the read replica if a failure occurs.

D.

Enable high availability (HA) for the database to make it regional.

Buy Now
Questions 5

Your application uses Cloud SQL for MySQL. Your users run reports on data that relies on near-real time; however, the additional analytics caused excessive load on the primary database. You created a read replica for the analytics workloads, but now your users are complaining about the lag in data changes and that their reports are still slow. You need to improve the report performance and shorten the lag in data replication without making changes to the current reports. Which two approaches should you implement? (Choose two.)

Options:

A.

Create secondary indexes on the replica.

B.

Create additional read replicas, and partition your analytics users to use different read replicas.

C.

Disable replication on the read replica, and set the flag for parallel replication on the read replica. Re-enable replication and optimize performance by setting flags on the primary instance.

D.

Disable replication on the primary instance, and set the flag for parallel replication on the primary instance. Re-enable replication and optimize performance by setting flags on the read replica.

E.

Move your analytics workloads to BigQuery, and set up a streaming pipeline to move data and update BigQuery.

Buy Now
Questions 6

Your organization has hundreds of Cloud SQL for MySQL instances. You want to follow Google-recommended practices to optimize platform costs. What should you do?

Options:

A.

Use Query Insights to identify idle instances.

B.

Remove inactive user accounts.

C.

Run the Recommender API to identify overprovisioned instances.

D.

Build indexes on heavily accessed tables.

Buy Now
Questions 7

During an internal audit, you realized that one of your Cloud SQL for MySQL instances does not have high availability (HA) enabled. You want to follow Google-recommended practices to enable HA on your existing instance. What should you do?

Options:

A.

Create a new Cloud SQL for MySQL instance, enable HA, and use the export and import option to migrate your data.

B.

Create a new Cloud SQL for MySQL instance, enable HA, and use Cloud Data Fusion to migrate your data.

C.

Use the gcloud instances patch command to update your existing Cloud SQL for MySQL instance.

D.

Shut down your existing Cloud SQL for MySQL instance, and enable HA.

Buy Now
Questions 8

Your company is evaluating Google Cloud database options for a mission-critical global payments gateway application. The application must be available 24/7 to users worldwide, horizontally scalable, and support open source databases. You need to select an automatically shardable, fully managed database with 99.999% availability and strong transactional consistency. What should you do?

Options:

A.

Select Bare Metal Solution for Oracle.

B.

Select Cloud SQL.

C.

Select Bigtable.

D.

Select Cloud Spanner.

Buy Now
Questions 9

You need to migrate a 1 TB PostgreSQL database from a Compute Engine VM to Cloud SQL for PostgreSQL. You want to ensure that there is minimal downtime during the migration. What should you do?

Options:

A.

Export the data from the existing database, and load the data into a new Cloud SQL database.

B.

Use Migrate for Compute Engine to complete the migration.

C.

Use Datastream to complete the migration.

D.

Use Database Migration Service to complete the migration.

Buy Now
Questions 10

Your team uses thousands of connected IoT devices to collect device maintenance data for your oil and gas customers in real time. You want to design inspection routines, device repair, and replacement schedules based on insights gathered from the data produced by these devices. You need a managed solution that is highly scalable, supports a multi-cloud strategy, and offers low latency for these IoT devices. What should you do?

Options:

A.

Use Firestore with Looker.

B.

Use Cloud Spanner with Data Studio.

C.

Use MongoD8 Atlas with Charts.

D.

Use Bigtable with Looker.

Buy Now
Questions 11

You are migrating a telehealth care company's on-premises data center to Google Cloud. The migration plan specifies:

PostgreSQL databases must be migrated to a multi-region backup configuration with cross-region replicas to allow restore and failover in multiple scenarios.

MySQL databases handle personally identifiable information (PII) and require data residency compliance at the regional level.

You want to set up the environment with minimal administrative effort. What should you do?

Options:

A.

Set up Cloud Logging and Cloud Monitoring with Cloud Functions to send an alert every time a new database instance is created, and manually validate the region.

B.

Set up different organizations for each database type, and apply policy constraints at the organization level.

C.

Set up Pub/Sub to ingest data from Cloud Logging, send an alert every time a new database instance is created, and manually validate the region.

D.

Set up different projects for PostgreSQL and MySQL databases, and apply organizational policy constraints at a project level.

Buy Now
Questions 12

You have deployed a Cloud SQL for SQL Server instance. In addition, you created a cross-region read replica for disaster recovery (DR) purposes. Your company requires you to maintain and monitor a recovery point objective (RPO) of less than 5 minutes. You need to verify that your cross-region read replica meets the allowed RPO. What should you do?

Options:

A.

Use Cloud SQL instance monitoring.

B.

Use the Cloud Monitoring dashboard with available metrics from Cloud SQL.

C.

Use Cloud SQL logs.

D.

Use the SQL Server Always On Availability Group dashboard.

Buy Now
Questions 13

Your retail organization is preparing for the holiday season. Use of catalog services is increasing, and your DevOps team is supporting the Cloud SQL databases that power a microservices-based application. The DevOps team has added instrumentation through Sqlcommenter. You need to identify the root cause of why certain microservice calls are failing. What should you do?

Options:

A.

Watch Query Insights for long running queries.

B.

Watch the Cloud SQL instance monitor for CPU utilization metrics.

C.

Watch the Cloud SQL recommenders for overprovisioned instances.

D.

Watch Cloud Trace for application requests that are failing.

Buy Now
Questions 14

Your organization works with sensitive data that requires you to manage your own encryption keys. You are working on a project that stores that data in a Cloud SQL database. You need to ensure that stored data is encrypted with your keys. What should you do?

Options:

A.

Export data periodically to a Cloud Storage bucket protected by Customer-Supplied Encryption Keys.

B.

Use Cloud SQL Auth proxy.

C.

Connect to Cloud SQL using a connection that has SSL encryption.

D.

Use customer-managed encryption keys with Cloud SQL.

Buy Now
Questions 15

You are running a transactional application on Cloud SQL for PostgreSQL in Google Cloud. The database is running in a high availability configuration within one region. You have encountered issues with data and want to restore to the last known pristine version of the database. What should you do?

Options:

A.

Create a clone database from a read replica database, and restore the clone in the same region.

B.

Create a clone database from a read replica database, and restore the clone into a different zone.

C.

Use the Cloud SQL point-in-time recovery (PITR) feature. Restore the copy from two hours ago to a new database instance.

D.

Use the Cloud SQL database import feature. Import last week's dump file from Cloud Storage.

Buy Now
Questions 16

Your organization deployed a new version of a critical application that uses Cloud SQL for MySQL with high availability (HA) and binary logging enabled to store transactional information. The latest release of the application had an error that caused massive data corruption in your Cloud SQL for MySQL database. You need to minimize data loss. What should you do?

Options:

A.

Open the Google Cloud Console, navigate to SQL > Backups, and select the last version of the automated backup before the corruption.

B.

Reload the Cloud SQL for MySQL database using the LOAD DATA command to load data from CSV files that were used to initialize the instance.

C.

Perform a point-in-time recovery of your Cloud SQL for MySQL database, selecting a date and time before the data was corrupted.

D.

Fail over to the Cloud SQL for MySQL HA instance. Use that instance to recover the transactions that occurred before the corruption.

Buy Now
Questions 17

Your organization is running a Firestore-backed Firebase app that serves the same top ten news stories on a daily basis to a large global audience. You want to optimize content delivery while decreasing cost and latency. What should you do?

Options:

A.

Enable serializable isolation in the Firebase app.

B.

Deploy a US multi-region Firestore location.

C.

Build a Firestore bundle, and deploy bundles to Cloud CDN.

D.

Create a Firestore index on the news story date.

Buy Now
Questions 18

You are designing a highly available (HA) Cloud SQL for PostgreSQL instance that will be used by 100 databases. Each database contains 80 tables that were migrated from your on-premises environment to Google Cloud. The applications that use these databases are located in multiple regions in the US, and you need to ensure that read and write operations have low latency. What should you do?

Options:

A.

Deploy 2 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-east1 and us-west1.

B.

Deploy 2 Cloud SQL instances in the us-central1 region, and create read replicas in us-east1 and us-west1.

C.

Deploy 4 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-central1, us-east1, and us-west1.

D.

Deploy 4 Cloud SQL instances in the us-central1 region, and create read replicas in us-central1, us-east1 and us-west1.

Buy Now
Questions 19

You are the primary DBA of a Cloud SQL for PostgreSQL database that supports 6 enterprise applications in production. You used Cloud SQL Insights to identify inefficient queries and now need to identify the application that is originating the inefficient queries. You want to follow Google-recommended practices. What should you do?

Options:

A.

Shut down and restart each application.

B.

Write a utility to scan database query logs.

C.

Write a utility to scan application logs.

D.

Use query tags to add application-centric database monitoring.

Buy Now
Questions 20

You are running an instance of Cloud Spanner as the backend of your ecommerce website. You learn that the quality assurance (QA) team has doubled the number of their test cases. You need to create a copy of your Cloud Spanner database in a new test environment to accommodate the additional test cases. You want to follow Google-recommended practices. What should you do?

Options:

A.

Use Cloud Functions to run the export in Avro format.

B.

Use Cloud Functions to run the export in text format.

C.

Use Dataflow to run the export in Avro format.

D.

Use Dataflow to run the export in text format.

Buy Now
Questions 21

You are choosing a new database backend for an existing application. The current database is running PostgreSQL on an on-premises VM and is managed by a database administrator and operations team. The application data is relational and has light traffic. You want to minimize costs and the migration effort for this application. What should you do?

Options:

A.

Migrate the existing database to Firestore.

B.

Migrate the existing database to Cloud SQL for PostgreSQL.

C.

Migrate the existing database to Cloud Spanner.

D.

Migrate the existing database to PostgreSQL running on Compute Engine.

Buy Now
Questions 22

You use Python scripts to generate weekly SQL reports to assess the state of your databases and determine whether you need to reorganize tables or run statistics. You want to automate this report but need to minimize operational costs and overhead. What should you do?

Options:

A.

Create a VM in Compute Engine, and run a cron job.

B.

Create a Cloud Composer instance, and create a directed acyclic graph (DAG).

C.

Create a Cloud Function, and call the Cloud Function using Cloud Scheduler.

D.

Create a Cloud Function, and call the Cloud Function from a Cloud Tasks queue.

Buy Now
Questions 23

Your project is using Bigtable to store data that should not be accessed from the public internet under any circumstances, even if the requestor has a valid service account key. You need to secure access to this data. What should you do?

Options:

A.

Use Identity and Access Management (IAM) for Bigtable access control.

B.

Use VPC Service Controls to create a trusted network for the Bigtable service.

C.

Use customer-managed encryption keys (CMEK).

D.

Use Google Cloud Armor to add IP addresses to an allowlist.

Buy Now
Questions 24

You are managing a Cloud SQL for PostgreSQL instance in Google Cloud. You have a primary instance in region 1 and a read replica in region 2. After a failure of region 1, you need to make the Cloud SQL instance available again. You want to minimize data loss and follow Google-recommended practices. What should you do?

Options:

A.

Restore the Cloud SQL instance from the automatic backups in region 3.

B.

Restore the Cloud SQL instance from the automatic backups in another zone in region 1.

C.

Check "Lag Bytes" in the monitoring dashboard for the primary instance in the read replica instance. Check the replication status using pg_catalog.pg_last_wal_receive_lsn(). Then, fail over to region 2 by promoting the read replica instance.

D.

Check your instance operational log for the automatic failover status. Look for time, type, and status of the operations. If the failover operation is successful, no action is necessary. Otherwise, manually perform gcloud sql instances failover .

Buy Now
Questions 25

You are deploying a new Cloud SQL instance on Google Cloud using the Cloud SQL Auth proxy. You have identified snippets of application code that need to access the new Cloud SQL instance. The snippets reside and execute on an application server running on a Compute Engine machine. You want to follow Google-recommended practices to set up Identity and Access Management (IAM) as quickly and securely as possible. What should you do?

Options:

A.

For each application code, set up a common shared user account.

B.

For each application code, set up a dedicated user account.

C.

For the application server, set up a service account.

D.

For the application server, set up a common shared user account.

Buy Now
Questions 26

You plan to use Database Migration Service to migrate data from a PostgreSQL on-premises instance to Cloud SQL. You need to identify the prerequisites for creating and automating the task. What should you do? (Choose two.)

Options:

A.

Drop or disable all users except database administration users.

B.

Disable all foreign key constraints on the source PostgreSQL database.

C.

Ensure that all PostgreSQL tables have a primary key.

D.

Shut down the database before the Data Migration Service task is started.

E.

Ensure that pglogical is installed on the source PostgreSQL database.

Buy Now
Questions 27

You are migrating an on-premises application to Google Cloud. The application requires a high availability (HA) PostgreSQL database to support business-critical functions. Your company's disaster recovery strategy requires a recovery time objective (RTO) and recovery point objective (RPO) within 30 minutes of failure. You plan to use a Google Cloud managed service. What should you do to maximize uptime for your application?

Options:

A.

Deploy Cloud SQL for PostgreSQL in a regional configuration. Create a read replica in a different zone in the same region and a read replica in another region for disaster recovery.

B.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Take periodic backups, and use this backup to restore to a new Cloud SQL for PostgreSQL instance in another region during a disaster recovery event.

C.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Create a cross-region read replica, and promote the read replica as the primary node for disaster recovery.

D.

Migrate the PostgreSQL database to multi-regional Cloud Spanner so that a single region outage will not affect your application. Update the schema to support Cloud Spanner data types, and refactor the application.

Buy Now
Questions 28

You are starting a large CSV import into a Cloud SQL for MySQL instance that has many open connections. You checked memory and CPU usage, and sufficient resources are available. You want to follow Google-recommended practices to ensure that the import will not time out. What should you do?

Options:

A.

Close idle connections or restart the instance before beginning the import operation.

B.

Increase the amount of memory allocated to your instance.

C.

Ensure that the service account has the Storage Admin role.

D.

Increase the number of CPUs for the instance to ensure that it can handle the additional import operation.

Buy Now
Questions 29

You need to perform a one-time migration of data from a running Cloud SQL for MySQL instance in the us-central1 region to a new Cloud SQL for MySQL instance in the us-east1 region. You want to follow Google-recommended practices to minimize performance impact on the currently running instance. What should you do?

Options:

A.

Create and run a Dataflow job that uses JdbcIO to copy data from one Cloud SQL instance to another.

B.

Create two Datastream connection profiles, and use them to create a stream from one Cloud SQL instance to another.

C.

Create a SQL dump file in Cloud Storage using a temporary instance, and then use that file to import into a new instance.

D.

Create a CSV file by running the SQL statement SELECT...INTO OUTFILE, copy the file to a Cloud Storage bucket, and import it into a new instance.

Buy Now
Questions 30

Your customer has a global chat application that uses a multi-regional Cloud Spanner instance. The application has recently experienced degraded performance after a new version of the application was launched. Your customer asked you for assistance. During initial troubleshooting, you observed high read latency. What should you do?

Options:

A.

Use query parameters to speed up frequently executed queries.

B.

Change the Cloud Spanner configuration from multi-region to single region.

C.

Use SQL statements to analyze SPANNER_SYS.READ_STATS* tables.

D.

Use SQL statements to analyze SPANNER_SYS.QUERY_STATS* tables.

Buy Now
Questions 31

You are a DBA on a Cloud Spanner instance with multiple databases. You need to assign these privileges to all members of the application development team on a specific database:

Can read tables, views, and DDL

Can write rows to the tables

Can add columns and indexes

Cannot drop the database

What should you do?

Options:

A.

Assign the Cloud Spanner Database Reader and Cloud Spanner Backup Writer roles.

B.

Assign the Cloud Spanner Database Admin role.

C.

Assign the Cloud Spanner Database User role.

D.

Assign the Cloud Spanner Admin role.

Buy Now
Questions 32

You work for a financial services company that wants to use fully managed database services. Traffic volume for your consumer services products has increased annually at a constant rate with occasional spikes around holidays. You frequently need to upgrade the capacity of your database. You want to use Cloud Spanner and include an automated method to increase your hardware capacity to support a higher level of concurrency. What should you do?

Options:

A.

Use linear scaling to implement the Autoscaler-based architecture

B.

Use direct scaling to implement the Autoscaler-based architecture.

C.

Upgrade the Cloud Spanner instance on a periodic basis during the scheduled maintenance window.

D.

Set up alerts that are triggered when Cloud Spanner utilization metrics breach the threshold, and then schedule an upgrade during the scheduled maintenance window.

Buy Now
Questions 33

You are the DBA of an online tutoring application that runs on a Cloud SQL for PostgreSQL database. You are testing the implementation of the cross-regional failover configuration. The database in region R1 fails over successfully to region R2, and the database becomes available for the application to process data. During testing, certain scenarios of the application work as expected in region R2, but a few scenarios fail with database errors. The application-related database queries, when executed in isolation from Cloud SQL for PostgreSQL in region R2, work as expected. The application performs completely as expected when the database fails back to region R1. You need to identify the cause of the database errors in region R2. What should you do?

Options:

A.

Determine whether the versions of Cloud SQL for PostgreSQL in regions R1 and R2 are different.

B.

Determine whether the database patches of Cloud SQI for PostgreSQL in regions R1 and R2 are different.

C.

Determine whether the failover of Cloud SQL for PostgreSQL from region R1 to region R2 is in progress or has completed successfully.

D.

Determine whether Cloud SQL for PostgreSQL in region R2 is a near-real-time copy of region R1 but not an exact copy.

Buy Now
Questions 34

Your organization is running a critical production database on a virtual machine (VM) on Compute Engine. The VM has an ext4-formatted persistent disk for data files. The database will soon run out of storage space. You need to implement a solution that avoids downtime. What should you do?

Options:

A.

In the Google Cloud Console, increase the size of the persistent disk, and use the resize2fs command to extend the disk.

B.

In the Google Cloud Console, increase the size of the persistent disk, and use the fdisk command to verify that the new space is ready to use

C.

In the Google Cloud Console, create a snapshot of the persistent disk, restore the snapshot to a new larger disk, unmount the old disk, mount the new disk, and restart the database service.

D.

In the Google Cloud Console, create a new persistent disk attached to the VM, and configure the database service to move the files to the new disk.

Buy Now
Questions 35

Your company's mission-critical, globally available application is supported by a Cloud Spanner database. Experienced users of the application have read and write access to the database, but new users are assigned read-only access to the database. You need to assign the appropriate Cloud Spanner Identity and Access Management (IAM) role to new users being onboarded soon. What roles should you set up?

Options:

A.

roles/spanner.databaseReader

B.

roles/spanner.databaseUser

C.

roles/spanner.viewer

D.

roles/spanner.backupWriter

Buy Now
Questions 36

Your application follows a microservices architecture and uses a single large Cloud SQL instance, which is starting to have performance issues as your application grows. in the Cloud Monitoring dashboard, the CPU utilization looks normal You want to follow Google-recommended practices to resolve and prevent these performance issues while avoiding any major refactoring. What should you do?

Options:

A.

Use Cloud Spanner instead of Cloud SQL.

B.

Increase the number of CPUs for your instance.

C.

Increase the storage size for the instance.

D.

Use many smaller Cloud SQL instances.

Buy Now
Questions 37

Your organization operates in a highly regulated industry. Separation of concerns (SoC) and security principle of least privilege (PoLP) are critical. The operations team consists of:

Person A is a database administrator.

Person B is an analyst who generates metric reports.

Application C is responsible for automatic backups.

You need to assign roles to team members for Cloud Spanner. Which roles should you assign?

Options:

A.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseReader for Person B

roles/spanner.backupWriter for Application C

B.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseReader for Person B

roles/spanner.backupAdmin for Application C

C.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseUser for Person B

roles/spanner databaseReader for Application C

D.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseUser for Person B

roles/spanner.backupWriter for Application C

Buy Now
Questions 38

You are running a large, highly transactional application on Oracle Real Application Cluster (RAC) that is multi-tenant and uses shared storage. You need a solution that ensures high-performance throughput and a low-latency connection between applications and databases. The solution must also support existing Oracle features and provide ease of migration to Google Cloud. What should you do?

Options:

A.

Migrate to Compute Engine.

B.

Migrate to Bare Metal Solution for Oracle.

C.

Migrate to Google Kubernetes Engine (GKE)

D.

Migrate to Google Cloud VMware Engine

Buy Now
Exam Name: Google Cloud Certified - Professional Cloud Database Engineer
Last Update: May 17, 2024
Questions: 132
Professional-Cloud-Database-Engineer pdf

Professional-Cloud-Database-Engineer PDF

$28  $80
Professional-Cloud-Database-Engineer Engine

Professional-Cloud-Database-Engineer Testing Engine

$33.25  $95
Professional-Cloud-Database-Engineer PDF + Engine

Professional-Cloud-Database-Engineer PDF + Testing Engine

$45.5  $130