FlashGenius Logo FlashGenius
Login Sign Up

GCP-PCA Practice Questions: Analyzing and optimizing technical and business processes Domain

Test your GCP-PCA knowledge with 10 practice questions from the Analyzing and optimizing technical and business processes domain. Includes detailed explanations and answers.

GCP-PCA Practice Questions

Master the Analyzing and optimizing technical and business processes Domain

Test your knowledge in the Analyzing and optimizing technical and business processes domain with these 10 practice questions. Each question is designed to help you prepare for the GCP-PCA certification exam with detailed explanations to reinforce your learning.

Question 1

A SaaS company provides a multi-tenant analytics platform hosted on Google Cloud. The architecture includes: - A front-end on Cloud Run - A backend API on GKE - A multi-tenant BigQuery dataset for analytics queries The company is experiencing: - Unpredictable BigQuery costs due to a few tenants running very large ad-hoc queries - Occasional query slowdowns affecting other tenants Business and technical requirements: - Maintain a shared analytics platform to avoid per-tenant infrastructure duplication - Provide predictable costs and performance isolation between tenants - Minimize operational overhead for the small SRE team - Avoid significant changes to the existing application code Which approach best optimizes cost and performance while meeting these requirements?

A) Create a separate BigQuery project and dataset for each tenant, and route queries to the appropriate project using service accounts, enabling per-tenant cost tracking and quotas at the project level.

B) Use BigQuery reservations with separate reservation assignments for different tenant tiers (e.g., standard vs premium), and implement per-tenant query cost controls using BigQuery custom quotas and monitoring, while keeping a shared dataset.

C) Export all BigQuery data to Cloud Storage daily and require tenants to run their heavy analytics using Dataproc clusters they pay for directly, reducing BigQuery usage and shifting costs to tenants.

D) Enable BigQuery BI Engine for the shared dataset to accelerate queries and rely on faster performance to reduce total costs and mitigate the impact of heavy queries from a few tenants.

Show Answer & Explanation

Correct Answer: B

Explanation:

Option B is best because it leverages BigQuery reservations and custom quotas to provide predictable capacity and performance isolation between tenant tiers while maintaining a shared dataset. Reservations allow you to allocate dedicated slots to different groups (e.g., premium tenants) and control how much capacity they can consume. Custom quotas and monitoring help detect and limit abusive or unusually heavy usage, improving cost predictability. This approach requires minimal changes to the application and keeps operational overhead manageable.

Why not A: Per-tenant projects and datasets provide strong isolation and cost tracking but introduce significant operational complexity (provisioning, IAM, schema management, monitoring across many projects). For a multi-tenant SaaS with many tenants, this does not align with minimizing operational overhead.

Why not C: Exporting data to Cloud Storage and forcing tenants to use Dataproc shifts complexity and cost management to tenants and degrades the SaaS experience. It also adds data latency and operational burden (exports, cluster management, support). This conflicts with the requirement to maintain a shared analytics platform and minimize operational overhead.

Why not D: BI Engine can accelerate certain types of queries, especially dashboard-style BI workloads, but it does not inherently provide cost controls or strong performance isolation between tenants. Heavy ad-hoc queries can still consume resources and impact others. This option addresses symptoms but not the underlying multi-tenant cost and isolation problem.

Domain alignment: The question focuses on analyzing multi-tenant cost and performance issues and applying BigQuery resource management features to optimize business and technical outcomes.

Question 2

A healthcare analytics startup processes sensitive patient data for hospitals in the EU. They currently run nightly ETL pipelines on a self-managed Hadoop cluster in a single EU region. The business wants: - Near-real-time analytics (latency under 5 minutes) for key dashboards - Strict adherence to EU data residency and healthcare compliance requirements - Reduced operational overhead and faster iteration on new data transformations - Predictable costs with the ability to scale up during peak hours and down during off-hours Which architecture change best optimizes their technical and business processes while meeting constraints?

A) Migrate the Hadoop cluster to Compute Engine in the same EU region and use autoscaling managed instance groups to scale the cluster based on CPU utilization, keeping the nightly batch model.

B) Replace the Hadoop cluster with a regional Dataproc cluster and schedule more frequent batch jobs (every 15 minutes) to approximate near-real-time analytics while using cluster autoscaling.

C) Implement a streaming pipeline using Pub/Sub and Dataflow in an EU region, write curated data to BigQuery with partitioned tables, and build dashboards directly on BigQuery.

D) Use Cloud Storage in an EU multi-region as a data lake, run scheduled Dataflow batch jobs every 10 minutes to process new files, and load results into Cloud SQL for dashboard queries.

Show Answer & Explanation

Correct Answer: C

Explanation:

The key drivers are near-real-time analytics, EU data residency, compliance, reduced ops overhead, and cost elasticity.

C is best: A streaming architecture with Pub/Sub and Dataflow in an EU region supports sub-5-minute latency for analytics. Dataflow is fully managed, reducing operational overhead compared to self-managed Hadoop. Writing to BigQuery (regional EU dataset) provides a compliant, highly scalable analytics store with strong integration to BI tools. Partitioned tables help control cost and performance. You can scale Dataflow jobs based on throughput and use autoscaling to match peak/off-peak demand, aligning cost with usage.

A is suboptimal: Simply lifting Hadoop to Compute Engine preserves the nightly batch model and doesn’t meet the near-real-time requirement. It also keeps high operational overhead (cluster management, tuning, upgrades). Autoscaling helps cost somewhat but doesn’t address the core latency and agility needs.

B is an improvement but still limited: Dataproc reduces some management overhead compared to self-managed Hadoop, but it’s still a cluster-based model. Running jobs every 15 minutes may not reliably meet a 5-minute latency target, especially under variable load. It also keeps a batch mindset rather than event-driven streaming, and operational complexity remains higher than with Dataflow.

D is partially valid but not optimal: Cloud Storage + frequent Dataflow batch jobs can reduce latency, but 10-minute schedules may not consistently meet the 5-minute requirement. Cloud SQL is not ideal for large-scale analytics workloads; it can become a performance and cost bottleneck for complex queries and many concurrent dashboard users. BigQuery is better suited for analytics at scale with simpler operations.

Therefore, C best aligns with the performance, compliance, cost, and operational requirements while modernizing the data processing model.

Question 3

A healthcare analytics company ingests sensitive patient data from multiple hospitals into Google Cloud for reporting and machine learning. The current architecture: - Uses Cloud Storage for raw data - Runs nightly Dataflow jobs to transform and load data into BigQuery - Provides dashboards to hospital users via Looker Studio New regulatory requirements mandate: - Data residency in the EU for all patient-identifiable data - Fine-grained access control so each hospital can only see its own data - End-to-end auditability of data access and transformations The company also wants to: - Minimize operational overhead - Avoid duplicating pipelines per hospital - Keep costs under control while data volume is growing 30% per quarter What should you recommend to optimize the architecture and processes while meeting the new requirements?

A) Create separate EU-based projects and BigQuery datasets per hospital, run separate Dataflow pipelines for each hospital, and use project-level IAM to restrict access.

B) Centralize all data in a single EU-based BigQuery dataset, use row-level security and authorized views to restrict access per hospital, and configure Cloud Audit Logs and Data Catalog for lineage and access auditing.

C) Store each hospital’s data in separate EU-based Cloud Storage buckets, run a single Dataflow pipeline that writes to separate BigQuery datasets per hospital, and control access using dataset-level IAM.

D) Use BigQuery Omni to keep data in each hospital’s on-premises environment, query it from a central project, and rely on network-level controls and VPC Service Controls for security and compliance.

Show Answer & Explanation

Correct Answer: B

Explanation:

Option B is best because it meets regulatory and business requirements while optimizing operations and cost. Centralizing data in a single EU-based BigQuery dataset ensures data residency. Row-level security and authorized views allow fine-grained, per-hospital access control without duplicating datasets or pipelines, which reduces operational overhead. Cloud Audit Logs provide access auditing, and Data Catalog (with lineage and tags) supports end-to-end auditability of data transformations and usage. A single, scalable pipeline is easier to maintain as data volume grows.

Option A meets isolation and residency but is operationally heavy and costly. Separate projects, datasets, and pipelines per hospital significantly increase management overhead, complicate deployments, and make global changes difficult. It also increases the risk of configuration drift and inconsistent policies.

Option C improves over A by using a single pipeline, but separate datasets per hospital still create more administrative overhead than necessary and complicate cross-hospital analytics if needed in the future. Dataset-level IAM is coarser than row-level security and may require more structural changes as requirements evolve.

Option D with BigQuery Omni is not aligned with the stated architecture and requirements. The data is already being ingested into Google Cloud, and Omni is designed for querying data where it resides (e.g., other clouds or on-prem). It would not simplify operations here and would likely increase complexity and cost. It also does not inherently solve fine-grained, per-hospital access control as cleanly as BigQuery’s built-in security features.

Question 4

A financial services company processes loan applications through a microservices-based system running on GKE. The workflow includes: - A public-facing API service - A credit scoring service - A document processing service that uses OCR - A risk analysis service that aggregates results The current design uses synchronous REST calls between services. During high load, the system experiences cascading latency and occasional timeouts, causing SLA violations and customer complaints. The business requirements are: - End-to-end processing time target: 5 minutes (not real-time) - High reliability and resilience to partial failures - Ability to prioritize premium customers’ applications - Minimal operational overhead for the platform team You are asked to optimize the architecture to meet these requirements while minimizing code changes. What should you do?

A) Introduce Cloud Pub/Sub to decouple services, using topics for each processing stage and subscriptions for each microservice, and implement dead-letter topics for failed messages.

B) Migrate the entire workflow to Cloud Composer and orchestrate each microservice as a task in a DAG, using retries and SLAs at the DAG level.

C) Use Cloud Tasks to queue requests for each microservice, with per-queue rate limits and retries, and keep the services communicating via HTTP callbacks.

D) Implement Cloud Run for all services with autoscaling based on request concurrency, and keep the synchronous REST pattern while increasing resource limits.

Show Answer & Explanation

Correct Answer: A

Explanation:

Option A is best because it directly addresses cascading latency and reliability by decoupling services with asynchronous messaging. Cloud Pub/Sub allows each microservice to process messages independently, absorb bursts, and avoid synchronous dependencies. Dead-letter topics improve operational visibility and resilience. Prioritization can be implemented with separate topics or subscriptions for premium customers, and the platform team benefits from a managed, low-ops messaging layer. The change can be made with relatively small code modifications (publishing and subscribing instead of direct REST calls) while preserving service boundaries.

Option B (Cloud Composer) is better suited for batch workflows and data pipelines, not high-throughput, event-driven microservice orchestration. It introduces additional operational complexity and is not optimized for per-request prioritization or large-scale, near-real-time processing.

Option C with Cloud Tasks can help manage retries and rate limiting, but it is more appropriate for background tasks initiated by user requests, not for fully decoupling a multi-stage microservice workflow. It also tends to maintain a request/response pattern and can become complex to manage across many services and priorities.

Option D improves scalability but keeps the synchronous coupling that causes cascading failures. Simply moving to Cloud Run and increasing resources does not address the architectural issue of synchronous dependencies and will likely still result in timeouts and SLA violations under heavy load.

Question 5

A media analytics company ingests clickstream events from multiple mobile apps worldwide. The current architecture is: - Events are sent to Pub/Sub. - A Dataflow streaming job enriches and aggregates events. - Results are written to BigQuery for analytics and to Cloud Storage for long-term archival. The business team complains that analytics dashboards are often delayed by 10–15 minutes, impacting their ability to react to live campaigns. The Dataflow job is configured with large window sizes and uses complex transformations. The CFO is concerned about rising Dataflow and BigQuery costs and wants to avoid a major rewrite of the pipeline. You are asked to optimize the architecture and processes to reduce end-to-end latency while keeping costs under control and minimizing operational complexity. What should you do?

A) Reduce the Dataflow window size and watermarking delay to process events more frequently, and enable BigQuery streaming inserts from Dataflow. Use BigQuery table partitioning and clustering to optimize query performance and cost.

B) Replace the Dataflow job with a custom streaming application on GKE that writes directly to BigQuery using the Storage Write API. Use aggressive autoscaling on the GKE cluster to handle peak loads.

C) Introduce Cloud Functions to pre-aggregate events from Pub/Sub and send summarized data to the existing Dataflow job at lower frequency, reducing Dataflow workload and BigQuery storage usage.

D) Configure Dataflow to write intermediate results to Cloud Storage every minute and schedule a BigQuery load job every 5 minutes to import the latest data into partitioned tables for analytics.

Show Answer & Explanation

Correct Answer: A

Explanation:

Option A is best because it directly addresses latency while leveraging existing managed services and minimizing architectural change. By reducing Dataflow window sizes and watermarking delays, you:
- Decrease the time Dataflow waits before emitting results, reducing end-to-end latency.
- Maintain the same overall pipeline structure, avoiding a costly rewrite.

Enabling BigQuery streaming inserts from Dataflow allows near-real-time data availability for dashboards. Using partitioned and clustered tables improves query performance and can reduce cost by:
- Scanning less data for time-bounded queries.
- Organizing data for common filter and aggregation patterns.

This approach aligns with the Well-Architected principles of performance efficiency and cost optimization while preserving operational simplicity.

Option B is technically feasible but suboptimal. Replacing Dataflow with a custom GKE-based streaming application increases operational overhead (cluster management, scaling logic, fault tolerance) and development effort. It contradicts the requirement to avoid a major rewrite and may not yield better cost or latency than a tuned Dataflow pipeline.

Option C adds complexity and may not significantly reduce latency. Introducing Cloud Functions for pre-aggregation creates another processing layer, increasing operational complexity and potential failure points. It may reduce some Dataflow and BigQuery costs but at the expense of flexibility and maintainability. It also does not directly address the core issue of large windows and watermark delays.

Option D still relies on batch-style loading into BigQuery, which inherently introduces latency due to the load job schedule and processing time. Writing intermediate results to Cloud Storage every minute and loading every 5 minutes will likely keep latency in the several-minute range and adds operational overhead for managing frequent load jobs.

Therefore, Option A provides the most balanced improvement in latency, cost control, and operational simplicity with minimal changes to the existing architecture.

Question 6

A healthcare SaaS provider hosts a multi-tenant patient management platform. They are migrating from a single-tenant, VM-based architecture on-premises to Google Cloud. Each customer (clinic or hospital) has its own database schema today. The company wants to reduce operational overhead and improve scalability while maintaining strong tenant isolation and meeting healthcare compliance requirements. Requirements: - PHI (Protected Health Information) must be encrypted at rest and in transit. - Each tenant must be logically isolated so that no tenant can access another tenant’s data. - The platform must support hundreds of small tenants and a few very large tenants with higher performance needs. - The operations team wants to minimize the number of database instances to manage while still being able to tune performance for large tenants. Constraints: - The company wants to avoid a full application rewrite and prefers to keep the existing per-tenant schema model for now. - They need a clear path to onboard new tenants quickly with minimal manual operations. Which database architecture on Google Cloud best balances isolation, manageability, and performance tuning needs?

A) Use a single multi-tenant Cloud SQL instance with one database per tenant and shared tables. Enforce tenant isolation using a tenant_id column and application-level access controls.

B) Use a single Cloud SQL instance with separate databases per tenant, each containing its own schema. Use database-level permissions to isolate tenants and configure read replicas for large tenants.

C) Use multiple Cloud SQL instances: one shared multi-tenant instance for small tenants (with separate databases per tenant) and dedicated Cloud SQL instances for very large tenants. Automate tenant provisioning with scripts or tooling.

D) Use Spanner with a single multi-tenant database and interleaved tables keyed by tenant_id. Use fine-grained IAM and row-level security to isolate tenants and scale horizontally as needed.

Show Answer & Explanation

Correct Answer: C

Explanation:

Option C provides a pragmatic balance between isolation, manageability, and performance tuning, while aligning with the existing per-tenant schema model.

Why C is best:
- A shared Cloud SQL instance for many small tenants (with separate databases per tenant) reduces the number of instances to manage, meeting the operations team’s goal.
- Dedicated Cloud SQL instances for very large tenants allow independent performance tuning (e.g., machine type, storage, replicas) without impacting smaller tenants.
- Per-tenant databases preserve the current per-tenant schema model, avoiding a full application rewrite.
- Logical isolation is improved compared to a single shared schema: database-level separation plus application controls reduce the risk of cross-tenant data access.
- Tenant onboarding can be automated by scripting database creation on the shared instance or provisioning a new dedicated instance for large tenants, supporting rapid onboarding.
- Cloud SQL provides encryption at rest by default and supports TLS for in-transit encryption, satisfying PHI requirements when combined with proper configuration and IAM.

Why not A:
- A single instance with one database per tenant but shared tables and tenant_id is effectively a shared schema model; it increases the risk of cross-tenant data exposure due to application bugs.
- It does not leverage database-level isolation, which is desirable for PHI and multi-tenant SaaS.
- Performance tuning per tenant is difficult because all tenants share the same tables and resources.

Why not B:
- A single Cloud SQL instance with separate databases per tenant improves isolation compared to A, but it still has limitations:
- All tenants share the same instance resources, making it hard to tune for very large tenants without overprovisioning for small ones.
- A noisy large tenant can impact performance for all others.
- While this is manageable for a moderate number of tenants, the requirement includes both hundreds of small tenants and a few very large ones, making a single instance a scaling and performance risk.

Why not D:
- Spanner is powerful and supports horizontal scaling, but it represents a significant architectural shift from per-tenant schemas to a shared, globally distributed database model.
- It likely requires more application changes (e.g., schema redesign, query changes) than the company is willing to undertake now, conflicting with the desire to avoid a full rewrite.
- Spanner is also typically more expensive and complex to operate than Cloud SQL for smaller workloads, which may not be justified for hundreds of small tenants.

Therefore, option C offers a hybrid approach: shared infrastructure for small tenants to minimize overhead, and dedicated instances for large tenants to allow targeted performance tuning and stronger isolation.

Question 7

A global retail company runs a monolithic Java application on Compute Engine managed instance groups behind a global external HTTP(S) load balancer. The application handles both customer-facing traffic and nightly batch processing. The operations team reports that during peak shopping hours, batch jobs slow down the user-facing API, causing latency SLO violations. The CIO wants to improve performance and reliability without significantly increasing operational overhead. The company has a moderate budget and wants to avoid a full rewrite in the short term. What should you do to optimize both technical and business processes while meeting these constraints?

A) Split the monolith into two separate instance groups: one for user-facing traffic and one for batch processing. Use separate backend services on the existing HTTP(S) load balancer and configure autoscaling policies tailored to each group. Schedule batch instance groups to scale down during peak hours and scale up during off-peak hours.

B) Migrate the entire monolithic application to a single regional GKE cluster with node autoscaling. Use Kubernetes HPA to scale pods based on CPU utilization and configure separate namespaces for batch and user-facing workloads.

C) Move the batch processing logic to Cloud Functions triggered by Cloud Scheduler. Keep the user-facing monolith on the existing instance group and reduce the size of the instances to save costs, relying on autoscaling to handle peak loads.

D) Containerize the monolith and deploy it to Cloud Run with maximum concurrency set to 1 to isolate requests. Use Cloud Scheduler to trigger batch endpoints and rely on Cloud Run autoscaling to handle both batch and user-facing workloads.

Show Answer & Explanation

Correct Answer: A

Explanation:

Option A is best because it optimizes both technical and business processes with minimal disruption and operational complexity. By separating batch and user-facing workloads into different managed instance groups, you can:
- Apply different autoscaling policies: aggressive scaling for user-facing traffic based on request load and latency, and scheduled or CPU-based scaling for batch jobs.
- Protect customer-facing SLOs by limiting or scheduling batch capacity during peak hours.
- Reuse existing patterns (HTTP(S) load balancer, MIGs) without a full platform migration, reducing risk and time-to-value.
- Improve cost efficiency by scaling batch capacity down when not needed.
This aligns with the Well-Architected principles of reliability (isolation of workloads), performance (right-sizing and scaling policies), and operational excellence (incremental change, reuse of existing tooling).

Option B is technically valid but suboptimal. Moving to GKE introduces significant operational overhead (cluster management, Kubernetes expertise, observability changes) for a team that currently runs MIGs. While namespaces and HPA help, they do not inherently isolate resource contention between batch and user-facing pods unless you also configure resource requests/limits, node pools, and potentially separate clusters. This is a larger transformation than needed for the stated short-term goal.

Option C is risky and incomplete. Moving batch logic to Cloud Functions may require substantial refactoring of a monolith, which contradicts the desire to avoid a full rewrite. Cloud Functions are not ideal for long-running or complex batch jobs and may hit execution limits. Reducing instance size to save costs could worsen performance during peaks, and relying solely on autoscaling may not fully protect user-facing SLOs if batch and user traffic still compete for shared resources.

Option D is also suboptimal. Cloud Run is well-suited for stateless services, but lifting a monolithic app directly into Cloud Run often requires significant changes (e.g., startup time, filesystem assumptions, long-running tasks). Setting concurrency to 1 greatly increases cost and may not scale efficiently for batch workloads. Running both batch and user-facing workloads on the same Cloud Run service still risks contention and complicates cost control.

Therefore, Option A provides the best balance of improved reliability, performance, cost control, and low migration risk.

Question 8

A media streaming company runs its recommendation engine on Google Cloud. The system ingests user interaction events, processes them, and updates personalized recommendations. Currently: • Events are ingested via Pub/Sub and processed by a Dataflow streaming job • Processed features are stored in Cloud Bigtable • Recommendation microservices on GKE read from Bigtable with p95 latency around 40 ms New business goals and constraints: • Launch in two additional regions to reduce latency for users globally • Maintain a p95 read latency under 50 ms for recommendations in all regions • Minimize cross-region data transfer costs • Ensure that recommendations remain eventually consistent across regions within a few minutes • Keep operational complexity manageable for a small SRE team What architecture change should you recommend for the data storage and access layer?

A) Keep a single regional Bigtable instance in the current region and use a global HTTP(S) Load Balancer to route read requests from all regions to the existing Bigtable instance. Rely on caching at the application layer to reduce latency.

B) Migrate from Bigtable to a multi-region Cloud Spanner instance and update the microservices to read from Spanner. Use Spanner’s synchronous replication to ensure strong consistency across regions.

C) Create separate regional Bigtable instances in each target region. Use Pub/Sub topics per region and regional Dataflow jobs to process events and write to the local Bigtable instance. Accept eventual consistency across regions.

D) Enable Bigtable replication across regions by creating additional clusters in the same Bigtable instance in each target region. Configure the microservices in each region to read from the local Bigtable cluster while writes are replicated asynchronously.

Show Answer & Explanation

Correct Answer: D

Explanation:

Option D best meets the latency, cost, consistency, and operational requirements:

• Bigtable multi-cluster routing with additional clusters in multiple regions allows local reads in each region, keeping p95 latency under 50 ms.
• Writes are replicated asynchronously across clusters, providing eventual consistency within minutes, which matches the business requirement.
• Cross-region data transfer is limited to replication traffic, which is more efficient than routing all reads cross-region.
• Operationally, this remains a single Bigtable instance with multiple clusters, which is simpler than managing multiple independent instances and pipelines.

Why the others are suboptimal:

A) A single regional Bigtable instance forces cross-region reads, increasing latency and cross-region egress costs. Application-level caching can help but is unlikely to consistently keep p95 latency under 50 ms for all users globally and adds complexity.

B) Cloud Spanner provides strong consistency and multi-region replication but is typically more expensive and may be overkill if strong consistency is not required. The requirement is eventual consistency within minutes, and migrating from Bigtable to Spanner plus schema redesign increases complexity and cost unnecessarily.

C) Separate regional Bigtable instances with independent pipelines provide local reads but require duplicating ingestion and processing pipelines per region, significantly increasing operational complexity. Ensuring consistent processing logic and monitoring across multiple pipelines is harder, and coordinating global behavior becomes complex compared to Bigtable’s built-in multi-cluster replication.

Question 9

A logistics company runs a route-optimization API on GKE in a single region. The API is latency-sensitive and must respond within 200 ms for 95% of requests. The business is expanding to new continents, and users in distant regions are experiencing higher latency. The company wants to: - Improve global user experience - Keep infrastructure and operations as simple as possible - Control costs while preparing for future growth - Maintain high availability (99.9% or better) The application is stateless, but it depends on a regional Cloud Memorystore instance for caching and a regional Cloud SQL instance for transactional data. Which architecture change best meets these goals?

A) Deploy additional GKE clusters, Cloud Memorystore instances, and Cloud SQL instances in each target region, implement active-active replication across all regions for both cache and database, and use global HTTP(S) Load Balancing with traffic steering based on client location.

B) Keep the existing regional GKE, Cloud Memorystore, and Cloud SQL deployment, front the API with Cloud CDN and global HTTP(S) Load Balancing, and configure aggressive caching of API responses at the edge to reduce perceived latency for users in distant regions.

C) Deploy additional GKE clusters in each target region for the API only, keep Cloud Memorystore and Cloud SQL in the original region, use global HTTP(S) Load Balancing with latency-based routing to direct users to the nearest API cluster, and rely on cross-region calls from the API to the original data stores.

D) Migrate the API to Cloud Run with multiple regional deployments, keep Cloud SQL in a primary region with read replicas in other regions, use global HTTP(S) Load Balancing with traffic steering to route reads to the nearest replica and writes to the primary, and use Cloud CDN only for static assets.

Show Answer & Explanation

Correct Answer: D

Explanation:

Option D best balances global performance, simplicity, cost, and availability.

Analysis of D:
- Cloud Run with multiple regional deployments simplifies operations compared to managing multiple GKE clusters, aligning with the desire for operational simplicity.
- Using Cloud SQL with a primary region and read replicas in other regions is a supported pattern that improves read latency for global users while maintaining a single write master.
- Global HTTP(S) Load Balancing with appropriate routing rules can direct traffic to the nearest regional Cloud Run service and read replica, improving latency while preserving consistency for writes.
- Cloud CDN for static assets offloads static content delivery, further improving user experience and reducing load on the backend.
- This design supports high availability and prepares for future growth with managed services.

Why A is suboptimal:
- Active-active replication for both cache and database across many regions is complex and expensive to implement and operate.
- Multi-region active-active for Cloud SQL is not straightforward; cross-region writes can introduce consistency and conflict issues.
- Managing multiple GKE clusters, caches, and databases in each region significantly increases operational overhead.

Why B is suboptimal:
- Cloud CDN is effective for cacheable, mostly static content, but a route-optimization API is typically dynamic and user-specific, making aggressive caching difficult without risking stale or incorrect responses.
- Latency from distant regions to a single backend region remains high; CDN cannot fully mask backend latency for dynamic requests.
- Does not address the underlying distance to the compute and data.

Why C is suboptimal:
- Deploying only the API in multiple regions while keeping data stores in a single region still requires cross-region calls for every request, which can dominate latency and may not meet the 200 ms requirement.
- Cross-region calls for every request also increase egress costs and can complicate failure scenarios.
- Multiple GKE clusters increase operational complexity compared to a managed platform like Cloud Run.

D uses managed, multi-regional deployments with appropriate data replication and routing to improve latency and availability while keeping the architecture relatively simple and cost-conscious.

Question 10

A healthcare provider is building a patient-facing portal on Google Cloud. The architecture includes: - A React front-end served via Cloud Storage and Cloud CDN - Backend APIs on Cloud Run - Patient data stored in Cloud SQL for MySQL Requirements and constraints: - Strict regulatory requirements for PHI (Protected Health Information) - All PHI must remain within a specific region - The portal must be highly available with an RTO of 1 hour and RPO of 5 minutes - The small DevOps team wants to minimize operational overhead and avoid managing complex DR runbooks - The business wants to control costs but is willing to pay for reliability and compliance Which design best optimizes for compliance, availability, and operational simplicity?

A) Deploy Cloud Run and Cloud SQL in a single region with high-availability configuration, enable automated backups and point-in-time recovery, and rely on restoring from backups to another region in case of a regional outage.

B) Deploy Cloud Run in two regions with global HTTP(S) Load Balancing, use Cloud SQL with cross-region replicas, and configure automatic failover to a replica in another region, ensuring PHI is replicated but never served outside the primary region.

C) Deploy Cloud Run and Cloud SQL in a single region with high-availability configuration, use Cloud Storage multi-region buckets for all PHI backups, and rely on Cloud CDN to serve cached data during outages.

D) Deploy Cloud Run in two regions with global HTTP(S) Load Balancing, use a single-region Cloud SQL instance with read replicas in the same region, and rely on Cloud Run’s multi-region deployment to meet the RTO and RPO requirements.

Show Answer & Explanation

Correct Answer: B

Explanation:

Option B is best because it balances compliance, availability, and operational simplicity. Cloud Run in two regions behind global HTTP(S) Load Balancing provides high availability for stateless services. Cloud SQL with cross-region replicas and configured automatic failover supports an RPO close to 5 minutes and RTO near 1 hour. PHI is replicated to another region for DR, but you can design the application and routing so that PHI is only served from the primary region under normal operations, satisfying the requirement that PHI remain within a specific region for serving while still enabling DR.

Why not A: A single-region deployment with backups and point-in-time recovery does not meet the RTO of 1 hour in the event of a regional outage. Restoring backups to another region is manual, slow, and operationally complex, contradicting the desire to avoid complex DR runbooks.

Why not C: Multi-region Cloud Storage for PHI backups may violate the requirement that PHI remain within a specific region, depending on regulatory interpretation. Cloud CDN is not appropriate for serving dynamic, personalized PHI and cannot replace the database during outages. This design does not provide a robust DR strategy for the database.

Why not D: Having Cloud Run in two regions improves availability for the stateless tier, but a single-region Cloud SQL instance with same-region replicas is still a single point of failure at the regional level. This design cannot meet the RTO/RPO requirements in the event of a regional outage and does not provide a DR database in another region.

Domain alignment: The question emphasizes analyzing compliance and availability requirements and choosing a design that optimizes reliability and operational simplicity while respecting regulatory constraints.

Ready to Accelerate Your GCP-PCA Preparation?

Join thousands of professionals who are advancing their careers through expert certification preparation with FlashGenius.

  • ✅ Unlimited practice questions across all GCP-PCA domains
  • ✅ Full-length exam simulations with real-time scoring
  • ✅ AI-powered performance tracking and weak area identification
  • ✅ Personalized study plans with adaptive learning
  • ✅ Mobile-friendly platform for studying anywhere, anytime
  • ✅ Expert explanations and study resources
Start Free Practice Now

Already have an account? Sign in here

About GCP-PCA Certification

The GCP-PCA certification validates your expertise in analyzing and optimizing technical and business processes and other critical domains. Our comprehensive practice questions are carefully crafted to mirror the actual exam experience and help you identify knowledge gaps before test day.

Continue Your Google Professional Cloud Architect (PCA) Prep on FlashGenius

Practice the highest-impact PCA domains with scenario-based questions, detailed explanations, and domain-wise improvement tracking.

PCA Domain Practice
Analyzing & Optimizing Technical and Business Processes
Cost optimization, process design, SDLC/CI-CD, governance, and stakeholder-driven trade-offs — PCA-style scenarios.
Start Domain Practice →
PCA Domain Practice
Managing & Provisioning a Cloud Solution Infrastructure
Compute, networking, storage, IAM, and scaling decisions — designed to mirror real PCA architecture trade-offs.
Practice Infrastructure Questions →
PCA Domain Practice
Designing & Planning a Cloud Solution Architecture
End-to-end architecture scenarios: requirements, trade-offs, migrations, HA/DR planning, and future-proofing.
Practice Architecture Questions →

Want a structured prep flow? Use Domain Practice first, then switch to Mixed Practice and Exam Simulation for full PCA readiness.

Explore FlashGenius PCA Prep →
PCA MASTER GUIDE

Google Professional Cloud Architect (PCA) — Ultimate 2026 Guide & 8-Week Study Plan

Go beyond practice questions. Learn how the exam is structured, how domains are weighted, and how to prepare efficiently using a proven, exam-aligned 8-week roadmap.

  • Official PCA exam domains & weightage explained
  • Realistic case-study strategy (Altostrat, Cymbal, EHR, KnightMotives)
  • Well-Architected Framework mapped to exam decisions
  • Common mistakes that cause PCA failures
  • 8-week, domain-by-domain study plan for busy professionals