Try 10 focused SnowPro Data Engineer DEA-C02 questions on Sharing and Replication, with explanations, then continue with IT Mastery.
Open the matching IT Mastery practice page for timed mocks, topic drills, progress tracking, explanations, and full practice.
Try SnowPro Data Engineer DEA-C02 on Web View full SnowPro Data Engineer DEA-C02 practice page
| Field | Detail |
|---|---|
| Exam route | SnowPro Data Engineer DEA-C02 |
| Topic area | Sharing, Replication, and Cross-Platform Delivery |
| Blueprint weight | 18% |
| Page purpose | Focused sample questions before returning to mixed practice |
Use this page to isolate Sharing, Replication, and Cross-Platform Delivery for SnowPro Data Engineer DEA-C02. Work through the 10 questions first, then review the explanations and return to mixed practice in IT Mastery.
| Pass | What to do | What to record |
|---|---|---|
| First attempt | Answer without checking the explanation first. | The fact, rule, calculation, or judgment point that controlled your answer. |
| Review | Read the explanation even when you were correct. | Why the best answer is stronger than the closest distractor. |
| Repair | Repeat only missed or uncertain items after a short break. | The pattern behind misses, not the answer letter. |
| Transfer | Return to mixed practice once the topic feels stable. | Whether the same skill holds up when the topic is no longer obvious. |
Blueprint context: 18% of the practice outline. A focused topic score can overstate readiness if you recognize the pattern too quickly, so use it as repair work before timed mixed sets.
These questions are original IT Mastery practice items aligned to this topic area. They are designed for self-assessment and are not official exam questions.
Topic: Sharing, Replication, and Cross-Platform Delivery
A data engineering team is deciding how to expose curated tables to another Snowflake account.
Consumer account: ANALYTICS_DEV
Cloud/region: same as provider account
Access needed: read-only on curated tables
Provider updates: must be visible without copying data
Independent writes in consumer: no
DR/failover objective: none
Based on the exhibit, what is the best next step?
Options:
A. Publish a listing with auto-fulfillment.
B. Replicate the database to the consumer account.
C. Create a reader account for the consumer team.
D. Create a direct share for the curated tables.
Best answer: D
Explanation: This is a collaboration requirement, not a replication or failover requirement. Because the consumer already has a Snowflake account and only needs read-only access to current provider-managed data in the same region, a direct share is the best fit.
Secure Data Sharing is the Snowflake pattern for collaboration when consumers need read-only access to live provider data. The exhibit explicitly rules out the usual reasons to copy data: there is no disaster recovery goal, no locality requirement, and no need for independent writes in the consumer account. Because the consumer account already exists and is in the same cloud and region, the provider can create a direct share, grant the curated objects to that share, and let the consumer query the latest data without storage duplication.
A listing can also distribute shared data, but for one known same-region consumer it is less direct than a standard share.
Topic: Sharing, Replication, and Cross-Platform Delivery
A team exposes customer profiles through an internal REST service that uses the Snowflake SQL API. After a masking policy was added to email and phone, the service still authenticates successfully, queries finish normally, and fresh rows appear in the API response. However, those two columns are always masked, while analysts using a different authorized role see cleartext values. What is the best next step?
Options:
A. Rework the SQL API OAuth integration
B. Restart the refresh task graph
C. Resize the virtual warehouse
D. Review the service role and masking policy conditions
Best answer: D
Explanation: This is a data access issue. Authentication works, queries complete, and fresh rows are returned, so the main problem is that masking policy evaluation for the service role is producing masked values.
The key distinction is where the symptom appears. Control plane integration problems usually prevent authentication, session creation, or API request execution. Pipeline execution problems usually show up as stale, missing, or incomplete data. Here, the service authenticates, the SQL runs successfully, and fresh rows are returned, but protected columns differ by role.
That pattern points to data access logic, specifically how the masking policy evaluates for the role used by the REST service. The next step is to verify the service role, any database-role or view-based access path it uses, and the masking policy condition that determines when cleartext is allowed.
Restarting tasks or resizing compute would not explain successful queries that return current rows with only specific columns masked.
Topic: Sharing, Replication, and Cross-Platform Delivery
A team loads and transforms order data with streams, tasks, and MERGE statements. An external web service retrieves the final ORDER_STATUS table through the Snowflake SQL API. During hourly backfills, the web service times out, but task runs finish successfully and ORDER_STATUS is current. Query history shows the web service SELECT statements queued on the same warehouse used by the backfill tasks. What is the best next step?
Options:
A. Replace SQL API calls with Snowpipe Streaming
B. Rewrite the backfill logic in Snowpark Python
C. Add a stream on ORDER_STATUS for the web service
D. Route SQL API queries to a dedicated warehouse
Best answer: D
Explanation: The pipeline is producing current data, so the transformation layer is working. The problem is warehouse contention for the application-facing SQL API queries, which are part of delivery design, so compute isolation is the right fix.
The deciding clue is that tasks complete successfully and the target table is already current. That means the main transformation logic is not failing. The timeout happens when the external service submits SELECT statements through the SQL API, and those statements are queued behind heavy backfill work on the same warehouse.
Rewriting the transforms or changing ingestion features would not address the actual bottleneck, which is queued delivery workload on shared compute.
SELECT statements from a consuming application.Topic: Sharing, Replication, and Cross-Platform Delivery
A team uses database replication to maintain a DR environment in another region. During a failover test, the replicated database is available and the table data is within the expected RPO, but the ingestion service cannot restart because the service user, role grants, and warehouse are missing in the secondary account. Which next step best addresses this problem?
Options:
A. Use a failover group for databases and required account objects.
B. Upsize the DR warehouse for recovery tasks.
C. Increase the database replication refresh frequency.
D. Rebuild ingestion with Snowpipe Streaming.
Best answer: A
Explanation: This symptom points to continuity, not freshness. The replicated data is already within the target RPO, so tuning ingestion or compute will not restore missing users, roles, grants, or warehouses in the secondary account.
Database replication helps keep replicated data objects available in another region, but it does not by itself solve every recovery dependency. Here, the data copy is already current enough, yet failover still cannot proceed because required account-level objects are absent. That makes this a continuity design issue.
A failover group is the appropriate Snowflake mechanism when recovery must include databases plus dependent account objects needed to run the workload after failover. By contrast, pipeline freshness tools such as faster ingestion, lower-latency CDC, or more warehouse compute address how quickly new data arrives or is transformed. They do not fix missing identities, grants, or warehouses in the DR account.
The key takeaway is to separate recovery readiness from data-latency tuning: this scenario needs broader replication scope, not a faster pipeline.
Topic: Sharing, Replication, and Cross-Platform Delivery
A data engineering team is exposing curated Snowflake data to an internal serverless REST service. The service must submit SQL statements and call a stored procedure over HTTPS, poll long-running requests asynchronously, and avoid installing JDBC or ODBC drivers on ephemeral compute. Which Snowflake design is the best fit?
Options:
A. Create an external function that routes requests through API Gateway
B. Publish the tables with Secure Data Sharing
C. Use the Snowflake SQL API from the service
D. Replicate the database to a separate account for the service
Best answer: C
Explanation: The Snowflake SQL API is designed for HTTPS-based programmatic access to Snowflake workloads. It fits stateless services well because it supports submitting SQL, checking status asynchronously, and keeping access control in Snowflake without adding driver management.
The key concept is choosing the Snowflake-native integration path that matches a REST-driven application pattern. For a serverless service that must call Snowflake over HTTPS and handle long-running work asynchronously, the Snowflake SQL API is the best fit. It lets the application submit SQL statements, including CALL statements for stored procedures, receive a handle for execution, and poll for completion without requiring JDBC or ODBC drivers on short-lived compute.
This matches the stem because it provides:
The closest distractors either solve a different problem or add unnecessary architecture instead of meeting the REST integration requirement directly.
Topic: Sharing, Replication, and Cross-Platform Delivery
A provider is considering secure sharing for consumers in another region. However, the secondary setup must keep replicated databases and supporting account objects so it can be promoted and continue operating if the primary account is unavailable. Which Snowflake feature meets this requirement?
Options:
A. Database replication
B. Secure data sharing
C. Reader accounts
D. Failover groups
Best answer: D
Explanation: Failover groups are for business continuity, not just consumer access. They replicate selected databases and supported account objects to a secondary account and allow promotion during an outage, which a sharing-based approach does not provide.
Secure sharing is a data-access mechanism: consumers can query provider-owned data, but they do not receive an independently replicated environment. When the requirement is a secondary account that can continue operating if the primary is unavailable, Snowflake needs replication with failover capability. Failover groups fit that need because they replicate selected databases plus supported account objects to a secondary account and support promotion of that secondary environment.
That is why a consumer-facing sharing answer is insufficient here.
Topic: Sharing, Replication, and Cross-Platform Delivery
A central Snowflake team distributes customer data to Finance and Sales in separate consumer accounts. To prepare for later cross-region delivery, an architect proposes distributing only raw tables and letting each team maintain its own secure views and task logic locally; analysts already report different active_customer counts, and each schema change triggers fixes in multiple accounts. Which next step best addresses the problem while preserving maintainability and cross-team clarity?
Options:
A. Centralize KPI logic in a producer-managed curated database and deliver that same layer to all teams.
B. Keep distributing raw tables and use task graphs in each consumer account to recreate local views after changes.
C. Replicate the raw database to more regions first, then let each team standardize its own views.
D. Create separate shared databases per team so each group can rename and organize objects independently.
Best answer: A
Explanation: The failure is semantic drift, not data movement. When each consumer account owns its own view logic, KPI definitions diverge and every upstream change must be implemented multiple times; a producer-managed curated layer keeps one authoritative definition for all teams.
When the stated need is maintainability and cross-team clarity, distributing raw tables and letting every team build its own secure views is the wrong architecture. Even if shared or replicated data is current, local modeling creates duplicate business logic, conflicting KPI definitions, and repeated remediation after every schema change. A better Snowflake pattern is to publish a producer-managed curated database, often with standardized secure views, and deliver that same governed layer to each consumer. Then business-rule updates and schema adjustments happen once, and all teams see the same meaning for active_customer. Replication can still help later with regional delivery, but it should move the curated layer rather than multiply independent semantic definitions.
Topic: Sharing, Replication, and Cross-Platform Delivery
A Snowflake team ingests partner data from three operational systems, replicates curated tables to another region, and shares them with external consumers. Pipelines finish on time, but source corrections overwrite prior values in the published tables, so consumers cannot reconstruct multi-year history of keys, relationships, and attributes across systems. Engineers propose adding more streams and faster tasks. What is the best next step?
Options:
A. Increase stream consumption and task cadence on the current-state tables
B. Add search optimization service to the published tables
C. Replace the published tables with dynamic tables that refresh more often
D. Redesign the integration layer with a historized model such as Data Vault
Best answer: D
Explanation: This is no longer mainly an ingestion-timing problem. The requirement is durable, multi-source, multi-year history, so the right move is to revisit the integration model with a historized pattern such as Data Vault before tuning streams, tasks, or refresh frequency.
When published tables overwrite prior values, increasing pipeline frequency only moves the same current-state design faster. The real issue is architectural: external consumers need long-term traceability of business keys, relationships, and descriptive changes across multiple source systems. A historized approach such as Data Vault is designed for that by separating keys, relationships, and changing attributes so history survives source corrections, late-arriving data, and cross-system integration.
In Snowflake, streams, tasks, and dynamic tables help implement ingestion and transformation, but they do not replace the need for the right core model. A strong pattern is to preserve integrated history in the core layer, then publish consumer-friendly marts or shared tables from that foundation. The closest distractor is faster CDC orchestration, but the broken part here is the model, not the load cadence.
Topic: Sharing, Replication, and Cross-Platform Delivery
A data engineering team lands source data as Parquet in cloud object storage. Snowflake must transform the data into curated datasets, and Spark and Trino users must query those same curated tables without a separate export pipeline. Which Snowflake-oriented architecture best fits?
Options:
A. Use external tables on the landing files
B. Use native tables with secure data sharing
C. Use native tables with dynamic tables
D. Use Iceberg tables for the curated layer
Best answer: D
Explanation: Iceberg tables are the best fit when curated data must be usable both in Snowflake and in other query engines. They balance storage, transformation, and delivery by keeping the curated layer in an open table format instead of creating a separate export path.
The key concept is choosing an architecture that supports cross-platform delivery without duplicating the curated dataset. Iceberg tables are designed for this pattern: data stays in object storage in an open table format, Snowflake can transform and manage the curated layer, and engines such as Spark or Trino can read the same tables.
In this scenario, the deciding factor is not just transformation automation but the delivery model. External tables expose files in place, but they are not a Snowflake-managed curated table architecture. Native tables work well for Snowflake-only consumption, and dynamic tables help maintain derived data, but both keep the curated layer inside Snowflake. The best trade-off here is an open table architecture that serves both Snowflake and non-Snowflake consumers.
Topic: Sharing, Replication, and Cross-Platform Delivery
A company uses secure data sharing from its primary Snowflake account to a reporting account. During a regional outage drill, the primary account became unavailable and reports against the shared database immediately failed. The business now requires a local secondary copy in another region and the ability to promote that secondary during failover. What is the best next step?
Options:
A. Replace the consumer account with a reader account
B. Publish the database through a listing
C. Keep the secure share and resize the consumer warehouse
D. Replicate to a secondary account and use a failover group
Best answer: D
Explanation: This requirement is about resilience and promoted failover, not just governed access. Secure data sharing provides live access without creating an independent copy, so it does not satisfy a need for a local secondary that can be promoted during an outage.
Secure data sharing and replication solve different problems. Secure sharing is best when consumers need governed, near-live access to provider data without managing extra copies. But it does not create an independently recoverable target for regional resilience or failover promotion. When the requirement is locality, disaster recovery, or promoted failover behavior, the right pattern is replication to a secondary account or region and management through a failover group.
Changing compute or publishing method may help access patterns, but neither creates a promotable secondary copy.
Use the SnowPro Data Engineer DEA-C02 Practice Test page for the full IT Mastery route, mixed-topic practice, timed mock exams, explanations, and web/mobile app access.
Try SnowPro Data Engineer DEA-C02 on Web View SnowPro Data Engineer DEA-C02 Practice Test
Read the SnowPro Data Engineer DEA-C02 Cheat Sheet on Tech Exam Lexicon, then return to IT Mastery for timed practice.