InterviewStack.io LogoInterviewStack.io

Data Quality and Edge Case Handling Questions

Practical skills and best practices for recognizing, preventing, and resolving real world data quality problems and edge cases in queries, analyses, and production data pipelines. Core areas include handling missing and null values, empty and single row result sets, duplicate records and deduplication strategies, outliers and distributional assumptions, data type mismatches and inconsistent formatting, canonicalization and normalization of identifiers and addresses, time zone and daylight saving time handling, null propagation in joins, and guarding against division by zero and other runtime anomalies. It also covers merging partial or inconsistent records from multiple sources, attribution and aggregation edge cases, group by and window function corner cases, performance and correctness trade offs at scale, designing robust queries and pipeline validations, implementing sanity checks and test datasets, and documenting data limitations and assumptions. At senior levels this expands to proactively designing automated data quality checks, monitoring and alerting for anomalies, defining remediation workflows, communicating trade offs to stakeholders, and balancing engineering effort against business risk.

HardSystem Design
0 practiced
Architect an automated data quality monitoring system to run checks (schema, row counts, null rates, anomalous distributions) for 10TB of daily data across 500 datasets. Specify components for check execution, storage of results, alerting, lineage tracking, and a remediation workflow that allows reopening and reprocessing of affected datasets.
HardTechnical
0 practiced
Write an advanced SQL query that computes a 7-day rolling conversion rate for users, partitioned by user timezone, while avoiding division-by-zero and ensuring deterministic results for users with intermittent activity. Assume tables exposures(user_id, ts, timezone) and conversions(user_id, ts, timezone). Use standard SQL and explain your timezone bucketing strategy.
HardTechnical
0 practiced
You're responsible for defining SLAs and error budgets for dataset freshness and accuracy across the analytics platform. Propose measurable SLOs for freshness, completeness, and accuracy, describe how to compute an error budget, and explain how teams should act when error budgets are exhausted.
MediumTechnical
0 practiced
Design a medium-complexity Spark Structured Streaming pattern to deduplicate events in real time using event_id while tolerating out-of-order arrivals and late events up to 24 hours. Describe state retention settings, watermark usage, and how to evict state for long-running workloads.
EasyTechnical
0 practiced
Explain null propagation in SQL joins. Given two tables customers(customer_id, email) and orders(order_id, customer_id, amount), describe how inner, left, right, and full joins affect null values and why nulls can silently propagate into aggregations. Provide a short SQL pattern to safely compute total spend per customer including those with zero spend.

Unlock Full Question Bank

Get access to hundreds of Data Quality and Edge Case Handling interview questions and detailed answers.

Sign in to Continue

Join thousands of developers preparing for their dream job.