Difficulty: Easy
Correct Answer: Correct
Explanation:
Introduction / Context:
Organizations often apply data scrubbing during ETL to standardize codes, remove duplicates, fix formats, and repair obvious errors. This question tests whether you understand the distinction between tactical cleansing and strategic data quality management at the source.
Given Data / Assumptions:
Concept / Approach:
Data scrubbing detects and corrects issues already present (for example, mapping “NY”, “N.Y.” to “NY”). While this raises quality for downstream analytics, it does not change the upstream processes that keep generating errors. Long-term solutions involve governing master/reference data, improving application validations, standardizing codes, and instituting stewardship and monitoring. Therefore, scrubbing is necessary but insufficient on its own.
Step-by-Step Solution:
Verification / Alternative check:
Track defect recurrence after ETL cleansing. If the same anomalies reappear each load, cleansing alone has not solved the problem—source-level fixes are required.
Why Other Options Are Wrong:
Common Pitfalls:
Relying on complex ETL mappings as a permanent crutch, which becomes costly and brittle. Always pair cleansing with upstream remediation and governance.
Final Answer:
Correct
Discussion & Comments