Why does duplicate (redundant) data frequently lead to a loss of data integrity in organizational databases?

Difficulty: Easy

Correct Answer: Both A & B.

Explanation:


Introduction / Context:
Duplicate data is a primary source of integrity problems. When multiple copies of the same fact exist, keeping them synchronized is difficult, which can undermine reports, analytics, and decision-making.



Given Data / Assumptions:

  • Different systems or tables may hold overlapping data elements.
  • Updates may happen in one place but not others.
  • Formats and validation rules may differ across copies.


Concept / Approach:
Redundancy increases the risk of anomalies: inconsistent formats (e.g., date vs. text), and conflicting values (e.g., mismatched customer addresses). Master data management and normalization reduce duplication and enforce a single source of truth.



Step-by-Step Solution:

Assess reasons integrity degrades: formatting inconsistencies and conflicting values.Acknowledge both causes frequently co-occur in redundant datasets.Select the option that includes both reasons.


Verification / Alternative check:
Data quality frameworks cite “consistency” and “accuracy” as top dimensions harmed by duplication.



Why Other Options Are Wrong:
Choosing only one reason is incomplete; both are typical and damaging. “Neither” contradicts widespread data quality evidence.



Common Pitfalls:
Assuming deduplication alone fixes integrity; harmonizing formats, validation rules, and governance is also required.



Final Answer:
Both A & B.

Discussion & Comments

No comments yet. Be the first to comment!
Join Discussion