FPD blogs

Data Quality Audits: How to Spot Weak Data and Fix It Before Reporting

Written by James Archibald | Mar 27, 2026 7:16:58 AM

The integrity of developmental outcomes and the efficacy of public policy in Africa rely on the quality of the evidence used to inform them. Within the field of Monitoring and Evaluation (M&E), a data quality audit serves as the primary mechanism for ensuring that reported results are accurate reflections of reality. As South Africa transitions toward complex governance models, such as the implementation of National Health Insurance (NHI), the demand for robust data quality audit M&E processes has never been greater.

Through its Advanced Certificate in Monitoring and Evaluation, the Foundation for Professional Development (FPD) equips professionals with the capacity to design M&E frameworks that withstand the scrutiny of both internal and external audits. This article examines the methodology for identifying weak data, the standards for conducting audits and the professional habits required to maintain audit-ready documentation.

The Multi-Dimensional Construct of Data Quality

Data quality is a sophisticated construct composed of several interdependent dimensions. Verification involves assessing accuracy (description of the real-world phenomenon), completeness (presence of all required data points), reliability (consistency over time), timeliness (availability for decision-making), precision (granularity), integrity (freedom from bias), and validity (appropriateness of the indicator).

To ensure indicators support high-quality data, the OECD advocates for the "CREAM" criteria, under which indicators should be:

  • Clear: Precise and unambiguous

  • Relevant: Appropriate to the subject at hand

  • Economic: Available at a reasonable cost

  • Adequate: Provide a sufficient basis to assess performance

  • Monitorable: Amenable to independent validation

 

FPD’s curriculum addresses these concepts by training students in indicator selection and definition to ensure frameworks can be practically monitored.

Identifying Symptoms of Weak Data

Spotting weak data requires a detective-like approach to diagnostics. In African M&E systems, unique pressures, such as the "digital divide" and political interference, create distinct patterns of data weakness.

Red Flags and Diagnostic Signals

When looking at data, be sure to look out for any of the following warning signs:

  • Improbable Regularity: Month-over-month identical numbers suggest data has been "estimated" or manufactured rather than recorded.

  • Missing Outliers: A lack of variance in national databases often indicates that outlier analysis is not being performed or that data is being "smoothed" to avoid questioning.

  • Logical Inconsistency: Triangulation should show links between related indicators. If "training attendance" rises but "service delivery" remains flat, a disconnect exists.

  • Data Fatigue: Communities surveyed repeatedly without seeing results may view M&E as a "donor box-ticking exercise," leading to a decline in meticulousness.

  • Digital-Paper Discrepancies: In South Africa, common errors occur during manual transcription from primary paper registers to digital systems such as DHIS2.

 

Methodology of the Data Quality Audit (DQA)

A DQA focuses on the veracity of the data management system itself. International agencies utilise two primary tools: the Data Quality Audit (DQA) for external accountability and the Routine Data Quality Assessment (RDQA) for internal capacity building.

Functional Components and the Audit Process

A system must establish five functional components to produce quality information: dedicated M&E structures and capabilities; standardised indicator definitions; consistent collection tools; clear data management processes; and integration with national systems.

A professional audit follows five phases:

  • Planning (Phase 1): Defining the scope and indicators to be audited.

  • System Assessment (Phase 2): Reviewing M&E plans, reporting forms and standard operating procedures (SOPs).

  • Trace-Back Exercise (Phase 3): Comparing reported numbers against primary source documents, verifying aggregation, and cross-checking against secondary sources.

  • Sampling (Phase 4): Utilising statistically appropriate sizes, such as the square root plus one of the population, to draw conclusions.

  • Remediation (Phase 5): Developing a Data Quality Improvement Plan (DQIP) focused on root-cause resolution, such as redesigning user interfaces or providing computer literacy training.

 

The South African Regulatory Context: DPME Standards

M&E professionals in South Africa operate under the Public Finance Management Act (PFMA) and the Department of Planning, Monitoring and Evaluation's standards. The DPME Guideline provides a framework for assessing evaluation quality. Evaluations receive ratings from 1.00 (Very Poor) to 5.00 (Excellent), with 3.00 as the minimum threshold for credibility in policy benchmarking.

A case study of the NHI rollout highlights that Universal Health Coverage (UHC) requires a unified national M&E framework. Currently, fragmentation between disease-specific systems makes it difficult for policymakers to track disparities in healthcare access.

Documentation Habits for Audit Readiness

A credible M&E report must be supported by an audit trail. This is a chronological record of every action affecting a dataset.

  • Contemporaneous Recording: Entering data immediately after the event reduces "accuracy debt" caused by memory-based entry.

  • Immutability and Logs: Digital systems must maintain change logs (who, what, when and why). Raw data should never be deleted; versioning is preferred.

  • The "Four-Eyes" Check: Implementing internal peer review catches errors before final reporting.

  • Metadata: Documenting the context, such as indicator definitions and data collection limitations, ensures findings are interpreted correctly.

 

Transitioning to Digital: Opportunities and Risks

Digital tools like ODK and KoBoToolbox provide real-time logic checks and GPS tagging to verify the geographic origin of data. However, poor user interface (UI) design can lead to "system bypass," where staff revert to unofficial paper notes, and "black box" logic can cause managers to trust dashboards that are incorrectly calculated. FPD’s certificate prepares professionals to navigate these technologies to optimise project outcomes.

Conclusion

The path to credible reporting begins with a commitment to data integrity. In an era of big data and AI integration, the foundational skills of verification, documentation and ethical inquiry remain essential. The fully online Advanced Certificate in Monitoring and Evaluation from FPD provides the academic foundation required to lead evidence-based reforms in public health, government, and NGOs.

FAQs

1. What is the difference between a DQA and an RDQA?

A Data Quality Audit (DQA) is usually an external verification conducted by donors to ensure accountability for reported results, which can affect funding. A Routine Data Quality Assessment (RDQA) is a more flexible, internal tool used by managers for ongoing supervision and to build staff capacity for data management.

2. Why is accuracy prioritised over other data dimensions during an audit?

While all dimensions are vital, accuracy is the bedrock of credibility. If data inaccurately represent reality, such as over-reporting programme graduates, any subsequent analysis of impact or value-for-money will be fundamentally flawed and misleading for policy decisions.

3. How should an M&E officer handle weak data that has already been collected?

Weak data requires a transparent process of cleaning and reconciliation. This involves identifying outliers through statistical analysis, conducting follow-up interviews to clarify discrepancies, and triangulating multiple sources to find the most likely true value. All such modifications must be documented in a change log to maintain the audit trail.

4. What red flags typically trigger an external data quality audit?

Auditors look for downtime indicators, such as sudden, improbable performance improvements, statistical outliers, or reports showing zero variance (identical numbers month over month), which may suggest the data has been fabricated to meet targets.

5. Does the FPD Advanced Certificate cover digital M&E tools?

Yes. The programme focuses on "using M&E tools to optimise project outcomes". It trains professionals to design frameworks applicable to modern digital contexts, including those integrated with national databases such as South Africa’s DHIS2.