SlideShare une entreprise Scribd logo
1  sur  49
MEASURE Evaluation
Data Quality Assessment
 Methodology and Tools


          USAID
      Washington D.C.
         July 19, 2012
Background

 ■ National Programs and Donor-funded projects are working towards
   achieving ambitious goals in the fight against HIV, TB and Malaria.

 ■ Measuring success and improving management of these initiatives is
   predicated on strong M&E systems that produce quality data regarding
   program implementation.

 ■ In the spirit of the “Three Ones”, the “Stop TB Strategy” and the “RBM
   Global Strategic Plan”, a multi-partner project* was launched 2006 to
   develop a joint Data Quality Audit (DQA) Tool.

 ■ The objective of this initiative is to provide a common approach for
   assessing and improving data quality. A single tool ensures that
   standards are harmonized and allows for joint implementation (between
   partners and with National Programs).




* Partners most directly involved include PEPFAR, USAID, WHO, Stop TB, the Global Fund and MEASURE
Outline
 Auditing
    DQA
        Evolution
        Methods and tools
        How and where applied
 Routine Data Quality Assurance and Capacity Building
    RDQA
        Purpose
        Methods and tools
        How and where applied
        Integration into Routine Practice
Objective of the Data Quality Audit (DQA) Tool




The Data Quality Audit (DQA) Tool is designed to:


 verify the quality of reported data for key indicators at
    selected sites; and


 assess the ability of data-management systems to collect
    manage and report quality data.
Conceptual Framework


                   Generally, the quality of reported data is dependent on the underlying data
                   management and reporting systems; stronger systems should produce
                   better quality data.
                                                                                                           Dimensions of Quality
                                                            QUALITY DATA                       Accuracy, Completeness, Reliability, Timeliness,
                                                                                                     Confidentiality, Precision, Integrity




                                                                                             Functional Components of a Data Management System
                                                                Data-Management and


                                    M&E Unit                                                            Needed to Ensure Data Quality
                                                                  Reporting System
REPORTING LEVELS




                                                                                      I         M&E Structure, Functions and Capabilities

                                                                                      II        Indicator Definitions and Reporting Guidelines
                           Intermediate Aggregation                                   III       Data-collection and Reporting Forms / Tools
                         Levels (e.g. Districts, Regions)
                                                                                      IV        Data Management Processes

                                                                                      V         Links with National Reporting System

                                                                                      VI        Data management processes
                                 Service Points
                                                                                      VII
                                                                                                              5
                                                                                                Data quality mechanisms and controls

                                                                                      VIII      Links with the national reporting system
DQA METHODOLOGY – 2 Protocols




■ The methodology for the DQA includes two (2) protocols:


1
    Assessment of Data
       Management         Qualitative assessment of the strengths and
         Systems          weaknesses of the data-collection and reporting
                          system.
         (Protocol 1)

2
                          Quantitative comparison of recounted to reported
     Data Verifications
                          data and review of timeliness, completeness and
         (Protocol 2)
                          availability of reports.
Typical Timeline


   PHASE 1             PHASE 2               PHASE 3                PHASE 4                  PHASE 5             PHASE 6

                                                                  Intermediate
Preparation and
                        M&E              Service Delivery         Aggregation                 M&E               Completion
   Initiation
                     Management               Sites /                levels                Management             (multiple
    (multiple           Unit              Organizations                                       Unit
                                                                   (eg. District,                                locations)
   locations)
                                                                     Region)

1. Select                                                                                 6. Draft initial
                                                                                                              7. Draft and
   Indicators and                                                                            findings and
                     3. Assess Data Management and Reporting Systems (Part 1)                                    discuss Audit
   Reporting                                                                                 conduct close-
                                                                                                                 Report
   Period                                                                                    out meeting

2. Notify           4. Select/Confirm
                                                                                                              8. Initiate
   Program and         Service
                                                                                                                 follow-up of
   desk review of      Delivery                    5. Trace and Verify Reported Results (Part 2)                 recommended
   Program             Points to be
                                                                                                                 actions
   documentation       visited




  ■      The DQA is implemented chronologically in 6 Phases.
  ■      Assessments and verifications will take place at every stage of the reporting system:
          - M&E Management Unit
          - Intermediate Aggregation Level (Districts, Regions)
          - Service Delivery Sites.
                                                                                             7
PROTOCOL 1:
Assessment of Data Management
          Systems




                      8
Intermediate
                                          M&E             Service Delivery
     PROTOCOL 1:                                                                 Aggregation
                                       Management              Sites /
                                                                                    levels
                                          Unit             Organizations
  Assessment of Data                                                           (eg. District, Region)

 Management Systems
                                      3. Assess Data Management and Reporting Systems (Part 1)




■ PURPOSE:          Identify potential risks to data quality created by the data-
                    management and reporting systems at:
                -     the M&E Management Unit;
                -     the Service Delivery Points;
                -     any Intermediary Aggregation Level (District or Region).

■ The DQA assesses both (1) the design; and (2) the implementation of the data-
  management and reporting systems.

■ The assessment covers 5 functional areas (HR, Training, Data Management
  Processes , etc.)
                                                                       9
SYSTEMS ASSESSMENT QUESTIONS BY FUNCTIONAL AREA

      Functional Areas                          Summary Questions

I       M&E Structure,        Are key M&E and data-management staff identified with
        Functions and     1
                              clearly assigned responsibilities?
        Capabilities
                              Have the majority of key M&E and data-management staff
                          2
                              received the required training?
II      Indicator             Has the Program/Project clearly documented (in writing)
        Definitions and   3   what is reported to who, and how and when reporting is
        Reporting             required?
        Guidelines
                              Are there operational indicator definitions meeting relevant
                          4   standards? And are they consistently followed by all
                              service points?
III     Data-collection       Are there standard data-collection and reporting forms that
        and Reporting     5
                              are systematically used?
        Forms and Tools
                              Are data recorded with sufficient precision/detail to
                          6
                              measure relevant indicators?
                              Are data maintained in accordance with international or
                          7
                              national confidentiality guidelines?
                              Are source documents kept and made available in
                          8
                              accordance with a written policy?
SYSTEMS ASSESSMENT QUESTIONS BY FUNCTIONAL AREA
     Functional Areas                       Summary Questions

IV     Data
       Management            Does clear documentation of collection, aggregation
                        9
       Processes             and manipulation steps exist?


                             Are data quality challenges identified and are
                        10
                             mechanisms in place for addressing them?


                             Are there clearly defined and followed procedures to
                        11
                             identify and reconcile discrepancies in reports?


                             Are there clearly defined and followed procedures to
                        12
                             periodically verify source data?

V      Links with
       National              Does the data collection and reporting system of the
       Reporting        13
                             Program/project link to the National Reporting System?
       System
PROTOCOL 2:
Data Verifications




                 13
Intermediate
                                          M&E             Service Delivery
                                                                                 Aggregation
                                       Management              Sites /
        PROTOCOL 2:                       Unit             Organizations
                                                                                    levels
                                                                                (eg. District, Region)
   Data Verifications

                                           5. Trace and Verify Reported Results (Part 2)




■ PURPOSE: Assess on a limited scale if Service Delivery Points and
  Intermediate Aggregation Sites are collecting and reporting data
  accurately and on time.

■ The trace and verification exercise will take place in two stages:
    -     In-depth verifications at the Service Delivery Points; and
    -     Follow-up verifications at the Intermediate Aggregation Levels (Districts,
          Regions) and at the M&E Unit.

                                                                       14
DQA Data Verifications

                                                                                             M&E Unit/National
                                                                                                 Monthly Report
                                                                                                                                                                       ILLUSTRATION
                                                                                           District 1        65
                                                                                           District 2        45
                                                                                           District 3        75
                                                                                           District 4        250
                                                                                            TOTAL            435




                       District 1                                       District 2                                     District 3                            District 4
                  Monthly Report                                     Monthly Report                                 Monthly Report                         Monthly Report
               SDS 1                45                             SDS 3             45                           SDS 4             75                SDP 5               50
               SDS 2                20                             TOTAL             45                       TOTAL                 75                SDP 6               200
               TOTAL                65                                                                                                                TOTAL               250




  Service Delivery Site 1                Service Delivery Site 2           Service Delivery Site 3                  Service Delivery Site 4     Service Delivery Site 5           Service Delivery Site 6
      Monthly Report                        Monthly Report                    Monthly Report                              Monthly Report            Monthly Report                    Monthly Report
  ARV Nb.          45                 ARV Nb.             20            ARV Nb.             45                      ARV Nb.              75    ARV Nb.            50              ARV Nb.          200



  Source                              Source                            Source                                      Source                      Source                            Source
Document 1                          Document 1                        Document 1                                  Document 1                  Document 1                        Document 1
Service Delivery Points – Data Verification


                  SERVICE DELIVERY POINT - 5 TYPES OF DATA VERIFICATIONS

     Verifications                              Description                               -

Verification no. 1:      Describe the connection between the delivery of
                         services/commodities and the completion of the source       In all cases
Description              document that records that service delivery.

Verification no. 2:
                         Review availability and completeness of all indicator
Documentation                                                                        In all cases
                         source documents for the selected reporting period.
    Review

                         Trace and verify reported numbers: (1) Recount the
Verification n . 3:
              o
                         reported numbers from available source documents; (2)
                                                                                     In all cases
Trace and Verification   Compare the verified numbers to the site reported
                         number; (3) Identify reasons for any differences.

Verification no. 4:      Perform “cross-checks” of the verified report totals with
                         other data-sources (eg. inventory records, laboratory       If feasible
Cross-checks             reports, etc.).

Verification no. 5:      Perform “spot checks” to verify the actual delivery of
                                                                                     If feasible
Spot checks              services or commodities to the target populations.
Summary Statistics of Data Verifications


         Total and Adjusted District Verification Factors from                                                                               % On Time Reports from DQA
         Total                  DQA
         Verification
         Factor
         IAL #4                                                                                                     M&E Unit
                                                                                                                                                                                         0.75
                                                                                                                    IAL #4
                                                                                       1,07                                                                                             0.73
                                                                                                                    IAL #3
         IAL #3
                                                                                                1.22                IAL #2               1                                                            0.88

                          1                                                      1,03                               IAL #1
         IAL #2                                                                                                                                                    0.47

                                                                                  1,07
                                                                                                                                                                                               0.80
         IAL #1                                                          0,95



                               0, 00   0, 20          0, 40    0,60      0,80           1, 00          1, 20                             0.00        0.20     0.40               0.60             0.80       1.00




                              % Available Reports from DQA                                                                                   % Complete Reports from DQA




                                                                                                                                                                    0.50
                                                                      0,75

M&E Unit                                                                                                                                                                                   0.67
                                                                                0,83
IAL #4                                                                                                          M&E Unit
IAL #3                                                                                                          IAL #4     1                                                                      0.72
           1                                                                            0,92
IAL #2                                                                                                          IAL #3
IAL #1                                                                                                          IAL #2
                                               0,50                                                                                                                              0.60
                                                                                                                IAL #1

                                                                         0,80                                                                                                                   0.70




           0,00         0,20              0,40                0,60               0,80                    1,00              0.00   0.10       0.20    0.30   0.40          0.50          0.60          0.70   0.80


                                                                                                                                                              17
Outputs of the DQA Implementation



■ Completed protocols and templates related to:
    -   Protocol 1: System Assessment; and
    -   Protocol 2: Data Verification.

■ Write-ups of observations, interviews, and conversations with key staff.

■ Preliminary findings and recommendation notes.

■ Final DQA Report, including:
    -   Summary of the evidence collected by the DQA team;
    -   Specific findings or gaps related to that evidence;
    -   Recommendations to improve data-management systems and overall data quality.

The Final DQA Report should also include summary statistics on:
1. Strengths and weaknesses of data-collection and reporting systems;
2. Recounting exercise performed on primary records and aggregated reports;
3. Available, On time and Complete Reports.                     18
Use of the DQA since 2008
Use of DQA (continued)

 Biggest user is the Global Fund
 Part of the grant renewal process
 Random and purposive selection of grants
 External audit teams – DQA IQC
    1st IQC 2008-2010
    2nd IQC 2011-2012
 Standardized output
 GF Reorganization
Global Fund DQAs 2008-2012
                                               Disease
  Year    Qtr
                      HIV/AIDS                   TB                   Malaria
  2008     4         Mali, Belarus           Philippines
                                                                Rwanda, China, LAC-
                                                                   Multi-country
  2009     1       Comoros Islands          China, Yemen
                                                                 (Americas – Andean
                                                                      Region)
  2009     2        Haiti, Burundi              Niger             Ghana, Indonesia

                   Mozambique,
                                                                Gambia, Vietnam, Sri
  2009     3     Dominican Republic,
                                                               Lanka, Nigeria, Zambia+
                     Tanzania

                Peru, Swaziland, South
                                                                Kenya, Cote d’Ivoire,
  2009     4       Africa, Thailand,          Tajikistan
                                                                    Madagascar
                       Guyana

                  Ethiopia, Paraguay,
  2010     1        Malawi, C.A.R.
                                             Uzbekistan          Uganda, Cameroon

                                                                Papua New Guinea,
                   Russia, Guinea-
  2010     2       Conakry, Eritrea
                                          Pakistan, Paraguay   Pakistan, South Sudan,
                                                                    Madagascar
  2011     2           Zambia                Kyrgyzstan             North Sudan
                  Angola, Cambodia,
                                                               Senegal, South Sudan,
  2011     4    Namibia, Guyana, India,      North Korea
                                                                       DRC
                  Myanmar, Ukraine
GF Audits by Indicator (2008-2010)
          Disease                   Program Area          Number of Audits
Malaria             Malaria Treatment                           13
HIV                 Current on ART                              12
Malaria             LLIN                                        10
HIV                 C&T HIV                                      9
HIV                 Care and support                             6
TB                  TB Detection                                 5
HIV                 PMTCT – preg. women on prophylaxis           5
Malaria             Malaria Training                             4
TB                  TB Treatment                                 4
HIV                 HIV Education                                4
Malaria             IPT                                          3
HIV                 Condoms                                      3
Malaria             Malaria Treatment - Community Based          2
Malaria             Malaria Testing                              2
HIV                 OVC                                          2
HIV                 HIV Training                                 1
TB                  MDR TB Treatment                             1
HIV                 TB/HIV                                       1
Total                                                           87
Use of DQA (continued)

 PEPFAR Countries
    2009 Country Operational Plan Guidance
    Mozambique, Ukraine, DRC, Uganda, Nigeria,
     Cote d’Ivoire, Dominican Republic
 Ad hoc use by Partners, other agencies
    UNICEF in India
 Adaptation for local use
    South Africa (ESI)
New DQA Dashboards
DQA Performance Analysis
                   National Level:

                   Performance Table:                                         Performance Scores                            Weighted Final Audit
                                                                                                                            Performance Score
                                                                      100%
                                   Sample Accuracy             77%
                                                                                                             100%
                                                                      90%
Values provided automatically




                                   OSDV Accuracy               92%

                                                                      80%                                        90%
                                   Availability                74%

                                                                      70%
                                   Timeliness                  52%                                               80%

                                                                      60%
                                   Completeness                84%                                                                         76.5%
                                                                                                                 70%
                                                                      50%
                                   Cross-Checks                79%
                                                                                                                 60%
                                   Service Site                       40%
                                                               65%
                                   Documentation Review
                                   M&E Structure, Functions           30%                                        50%
                                                               2.40
                                   and Capabilities
Qualitative - Enter these values




                                   Indicator Definitions and          20%
                                                               2.50                                              40%
                                   Reporting Guidelines
                                   Data-collection and                10%
                                                               2.30
                                   Reporting Forms / Tools                                                       30%
                                   Data Management                     0%
                                                               1.90
                                   Processes
                                   Links with National
                                                               2.80                                     Excellent      = PS 91.0 - 100.0           Poor   = PS 61.0 - 70.0
                                   Reporting System

                                                                                                         Good          = PS 81.0 - 90.0       Very Poor   = PS < 60.0


                                                                                                          Fair         = PS 71.0 - 80.0
Evolution of RDQA tool

 Need expressed for a tool for
  programs/projects to use for self assessment,
  improved data quality, capacity building and
  to prepare for external audits
 New generic, simplified tool adapted to the
  country, program and/or project
 Routine use
 Quantify and track data quality performance
  over time
 Identify and respond to data quality problems
Use of the Auditing and Self Assessment/Routine Versions


        Auditing version…          Capacity Building version…

 Assessment by funding        Simplified from the auditing
  agency                        version

 Standardized approach to     Flexible use by program
  auditing
                               Self-assessment
 Conducted by an external
  audit team                   Focus is on identifying
                                weaknesses for system
 Relatively expensive          strengthening

 Limited opportunities for    Should be used routinely – data
  capacity building             quality performance indicators
                                monitored over time
RDQA Modifications

 No implied sampling methodology
 No aggregate sample statistics
 Quantitative and qualitative assessments merged
  into the same tool
 Dashboards by site and level to facilitate data use
 Easily extract data to link with data systems
 Action plans for system strengthening
Objectives of the RDQA


OBJECTIVES:

  •   VERIFY
       • the quality of reported data for key indicators at selected
         sites; and the ability of data-management systems to
         collect, manage and report quality data.

  •   IDENTIFY
       • weaknesses in the data management system and
       • interventions for system strengthening.

  •   MONITOR
       • data quality performance over time and
       • capacity to produce good quality data
Uses of the RDQA

The RDQA is designed to be flexible in use and can serve
  multiple purposes:
 Routine data quality checks as part of on-going
  supervision
 Initial and follow-up assessments of data management
  and reporting systems – measure performance
  improvement over time:
 Strengthening program staff’s capacity in data
  management and reporting:
 Preparation for a formal data quality audit·
 External assessment by partners, other stakeholders
RDQA Outputs

1.   Strength of the M&E System
     evaluation based on a review of the Program/project’s data management
     and reporting system, including responses to overall summary questions
     on how well the system is designed and implemented;


2.   Verification Factors
     generated from the trace and verify recounting exercise performed on
     primary records and/or aggregated reports (i.e. the ratio of the recounted
     value of the indicator to the reported value);


3.   Available, On time and Complete Reports
     percentages calculated at the Intermediate Aggregation Levels and the
     M&E Unit).


4.   Action Plan for System Strengthening for each site and level
     assessed.
RDQA Dashboards
RDQA Summary Statistics – Level Specific Dashboard


Part 4: DASHBOARD: National Level - M&E Unit


               Data Management Assessment -                                                    Data and Verifications -                                                   Reporting Performance -
                  National Level - M&E Unit                                                   National Level - M&E Unit                                                   National Level - M&E Unit

                        I - M&E Structure,                            120%
                           Functions and
                            Capabilities
                                                                                                    106%                                               % Complete                                     82%
                              3                                       100%
                                                                                                                     102%
                                                                                                                                       105%

                                                                                  94%
                              2
                                                     II- Indicator    80%
  V - Links with
                              1
                                                    Definitions and
     National
                                                      Reporting
Reporting System                                                      60%                                                                              % On Time                 41%
                                                      Guidelines
                              0

                                                                      40%



                                              III - Data-collection   20%
  IV- Data Management
                                             and Reporting Forms                                                                                       % Available                                           92%
        Processes
                                                     and Tools
                                                                       0%
                                                                             ANC Registered   ANC Counseled and   ANC Received   ANC Tested Positive
                                                                                                   Tested           Results                                          0%   20%      40%      60%        80%         100%
Global Dashboard
                                   Data Management Assessment -                                                             System Assessment Results by Level of the Reporting System
                                          Overall Average
                                                                                                      3.00
                                              M&E Structure,
                                              Functions and                                           2.50
                                               Capabilities

  RDQA                                               3
                                                                                                      2.00



 Summary
                                                     2
                                                                                                      1.50
                   Links with                                                 Indicator Definitions
               National Reporting                    1                           and Reporting

Statistics –        System                                                         Guidelines         1.00
                                                     0
                                                                                                      0.50

  Global                                                                                              0.00


Dashboard
                               Data                                    Data-collection                       National M&E Unit - Cotonou
                            Management                                 and Reporting
                             Processes                                  Forms / Tools
                                                                                                                  M&E Structure,         Indicator Definitions       Data-collection   Data              Links with
                                                                                                                  Functions and           and Reporting              and Reporting     Management         National Reporting
                                                                                                                  Capabilities           Guidelines                  Forms / Tools     Processes         System




                                  Data Verifications - Overall Average                                               Data Verifications - Verification Factors by Level of the Reporting System
                                              by Indicator
                140%                                                                                  1.4

                                                                                                      1.2
                120%                                                                124%

                                              112%                112%                                  1
                100%


                             90%
                                                                                                      0.8
                 80%

                                                                                                      0.6
                 60%
                                                                                                      0.4
                 40%
                                                                                                      0.2
                 20%
                                                                                                        0
                                                                                                               Service Site Average               District Average               Regional Average        National M&E Unit - Cotonou
                  0%
                       ANC Registered     ANC Counseled        ANC Received       ANC Tested
                                            and Tested           Results           Positive                        ANC Registered         ANC Counseled and Tested           ANC Received Results      ANC Tested Positive



                           Reporting Performance - Overall Average                                                                        Reporting Performance by Reporting Level
                                                                                                       1
                                                                                                      0.9
                                                                                                      0.8
                Completeness                                                        132%
                                                                                                      0.7

                                                                                                      0.6
                                                                                                      0.5
                   Timeliness                   51%                                                   0.4
                                                                                                      0.3
                                                                                                      0.2

                                                                                                      0.1
                   Availability                                  86%
                                                                                                       0
                                                                                                                      District Average                           Regional Average                   National M&E Unit - Cotonou

                                  0%    20%   40%        60%     80%      100%    120%     140%
                                                                                                                                              Availability             Timeliness           Completeness
RDQA
 Summary
Statistics –
    Site
  specific
qualitative
  results
RDQA Output
- System
Strengthening
Action Plan
Data Export
Tab –
Easily
Integrate
Results
from
different
workbooks
Growing family of RDQA Tools
                                                              Data Verifications - Service Site Summary
 RDQA traditional version:                                    Trend in Verification Ratio by Indicator

  one indicator per                              3.5


  workbook, x-sectional                           3

                                                 2.5




                               Health Facility
 RDQA Multi-indicator                            2


  version: up to four                            1.5



  indicators for the same                         1

                                                 0.5
  program area in x-section                       0
                                                                Q1                     Q2                 Q3                     Q4

 RDQA Longitudinal                              CS Parakou
                                                                                         Percent Accuracy
                                                                                                   Maternite de bonne saveur Parakou

  version: one indicator for                     CS Immacule Conception de Kalale                  Hopital St. Louis de Kalale


  up to four points in time.                     Hopital Regionale de Savalou                      CS de Zongo - Savalou


                                                 Centre Hopitalier de Grand Quelqu'un d'Abomey     CS Behanzin de Abomey


 RDQA training materials
 PMTCT and TB RDQA
Countries where RDQA has been used or is
 currently being implemented
 Nigeria              •   South Africa
 Kenya                •   Lesotho
 Tanzania             •   Swaziland
 Cote d’Ivoire        •   DRC
 Guyana               •   India
                       •   Botswana
 Haiti
                       •   Global Fund On Site
 Mozambique               Data Verification
 Zimbabwe                 (OSDV) by LFAs in
                           many countries
 Uganda
 Dominican Republic
 Peru
 Paraguay
Now we know the quality of the
data. Now what?
 Standardization:
    Indicators
    data collection and reporting tools
    data management and indicator compilation
     practices
 Capacity building to address identified needs
    MEval Data quality assurance curriculum
 Formalization of RDQA into routine practice
 Monitor data quality performance indicators over
  time. What is the trend?
Integrating Routine Data Quality Assurance into
Standard Operating Procedures

 Explicitly State:
    Who will do what?
    When they will do it?
    What is the output?
    Where will it happen?
    What is the required follow up?
    Who will verify that it is happening?
 Formalize in official Program documentation
 Incorporate into pre-service/in-service training
Integration of RDQA into SOPs
 Example from
  Botswana
Key Success Factors for Data Quality
1. Functioning information systems
2. Clear definition of indicators consistently used at all levels

3. Description of roles and responsibilities at all levels

4. Specific reporting timelines

5. Standard/compatible data-collection and reporting forms/tools with clear
    instructions
6. Documented data review procedures to be performed at all levels

7. Steps for addressing data quality challenges (missing data, double-
    counting, lost to follow up, …)

8. Storage policy and filling practices that allow retrieval of documents for
    auditing purposes (leaving an audit trail)
MEASURE Evaluation is funded by the U.S. Agency for
International Development (USAID) through Cooperative Agreement
    GPO-A-00-03-00003-00 and is implemented by the Carolina
 Population Center at the University of North Carolina in partnership
       with Futures Group, ICF International, John Snow, Inc.,
      Management Sciences for Health, and Tulane University.

        Visit us online at http://www.cpc.unc.edu/measure.

Contenu connexe

Tendances

Tableau 7.0 prsentation
Tableau 7.0 prsentationTableau 7.0 prsentation
Tableau 7.0 prsentationinam_slides
 
Assessing M&E Systems For Data Quality
Assessing M&E Systems For Data QualityAssessing M&E Systems For Data Quality
Assessing M&E Systems For Data QualityMEASURE Evaluation
 
chiranjivi yojna.pptx
chiranjivi yojna.pptxchiranjivi yojna.pptx
chiranjivi yojna.pptxSaurabh Gupta
 
Anemia mukth bharath
Anemia mukth bharathAnemia mukth bharath
Anemia mukth bharathBalaKumar263
 
Health Information System in the Nepal: Barriers to the adoption of Electron...
 Health Information System in the Nepal: Barriers to the adoption of Electron... Health Information System in the Nepal: Barriers to the adoption of Electron...
Health Information System in the Nepal: Barriers to the adoption of Electron...Nirmal Sharma
 
Data Modeling, Data Governance, & Data Quality
Data Modeling, Data Governance, & Data QualityData Modeling, Data Governance, & Data Quality
Data Modeling, Data Governance, & Data QualityDATAVERSITY
 
Data Architecture Best Practices for Advanced Analytics
Data Architecture Best Practices for Advanced AnalyticsData Architecture Best Practices for Advanced Analytics
Data Architecture Best Practices for Advanced AnalyticsDATAVERSITY
 
The data quality challenge
The data quality challengeThe data quality challenge
The data quality challengeLenia Miltiadous
 
How to Realize Benefits from Data Management Maturity Models
How to Realize Benefits from Data Management Maturity ModelsHow to Realize Benefits from Data Management Maturity Models
How to Realize Benefits from Data Management Maturity ModelsKingland
 
A Data Management Maturity Model Case Study
A Data Management Maturity Model Case StudyA Data Management Maturity Model Case Study
A Data Management Maturity Model Case StudyDATAVERSITY
 
Session 6 a DHIS2 : Overview and Implementation in West Africa
Session 6 a DHIS2 : Overview and Implementation in West AfricaSession 6 a DHIS2 : Overview and Implementation in West Africa
Session 6 a DHIS2 : Overview and Implementation in West AfricaCOP_HHA
 
National Population Policy
National Population Policy National Population Policy
National Population Policy Vedantha Vinod
 
The Role of Data Governance in a Data Strategy
The Role of Data Governance in a Data StrategyThe Role of Data Governance in a Data Strategy
The Role of Data Governance in a Data StrategyDATAVERSITY
 
Master Data Management's Place in the Data Governance Landscape
Master Data Management's Place in the Data Governance Landscape Master Data Management's Place in the Data Governance Landscape
Master Data Management's Place in the Data Governance Landscape CCG
 
Data Demand and Use Training Materials
Data Demand and Use Training MaterialsData Demand and Use Training Materials
Data Demand and Use Training MaterialsMEASURE Evaluation
 
Data Governance
Data GovernanceData Governance
Data GovernanceSambaSoup
 

Tendances (20)

Tableau 7.0 prsentation
Tableau 7.0 prsentationTableau 7.0 prsentation
Tableau 7.0 prsentation
 
Assessing M&E Systems For Data Quality
Assessing M&E Systems For Data QualityAssessing M&E Systems For Data Quality
Assessing M&E Systems For Data Quality
 
chiranjivi yojna.pptx
chiranjivi yojna.pptxchiranjivi yojna.pptx
chiranjivi yojna.pptx
 
Power bi
Power biPower bi
Power bi
 
Anemia mukth bharath
Anemia mukth bharathAnemia mukth bharath
Anemia mukth bharath
 
Health Information System in the Nepal: Barriers to the adoption of Electron...
 Health Information System in the Nepal: Barriers to the adoption of Electron... Health Information System in the Nepal: Barriers to the adoption of Electron...
Health Information System in the Nepal: Barriers to the adoption of Electron...
 
Data Quality in Healthcare: An Important Challenge
Data Quality in Healthcare: An Important ChallengeData Quality in Healthcare: An Important Challenge
Data Quality in Healthcare: An Important Challenge
 
Data Modeling, Data Governance, & Data Quality
Data Modeling, Data Governance, & Data QualityData Modeling, Data Governance, & Data Quality
Data Modeling, Data Governance, & Data Quality
 
Data Architecture Best Practices for Advanced Analytics
Data Architecture Best Practices for Advanced AnalyticsData Architecture Best Practices for Advanced Analytics
Data Architecture Best Practices for Advanced Analytics
 
The data quality challenge
The data quality challengeThe data quality challenge
The data quality challenge
 
Data Quality Presentation.ppt
Data Quality Presentation.pptData Quality Presentation.ppt
Data Quality Presentation.ppt
 
How to Realize Benefits from Data Management Maturity Models
How to Realize Benefits from Data Management Maturity ModelsHow to Realize Benefits from Data Management Maturity Models
How to Realize Benefits from Data Management Maturity Models
 
A Data Management Maturity Model Case Study
A Data Management Maturity Model Case StudyA Data Management Maturity Model Case Study
A Data Management Maturity Model Case Study
 
Data analytics and powerbi intro
Data analytics and powerbi introData analytics and powerbi intro
Data analytics and powerbi intro
 
Session 6 a DHIS2 : Overview and Implementation in West Africa
Session 6 a DHIS2 : Overview and Implementation in West AfricaSession 6 a DHIS2 : Overview and Implementation in West Africa
Session 6 a DHIS2 : Overview and Implementation in West Africa
 
National Population Policy
National Population Policy National Population Policy
National Population Policy
 
The Role of Data Governance in a Data Strategy
The Role of Data Governance in a Data StrategyThe Role of Data Governance in a Data Strategy
The Role of Data Governance in a Data Strategy
 
Master Data Management's Place in the Data Governance Landscape
Master Data Management's Place in the Data Governance Landscape Master Data Management's Place in the Data Governance Landscape
Master Data Management's Place in the Data Governance Landscape
 
Data Demand and Use Training Materials
Data Demand and Use Training MaterialsData Demand and Use Training Materials
Data Demand and Use Training Materials
 
Data Governance
Data GovernanceData Governance
Data Governance
 

En vedette

Data quality and data profiling
Data quality and data profilingData quality and data profiling
Data quality and data profilingShailja Khurana
 
Data quality architecture
Data quality architectureData quality architecture
Data quality architectureanicewick
 
Data Quality Rules introduction
Data Quality Rules introductionData Quality Rules introduction
Data Quality Rules introductiondatatovalue
 
Data Quality Dashboards
Data Quality DashboardsData Quality Dashboards
Data Quality DashboardsWilliam Sharp
 
Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...
Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...
Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...MEASURE Evaluation
 
Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...
Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...
Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...MEASURE Evaluation
 
Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...
Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...
Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...MEASURE Evaluation
 
SUSLA IE Process: Writing Assessment Results
SUSLA IE Process: Writing Assessment Results SUSLA IE Process: Writing Assessment Results
SUSLA IE Process: Writing Assessment Results Cleopatra Allen
 
VC M&E Module 6 - Developing Analysis and Feedback Loops
VC M&E Module 6 - Developing Analysis and Feedback LoopsVC M&E Module 6 - Developing Analysis and Feedback Loops
VC M&E Module 6 - Developing Analysis and Feedback LoopsCARE Economic Development Unit
 
Automated Quality Controls with FME
Automated Quality Controls with FMEAutomated Quality Controls with FME
Automated Quality Controls with FMESafe Software
 
Basic Concepts in Measurement and Evaluation
Basic Concepts in Measurement and EvaluationBasic Concepts in Measurement and Evaluation
Basic Concepts in Measurement and EvaluationChacha Bugais
 
Open Source Power Quality Data Visualization
Open Source Power Quality Data VisualizationOpen Source Power Quality Data Visualization
Open Source Power Quality Data VisualizationGrid Protection Alliance
 
LDIF Lightening Talk
LDIF Lightening TalkLDIF Lightening Talk
LDIF Lightening TalkWilliam Smith
 
Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...
Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...
Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...Beniamino Murgante
 
Assessment & adjustment for data quality used in the South African DISTRICT ...
Assessment & adjustment for data quality used in the South African DISTRICT ...Assessment & adjustment for data quality used in the South African DISTRICT ...
Assessment & adjustment for data quality used in the South African DISTRICT ...Routine Health Information NetwOrk (RHINO)
 
2014 review of data quality assessment methods
2014 review of data quality assessment methods2014 review of data quality assessment methods
2014 review of data quality assessment methodsRoger Zapata
 
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality AssessmentLeveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality AssessmentUmair ul Hassan
 

En vedette (20)

Data quality and data profiling
Data quality and data profilingData quality and data profiling
Data quality and data profiling
 
Data quality architecture
Data quality architectureData quality architecture
Data quality architecture
 
Data Quality Rules introduction
Data Quality Rules introductionData Quality Rules introduction
Data Quality Rules introduction
 
Data Quality Definitions
Data Quality DefinitionsData Quality Definitions
Data Quality Definitions
 
Data Quality Dashboards
Data Quality DashboardsData Quality Dashboards
Data Quality Dashboards
 
Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...
Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...
Collecting the PEPFAR OVC MER Essential Survey Indicators: Frequently Asked Q...
 
Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...
Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...
Data Quality Review (DQR) Methods and Tools: Holistic, Country-Led Data Qual...
 
Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...
Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...
Strengthening M&E System Among Partners Implementing HIV/AIDS Projects in Tan...
 
SUSLA IE Process: Writing Assessment Results
SUSLA IE Process: Writing Assessment Results SUSLA IE Process: Writing Assessment Results
SUSLA IE Process: Writing Assessment Results
 
VC M&E Module 6 - Developing Analysis and Feedback Loops
VC M&E Module 6 - Developing Analysis and Feedback LoopsVC M&E Module 6 - Developing Analysis and Feedback Loops
VC M&E Module 6 - Developing Analysis and Feedback Loops
 
Data journalism: Data rules, while data rule
Data journalism: Data rules, while data ruleData journalism: Data rules, while data rule
Data journalism: Data rules, while data rule
 
Automated Quality Controls with FME
Automated Quality Controls with FMEAutomated Quality Controls with FME
Automated Quality Controls with FME
 
Basic Concepts in Measurement and Evaluation
Basic Concepts in Measurement and EvaluationBasic Concepts in Measurement and Evaluation
Basic Concepts in Measurement and Evaluation
 
Open Source Power Quality Data Visualization
Open Source Power Quality Data VisualizationOpen Source Power Quality Data Visualization
Open Source Power Quality Data Visualization
 
LDIF Lightening Talk
LDIF Lightening TalkLDIF Lightening Talk
LDIF Lightening Talk
 
Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...
Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...
Data Usability Assessment for Remote Sensing Data: Accuracy of Interactive Da...
 
Assessment & adjustment for data quality used in the South African DISTRICT ...
Assessment & adjustment for data quality used in the South African DISTRICT ...Assessment & adjustment for data quality used in the South African DISTRICT ...
Assessment & adjustment for data quality used in the South African DISTRICT ...
 
2014 review of data quality assessment methods
2014 review of data quality assessment methods2014 review of data quality assessment methods
2014 review of data quality assessment methods
 
LDQ 2014 DQ Methodology
LDQ 2014 DQ MethodologyLDQ 2014 DQ Methodology
LDQ 2014 DQ Methodology
 
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality AssessmentLeveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment
 

Similaire à MEASURE Evaluation Data Quality Assessment Methodology and Tools

Effectively managing project performance reporting
Effectively managing project performance reportingEffectively managing project performance reporting
Effectively managing project performance reportingMindtree Ltd.
 
PEPFAR Vietnam Indicator Database
PEPFAR Vietnam Indicator DatabasePEPFAR Vietnam Indicator Database
PEPFAR Vietnam Indicator DatabaseMEASURE Evaluation
 
What is your PMI? A Model for Assessing the Maturity of Performance Manageme...
What is your PMI? A Model for Assessing the Maturity of Performance Manageme...What is your PMI? A Model for Assessing the Maturity of Performance Manageme...
What is your PMI? A Model for Assessing the Maturity of Performance Manageme...Mika Aho
 
Solvency - II Programme Setup
Solvency - II Programme SetupSolvency - II Programme Setup
Solvency - II Programme Setupgainline
 
Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...
Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...
Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...MEASURE Evaluation
 
Data Quality Management: Cleaner Data, Better Reporting
Data Quality Management: Cleaner Data, Better ReportingData Quality Management: Cleaner Data, Better Reporting
Data Quality Management: Cleaner Data, Better Reportingaccenture
 
Managing a public cloud
Managing a public cloudManaging a public cloud
Managing a public cloudInterop
 
How to Conduct A Nationally Owned Governance Assessment
 How to Conduct A Nationally Owned Governance Assessment  How to Conduct A Nationally Owned Governance Assessment
How to Conduct A Nationally Owned Governance Assessment Governance Asssessment Portal
 
Systems Lifecycle workbook
Systems Lifecycle workbookSystems Lifecycle workbook
Systems Lifecycle workbookMISY
 
Solvency II - Systems & Data Workstream
Solvency II - Systems & Data WorkstreamSolvency II - Systems & Data Workstream
Solvency II - Systems & Data Workstreamgainline
 
Prime Dimensions Capabilities
Prime Dimensions CapabilitiesPrime Dimensions Capabilities
Prime Dimensions Capabilitiesdrowan
 
Software QA Metrics Dashboard Benchmarking
Software QA Metrics Dashboard BenchmarkingSoftware QA Metrics Dashboard Benchmarking
Software QA Metrics Dashboard BenchmarkingJohn Carter
 
CA Quality Management System
CA Quality Management SystemCA Quality Management System
CA Quality Management SystemWahyu Prasetianto
 

Similaire à MEASURE Evaluation Data Quality Assessment Methodology and Tools (20)

Effectively managing project performance reporting
Effectively managing project performance reportingEffectively managing project performance reporting
Effectively managing project performance reporting
 
PEPFAR Vietnam Indicator Database
PEPFAR Vietnam Indicator DatabasePEPFAR Vietnam Indicator Database
PEPFAR Vietnam Indicator Database
 
What is your PMI? A Model for Assessing the Maturity of Performance Manageme...
What is your PMI? A Model for Assessing the Maturity of Performance Manageme...What is your PMI? A Model for Assessing the Maturity of Performance Manageme...
What is your PMI? A Model for Assessing the Maturity of Performance Manageme...
 
How to implement interoperability
How to implement interoperabilityHow to implement interoperability
How to implement interoperability
 
Tatiana Stebakova
Tatiana StebakovaTatiana Stebakova
Tatiana Stebakova
 
IBM GPRA
IBM GPRAIBM GPRA
IBM GPRA
 
Solvency - II Programme Setup
Solvency - II Programme SetupSolvency - II Programme Setup
Solvency - II Programme Setup
 
Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...
Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...
Results of the Piloting the Toolkit for Rapid Monitoring of AIDS Referral Sys...
 
Data Quality Management: Cleaner Data, Better Reporting
Data Quality Management: Cleaner Data, Better ReportingData Quality Management: Cleaner Data, Better Reporting
Data Quality Management: Cleaner Data, Better Reporting
 
Managing a public cloud
Managing a public cloudManaging a public cloud
Managing a public cloud
 
ISO 20K is Not Scary
ISO 20K is Not ScaryISO 20K is Not Scary
ISO 20K is Not Scary
 
Guide dogs
Guide dogsGuide dogs
Guide dogs
 
How to Conduct A Nationally Owned Governance Assessment
 How to Conduct A Nationally Owned Governance Assessment  How to Conduct A Nationally Owned Governance Assessment
How to Conduct A Nationally Owned Governance Assessment
 
Systems Lifecycle workbook
Systems Lifecycle workbookSystems Lifecycle workbook
Systems Lifecycle workbook
 
iso 20000
iso 20000iso 20000
iso 20000
 
Solvency II - Systems & Data Workstream
Solvency II - Systems & Data WorkstreamSolvency II - Systems & Data Workstream
Solvency II - Systems & Data Workstream
 
Ops A La Carte SPC Seminar
Ops A La Carte SPC SeminarOps A La Carte SPC Seminar
Ops A La Carte SPC Seminar
 
Prime Dimensions Capabilities
Prime Dimensions CapabilitiesPrime Dimensions Capabilities
Prime Dimensions Capabilities
 
Software QA Metrics Dashboard Benchmarking
Software QA Metrics Dashboard BenchmarkingSoftware QA Metrics Dashboard Benchmarking
Software QA Metrics Dashboard Benchmarking
 
CA Quality Management System
CA Quality Management SystemCA Quality Management System
CA Quality Management System
 

Plus de MEASURE Evaluation

Managing missing values in routinely reported data: One approach from the Dem...
Managing missing values in routinely reported data: One approach from the Dem...Managing missing values in routinely reported data: One approach from the Dem...
Managing missing values in routinely reported data: One approach from the Dem...MEASURE Evaluation
 
Use of Routine Data for Economic Evaluations
Use of Routine Data for Economic EvaluationsUse of Routine Data for Economic Evaluations
Use of Routine Data for Economic EvaluationsMEASURE Evaluation
 
Routine data use in evaluation: practical guidance
Routine data use in evaluation: practical guidanceRoutine data use in evaluation: practical guidance
Routine data use in evaluation: practical guidanceMEASURE Evaluation
 
Tuberculosis/HIV Mobility Study: Objectives and Background
Tuberculosis/HIV Mobility Study: Objectives and BackgroundTuberculosis/HIV Mobility Study: Objectives and Background
Tuberculosis/HIV Mobility Study: Objectives and BackgroundMEASURE Evaluation
 
How to improve the capabilities of health information systems to address emer...
How to improve the capabilities of health information systems to address emer...How to improve the capabilities of health information systems to address emer...
How to improve the capabilities of health information systems to address emer...MEASURE Evaluation
 
LCI Evaluation Uganda Organizational Network Analysis
LCI Evaluation Uganda Organizational Network AnalysisLCI Evaluation Uganda Organizational Network Analysis
LCI Evaluation Uganda Organizational Network AnalysisMEASURE Evaluation
 
Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...
Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...
Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...MEASURE Evaluation
 
Understanding Referral Networks for Adolescent Girls and Young Women
Understanding Referral Networks for Adolescent Girls and Young WomenUnderstanding Referral Networks for Adolescent Girls and Young Women
Understanding Referral Networks for Adolescent Girls and Young WomenMEASURE Evaluation
 
Data for Impact: Lessons Learned in Using the Ripple Effects Mapping Method
Data for Impact: Lessons Learned in Using the Ripple Effects Mapping MethodData for Impact: Lessons Learned in Using the Ripple Effects Mapping Method
Data for Impact: Lessons Learned in Using the Ripple Effects Mapping MethodMEASURE Evaluation
 
Local Capacity Initiative (LCI) Evaluation
Local Capacity Initiative (LCI) EvaluationLocal Capacity Initiative (LCI) Evaluation
Local Capacity Initiative (LCI) EvaluationMEASURE Evaluation
 
Development and Validation of a Reproductive Empowerment Scale
Development and Validation of a Reproductive Empowerment ScaleDevelopment and Validation of a Reproductive Empowerment Scale
Development and Validation of a Reproductive Empowerment ScaleMEASURE Evaluation
 
Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...
Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...
Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...MEASURE Evaluation
 
Using Most Significant Change in a Mixed-Methods Evaluation in Uganda
Using Most Significant Change in a Mixed-Methods Evaluation in UgandaUsing Most Significant Change in a Mixed-Methods Evaluation in Uganda
Using Most Significant Change in a Mixed-Methods Evaluation in UgandaMEASURE Evaluation
 
Lessons Learned In Using the Most Significant Change Technique in Evaluation
Lessons Learned In Using the Most Significant Change Technique in EvaluationLessons Learned In Using the Most Significant Change Technique in Evaluation
Lessons Learned In Using the Most Significant Change Technique in EvaluationMEASURE Evaluation
 
Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...
Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...
Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...MEASURE Evaluation
 
Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...
Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...
Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...MEASURE Evaluation
 
Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...
Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...
Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...MEASURE Evaluation
 
Lessons learned in using process tracing for evaluation
Lessons learned in using process tracing for evaluationLessons learned in using process tracing for evaluation
Lessons learned in using process tracing for evaluationMEASURE Evaluation
 
Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...
Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...
Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...MEASURE Evaluation
 
Sustaining the Impact: MEASURE Evaluation Conversation on Health Informatics
Sustaining the Impact: MEASURE Evaluation Conversation on Health InformaticsSustaining the Impact: MEASURE Evaluation Conversation on Health Informatics
Sustaining the Impact: MEASURE Evaluation Conversation on Health InformaticsMEASURE Evaluation
 

Plus de MEASURE Evaluation (20)

Managing missing values in routinely reported data: One approach from the Dem...
Managing missing values in routinely reported data: One approach from the Dem...Managing missing values in routinely reported data: One approach from the Dem...
Managing missing values in routinely reported data: One approach from the Dem...
 
Use of Routine Data for Economic Evaluations
Use of Routine Data for Economic EvaluationsUse of Routine Data for Economic Evaluations
Use of Routine Data for Economic Evaluations
 
Routine data use in evaluation: practical guidance
Routine data use in evaluation: practical guidanceRoutine data use in evaluation: practical guidance
Routine data use in evaluation: practical guidance
 
Tuberculosis/HIV Mobility Study: Objectives and Background
Tuberculosis/HIV Mobility Study: Objectives and BackgroundTuberculosis/HIV Mobility Study: Objectives and Background
Tuberculosis/HIV Mobility Study: Objectives and Background
 
How to improve the capabilities of health information systems to address emer...
How to improve the capabilities of health information systems to address emer...How to improve the capabilities of health information systems to address emer...
How to improve the capabilities of health information systems to address emer...
 
LCI Evaluation Uganda Organizational Network Analysis
LCI Evaluation Uganda Organizational Network AnalysisLCI Evaluation Uganda Organizational Network Analysis
LCI Evaluation Uganda Organizational Network Analysis
 
Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...
Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...
Using Organizational Network Analysis to Plan and Evaluate Global Health Prog...
 
Understanding Referral Networks for Adolescent Girls and Young Women
Understanding Referral Networks for Adolescent Girls and Young WomenUnderstanding Referral Networks for Adolescent Girls and Young Women
Understanding Referral Networks for Adolescent Girls and Young Women
 
Data for Impact: Lessons Learned in Using the Ripple Effects Mapping Method
Data for Impact: Lessons Learned in Using the Ripple Effects Mapping MethodData for Impact: Lessons Learned in Using the Ripple Effects Mapping Method
Data for Impact: Lessons Learned in Using the Ripple Effects Mapping Method
 
Local Capacity Initiative (LCI) Evaluation
Local Capacity Initiative (LCI) EvaluationLocal Capacity Initiative (LCI) Evaluation
Local Capacity Initiative (LCI) Evaluation
 
Development and Validation of a Reproductive Empowerment Scale
Development and Validation of a Reproductive Empowerment ScaleDevelopment and Validation of a Reproductive Empowerment Scale
Development and Validation of a Reproductive Empowerment Scale
 
Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...
Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...
Sustaining the Impact: MEASURE Evaluation Conversation on Maternal and Child ...
 
Using Most Significant Change in a Mixed-Methods Evaluation in Uganda
Using Most Significant Change in a Mixed-Methods Evaluation in UgandaUsing Most Significant Change in a Mixed-Methods Evaluation in Uganda
Using Most Significant Change in a Mixed-Methods Evaluation in Uganda
 
Lessons Learned In Using the Most Significant Change Technique in Evaluation
Lessons Learned In Using the Most Significant Change Technique in EvaluationLessons Learned In Using the Most Significant Change Technique in Evaluation
Lessons Learned In Using the Most Significant Change Technique in Evaluation
 
Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...
Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...
Malaria Data Quality and Use in Selected Centers of Excellence in Madagascar:...
 
Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...
Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...
Evaluating National Malaria Programs’ Impact in Moderate- and Low-Transmissio...
 
Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...
Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...
Improved Performance of the Malaria Surveillance, Monitoring, and Evaluation ...
 
Lessons learned in using process tracing for evaluation
Lessons learned in using process tracing for evaluationLessons learned in using process tracing for evaluation
Lessons learned in using process tracing for evaluation
 
Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...
Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...
Use of Qualitative Comparative Analysis in the Assessment of the Actionable D...
 
Sustaining the Impact: MEASURE Evaluation Conversation on Health Informatics
Sustaining the Impact: MEASURE Evaluation Conversation on Health InformaticsSustaining the Impact: MEASURE Evaluation Conversation on Health Informatics
Sustaining the Impact: MEASURE Evaluation Conversation on Health Informatics
 

Dernier

Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Mark Goldstein
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxGenerative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxfnnc6jmgwh
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...itnewsafrica
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkPixlogix Infotech
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfIngrid Airi González
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationKnoldus Inc.
 
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfpanagenda
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI AgeCprime
 
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better StrongerModern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better Strongerpanagenda
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integrationmarketing932765
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch TuesdayIvanti
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Nikki Chapple
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersNicole Novielli
 

Dernier (20)

Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxGenerative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App Framework
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdf
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog Presentation
 
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI Age
 
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better StrongerModern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch Tuesday
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
Microsoft 365 Copilot: How to boost your productivity with AI – Part one: Ado...
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software Developers
 

MEASURE Evaluation Data Quality Assessment Methodology and Tools

  • 1. MEASURE Evaluation Data Quality Assessment Methodology and Tools USAID Washington D.C. July 19, 2012
  • 2. Background ■ National Programs and Donor-funded projects are working towards achieving ambitious goals in the fight against HIV, TB and Malaria. ■ Measuring success and improving management of these initiatives is predicated on strong M&E systems that produce quality data regarding program implementation. ■ In the spirit of the “Three Ones”, the “Stop TB Strategy” and the “RBM Global Strategic Plan”, a multi-partner project* was launched 2006 to develop a joint Data Quality Audit (DQA) Tool. ■ The objective of this initiative is to provide a common approach for assessing and improving data quality. A single tool ensures that standards are harmonized and allows for joint implementation (between partners and with National Programs). * Partners most directly involved include PEPFAR, USAID, WHO, Stop TB, the Global Fund and MEASURE
  • 3. Outline  Auditing  DQA  Evolution  Methods and tools  How and where applied  Routine Data Quality Assurance and Capacity Building  RDQA  Purpose  Methods and tools  How and where applied  Integration into Routine Practice
  • 4. Objective of the Data Quality Audit (DQA) Tool The Data Quality Audit (DQA) Tool is designed to:  verify the quality of reported data for key indicators at selected sites; and  assess the ability of data-management systems to collect manage and report quality data.
  • 5. Conceptual Framework Generally, the quality of reported data is dependent on the underlying data management and reporting systems; stronger systems should produce better quality data. Dimensions of Quality QUALITY DATA Accuracy, Completeness, Reliability, Timeliness, Confidentiality, Precision, Integrity Functional Components of a Data Management System Data-Management and M&E Unit Needed to Ensure Data Quality Reporting System REPORTING LEVELS I M&E Structure, Functions and Capabilities II Indicator Definitions and Reporting Guidelines Intermediate Aggregation III Data-collection and Reporting Forms / Tools Levels (e.g. Districts, Regions) IV Data Management Processes V Links with National Reporting System VI Data management processes Service Points VII 5 Data quality mechanisms and controls VIII Links with the national reporting system
  • 6. DQA METHODOLOGY – 2 Protocols ■ The methodology for the DQA includes two (2) protocols: 1 Assessment of Data Management Qualitative assessment of the strengths and Systems weaknesses of the data-collection and reporting system. (Protocol 1) 2 Quantitative comparison of recounted to reported Data Verifications data and review of timeliness, completeness and (Protocol 2) availability of reports.
  • 7. Typical Timeline PHASE 1 PHASE 2 PHASE 3 PHASE 4 PHASE 5 PHASE 6 Intermediate Preparation and M&E Service Delivery Aggregation M&E Completion Initiation Management Sites / levels Management (multiple (multiple Unit Organizations Unit (eg. District, locations) locations) Region) 1. Select 6. Draft initial 7. Draft and Indicators and findings and 3. Assess Data Management and Reporting Systems (Part 1) discuss Audit Reporting conduct close- Report Period out meeting 2. Notify 4. Select/Confirm 8. Initiate Program and Service follow-up of desk review of Delivery 5. Trace and Verify Reported Results (Part 2) recommended Program Points to be actions documentation visited ■ The DQA is implemented chronologically in 6 Phases. ■ Assessments and verifications will take place at every stage of the reporting system: - M&E Management Unit - Intermediate Aggregation Level (Districts, Regions) - Service Delivery Sites. 7
  • 8. PROTOCOL 1: Assessment of Data Management Systems 8
  • 9. Intermediate M&E Service Delivery PROTOCOL 1: Aggregation Management Sites / levels Unit Organizations Assessment of Data (eg. District, Region) Management Systems 3. Assess Data Management and Reporting Systems (Part 1) ■ PURPOSE: Identify potential risks to data quality created by the data- management and reporting systems at: - the M&E Management Unit; - the Service Delivery Points; - any Intermediary Aggregation Level (District or Region). ■ The DQA assesses both (1) the design; and (2) the implementation of the data- management and reporting systems. ■ The assessment covers 5 functional areas (HR, Training, Data Management Processes , etc.) 9
  • 10. SYSTEMS ASSESSMENT QUESTIONS BY FUNCTIONAL AREA Functional Areas Summary Questions I M&E Structure, Are key M&E and data-management staff identified with Functions and 1 clearly assigned responsibilities? Capabilities Have the majority of key M&E and data-management staff 2 received the required training? II Indicator Has the Program/Project clearly documented (in writing) Definitions and 3 what is reported to who, and how and when reporting is Reporting required? Guidelines Are there operational indicator definitions meeting relevant 4 standards? And are they consistently followed by all service points? III Data-collection Are there standard data-collection and reporting forms that and Reporting 5 are systematically used? Forms and Tools Are data recorded with sufficient precision/detail to 6 measure relevant indicators? Are data maintained in accordance with international or 7 national confidentiality guidelines? Are source documents kept and made available in 8 accordance with a written policy?
  • 11. SYSTEMS ASSESSMENT QUESTIONS BY FUNCTIONAL AREA Functional Areas Summary Questions IV Data Management Does clear documentation of collection, aggregation 9 Processes and manipulation steps exist? Are data quality challenges identified and are 10 mechanisms in place for addressing them? Are there clearly defined and followed procedures to 11 identify and reconcile discrepancies in reports? Are there clearly defined and followed procedures to 12 periodically verify source data? V Links with National Does the data collection and reporting system of the Reporting 13 Program/project link to the National Reporting System? System
  • 12.
  • 14. Intermediate M&E Service Delivery Aggregation Management Sites / PROTOCOL 2: Unit Organizations levels (eg. District, Region) Data Verifications 5. Trace and Verify Reported Results (Part 2) ■ PURPOSE: Assess on a limited scale if Service Delivery Points and Intermediate Aggregation Sites are collecting and reporting data accurately and on time. ■ The trace and verification exercise will take place in two stages: - In-depth verifications at the Service Delivery Points; and - Follow-up verifications at the Intermediate Aggregation Levels (Districts, Regions) and at the M&E Unit. 14
  • 15. DQA Data Verifications M&E Unit/National Monthly Report ILLUSTRATION District 1 65 District 2 45 District 3 75 District 4 250 TOTAL 435 District 1 District 2 District 3 District 4 Monthly Report Monthly Report Monthly Report Monthly Report SDS 1 45 SDS 3 45 SDS 4 75 SDP 5 50 SDS 2 20 TOTAL 45 TOTAL 75 SDP 6 200 TOTAL 65 TOTAL 250 Service Delivery Site 1 Service Delivery Site 2 Service Delivery Site 3 Service Delivery Site 4 Service Delivery Site 5 Service Delivery Site 6 Monthly Report Monthly Report Monthly Report Monthly Report Monthly Report Monthly Report ARV Nb. 45 ARV Nb. 20 ARV Nb. 45 ARV Nb. 75 ARV Nb. 50 ARV Nb. 200 Source Source Source Source Source Source Document 1 Document 1 Document 1 Document 1 Document 1 Document 1
  • 16. Service Delivery Points – Data Verification SERVICE DELIVERY POINT - 5 TYPES OF DATA VERIFICATIONS Verifications Description - Verification no. 1: Describe the connection between the delivery of services/commodities and the completion of the source In all cases Description document that records that service delivery. Verification no. 2: Review availability and completeness of all indicator Documentation In all cases source documents for the selected reporting period. Review Trace and verify reported numbers: (1) Recount the Verification n . 3: o reported numbers from available source documents; (2) In all cases Trace and Verification Compare the verified numbers to the site reported number; (3) Identify reasons for any differences. Verification no. 4: Perform “cross-checks” of the verified report totals with other data-sources (eg. inventory records, laboratory If feasible Cross-checks reports, etc.). Verification no. 5: Perform “spot checks” to verify the actual delivery of If feasible Spot checks services or commodities to the target populations.
  • 17. Summary Statistics of Data Verifications Total and Adjusted District Verification Factors from % On Time Reports from DQA Total DQA Verification Factor IAL #4 M&E Unit 0.75 IAL #4 1,07 0.73 IAL #3 IAL #3 1.22 IAL #2 1 0.88 1 1,03 IAL #1 IAL #2 0.47 1,07 0.80 IAL #1 0,95 0, 00 0, 20 0, 40 0,60 0,80 1, 00 1, 20 0.00 0.20 0.40 0.60 0.80 1.00 % Available Reports from DQA % Complete Reports from DQA 0.50 0,75 M&E Unit 0.67 0,83 IAL #4 M&E Unit IAL #3 IAL #4 1 0.72 1 0,92 IAL #2 IAL #3 IAL #1 IAL #2 0,50 0.60 IAL #1 0,80 0.70 0,00 0,20 0,40 0,60 0,80 1,00 0.00 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 17
  • 18. Outputs of the DQA Implementation ■ Completed protocols and templates related to: - Protocol 1: System Assessment; and - Protocol 2: Data Verification. ■ Write-ups of observations, interviews, and conversations with key staff. ■ Preliminary findings and recommendation notes. ■ Final DQA Report, including: - Summary of the evidence collected by the DQA team; - Specific findings or gaps related to that evidence; - Recommendations to improve data-management systems and overall data quality. The Final DQA Report should also include summary statistics on: 1. Strengths and weaknesses of data-collection and reporting systems; 2. Recounting exercise performed on primary records and aggregated reports; 3. Available, On time and Complete Reports. 18
  • 19. Use of the DQA since 2008
  • 20. Use of DQA (continued)  Biggest user is the Global Fund  Part of the grant renewal process  Random and purposive selection of grants  External audit teams – DQA IQC  1st IQC 2008-2010  2nd IQC 2011-2012  Standardized output  GF Reorganization
  • 21. Global Fund DQAs 2008-2012 Disease Year Qtr HIV/AIDS TB Malaria 2008 4 Mali, Belarus Philippines Rwanda, China, LAC- Multi-country 2009 1 Comoros Islands China, Yemen (Americas – Andean Region) 2009 2 Haiti, Burundi Niger Ghana, Indonesia Mozambique, Gambia, Vietnam, Sri 2009 3 Dominican Republic, Lanka, Nigeria, Zambia+ Tanzania Peru, Swaziland, South Kenya, Cote d’Ivoire, 2009 4 Africa, Thailand, Tajikistan Madagascar Guyana Ethiopia, Paraguay, 2010 1 Malawi, C.A.R. Uzbekistan Uganda, Cameroon Papua New Guinea, Russia, Guinea- 2010 2 Conakry, Eritrea Pakistan, Paraguay Pakistan, South Sudan, Madagascar 2011 2 Zambia Kyrgyzstan North Sudan Angola, Cambodia, Senegal, South Sudan, 2011 4 Namibia, Guyana, India, North Korea DRC Myanmar, Ukraine
  • 22. GF Audits by Indicator (2008-2010) Disease Program Area Number of Audits Malaria Malaria Treatment 13 HIV Current on ART 12 Malaria LLIN 10 HIV C&T HIV 9 HIV Care and support 6 TB TB Detection 5 HIV PMTCT – preg. women on prophylaxis 5 Malaria Malaria Training 4 TB TB Treatment 4 HIV HIV Education 4 Malaria IPT 3 HIV Condoms 3 Malaria Malaria Treatment - Community Based 2 Malaria Malaria Testing 2 HIV OVC 2 HIV HIV Training 1 TB MDR TB Treatment 1 HIV TB/HIV 1 Total 87
  • 23. Use of DQA (continued)  PEPFAR Countries  2009 Country Operational Plan Guidance  Mozambique, Ukraine, DRC, Uganda, Nigeria, Cote d’Ivoire, Dominican Republic  Ad hoc use by Partners, other agencies  UNICEF in India  Adaptation for local use  South Africa (ESI)
  • 25. DQA Performance Analysis National Level: Performance Table: Performance Scores Weighted Final Audit Performance Score 100% Sample Accuracy 77% 100% 90% Values provided automatically OSDV Accuracy 92% 80% 90% Availability 74% 70% Timeliness 52% 80% 60% Completeness 84% 76.5% 70% 50% Cross-Checks 79% 60% Service Site 40% 65% Documentation Review M&E Structure, Functions 30% 50% 2.40 and Capabilities Qualitative - Enter these values Indicator Definitions and 20% 2.50 40% Reporting Guidelines Data-collection and 10% 2.30 Reporting Forms / Tools 30% Data Management 0% 1.90 Processes Links with National 2.80 Excellent = PS 91.0 - 100.0 Poor = PS 61.0 - 70.0 Reporting System Good = PS 81.0 - 90.0 Very Poor = PS < 60.0 Fair = PS 71.0 - 80.0
  • 26.
  • 27.
  • 28.
  • 29.
  • 30. Evolution of RDQA tool  Need expressed for a tool for programs/projects to use for self assessment, improved data quality, capacity building and to prepare for external audits  New generic, simplified tool adapted to the country, program and/or project  Routine use  Quantify and track data quality performance over time  Identify and respond to data quality problems
  • 31. Use of the Auditing and Self Assessment/Routine Versions Auditing version… Capacity Building version…  Assessment by funding  Simplified from the auditing agency version  Standardized approach to  Flexible use by program auditing  Self-assessment  Conducted by an external audit team  Focus is on identifying weaknesses for system  Relatively expensive strengthening  Limited opportunities for  Should be used routinely – data capacity building quality performance indicators monitored over time
  • 32. RDQA Modifications  No implied sampling methodology  No aggregate sample statistics  Quantitative and qualitative assessments merged into the same tool  Dashboards by site and level to facilitate data use  Easily extract data to link with data systems  Action plans for system strengthening
  • 33. Objectives of the RDQA OBJECTIVES: • VERIFY • the quality of reported data for key indicators at selected sites; and the ability of data-management systems to collect, manage and report quality data. • IDENTIFY • weaknesses in the data management system and • interventions for system strengthening. • MONITOR • data quality performance over time and • capacity to produce good quality data
  • 34. Uses of the RDQA The RDQA is designed to be flexible in use and can serve multiple purposes:  Routine data quality checks as part of on-going supervision  Initial and follow-up assessments of data management and reporting systems – measure performance improvement over time:  Strengthening program staff’s capacity in data management and reporting:  Preparation for a formal data quality audit·  External assessment by partners, other stakeholders
  • 35. RDQA Outputs 1. Strength of the M&E System evaluation based on a review of the Program/project’s data management and reporting system, including responses to overall summary questions on how well the system is designed and implemented; 2. Verification Factors generated from the trace and verify recounting exercise performed on primary records and/or aggregated reports (i.e. the ratio of the recounted value of the indicator to the reported value); 3. Available, On time and Complete Reports percentages calculated at the Intermediate Aggregation Levels and the M&E Unit). 4. Action Plan for System Strengthening for each site and level assessed.
  • 37. RDQA Summary Statistics – Level Specific Dashboard Part 4: DASHBOARD: National Level - M&E Unit Data Management Assessment - Data and Verifications - Reporting Performance - National Level - M&E Unit National Level - M&E Unit National Level - M&E Unit I - M&E Structure, 120% Functions and Capabilities 106% % Complete 82% 3 100% 102% 105% 94% 2 II- Indicator 80% V - Links with 1 Definitions and National Reporting Reporting System 60% % On Time 41% Guidelines 0 40% III - Data-collection 20% IV- Data Management and Reporting Forms % Available 92% Processes and Tools 0% ANC Registered ANC Counseled and ANC Received ANC Tested Positive Tested Results 0% 20% 40% 60% 80% 100%
  • 38. Global Dashboard Data Management Assessment - System Assessment Results by Level of the Reporting System Overall Average 3.00 M&E Structure, Functions and 2.50 Capabilities RDQA 3 2.00 Summary 2 1.50 Links with Indicator Definitions National Reporting 1 and Reporting Statistics – System Guidelines 1.00 0 0.50 Global 0.00 Dashboard Data Data-collection National M&E Unit - Cotonou Management and Reporting Processes Forms / Tools M&E Structure, Indicator Definitions Data-collection Data Links with Functions and and Reporting and Reporting Management National Reporting Capabilities Guidelines Forms / Tools Processes System Data Verifications - Overall Average Data Verifications - Verification Factors by Level of the Reporting System by Indicator 140% 1.4 1.2 120% 124% 112% 112% 1 100% 90% 0.8 80% 0.6 60% 0.4 40% 0.2 20% 0 Service Site Average District Average Regional Average National M&E Unit - Cotonou 0% ANC Registered ANC Counseled ANC Received ANC Tested and Tested Results Positive ANC Registered ANC Counseled and Tested ANC Received Results ANC Tested Positive Reporting Performance - Overall Average Reporting Performance by Reporting Level 1 0.9 0.8 Completeness 132% 0.7 0.6 0.5 Timeliness 51% 0.4 0.3 0.2 0.1 Availability 86% 0 District Average Regional Average National M&E Unit - Cotonou 0% 20% 40% 60% 80% 100% 120% 140% Availability Timeliness Completeness
  • 39. RDQA Summary Statistics – Site specific qualitative results
  • 42. Growing family of RDQA Tools Data Verifications - Service Site Summary  RDQA traditional version: Trend in Verification Ratio by Indicator one indicator per 3.5 workbook, x-sectional 3 2.5 Health Facility  RDQA Multi-indicator 2 version: up to four 1.5 indicators for the same 1 0.5 program area in x-section 0 Q1 Q2 Q3 Q4  RDQA Longitudinal CS Parakou Percent Accuracy Maternite de bonne saveur Parakou version: one indicator for CS Immacule Conception de Kalale Hopital St. Louis de Kalale up to four points in time. Hopital Regionale de Savalou CS de Zongo - Savalou Centre Hopitalier de Grand Quelqu'un d'Abomey CS Behanzin de Abomey  RDQA training materials  PMTCT and TB RDQA
  • 43. Countries where RDQA has been used or is currently being implemented  Nigeria • South Africa  Kenya • Lesotho  Tanzania • Swaziland  Cote d’Ivoire • DRC  Guyana • India • Botswana  Haiti • Global Fund On Site  Mozambique Data Verification  Zimbabwe (OSDV) by LFAs in many countries  Uganda  Dominican Republic  Peru  Paraguay
  • 44. Now we know the quality of the data. Now what?  Standardization:  Indicators  data collection and reporting tools  data management and indicator compilation practices  Capacity building to address identified needs  MEval Data quality assurance curriculum  Formalization of RDQA into routine practice  Monitor data quality performance indicators over time. What is the trend?
  • 45. Integrating Routine Data Quality Assurance into Standard Operating Procedures  Explicitly State:  Who will do what?  When they will do it?  What is the output?  Where will it happen?  What is the required follow up?  Who will verify that it is happening?  Formalize in official Program documentation  Incorporate into pre-service/in-service training
  • 46. Integration of RDQA into SOPs  Example from Botswana
  • 47. Key Success Factors for Data Quality 1. Functioning information systems 2. Clear definition of indicators consistently used at all levels 3. Description of roles and responsibilities at all levels 4. Specific reporting timelines 5. Standard/compatible data-collection and reporting forms/tools with clear instructions 6. Documented data review procedures to be performed at all levels 7. Steps for addressing data quality challenges (missing data, double- counting, lost to follow up, …) 8. Storage policy and filling practices that allow retrieval of documents for auditing purposes (leaving an audit trail)
  • 48.
  • 49. MEASURE Evaluation is funded by the U.S. Agency for International Development (USAID) through Cooperative Agreement GPO-A-00-03-00003-00 and is implemented by the Carolina Population Center at the University of North Carolina in partnership with Futures Group, ICF International, John Snow, Inc., Management Sciences for Health, and Tulane University. Visit us online at http://www.cpc.unc.edu/measure.

Notes de l'éditeur

  1. Another way to assess data quality is through a data quality audit.
  2. This chart shows visually that the trace and verify protocol starts with data at the level of service delivery (either in a health facility or a community based program) and how the data are “traced” to the “intermediate aggregation level” (in this case a district) and then to the M&amp;E Unit Level (in this case the national level).
  3. These charts show the summary statistics that are automatically generated from the trace and verify protocol of the data quality audit tool. They show for the reporting period and indicator audited, the: Verification factor (how the recounted data compare to the reported data) Availability of Reports Timeliness of Reports Completeness of Reports
  4. These charts show the summary statistics that are automatically generated from the trace and verify protocol of the data quality audit tool. They show for the reporting period and indicator audited, the: Verification factor (how the recounted data compare to the reported data) Availability of Reports Timeliness of Reports Completeness of Reports