Performance Improvement 1997. Appendix C. Senior Editorial Advisors' Review Criteria for Assessing Program Evaluations

02/01/1997

Overall Significance

  • The study addresses a significant issue of policy relevance.
  • Evaluation findings are likely to be useful.

Conceptual Criteria

Conceptual Foundations

  • A literature review is included.
  • The project is shown to be logically based on previous findings; the report uses either theory or models, or both.
  • The program assumptions are stated.
  • The evaluation draws from any previous evaluation.
  • The report is linked with a program and describes the program.
  • The report presents multiple perspectives.
  • Multiple relevant stakeholders are consulted and involved.
  • The timing is appropriate because the program is ready for evaluation.

Questions for Evaluation

  • The aims of the evaluation are clear, well-specified, and testable.
  • The questions are feasible, significant, linked to the program, appropriate for the resources and audience, and derive logically from the conceptual foundations.
  • The questions show ingenuity and creativity.

Findings and Interpretation

  • The conclusions are justified by the analyses.
  • The summary does not go beyond what the data will support.
  • The appropriate qualifiers are stated.
  • The conclusions fit the entire analysis.
  • Equivocal findings are handled appropriately.
  • The initial questions are answered.
  • The interpretation ties in with the conceptual foundation.
  • The report notes that the findings are either consistent with or deviate from the relevant literature.
  • The presentation is understandable.
  • The results have practical significance.
  • The extent of program implication is assessed.

Recommendations

  • The recommendations follow from findings, are worth carrying out, and are affordable, timely, feasible, useful, and appropriate.
  • The recommendations are shown to be relevant to the questions asked.
  • The breadth of specificity of the recommendations is addressed.
  • Any recommendations for either future evaluations or improvements or both are clearly presented.

Methods

Evaluation Design

  • Design considerations include overall appropriateness, soundness, funding and time constraints, generalizability, applicability for cultural diversity, assessment of the extent of program delivery, validity, feasibility for data collection, reliability of selected measurements, multiple measures of key concepts, and appropriateness of the sample.
  • Variables are clearly specified and fit with the questions and concepts.
  • The design permits measuring the extent of implementation of the program and answering the evaluation questions.

Data Collection

  • Data are collected using appropriate units of measurement for analysis, controls for participant selection and assignment bias, and proper handling of missing data and attrition.
  • Data collection is characterized by use of an appropriate comparison group of control; adequate sample size, response rate, and information about the sample; a plan and methods that are faithful to the plan; attention to and cooperation with the relevant community; project confidentiality; and consistency.
  • The quality of the data (including the quality of any extant data sets used in the study) and the efficiency of sampling are addressed.
  • The data collection is appropriate to the evaluation questions.

Data Analysis

  • The data analysis addresses the handling of attrition, the matching of the analysis to the design, the use of appropriate statistical controls, the use of methodology and levels of measurement appropriate to the type of data; and estimation of effect size.
  • The analysis shows sensitivity to cultural categories.
  • The analysis makes appropriate generalizability of inferences.
  • The chosen analysis type is simple and efficient.

Cross-Cutting Factors

The following are cross-cutting factors that are likely to be important at all stages of a report: clarity, presentation, operation at a state-of-the-art level, appropriateness, understandability, innovation, generalizability, efficiency of approach, logical relationships, and discussion of the report's limitations. The report should also address ethical issues, possible perceptual bias, cultural diversity, and any gaps in study execution.