Monitoring evaluation

download Monitoring evaluation

of 22

  • date post

    05-Dec-2014
  • Category

    Technology

  • view

    322
  • download

    2

Embed Size (px)

description

 

Transcript of Monitoring evaluation

  • 1. Monitoring Evaluation

2. Introduction Monitoring Evaluation relation to *Program management within an organization. Evaluative objective provides info to ensure that programs are working and they contribute to success Monitoring Form- associated with the allocation of the resources M& E is part of the total quality management and quality assurance thrusts. Quality assurance is motivated by the need for governments to be seen to deliver high quality services. 3. Summary of Monitoring Evaluation (Form D) Dimension Properties Orientation Assessing Program processes and outcomes, for fine- tuning and to account for Program resources Typical Issues Is the program reaching the target population? Is implementation meeting program objectives and benchmarks? How is implementation going between sites? .compared with a month ago? How can we finetune this Program to make it more efficient? to make it more effective? Is there a Program site which needs more attention to ensure more effective delivery State of program Settled. Program plan is in place Major focus Delivery and outcomes Timing (vis--vis) Program delivery During delivery Key Approaches Component Analysis Devolved performance assessment Systems Analysis Assembly of evidence Meaningful use of valid performance measures Quantitative indicators MIS 4. Key Approaches to Monitoring Evaluation Component Analysis - Senior management select a component of the Program for systematic analysis and review. Devolved performance assessment- Senior management encourage all components of a Program to assess their performance on a regular basis. Systems Analysis A program which is centrally specified and disseminated for implementation to a large number of sites. 5. Component Analysis Senior management select a component of the Program Assess that component in terms of its own objectives and overall goals of the Program The selection of the component is made on the grounds of concern* Key Assumptions: Senior Management Has sufficient overview of the organisation Has the power to direct the evaluation unit to address the issue Is a major audience for the evaluation findings 6. Devolved performance assessment Senior management encourages all components of a Program to assess their performance on a regular basis. Senior management receives these reports and, using appropriate criteria, makes judgments on the contribution of each component Senior management provides guidelines, resources, and principles for judging 7. Systems Analysis Applies to program which is centrally specified and disseminated for implementation to a large number of sites. Program specification includes important goals. Guidelines are provided for field staff; Field staff have little or no say in Program specification or implementation plans. Evaluation scenario: a set of important outcomes to be defined and made operational. Using a centralized evaluation unit Relating differences in attainment of the outcomes 8. Key Evaluation Questions Is the program reaching the target population? Is it being implemented in the ways specified Ia it effective How much does it cost? What are the costs relative to its effectiveness 9. Monitoring Evaluation: Trends and Case Examples Evaluands in program monitoring Focus is evaluation within big P Programs which are ongoing. Private sector- Ex: Training and development Program of a large regional bank Public sector Intellectual Disabilities Services Program Common: Contain mission statements Designed to translate aspects of policy into tangible outcomes Centrally planned or financed Prime responsibility of Senior management Ongoing, subject to modification 10. Elements of Program Management A strategic plan for the implementation of relevant aspects of government. Strategic planning the process by which an organisation creates a vision of its future and develops the necessary structure, resources, procedures and operations to achieve that future. A program structure Management arrangements The use of the MIS MIS used as a basis for decisions The use of MIS when considering fundamental issues 11. Large P Program evaluation Unique characteristics: 1. There is a strong emphasis on outcomes. 2. Programs are ongoing and there is a need for evaluative information over time. 3. Many Programs are designed to provide goods or services rather than promote changes in behaviour 4. Evaluative data are often processed and reported in simple but logical ways. 5. Senior management, inparticular, may require gross or aggregated information 12. Assembling Evidence for Monitoring Essential to use full range of data collection and analysis techniques Indicators need to be at least part of the data collection 13. Indicators as Evidence Key feature of indicators used continually to inform decisions designed to alter the state of the social system affecting them.. Can be used as statements about the effectiveness of organisations. Must be used in: comparing Program trends at different points in time (Monitoring) To compare the performance of a Program to an acceptable set of standards or goals. To compare the implementation of the same Program at different sites or locations 14. Types of indicators Appropriateness match between current community and government priorities and Program objectives; Efficiency the relative cost of achieving positive impacts via the program under consideration Effectiveness match between Program outcomes and Program objectives 15. Education for Sustainable Development An Expert Review of Processes and Learning Prof. Daniella Tilbury (University of Gloucestershire, United Kingdom) is the author of this publication commissioned by UNESCO. UNESCO 2011 Section for Education for Sustainable Development Division of Education for Peace and Sustainable Development UNESCO 7, Place de Fontenoy 75352 Paris 07 SP France Designed and printed at UNESCO Paris, France 16. Introduction Monitoring and evaluation (M&E) is an integral part of education programme planning and implementation. The United Nations Decade of Education for Sustainable Development (DESD, 2005-2014) is an endeavor that aims to reorient education policy, practice and investment to address sustainability. This publication endeavors to identify which commonly accepted learning processes are aligned with ESD and should be promoted through ESD-related programmes and activities. It also seeks to examine which learning opportunities contribute to sustainable development. 17. Summary The United Nations Decade in Education for Sustainable Development (DESD, 2005-2014) is a global movement which seeks to transform education policy, investment and practice. If it is successful, the DESD could change not only education but also the quality of life for many people across the globe. Key Objectives: i) Which commonly accepted learning processes are aligned with ESD and should be promoted through ESD activities? ii) What are ESD and related learning opportunities contributing to sustainable development? 18. ESD learning frameworks and processes The review has identified that certain key processes underpin ESD frameworks and practices. These include: < processes of collaboration and dialogue (including multi- stakeholder and intercultural dialogue); < processes which engage the whole system; < processes which innovate curriculum as well as teaching and learning experiences; and, < processes of active and participatory learning. 19. Learning for ESD defined Learning for ESD refers to what has been learnt and is learned by those engaged in ESD, including learners, facilitators, coordinators as well as funders. Often learning is interpreted as the gaining of knowledge, values and theories related to sustainable development but, as this review indicates, that ESD learning also refers to: < learning to ask critical questions; < learning to clarify ones own values; < learning to envision more positive and sustainable futures; < learning to think systemically; < learning to respond through applied learning; and, < learning to explore the dialectic between tradition and innovation. 20. Critical Lesson through the review It is difficult to access data on ESD processes and learning opportunities as these are rarely documented There is a noticeable lack of data to show how these objectives and outcomes are achieved. This relatively new field is only at the very earliest stages of generating the type of comparative and evaluative overview that provides a picture of effective processes and approaches. The study recommends that during Phase II i) data collection processes focus on actual experiences rather than reviews of the literature; and ii) data collection tools are based on tightly- focused questions that will capture greater detail about learning processes 21. Critical Question What is the extent and the depth of connection between the choice of processes in ESD initiatives and actual contributions to sustainable development? the level of evaluative assessment within the literature is in its infancy, the outcomes themselves are so varied and feature at multiple levels external review of case study findings, anecdotal evidence from individual programme evaluations and the reflections of programme leaders seems to suggest that there are links that should be explored in more detail. 22. Contribution to sustainable development this review presents a timely opportunity to consider the areas in which change is emerging The case studies reviewed in this document suggest that it is possible to map a wide range of contributions through ESD The review unpacks and categorizes the range of potential contributions and some of the themes and priorities that are apparent across these key initiatives. It has developed a template which could be