Collaboration for Impact’s Complexity and Evaluation Conference 2018: 3-4 May 2018
How Do You Measure Social Change?
Are you leading collective impact projects, addressing complex social problems or striving for population-level change? Are you regularly measuring the impact of your initiatives?
Collaboration for Impact’s Complexity and Evaluation Conference 2018 – taking place 3-4 May at the Melbourne Business School and featuring international evaluation experts Mark Cabaj and Kate McKegg – will equip you with tools and strategies to “unpack” complex systems change, and strengthen your approach to evaluating it.
Register for this conference if you are looking for:
- Useable approaches to ‘frame’ systems change as a first step to evaluating it;
- Support frameworks to help you understand where you are along the path to systems change, and what you need to make continual progress;
- Practical tools you can use tomorrow with collaborators or in your evaluation work; and
- Real-time support in your work of evaluating systems change.
UKES Annual Evaluation Conference: 2-5 May 2018
The theme of the 2018 United Kingdom Evaluation Society (UKES)' Annual Evaluation Conference is: The Quality of Evidence from Evaluation - demand, supply, and use. Fundamentally, evaluation should provide information that is credible and useful, enabling the incorporation of lessons learned into decision-making processes. This year’s theme focuses on quality throughout the evaluation cycle. By framing the theme around the evaluation cycle, we can consider the demand for, production and uptake of high quality evidence from evaluations.
For more information and to register please click here.
Canadian Evaluation Society 2018 Conference: 27-29 May 2018
The conference will be held in Calgary Alberta May 27-29, 2018 on the theme of "Co-Creation in Evaluation".
To keep up to date on the latest CES Conference news please visit http://c2018.evaluationcanada.ca/
Aotearoa New Zealand Evaluation Association Conference 2018: 16-19 July 2018
Venue: Waipuna Hotel & Conference Centre, 58 Waipuna Rd, Mount Wellington, Auckland 1060
Theme: "Evaluation for change: Change for evaluation – He arotake rereketanga: Kia rereke te arotaketanga"
As evaluators, our purpose is to show what’s working, for whom and why, to enable positive change for society, ecologies, organisations, communities, whānau, families and individuals. Evaluation can provide a grounded calmness and solid foundation for exploring what’s been and what might be. But pausing and reflecting, exploring uncertainties and challenges, might be seen as anathema to ‘action’ and ‘discovery’ oriented’ practitioners who might prefer to focus on new opportunities and novel ways of doing things. But the process of creating change, the outcomes and its impacts, touch everyone and sometimes in unexpected ways. Are we responsible, as Professor Viviane Robinson asserts, for “…too much change and not enough improvement”?
To learn more about our 2018 Conference and the theme please visit the ANZEA website.
13th European Evaluation Society Biennal Conference: 1-5 October 2018
Abstract submission for conference papers and pre-conference workshops are now open and will close on 15 March 2018. To submit please visit the website at http://www.ees2018.eu/abstract-submission-guidelines.htm and for the workshop program: http://www.ees2018.eu/application-form.htm
Registrations for the conference, which will be held in Thessaloniki, Greece are now open with early bird registrations closing 15 June 2018. Click here for the guidelines and to register. The conference theme is Evaluation for More Resilient Societies.
Developmental Evaluation Convening: 25-29 April 2018
Developmental evaluation convening, April 25-29, Bali Indonesia. A rare opportunity to learn from internationally leading developmental evaluators, Kate McKegg, Nan Wehipeihana and Nora Murphy. Over the course of a week you will grow your practice-based knowledge and skills, deepen your understanding of developmental evaluation and become part of a global developmental evaluation communnity. Find out more by clicking here.
Clear Horizon upcoming training courses
(20% off for AES members)
Clear Horizon is widely known for conducting evaluation research and being at the cutting edge of contemporary theory and practice with regard to monitoring and evaluation. Clear Horizon prides itself in using innovative solutions and best practice approaches. In 2006 Clear Horizon received the National Award for Evaluation Development from the Australasian Evaluation Society and now offer training in a number of our flagship techniques.
We are experienced and able trainers with strong facilitation skills. We believe in offering training that leads to real application. Our courses include a range of learning experiences including experiential learning, practical application and engaging discussions. Clear Horizon is the preferred provider for training in monitoring and evaluation for the Department of Sustainability and Environment, Victoria and also Federal Government.
Please visit the Clear Horizone website by clicking here
Evaluations that Make a Difference: Stories from around the world
This project is supported by EvalPartners, the African Development Bank and the Inter-American Development Bank
Evaluations that Make a Difference: Stories from around the world is one of the first pieces of systematic research looking at factors that contribute to high quality evaluations that are used by stakeholders to improve programs and improve people's lives. This initiative collected stories about evaluations that made a difference, not only from the perspective of the evaluators but also from the commissioners and users. The stories in this collection tell powerful stories about the findings in the evaluations and the ways the evaluations contributed to the impact of the programs.
The report can be accessed at: https://evaluationstories.wordpress.com/evaluation-story-publications/
Some outstanding news is that this research supports what many of the wiser evaluators already knew. Evaluations can make a difference if evaluations:
- Focus on making an impact
- Give voice to the voiceless
- Provide credible evidence
- Use a positive approach
- Ensure users and intended beneficiaries are actively engaged
- Embed evaluation within the programme
- Are recognized as important
- Have a champion within program
This is just the beginning. Evaluators need to take a systematic approach to collecting, analyzing and using information to learn more about whether evaluations are making a difference and what factors are the most important. Yes, we need to evaluate our own work more often. Hopefully, this will inspire others to think more critically about evaluation design and implementation and to do more research into what works.
International Organization for Collaborative Outcome Management (IOCOM)
IOCOM facilitates collaboration between professionals in outcome management and development, monitoring and evaluation associations, societies and networks to strengthen the discipline in the world.
Please visit our website.