Learning and Evaluation/News/2014

From Meta, a Wikimedia project coordination wiki
Learning and Evaluation

2014 news archive[edit]

Today, we launch our second round of voluntary reporting. We invite all program leaders and evaluators to participate in, the most epic data collection and analysis of Wikimedia programs we've done so far as a movement.

As we did in the Fall of 2013, once again we cordially request for you to share the data you have been collecting to evaluate your programs (i.e., Edit-a-thons, Editing Workshops, On-wiki Writing Contests, Photo Events, etc.). To best prepare, review the reporting items and start gathering that data you have filed away about your programs since September 1, 2013! You can ask questions directly on the reporting form preview or on our portal talk page. When you are ready to begin, you will find the reporting collector by clicking here.

On the IV Annual Meeting of Iberocoop, Wikimedia organizations from Latin America, Portugal, Spain and Italy got together to discuss how to effectively plan and define the strategic lines of work on a regional level, for the movement. Read the blogpost to find out more, comment and share!

In this new blogpost, we share the spirit behind international congresses, why they matter to the movement and how to receive economic and professional support from Wikimedia Foundation in planning and executing a cross-border gatherings. Comment and share!

Microgrants is a field yet under review by the Wikimedia movement. In this blogpost, community members tell their experience with Arabic Wikimedia community and what they learned from working together. Read and share!

Storytelling is a skill that few people have; however, when looking closely, the elements that make a good story can be identified very easily, and anyone can learn how to acquire this skill. In this series, we'll share with community members different aspects of good storytelling, and how to use reporting to bring out the stories in the movement. Learn more on the Meta page and read the blogpost!

The divide between quantitative and qualitative information and evaluation methods is theoretical and methodological. How to evaluate in a mix-methods world? Find out more of this new resource in the Evaluation portal, how it relates to Data Collection and Global Metrics. Read and share the new blogpost!

Starting this month, the Grantmaking team is putting into effect a set of Global Metrics, that will help us all understand, appreciate and be accountable for some of the work being done by Wikimedia communities worldwide. Read more in the blogpost! Comment, ask, share and learn!

How was the Grantmaking team involved in Wikimania 2014? What are some of the new challenges we face now? Find out more in the blogpost!

This category in Commons has the icons we use to build infographics. Program leaders are welcome to use these resources to make their own data visualization work.

This infographic was designed with data pulled from Evaluation Report (beta), the first round of beta reporting from Wikimedia program leaders in different parts of the world. It will be part of the Program Evaluation & Design summative poster of Year 1 reporting, presented at Wikimania 2014 in London.

On the online series Beyond Wikimetrics, Jonathan Morgan will guide participants on how to expand their wiki-research capabilities by accessing data directly through Databases and APIs. Read the blogpost to learn more about the upcoming virtual meet-ups!

Wikimedians in Residence: a journey of discovery

In this blogpost, Daria Cybulska (Wikimedia UK) shares the evaluation process that took 6 months of research on the reknown program, Wikimedians in Residence. Together with the Learning and Evaluation team, we worked to build an infographic that brings the report's data forward.

Find here different resources to make better surveys. Do you normally use surveys to evaluate programs? Share your experience with us!

Read more about what motivates a metrics discussion within the community. Share the blog and join the conversation!

We’d like to hear from the community of Wikimedia, program leaders, evaluators, volunteers, and other stakeholders in program evaluation and design, about your perspectives on the evaluation initiative efforts.

Maria Cruz has been hired as our new Program Evaluation Community Coordinator.

Eval report cover page

A summary of the highlights from the Evaluation Reportsbeta submitted last year.

An overview of tips and strategies for writing survey questions and developing a questionnaire.

The evaluation report of the Wikipedia Education Program has been released. Read and submit any comments!

Impact report released, covering 1st round of grants in 2013.

Find tracking and reporting tools: Forms you can copy and use and link for sharing your initial program reporting with our team! Request for comment and translations.

Evaluation report released. Request for comment.

Evaluation report released. Request for comment.