Every conflict resolution program coordinator has faced the budget meeting where their program is described as "valuable but hard to quantify." That description is a precursor to a funding cut. When administrators must choose between programs in a resource-constrained environment, programs with clear, compelling impact data survive; programs that rely on anecdote and goodwill do not. Measurement is not bureaucratic box-checking—it is the evidence base that protects the students who depend on these programs.
The case for measurement extends beyond program survival. Rigorous evaluation tells you what is working and what is not, enabling continuous improvement that generic goodwill never generates. A peer mediation program that measures resolution rates over time will detect if those rates are declining and investigate why; a program that does not measure will simply drift toward ineffectiveness without anyone noticing until a crisis makes the problem undeniable.
Measurement also builds credibility with skeptical stakeholders—teachers who are not sure the program is worth the class-release time, parents who wonder whether their child's conflict was handled appropriately, and school board members who see the line item and ask what they are getting for it. A well-constructed quarterly impact report answers all of these stakeholders' questions before they become objections.
The good news is that most of the data needed for a robust program evaluation already exists in your school's information systems. The task is not generating new data from scratch—it is connecting existing data streams, establishing baselines, and creating the reporting infrastructure to tell a coherent story over time.


