School administrator reviewing data on a tablet, representing program evaluation
← Back to Blog
🏫 K–12 Schools

How to Measure the Impact of Conflict Resolution Programs in K-12 Schools

April 14, 2025·10 min readprogram evaluationschool dataconflict resolution

Why Measurement Is Not Optional: The Stakes for Program Survival

Counselor reviewing program data on a laptop in a school office

Data-driven program evaluation is the foundation of long-term program sustainability.

Every conflict resolution program coordinator has faced the budget meeting where their program is described as "valuable but hard to quantify." That description is a precursor to a funding cut. When administrators must choose between programs in a resource-constrained environment, programs with clear, compelling impact data survive; programs that rely on anecdote and goodwill do not. Measurement is not bureaucratic box-checking—it is the evidence base that protects the students who depend on these programs.

The case for measurement extends beyond program survival. Rigorous evaluation tells you what is working and what is not, enabling continuous improvement that generic goodwill never generates. A peer mediation program that measures resolution rates over time will detect if those rates are declining and investigate why; a program that does not measure will simply drift toward ineffectiveness without anyone noticing until a crisis makes the problem undeniable.

Measurement also builds credibility with skeptical stakeholders—teachers who are not sure the program is worth the class-release time, parents who wonder whether their child's conflict was handled appropriately, and school board members who see the line item and ask what they are getting for it. A well-constructed quarterly impact report answers all of these stakeholders' questions before they become objections.

The good news is that most of the data needed for a robust program evaluation already exists in your school's information systems. The task is not generating new data from scratch—it is connecting existing data streams, establishing baselines, and creating the reporting infrastructure to tell a coherent story over time.

Disciplinary Incident Metrics: The Core Quantitative Story

Disciplinary incident data is the most straightforward quantitative indicator of conflict resolution program impact. Track the following at minimum: total office referrals per month, referrals specifically categorized as interpersonal conflict or altercations, in-school suspension assignments, out-of-school suspension assignments, and expulsions. Disaggregate all of these by grade level, race, gender, and disability status to detect whether the program is serving all students equitably or concentrating benefits in particular groups.

Establish a pre-program baseline using at least one full academic year of historical data before the program launched (or before the current reporting period). Without a baseline, you can report absolute numbers but not change—and change is what demonstrates impact. If your program launched mid-year, use that year's second-semester data as the baseline for the following year's comparison.

Be precise about what the data does and does not show. A decline in office referrals could indicate that conflicts are being resolved earlier (the hoped-for outcome), but it could also indicate underreporting by teachers, changes in administrative culture, or demographic shifts in the student population. Triangulating disciplinary data with other metrics—climate survey results, attendance data, mediator case logs—strengthens the causal story. For a specific framework focused on disciplinary incidents, see our guide on reducing disciplinary incidents in schools.

Suspension Rates: The Metric Administrators and School Boards Watch Most Closely

Suspension rates carry particular weight in the current policy environment. Federal guidance, state accountability frameworks, and community advocacy have made suspension disparities—particularly racial disparities—one of the most scrutinized metrics in public education. A conflict resolution program that demonstrably reduces suspension rates, especially for students of color who are disproportionately suspended for subjective offenses, is making a compelling case to every stakeholder simultaneously.

Track suspension rates as a percentage of enrollment (not raw counts) to enable year-over-year comparisons that account for enrollment changes. Track both overall rates and rates disaggregated by race, disability status, grade level, and referral reason. If your program includes specific components designed to reduce disparate discipline—such as restorative alternatives to suspension—track outcomes for those specific students separately to demonstrate the program mechanism, not just the aggregate effect.

Consider the distinction between suspension rate reduction and suspension avoidance. If a student who would previously have been suspended receives a restorative consequence that is more labor-intensive but produces better outcomes, the suspension rate declines but the school's commitment to accountability is maintained. Communicating this distinction clearly to board members and community members—who may interpret declining suspensions as lowered standards—is an important narrative management task for program coordinators.

Multi-year suspension rate tracking is particularly powerful. A program that shows a consistent year-over-year downward trend over three to five years, with corresponding improvements in attendance and academic performance, is making a case that is very difficult to dismiss. Building this longitudinal record starts with the decision to measure consistently, beginning now.

School Climate Surveys: The Perceptual Evidence

Students completing school climate surveys on tablets

Validated climate surveys capture the perceptual dimension of school safety that disciplinary data misses.

Disciplinary data captures behavior. Climate surveys capture perception—the felt experience of school as a safe, supportive place. Both dimensions matter, and they can diverge in ways that are instructive. A school can have declining disciplinary incidents while students still feel unsafe, or improving climate survey scores while administrative data shows no change. Tracking both provides a richer and more honest picture of program impact.

Validated school climate survey instruments include the Authoritative School Climate Survey (ASCS), the Comprehensive School Climate Inventory (CSCI), and the Conditions for Learning survey used in many state accountability frameworks. Using a validated instrument rather than a custom survey ensures that your data is comparable to national benchmarks and is more credible to external stakeholders. Most validated instruments can be administered in 15–20 minutes and generate subscale scores across dimensions including safety, respect, connectedness, and adult support.

Survey timing and frequency matter. Administering at the same point each year (e.g., March) controls for seasonal variation and produces clean year-over-year comparisons. Administering twice per year—fall baseline and spring follow-up—allows you to detect within-year change and attribute it more confidently to program activities that took place in between. Share survey results with staff and students: transparency about what the data shows builds trust and engagement with the improvement process.

Attendance as an Indirect Conflict Resolution Metric

Chronic absenteeism—defined as missing 10% or more of school days—is strongly correlated with school safety and conflict concerns. Students who feel unsafe, socially excluded, or persistently in conflict with peers or teachers avoid school. Conversely, students who experience school as a safe, connected environment attend more consistently. Tracking attendance data alongside conflict resolution metrics provides a powerful indirect indicator of program impact.

The specific attendance metrics most relevant to conflict resolution program evaluation include: chronic absenteeism rates overall and disaggregated by subgroup, attendance patterns among students who have participated in formal conflict resolution or peer mediation, and absenteeism rates in the days and weeks following documented conflict incidents (a spike in post-conflict absenteeism indicates unresolved fear or shame).

Attendance data is also useful for equity analysis. If your conflict resolution program is reducing absenteeism for some student groups but not others, that differential effect points toward either a differential reach problem (certain students are not being connected to program services) or a differential relevance problem (the program's approach resonates more with some cultural communities than others). Either finding is actionable and important.

Present attendance data in your impact reports alongside the disciplinary metrics rather than in a separate section. The story of a student who once missed 20 days due to peer conflict and who attended consistently after participating in peer mediation is more compelling than any aggregate statistic—and real cases like this can be shared (with appropriate privacy protections) to humanize quantitative findings for board members and community audiences.

Data Collection Tools: Building Infrastructure Without Overwhelm

The data collection infrastructure for program evaluation need not be elaborate to be effective. The key requirements are consistency (the same data collected in the same way at the same intervals), accessibility (data that can be retrieved and analyzed without heroic effort), and integration (data streams that can be connected to tell a coherent story). These requirements can be met with a combination of your existing student information system, a simple case tracking spreadsheet, and a validated climate survey instrument.

Mediator case logs are a program-specific data collection tool that most peer mediation and conflict resolution programs underutilize. A structured case log captures: date, referred-by, nature of dispute (using standardized categories), number of sessions, outcome (resolved/partially resolved/referred to adult), and brief mediator notes. Aggregating case logs monthly produces a rich dataset of program activity that complements administrative records and captures cases that never rose to the level of a formal office referral.

Post-session participant surveys are another high-value, low-burden data collection tool. A five-question survey administered immediately after a mediation session (Was the process fair? Did you feel heard? Are you more optimistic about this situation? Would you use this process again? Comments?) generates real-time feedback that improves practice and provides testimonial data for impact reporting. Keep surveys brief enough to be completed in three minutes—longer instruments depress completion rates dramatically.

Integrating Data Streams Into a Unified Impact Picture

A quarterly impact report that integrates all data streams—disciplinary incidents, suspension rates, climate survey scores, attendance trends, case log summaries, and participant satisfaction—is more convincing than any single metric presented in isolation. The integration tells a story: this program is reaching students, changing behavior, and creating a school environment where more students feel safe and connected.

Build a simple dashboard template (a one-page or two-page document with consistent sections) that you populate quarterly. The consistency of format matters as much as the content: administrators and board members who see the same report structure every quarter develop familiarity with the data and can detect trends themselves, reducing your explanatory burden over time. Consistency also signals that the program operates with rigor and professionalism.

Reporting to the School Board: Translating Data Into Decisions

School board members are not, in general, educational researchers. They are community representatives making policy and budget decisions for their district, and they need information presented in a format that enables decision-making rather than demonstrating methodological sophistication. Effective board reporting prioritizes clarity, relevance, and narrative over technical precision.

Lead with the headline metric: the finding most likely to matter to a board member and their constituents. "Disciplinary referrals fell 28% in the two years since we launched peer mediation, and our suspension rate for Black and Hispanic students—which had been twice the district average—has narrowed to parity" is a board-ready headline. The technical details of how that finding was derived belong in an appendix, available for board members who want to dig deeper but not required for decision-making.

Cost-per-outcome comparisons are powerful for budget audiences. If the annual cost of your conflict resolution program is $40,000 and you have eliminated 60 out-of-school suspensions that each cost the district approximately $800 in lost state funding and substitute teacher costs, you have generated $48,000 in savings against a $40,000 investment—a net positive return before counting any of the social or academic benefits. These calculations require assumptions, but conservative, well-documented assumptions are persuasive to budget-conscious administrators.

Always pair quantitative findings with qualitative illustrations. A 30-second student testimonial video, a brief anonymized case narrative, or a teacher quote about observed change in classroom climate gives board members something to feel as well as understand. Decisions are rarely made on data alone; the combination of rigorous evidence and human narrative is the most persuasive possible combination.

Longitudinal Tracking: Why the First Year's Data Is Just the Beginning

A school leader presenting data trends on a screen to faculty

Longitudinal tracking transforms a year's results into a compelling multi-year story of program impact.

Single-year program evaluation data answers the question "What happened this year?" Longitudinal data—three, five, or ten years of consistent measurement—answers the question "Is this program working over time and worth continued investment?" The latter question is what determines program survival, and it can only be answered by programs that have been measuring consistently for long enough to show a trend.

Design your evaluation system from the outset to support longitudinal tracking. This means using the same metrics, the same instruments, and the same administration protocols year after year, even when temptation arises to switch to a better survey or a more interesting metric. Consistency sacrifices some sensitivity to current best practices in exchange for the ability to show long-term trends—a trade that is almost always worth making.

Cohort tracking is a particularly powerful longitudinal method: following a cohort of students who participated in conflict resolution services as sixth graders and tracking their disciplinary, attendance, and academic trajectories through eighth and eventually twelfth grade. This approach can demonstrate not just immediate behavioral effects but longer-term developmental outcomes, including graduation rates and post-secondary enrollment—outcomes that resonate deeply with both educational leaders and community stakeholders.

Connect your longitudinal data to state and national benchmarks wherever possible. If your district's suspension rate is declining while statewide rates are flat or rising, that comparison is evidence that your program is driving real change rather than simply benefiting from demographic or economic trends. Benchmarked data is also more credible to external audiences, including grant funders and policy researchers who may amplify your results.

Common Measurement Mistakes and How to Avoid Them

The most common measurement mistake in school conflict resolution evaluation is measuring activity instead of outcomes. Reporting that the peer mediation program conducted 150 sessions this year describes what the program did, not what it achieved. Disciplinary incident rates, climate survey scores, and attendance data describe what changed in the school as a result of what the program did. Activity metrics are easy to collect and useful for monitoring program reach; outcome metrics are what justify program investment.

A second common mistake is attribution overreach—claiming that the conflict resolution program caused every positive change in the school's disciplinary or climate data. Responsible evaluation acknowledges confounding factors: a new principal with a different discipline philosophy, significant demographic shifts in the student population, a particularly skilled or ineffective cohort of teachers. These factors do not negate your program's contribution, but acknowledging them demonstrates intellectual honesty that builds credibility with sophisticated audiences.

Third, many programs make the mistake of discontinuing measurement during program stress—budget crunches, staff turnover, or high-conflict school years when documentation feels like an unaffordable luxury. These are precisely the moments when measurement matters most, because they provide the evidence needed to advocate for resources and justify the program's value to skeptical administrators. Build data collection into program protocols so that it happens automatically, not as a discretionary addition to already-stretched bandwidth.

Finally, avoid the trap of measuring only what you can easily count. The most important outcomes of conflict resolution programs—students' sense of safety, their capacity for empathy, their confidence in navigating difficult relationships—are harder to quantify than suspension rates. Climate surveys, focus groups, and qualitative case documentation capture these dimensions. A rigorous program evaluation uses both quantitative and qualitative evidence, recognizing that numbers tell the magnitude of change while stories tell its meaning. For tools to support this balanced evaluation approach, explore WeUnite for Schools.

📺 Watch & Learn

Video: measuring school conflict resolution program impact data collection

Deepen your understanding with this curated video on the topic.

▶ Watch on YouTube

More From the Blog

10 Examples of Inclusive Language
🏢 Enterprise

10 Examples of Inclusive Language

Explore 10 powerful examples of inclusive language for workplaces, schools, and families. Learn before/after phrasing to foster respect and understanding.

May 3, 2026 · 20 min read