Highest impact - news

What Is High-Impact Tutoring? The Research, the Requirements, and What It Takes to Deliver It

March 28, 2026
What Is High-Impact Tutoring? The Research, the Requirements, and What It Takes to Deliver It

The term is everywhere. The definition matters more than most districts realize.

Ayush Agarwal
Co-Founder and CEO

What Is High-Impact Tutoring? The Research, the Requirements, and What It Takes to Deliver It

The term is everywhere. The definition matters more than most districts realize.

Every district administrator has heard it by now. High-impact tutoring. It appears in state guidance documents, federal recovery frameworks, board presentations, and vendor decks. It has become, in the span of a few years, the dominant phrase in K-12 learning recovery — the intervention every superintendent is expected to have an opinion on.

But ask ten district leaders to define it precisely, and you will get ten different answers. Some think it means tutoring that works. Some assume it means one-on-one. Some believe any small-group instruction qualifies. A few have absorbed the marketing of full-service tutoring vendors and believe it requires outsourcing entirely.

None of these are quite right. And the imprecision carries real cost: districts are spending millions of dollars on programs they call high-impact tutoring that the research would not recognize as such — and getting proportionally weaker results.

This post is a precise answer to a deceptively simple question. What is high-impact tutoring, actually? What does the research say works, how specifically does it need to be delivered, and what does it take operationally to do it at scale?

Where the Term Comes From

The phrase "high-impact tutoring" was not invented by a vendor. It was shaped by researchers — and one researcher and institution in particular did more to define and disseminate it than any other.

Susanna Loeb, Professor of Education at Stanford's Graduate School of Education and founding executive director of the National Student Support Accelerator (NSSA), is the most influential voice in establishing what high-impact tutoring is and is not. Loeb, who previously directed the Annenberg Institute for School Reform at Brown University, launched the NSSA explicitly to translate the tutoring research base into guidance that practitioners could actually use. In a 2021 working paper co-authored with Carly D. Robinson — "High-Impact Tutoring: State of the Research and Priorities for Future Learning" — the framework began to crystallize. That paper, published through the Annenberg Institute, is the intellectual foundation that most serious HIT programs now build on.

Alongside Loeb's work, Matthew Kraft at Brown University and Beth Schueler at the University of Virginia have produced the most rigorous meta-analytic work on tutoring at scale. Their 2024 analysis, co-authored with Grace Falken, synthesized 282 randomized controlled trials — the largest such review in the field — and produced findings that every district administrator planning or running a tutoring program should understand.

The broader research community — Robinson at Stanford, Schueler at UVA, Kraft at Brown — together with the NSSA and the EdResearch for Action series, have converged on a set of design principles that separate high-impact tutoring from tutoring that is merely well-intentioned.

What the Evidence Actually Shows

Before getting into the design principles, it's worth understanding the magnitude of what the research supports — and the conditions under which it applies.

Tutoring is the most effective school-based academic intervention we know of. That is not a marketing claim; it is the consensus of the literature. A 2023 meta-analysis by Nickow, Oreopoulos, and Quan covering 89 randomized controlled trials found that tutoring improves student achievement by an average of 0.29 standard deviations — the equivalent of roughly four additional months of learning for the typical elementary-age student. That outperforms class size reduction, extended school days, summer school, and virtually every other intervention in the comparison set.

But here is the finding that changes the conversation: the effects are not uniform, and they decline sharply at scale.

Kraft, Schueler, and Falken's 2024 meta-analysis documented this with uncomfortable clarity. For small programs serving fewer than 100 students, average effect sizes were around +0.44 standard deviations — impressive by any measure. For programs serving 100 to 399 students, that dropped to +0.30. For programs serving 400 to 999 students: +0.21. For programs serving more than 1,000 students: +0.16.

The researchers identified that pooled effect sizes were only a third to half as large for large-scale programs as for the full sample — and that this decline was driven by stark drops in effectiveness as program scale increased.

That is not an argument against tutoring at scale. Even a +0.16 effect size is stronger than most educational interventions. The finding is an argument about how tutoring must be designed and managed when it scales — and about what happens to quality when the operational infrastructure cannot keep up with program growth.

Critically, Kraft and colleagues identified that a bundled package of recommended design features can partially inoculate programs from these attenuated effects at scale. In other words: the research tells us exactly what needs to stay true when a program gets bigger, and warns us that most programs fail to maintain it.

Those design features are what we call the HIT model. Here is what they require.

The Six Design Principles of High-Impact Tutoring

The clearest synthesis of HIT design principles comes from EdResearch for Action Brief #30 — "Design Principles for Accelerating Student Learning With High-Impact Tutoring," authored by Robinson, Kraft, Loeb, and Schueler, last updated in June 2024. Here is what it establishes.

1. Frequency and Duration: At Least Three Sessions Per Week

High-impact tutoring works best when it is embedded in schools during the day, with consistent sessions taking place for at least 30 minutes at a time and at a minimum of three days a week. This is not a guideline — it is the threshold below which the evidence base does not hold.

After-school tutoring programs, weekend programs, and drop-in tutoring may have value, but they are not what the HIT research supports. The effect sizes in the literature come from programs with meaningful dosage. A student who attends one 45-minute session per week is not receiving high-impact tutoring. She is receiving tutoring. The distinction matters enormously when districts are making funding decisions based on expected outcomes.

2. Group Size: No More Than Three Students Per Tutor

Successful tutoring programs are characterized by small group sizes of no more than three students per tutor. The research is strongest for 1:1 and 2:1 ratios; the evidence base on 3:1 is thinner and shows meaningfully smaller effects.

Kraft and Lovison's 2024 pilot study, which compared 1:1 and 3:1 ratios in an online tutoring program, found that both experimental estimates and tutor survey responses provided suggestive evidence that 1:1 tutoring is more effective than 3:1 tutoring in an online setting, with small-group online tutoring presenting additional challenges for personalizing instruction, developing relationships, fostering participation, and managing student behavior.

Many programs rationalize larger groups on cost grounds. The research is clear that this tradeoff directly reduces impact.

3. During the School Day

One of the most consistent findings in the HIT literature is that tutoring delivered during the school day outperforms tutoring delivered after school. The mechanisms are intuitive: students are more likely to attend, less likely to be fatigued, and can connect the tutoring content more directly to classroom instruction.

As Loeb has written, in-class high-impact tutoring is the most accessible and effective option — not because virtual or after-school models can't work, but because school-day delivery removes the most common barriers to consistent attendance. And consistent attendance, as we will discuss, is foundational.

4. Consistent Tutor-Student Relationships

Strong tutor-student relationships are a cornerstone of effective high-impact tutoring. The research is unambiguous on this: programs that rotate tutors, assign students to whoever is available, or treat the tutor as interchangeable produce weaker results. The learning relationship is not incidental to the intervention — it is the intervention.

This is why Loeb and the NSSA emphasize relationship-based tutoring so consistently. The academic gains come partly from the instruction itself and partly from the motivational and trust effects of a stable, personalized relationship with an adult who is there specifically for that student.

5. Curriculum Alignment and High-Quality Materials

High-quality materials aligned with the curriculum are a defining feature of effective programs. Tutoring that operates in isolation from what students are learning in their classrooms misses the reinforcement effect that makes dosage translate into retention. Tutors working with aligned materials are not doing remediation in a vacuum — they are extending and deepening the classroom experience.

6. Data-Driven Practice and Tutor Support

Programs are more effective when tutors receive proper training and ongoing support, and when data-informed practices are used. This encompasses everything from initial tutor training to ongoing coaching, progress monitoring, and the use of session-level data to adapt instruction.

This is the design principle most frequently honored in the breach. Districts often invest in curriculum and tutor recruitment, but treat data infrastructure as an afterthought. The result is programs that cannot tell you, in real time, whether they are working — and for which students.

The Principle No One Talks About: Attendance

Embedded across all six design principles is a dependency that the research is increasingly explicit about: none of it works if students don't show up consistently.

High-impact tutoring derives its power from cumulative exposure — the same student, with the same tutor, building knowledge and relationship over time. A student who misses a third of her sessions is not receiving a reduced-dosage HIT program. She is receiving something categorically different.

Loeb has been direct on this point. In a discussion of what distinguished programs that achieved strong outcomes from those that underperformed, attendance emerged as the variable that explained more variation than almost any design feature. A program can have the right tutor-student ratio, the right curriculum, and the right scheduling — and still fail if attendance is not actively managed.

This is not primarily a family engagement problem. It is an operations problem. Students miss sessions because of scheduling conflicts that were never caught. Because a tutor was absent and no substitute was arranged. Because a room change was never communicated. Because no one was tracking absences and flagging them before they became patterns.

The Scale Problem — and Why It's Fundamentally Operational

Return to the Kraft-Schueler finding: effects decline sharply as programs scale. The researchers explored four hypotheses for why this happens, and the most compelling involves program implementation quality.

As Loeb has observed, "the desire to increase scale comes with a push to decrease quality." That problem quickly cascades into other issues: too many students per tutor, too few sessions per week, and insufficient coaching for tutors.

This is the pattern. A program is designed well and delivers strong results at pilot scale. A district sees the results and expands. The expansion requires more tutors, more schedules, more coordination, more reporting. The tools that worked for 80 students — spreadsheets, email chains, shared folders — buckle under the weight of 800. Coordinators spend more time on logistics and less time on quality. Group sizes creep up. Session frequency drifts down. Tutor-student consistency erodes as scheduling errors multiply. The design principles that defined the program survive on paper and in presentations, but not in practice.

Research examining school-level barriers to scaling high-impact tutoring identified major obstacles including time and space constraints, tutor supply and quality challenges, the need for updated data systems, and school-level costs — while teacher buy-in emerged as the key facilitating factor.

Notice what is in that list: data systems. Operational infrastructure. The research on HIT is not only about what works — it is increasingly about what breaks when you try to grow what works.

This is why the question of tutoring management is not a software question. It is a fidelity question. The design principles of high-impact tutoring are only as good as the operational infrastructure that enforces them at scale.

What HIT Fidelity Requires Operationally

Consider what it takes to maintain fidelity to the six design principles across a 500-student program spanning multiple schools and multiple providers:

Session frequency requires scheduling that accounts for each student's school calendar, class schedule, and provider availability — and that flags when a student drops below the three-sessions-per-week threshold before it becomes a pattern.

Group size requires roster management that prevents a tutor from being assigned more than three students and that automatically adjusts when a student transfers or a tutor changes availability.

School-day delivery requires integration with school schedules and space management — the kind of coordination that falls apart immediately when it's managed across email and spreadsheet tabs.

Consistent relationships requires matching logic that persists across weeks and months, and alerts when a tutor-student pairing is disrupted so it can be re-established quickly.

Curriculum alignment requires data sharing between providers and district instructional staff — which requires standardized data collection in the first place.

Data-driven practice requires session-level attendance and progress data that is current, accessible, and synthesized across providers — not reconstructed from disparate files at the end of a semester.

None of this is possible at scale with manual coordination. It requires infrastructure purpose-built for the operational complexity of tutoring programs.

The Honest Summary

High-impact tutoring is the most evidence-supported academic intervention available to K-12 districts. Susanna Loeb, Matthew Kraft, Carly Robinson, Beth Schueler, and the broader research community have documented this with unusual rigor and consistency. The design principles are well-established and non-negotiable: frequent sessions, small groups, school-day delivery, consistent relationships, aligned curriculum, and data-informed practice.

The challenge is not knowing what works. The challenge is maintaining fidelity to what works as programs grow.

That is where most districts are failing — not from lack of intention, but from lack of infrastructure. The spreadsheet that managed the pilot cannot manage the program. The coordinator who knew every tutor by name cannot hold the complexity of 600 sessions per week in her head. The reporting that satisfied the board in year one cannot satisfy a state auditor in year three.

High-impact tutoring is a research claim and an operational commitment. The research defines the model. The operations determine whether the model survives contact with reality.

Sierra TMS is built specifically for the operational demands of high-impact tutoring — scheduling, attendance, rostering, reporting, and provider management at scale. to see how districts are using infrastructure to protect program fidelity as they grow.

Further Reading

Robinson, C.D. & Loeb, S. (2021). High-Impact Tutoring: State of the Research and Priorities for Future Learning. Annenberg Institute at Brown University. EdWorkingPaper 21-384.

Robinson, C.D., Kraft, M.A., Loeb, S. & Schueler, B. (2024). Design Principles for Accelerating Student Learning with High-Impact Tutoring. EdResearch for Action, Brief #30.

Kraft, M.A., Schueler, B.E. & Falken, G. (2024). What Impacts Should We Expect from Tutoring at Scale? Exploring Meta-Analytic Generalizability. Annenberg Institute at Brown University. EdWorkingPaper 24-1031.

National Student Support Accelerator: studentsupportaccelerator.org

Tags: high-impact tutoring, HIT research, K-12 tutoring, tutoring program design, Susanna Loeb, NSSA, tutoring at scale, tutoring management Meta description: High-impact tutoring is the most evidence-supported K-12 intervention we have — but only when delivered with fidelity to specific design principles. Here's what the research actually says, and what it takes operationally to deliver it at scale. Target keyword: high-impact tutoring Secondary keywords: high-impact tutoring definition, high-impact tutoring research, HIT tutoring K-12, tutoring program design principles, tutoring at scale Recommended internal links: What Is a Tutoring Management System? | The Hidden Costs of Running a Tutoring Program on Spreadsheets | How to Prove ROI on Your Tutoring Investment Author note: Citations are from primary research sources. All claims attributed to named researchers reflect published or publicly available work. Update annually as NSSA and EdResearch for Action release new briefs.