A teacher reviewing progress data while a therapy dog sits with a reading student

Reading Program Success Metrics

12 min read
Share:

Beyond fluency scores and word counts lies a more complete picture of therapy dog reading program impact. Here's how we measure success across every dimension that matters.

# Reading Program Success Metrics

The grant review committee wanted numbers. "We need data," the program officer explained. "Fluency improvements, reading level changes, standardized assessment results. Without quantitative outcomes, we can't justify funding." This request鈥攔easonable from a funder's perspective鈥攈ighlighted a fundamental tension in therapy dog reading programs: the outcomes that matter most aren't always the outcomes that measure most easily.

Yes, we track fluency scores. Our participants show average improvements of 15% over six-month participation periods, compared to 8% for matched non-participants. We track reading levels, comprehension assessments, and every quantifiable metric that funders and school administrators want to see.

But if we only tracked these metrics, we'd miss the most important outcomes: the child who now reads voluntarily at home, the student who no longer cries before reading time, the teenager who's joined a book club after years of reading avoidance. These transformations matter more than fluency numbers鈥攜et they resist the quantification that accountability systems demand.

This guide presents our comprehensive approach to measuring therapy dog reading program success. We've developed frameworks that satisfy stakeholder needs for quantifiable outcomes while capturing the qualitative transformations that define genuine success. Programs can adapt these frameworks to their own contexts, creating measurement systems that serve accountability without sacrificing meaning.

The Outcome Domains Framework

Rather than treating reading improvement as a single outcome, we assess four interconnected domains that together represent comprehensive success:

**Technical Reading Skill** includes the traditional metrics: decoding accuracy, reading fluency (words correct per minute), and comprehension. These objective measures provide the quantifiable data that external stakeholders require and allow comparison with non-program populations. They represent necessary but not sufficient indicators of success.

**Reading Behavior** tracks what children actually do with reading outside of sessions. Voluntary reading time, library visits, book ownership, and reading choices in unstructured moments indicate whether improved skills translate into reading habits. A child who reads better during assessment but never reads independently has achieved less than outcome scores suggest.

**Reading Relationship** assesses how children feel about reading鈥攖heir confidence, anxiety, enjoyment, and self-perception as readers. These emotional and identity dimensions often determine whether skill gains persist and whether children become lifelong readers or abandon reading as soon as external pressure lifts.

**Reading Engagement** measures participation in reading-related activities: classroom engagement during reading instruction, willingness to read aloud, discussion of books, and social integration around reading. These social-academic indicators reveal whether children are becoming active participants in reading communities or remaining isolated from reading culture.

Each domain contributes essential information; none alone tells the full story. A child might show skill improvement without behavior change, or behavior change without skill improvement. Comprehensive assessment across all domains reveals which children are experiencing genuine transformation versus partial progress.

Measuring Technical Reading Skill

Technical skill assessment uses standardized tools that allow comparison across children and time points. We administer these assessments at program entry, midpoint (typically three months), and exit (typically six months), creating trajectory data rather than just before/after snapshots.

**Oral Reading Fluency (ORF)** measures words correct per minute during grade-level passage reading. This widely-used metric provides quick assessment that's well-established in educational research. We use grade-level passages matched to each child's current instructional level rather than a single passage for all children, ensuring that measurements reflect actual reading rather than passage-specific factors.

Captain's handlers track ORF for every child in his program. The data reveals clear patterns: most children show modest improvement (5-10 words per minute) during the first eight weeks, followed by accelerated improvement (10-20 words per minute) as comfort and confidence increase. This pattern鈥攕low initial progress followed by faster growth鈥攃haracterizes successful therapy dog intervention.

**Reading Accuracy** measures the percentage of words read correctly, independent of speed. Some children improve accuracy before fluency; others improve fluency before accuracy. Tracking both reveals different developmental patterns that inform session planning.

**Reading Comprehension** assessment uses standardized instruments appropriate for each child's level. We use both passage-based comprehension questions and retelling assessments, capturing different aspects of understanding. Some children comprehend better than they can articulate; retelling reveals this pattern while question-answering misses it.

**Assessment Cautions**: We never assess during therapy dog sessions, which would contaminate the low-stress environment that makes our intervention effective. Assessments occur separately, typically administered by classroom teachers or reading specialists, minimizing assessment anxiety while maintaining measurement validity.

Measuring Reading Behavior

Behavior data comes primarily from parent and teacher reports, supplemented by child self-report and handler observations. We've developed simple tracking tools that capture meaningful behavioral changes without creating excessive documentation burden.

**Parent Weekly Logs** ask families to note reading-related behaviors observed during the week: independent reading (duration, not page count), library visits, book requests, reading discussions, and reading in preference to other activities. The log uses checkboxes for quick completion with optional notes for context. Parents complete logs throughout program participation, creating behavioral trajectory data.

Maya's parents completed weekly logs throughout her six months with Rosie, our Cocker Spaniel. The data told a clear story: Week 1 showed zero independent reading behaviors. Week 8 showed first independent reading (five minutes before bed). Week 16 showed daily reading initiation. Week 24 showed library card acquisition and first independent library trip. This behavioral progression鈥攙isible in simple checkbox data鈥攄ocumented transformation that skill scores alone couldn't capture.

**Teacher Behavior Checklists** track classroom reading engagement monthly: participation in reading instruction, independent reading during unstructured time, engagement with classroom library, and reading-related social behaviors. Teachers complete these brief checklists (under five minutes) using familiar behavioral indicators.

**Child Reading Diaries** for older children (typically third grade and up) provide self-reported reading experiences. Children note what they read, where, and how it made them feel. These diaries aren't meant for accurate data collection鈥攃hildren aren't reliable reporters of their own behavior鈥攂ut reveal self-perception changes that adult reports miss.

**Handler Session Notes** document in-session behaviors that predict out-of-session change: book selection enthusiasm, session anticipation, reading persistence through difficulty, and engagement patterns over time. Handlers complete brief structured notes after each session, creating rich longitudinal data.

Measuring Reading Relationship

Assessing how children feel about reading requires tools designed for emotional and identity dimensions鈥攁reas where traditional reading assessments don't apply.

**Pre/Post Reading Attitude Survey** uses validated instruments adapted for our population. Children rate agreement with statements like "I enjoy reading," "Reading is hard for me," "I am a good reader," and "I feel scared when I have to read out loud." Young children respond using pictorial scales; older children use numeric ratings. Survey administration at entry and exit provides change scores for each attitude dimension.

**Reading Self-Portrait** is a projective technique where children draw themselves reading and then describe their drawings. Analysis examines emotional content (expressions, context, body language), reader identity elements (books prominent vs. minimized, reading as central vs. peripheral), and narrative themes in descriptions. This qualitative tool captures self-perceptions that survey instruments miss.

Tucker's handler maintains a gallery of reading self-portraits from children in his program. The progression is often striking: early drawings showing frowning figures, crumpled papers, books pushed away; later drawings showing smiling readers, comfortable settings, books held close. These visual transformations document internal change that children might not articulate directly.

**Reading Confidence Interview** uses brief, structured conversations to assess children's beliefs about themselves as readers. "Tell me about yourself as a reader. What's easy for you? What's hard? What do you wish was different?" Handlers conduct these interviews at entry, midpoint, and exit, recording and coding responses for analysis.

**Anxiety Indicators Tracking** documents visible signs of reading-related anxiety during sessions: fidgeting, avoidance behaviors, physical tension, tears, and negative self-talk. Handlers rate anxiety level at each session using a simple scale, creating trajectory data that reveals anxiety reduction over time.

Measuring Reading Engagement

Social and academic engagement around reading indicates whether children are becoming reading community members鈥攑articipants in the shared culture of literacy that supports lifelong reading.

**Classroom Engagement Measures** track participation during reading instruction: responding to questions, volunteering to read aloud, contributing to discussions, and engaging with reading-related activities. Teachers rate these behaviors monthly, providing trajectory data that reveals engagement changes over time.

**Social Reading Indicators** assess peer-related reading behaviors: recommending books to friends, discussing books socially, participating in book-related activities (clubs, challenges, shared reading). Parent and teacher reports capture these behaviors that indicate reading's integration into children's social lives.

**Library Usage Data** provides behavioral evidence of reading engagement. We track library card acquisition, visit frequency, and borrowing patterns for program participants. Increases in library usage suggest reading integration into children's lives beyond program sessions.

Luna, our Border Collie mix, works with children from a school that shares library data with permission. Usage statistics showed that program participants increased library visits by 47% compared to pre-program baselines, while non-participants showed only 12% increase over the same period. This behavioral data corroborated parent reports of increased reading engagement.

**Handler Engagement Ratings** document in-session engagement: enthusiasm about book selection, spontaneous sharing of reading experiences outside sessions, and investment in reading progression. These ratings, completed after each session, track engagement trajectories that predict sustained reading behavior.

Analyzing and Interpreting Outcome Data

Data collection serves accountability only if data is analyzed meaningfully and communicated effectively. Our analysis approach emphasizes patterns, trajectories, and contextualized interpretation.

**Individual Trajectory Analysis** examines each child's progress across all four domains. Some children show rapid initial progress that plateaus; others show slow initial progress that accelerates. Some improve skills before behaviors; others change behaviors before skills. Individual patterns inform session planning and help families understand their children's unique developmental paths.

**Cohort Analysis** compares outcomes across groups of children with similar starting profiles, revealing which children benefit most from intervention and which might need additional support. We've identified that children with high anxiety and moderate skills often show the largest gains; children with low anxiety and low skills often show smaller gains (though still positive). These patterns suggest that anxiety reduction is a primary mechanism of our intervention's effectiveness.

**Comparison Group Analysis** (when ethical and possible) compares program participants to similar children who didn't participate. These comparisons demonstrate program impact beyond what normal development would produce and satisfy funders' needs for evidence that investment produces outcomes.

**Qualitative Pattern Analysis** examines narrative data鈥攑arent notes, handler observations, child interviews鈥攆or themes that numbers miss. These themes often explain quantitative patterns: why certain children progress faster, what factors predict sustained engagement, how families experience changes at home. Qualitative data provides the context that makes quantitative data meaningful.

Reporting Outcomes to Different Stakeholders

Different audiences need different information presented in different ways. Effective reporting adapts to stakeholder needs while maintaining accuracy and honest acknowledgment of limitations.

**Reports to Funders and Administrators** emphasize quantitative outcomes with appropriate statistical analysis. We present effect sizes, confidence intervals, and comparison group analyses when available. We contextualize numbers with representative case examples and acknowledge what data shows and doesn't show. Honest reporting builds credibility that survives scrutiny.

**Reports to Parents** translate metrics into everyday language. "Maya's reading fluency improved from 52 to 67 words per minute" becomes "Maya now reads more smoothly and quickly, making fewer stops and starts." We emphasize progress in terms parents can observe at home and provide suggestions for supporting continued growth.

**Reports to Teachers** emphasize classroom-relevant outcomes: participation changes, engagement during reading instruction, and social reading behaviors. We share specific strategies that have worked during therapy sessions that teachers might adapt for classroom use.

**Reports to Children** focus on growth and effort rather than comparison or judgment. "Look at this鈥攜ou read 15 more words per minute than you did in September! That's amazing growth from all your practice with Biscuit." We celebrate specific achievements and set new goals collaboratively.

Honest Acknowledgment of Limitations

No assessment system is perfect, and honest acknowledgment of limitations builds credibility while preventing overinterpretation of data.

**Measurement limitations** include the inherent subjectivity of behavioral and emotional assessment, the constraints of parent and teacher report accuracy, and the limited precision of standardized reading tests. We don't claim our data captures reality perfectly鈥攚e claim it captures meaningful signals amid inevitable noise.

**Attribution limitations** complicate interpretation. Children participating in therapy dog programs often receive other interventions simultaneously; isolating program-specific effects requires research designs most programs can't implement. We're careful about causation claims, noting that outcomes occurred during participation without necessarily claiming participation caused outcomes.

**Generalization limitations** restrict how broadly findings apply. Our participants aren't random samples; they're children whose families sought out our program, introducing selection bias. What works for children who self-select may not work identically for children who don't.

**Long-term limitations** acknowledge that we don't fully understand outcome durability. We track six-month outcomes comprehensively, with limited follow-up at three and six months post-program. Whether gains persist for years remains uncertain.

The Bottom Line: What Success Looks Like

After all the metrics, frameworks, and analysis, what does genuine success actually look like? It looks like Marcus, who hid his reading workbook under his mattress, now asking his parents to buy him books. It looks like Jordan, whose fluency scores plateaued but who joined a book club and now talks about books with friends. It looks like Alex, who spoke his first words to a therapy dog and now reads aloud to his class.

Success looks like children who don't just read better, but read more. Who don't just score higher, but feel differently about reading. Who don't just participate in our program, but become participating members of reading communities that will support them for life.

The metrics we track help us demonstrate this success to stakeholders who need evidence. But the metrics serve success鈥攖hey don't define it. The definition lives in the children whose relationships with reading transformed because a patient dog listened without judgment, because skilled handlers created safe space for risk-taking, because our program gave them something they needed that traditional intervention hadn't provided.

When we report our numbers to funders, we include the stories. When we analyze our data, we remember the children behind each data point. The numbers matter鈥攖hey prove our work deserves support鈥攂ut the children matter more.

Charts showing reading improvement metrics at a therapy dog program
馃惥

Paws & Pages Team

The Paws & Pages team is dedicated to building confident readers through the unconditional love of therapy dogs. Our team of educators, trainers, and volunteers share tips, stories, and resources to support literacy and the human-animal bond.

View all posts

You Might Also Enjoy

Diverse children reading together with therapy dogs in an accessible space

Creating Inclusive Reading Sessions

Every child deserves access to the magic of reading with therapy dogs. Here's how Paws & Pages designs sessions that welcome learners of all abilities, backgrounds, and needs.

15 min read
Children of various ages reading to a therapy dog in a library setting

The Best Age to Start Reading Programs

Parents often ask when their child should start reading with therapy dogs. The answer reveals something important about how children learn鈥攁nd why there's no such thing as 'too early' or 'too late.'

12 min read
A child selecting a book while a therapy dog waits attentively

How to Choose the Right Book Level

The Goldilocks zone of reading difficulty鈥攃hallenging enough to promote growth but easy enough to allow success鈥攊s crucial for therapy dog sessions. Here's how to find it.

14 min read