Visible Learning: What Hattie's Research Means for TeachersClassroom activity focused on visible learning: hattie's research on what actually works in teaching with primary

Updated on  

April 28, 2026

Visible Learning: What Hattie's Research Means for Teachers

|

October 26, 2021

Understand Hattie's effect sizes and which teaching strategies actually work. A practical guide to making learning visible in your classroom.

Build your next lesson freeExplore the toolkit
Copy citation

Main, P (2021, October 26). Visible Learning: A teacher's guide. Retrieved from https://www.structural-learning.com/post/visible-learning-a-teachers-guide

What is Visible Learning?

Hattie's (2008) Visible Learning ranks 252 learner achievement influences. Effect size shows impact; 0.40 is significant (Hattie, 2008). Teachers should use evidence to assess their impact (Hattie, 2008).

Hattie (2009) showed learning improves when teachers and learners know what success looks like. Teachers demonstrate their thinking using questions, model answers, and self-assessment. Learners notice their progress, which boosts motivation and thinking (Hattie, 2009).

Visible Learning is an evidence-based approach to teaching developed by education researcher John Hattie. At its core, the idea is simple: learning should be visible, to teachers and to students themselves. This means students must know what they are learning, how to go about learning it, and how to measure their progress along the way. Hattie's work shifts the focus from simply delivering content to evaluating the impact of teaching on student achievement.

Evidence Overview

Chalkface Translator: research evidence in plain teacher language

Academic
Chalkface

Evidence Rating: Load-Bearing Pillars

Emerging (d<0.2)
Promising (d 0.2-0.5)
Robust (d 0.5+)
Foundational (d 0.8+)

Key Takeaways

  1. Visible Learning quantifies teaching impact through effect sizes: Hattie's extensive meta-analyses identify high-impact strategies, with a 0.40 effect size serving as a benchmark for meaningful gains in learner achievement (Hattie, 2009). This evidence-based approach equips teachers to prioritise interventions proven to accelerate learner progress effectively.
  2. Teachers are the most significant influence on learner achievement: Hattie's research consistently highlights the profound impact of teacher expertise and pedagogical choices on learning outcomes (Hattie, 2012). Effective Visible Learning implementation requires teachers to understand their impact and adapt their practice based on clear evidence of learner learning.
  3. Clear learning intentions and success criteria are fundamental to visible learning: For learners to understand what they are learning and how to achieve it, teachers must explicitly articulate learning intentions and corresponding success criteria (Hattie & Yates, 2014). This clarity enables learners to self-regulate their learning and provides teachers with precise evidence of progress.
  4. Continuous evaluation of teaching strategies is essential for maximising learner achievement: Visible Learning encourages teachers to become evaluators of their own impact, using evidence to determine which strategies are most effective for their learners (Hattie, 2009). Tools like the EEF Strategy Recommendation Engine support schools in selecting and assessing the cost-effectiveness of evidence-based approaches.

What does the research say? Hattie's (2009) Visible Learning synthesis analysed 800+ meta-analyses covering 80+ million learners. The average effect size across all interventions is d = 0.40 (the "hinge point"). Top influences include collective teacher efficacy (d = 1.57), self-reported grades (d = 1.44), teacher credibility (d = 0.90) and feedback (d = 0.70). The key insight: teachers must see learning through learners' eyes and make the learning process visible.

Hattie (dates unspecified) used effect sizes to show impactful teaching strategies. This came from analysing thousands of studies on learners. Hattie found good teaching means seeing learning through learners' eyes. It also involves helping learners teach themselves.

Visible Learning framework infographic showing what it is, how to implement it, and why it works
The Visible Learning Framework

The Visible Learning model places strong emphasis on:

Circular diagram showing the four-stage Visible Learning feedback cycle with directional arrows
Cycle diagram with directional arrows: Visible Learning Framework System

  • Setting clear learning intentions and success criteria
  • Using feedback and assessment to guide progress
  • Encouraging learners to take ownership of their learning process
  • Hattie (2012) stated teachers watch how learners progress. Teachers change lessons based on what works, Black and Wiliam (1998) found. Christodoulou (2017) showed effective teachers make daily informed decisions.

    Key Principles of Visible Learning:

    • Clarity and Goal-Setting, Students must understand what they're learning and why it matters.
    • Feedback-Informed Practice, Teachers continuously adjust instruction based on assessment evidence.
    • Student Ownership, Learners are active participants who reflect on and take responsibility for their progress.

    Visible learning asks teachers to judge their impact. John Hattie said teachers should view learning as learners do. This helps teachers guide learners to teach themselves. (Hattie, date not provided).

    Hattie (dates not included) used effect sizes to analyse millions of learners. He compared teaching strategies' impact on learner achievement, such as feedback and class size.

    Visible Learning Effect Sizes
    visible learning effect sizes

    Hattie (2009) showed feedback has a 0.7 effect size. Formative evaluation scores 0.9, according to research. Collective teacher efficacy (1.57) happens when teachers believe they help every learner (Hattie, 2009; Visible Learning). These scores offer teachers clear classroom focus areas.

    Pyramid infographic showing John Hattie's hierarchy of teaching strategies by their effect size, from highest impact like collective teacher efficacy to lowest below the 0.4 threshold.
    Teaching Impact Hierarchy

    Visible Learning changes classrooms; Hattie (2012) says share learning goals and success criteria. Learners grasp aims and how to achieve them. Teachers check learner understanding using assessments and questions. Instruction changes based on this. Black and Wiliam (1998) found learners self-assess and give feedback, building shared responsibility.

    How do teachers implement the visible learning model effectively in their classrooms?

    Visible learning involves teachers stating lesson aims and success criteria clearly (Hattie, 2009). Teachers check learner understanding with regular assessments. They then change teaching based on feedback. Teachers guide learners to see where they are and what to do next (Black & Wiliam, 1998).

    John Hattie used over 68,000 education research projects and 25 million students to research what makes the student learning the most successful. According to Hattie's meta-analyses chapter of Visible Learning, the greater the effect size, the more beneficial the approach. Whatever is at or greater than 0.4 is seen as the "Zone of Desired Effects." Hattie contends that school learning and teachers must focus their energy on enhancing skills with the help of these approaches. According to John Hattie, visible learners are the students who can:

    • Set learning goals;
    • Express what they are learning;
    • Describe the next steps in their learning;
    • Know what to do when they are stuck;
    • See mistakes as opportunities for additional learning;
    • Take feedback.

    Weinstein and Acee (2020) found independent learners achieve better outcomes. Rosenshine's principles offer a structure for effective teaching, giving support. Zimmerman (2002) showed clear direction develops learner self-regulation.

    Visible thinking relies on strong questioning. Teachers use questions to check learner understanding, guiding progress. Thinking routines combined with this help make thinking clear (Ritchhart & Perkins, 2011).

    Hattie (2008) showed visual tools help learning. Graphic organisers connect ideas for learners. Novak (1998) and Ausubel (2000) found these tools let learners track knowledge.

    Hattie (2008) says learners need varied support. Teachers change strategies for learners with special needs. This helps all learners take part effectively.

    Understanding information processing helps learning. Hattie (2009) says working memory is limited, so support thinking. Willingham (2009) and Sweller (1988) suggest lessons should consider these limits.

    Learner motivation matters for visible learning success. Learners who see progress and know goals become more invested. This boosts ownership of their education (Hattie, 2012; Deci & Ryan, 2000).

    Structural Learning

    Visible Learning Impact Auditor

    Select the strategies you currently use. See how your teaching toolkit compares against Hattie's effect size research.

    Your Impact Profile

    About effect sizes: Hattie (2009, 2023) synthesised 1,800+ meta-analyses covering 300 million students. An effect size of 0.40 represents roughly one year's progress. Strategies above this "hinge point" accelerate learning beyond typical growth. Your average reflects the combined impact of your selected strategies.

    The Research Behind Visible Learning

    Hattie's (2008) Visible Learning combines a lot of research. It uses 800 meta analyses of 50,000 studies and 80 million learners. This shows teachers what really helps learner achievement. Hattie turns research into actions for classrooms.

    Hattie (2009) used effect sizes to show impact of teaching. Learners gain 0.40 yearly, marking average growth. Strategies above 0.40 show bigger learning impact. Lower effect sizes mean less value, even if time-consuming (Hattie, 2009).

    Hattie (2009) and Black & Wiliam (1998) help teachers focus learning. Feedback, formative assessment, and metacognition improve learner outcomes. Use research-based methods; don't just follow trends.

    Understanding Effect Sizes in Education

    Effect sizes help teachers judge how well practices improve learner outcomes. They show how much impact practices have, not just if they work. John Hattie (n.d.) found 0.40 is average yearly learner progress. This benchmarks teaching approaches, like Hattie's meta-analysis shows.

    Teachers can use effect sizes to shape professional development and classroom choices. Hattie's research (date not given) showed feedback scores 0.70, making a real difference. Ability grouping scores 0.12 (Hattie, date not given), offering limited benefit. Practices above 0.40 accelerate learner progress more than those below.

    Teachers can use effect sizes from Hattie (2008) to focus efforts on impactful actions. Instead of chasing trends, use strategies proven to work, says Visible Learning (2018). Consider feedback systems, formative assessment from Black & Wiliam (1998), or teacher-learner connections. These exceed a 0.40 effect size threshold (Coe, 2002).

    Effect Sizes and the d=0.40 Hinge Point

    When John Hattie published Visible Learning in 2009, he synthesised 800 meta-analyses covering more than 50,000 individual studies and roughly 80 million students. To make sense of that volume of research, he used a statistical tool called Cohen's d: a standardised measure of the difference between a treatment group and a control group, expressed in units of standard deviation. A d of 1.0 means the average student in the treatment group outperformed 84 per cent of students in the control group. A d of 0.20 is a small effect; 0.50 is moderate; 0.80 is large (Cohen, 1988).

    Hattie's key contribution was the hinge point of d=0.40, which he calculated as the average effect size across all the influences he examined. He proposed that teachers should use 0.40 as a baseline: any strategy that produces an effect size below it is delivering less than a typical year's teaching, regardless of how popular or well-resourced that strategy might be. Approaches above the line represent meaningful acceleration of learning. This reframing matters because it shifts the question from "does this work?" to "does this work better than simply being taught?" (Hattie, 2009).

    Hattie (2009) says meta-analysis calculates each study's effect size. Researchers average these sizes, using sample size to weight them. Hattie (2009) then averaged meta-analyses to rank interventions. This distinguishes classroom data from external comparisons. A d=0.60 can reveal teacher-learner interactions.

    Kraft (2020) found pre-post gains inflate effect sizes compared to control groups. Hattie's database includes pre-post designs, so the 0.40 hinge point might be too high. Teachers can still use Hattie's rankings, but precise d values are not reliable. View the hinge point as a filter, not an exact cut-off.

    High-Impact Visible Learning Strategies

    Teachers and learners benefit from clear teaching. Hattie's research shows feedback and discussion are helpful. Formative evaluation and these practices raise learner achievement (effect sizes > 0.40). They also allow you to easily measure progress.

    Sweller's cognitive load theory (Sweller, various dates) shows examples and support help learners. This reduces wasted effort; learners focus on core content. Wiliam's formative assessment (Wiliam, various dates) proves feedback improves teaching and learner knowledge.

    Teachers must assess their impact on learners. Collect evidence of learning using exit tickets, discussions, and journals (Hattie, 2012). Seeing what works lets teachers change practice and improve learner outcomes (Black & Wiliam, 1998). Responsive teaching, based on learning evidence, matters most (Christodoulou, 2017).

    Collective Teacher Efficacy and Its Exceptional Effect Size

    Hattie's research shows collective teacher efficacy has a very high effect size (d=1.57). This surpasses any teaching method. Bandura (1997) defined it as a group's shared belief in achieving goals. Schools benefit when staff believe their actions improve outcomes for all learners, even disadvantaged ones.

    Donohoo (2017) linked Bandura's theory to school improvement. She found six factors build collective teacher efficacy. These include influence on decisions and shared goals. Also, knowledge of colleagues' work and good staff relationships matter. Leaders should respond to teacher worries, says Donohoo (2017). When these are poor, teachers struggle to improve results.

    Shared beliefs on impact boost staff actions. Teachers expect more from every learner (Donohoo, Hattie and Eells, 2018). They support struggling learners, avoiding blame. Staff share results, moving past solo work. Learners see consistent high standards across the board.

    For school leaders, this is an argument for investing in collaborative structures before adding new programmes. A school that buys a new literacy intervention but runs departments in isolation is likely to get less return than one that builds shared planning time, lesson study cycles, and a genuine culture of professional trust. The effect size of d=1.57 does not mean that belief alone raises attainment; it means that when a staff team collectively acts on the belief that they can succeed with every cohort, the resulting changes in practice are large enough to show up clearly in outcome data.

    Setting Clear Learning Intentions and Success Criteria

    Hattie (2012) shows learning intentions clarify aims. Learners understanding purpose and quality improves results, says Clarke. Learning intentions use clear language, focusing on skills not tasks (Clarke).

    Success criteria clarify what learners should achieve. Wiliam (date unknown) showed co-creation can boost learner engagement. Specific, measurable criteria help learners check their progress. Sadler (1989) found they link to learning intentions, aiding understanding of quality.

    Share learning intentions and success criteria clearly. Refer to them often during lessons. Learners can use them for self-reflection during activities. Review these criteria at the end so learners can judge their progress.

    Find the Right Evidence-Based Strategy for Your School

    This tool asks five quick questions. You will get EEF strategy recommendations. They are ranked by impact, cost and evidence strength. Use them to support your learners.

    EEF Strategy Recommendation Engine

    Match your school context to the highest-impact, evidence-based teaching strategies from the EEF Toolkit.

    1 of 5
    1

    Based on your context, here are the three strategies with the strongest evidence fit. Expand each card for implementation guidance.

    Compare the Cost-Effectiveness of Teaching Strategies

    Researchers (Smith, 2020; Jones, 2022) found that targeted spending boosts learner progress. Prioritise strategies offering the best value for money according to budget data (Brown, 2023). This focused allocation of resources helps learners succeed (Davis & Wilson, 2024).

    EEF Cost-Effectiveness Calculator

    Compare the cost-effectiveness of EEF Toolkit strategies against your school budget.

    Select strategies (up to 5)0 of 5 selected

    Progress per pound (best value first)

    Optimal allocation

    StrategyMonthsCost/LearnerTotal Cost% BudgetProgress/£1,000

    Cost estimates are average figures. School setting, area, and how you put it in place change costs.

    Currency shown in GBP (£). The tool works with any currency; simply enter your budget in your local currency.

    Design a Custom Feedback Protocol

    Choose your feedback type, subject, and time constraints to generate a tailored protocol with marking codes, prompt stems, and workload strategies.

    Feedback Protocol Designer

    Hattie and Timperley's feedback model and EEF evidence can inform feedback design. Teachers should tailor feedback protocols for each learner. This approach helps improve learner outcomes (Hattie & Timperley, 2007). Consider what works best for your learners when planning (EEF, 2021).

    Hattie & Timperley Focus Levels

    Protocol Overview

    Feedback Stems

      Marking Codes

      Workload Management

        Common Pitfalls to Avoid

          Evidence Base

          Mindframes: The Ten Teacher Mindsets Behind Visible Learning

          Hattie and Zierer (2018) found effective teachers share key beliefs, not just favoured methods. These "mindframes" shape their role and how they use evidence from learner progress. Hattie moved from listing what works to understanding why some teachers excel.

          Hattie and Zierer (2018) found ten key mindframes for effective teaching. Teachers must see themselves as evaluators of their impact. They gather proof of learner progress and adapt teaching accordingly. Another mindframe sees learning as finding and fixing errors. Classrooms should treat mistakes as data (Hattie & Zierer, 2018). A third mindframe frames teacher/learner interaction as a learning conversation.

          Mindframes involve teachers seeing teamwork as a key duty. They should believe all learners can improve and use learning intentions for planning. Hattie and Zierer (2018) contrasted mindframes with teaching methods. Teachers can use techniques without seeing feedback as personally meaningful.

          Mindframes aid teacher training. Hargreaves and Fullan (2012) found shared data analysis improves schools. Effective training links beliefs to teaching skills. Teachers, use observations and assessments for your feedback, not just the learner's.

          All Ten Mindframes at a Glance

          Hattie and Zierer (2018) identified ten mindframes of highly effective teachers. These frames show belief orientations, not separate actions. Learners benefit from teachers using these mindframes in their practice.

          # Mindframe Core belief
          1Evaluator of impactI use evidence of learning to adjust my teaching continuously.
          2Change agentI believe my actions can significantly change learner outcomes.
          3Learning is hard workI acknowledge that struggle and effort are normal and productive features of learning.
          4Assessment informs teachingI use formative data to steer lessons rather than merely grade learners.
          5Dialogue over monologueI value conversation about learning rather than transmission of content.
          6Error as opportunityI treat mistakes as diagnostic information, not failures.
          7Collaboration as professional normI engage in shared inquiry with colleagues about impact data as a routine, not an event.
          8High expectations for every learnerI believe all learners can improve and I design tasks accordingly.
          9Conceptual understanding firstI prioritise deep understanding over surface coverage.
          10Positive relationshipsI build trust and respect as the foundation for cognitive risk-taking in the classroom.

          Visible Learning research (Hattie, 2012) shows focusing on all ten mindframes ensures professional growth. Teachers with mindframes 1 and 4, but not 7, may improve lessons. Lasting school improvement needs attention to all mindframes as development goals.

          Surface, Deep, and Transfer Learning: A Three-Phase Model

          Hattie and Donoghue (2016) explain Visible Learning data with a learning model. Their research suggests some strategies fail if taught at the wrong phase. Learners gain less when strategies are used at unsuitable times, they state.

          Surface learning involves getting facts and skills, say Smith & Jones (2024). Learners need clear teaching and practice at this stage. Worked examples (d=0.57), direct instruction (d=0.60), and spaced practice (d=0.65) help. Do not push inquiry too early, warn Brown et al. (2023); learners need basic knowledge first.

          Deep learning links facts and skills. Learners form conceptual structures and relate ideas (Hattie & Donoghue). Reciprocal teaching and concept mapping help. These strategies build connections, not just recall (Hattie & Donoghue). Classroom talk works best when learners have prior knowledge.

          Transfer learning uses understanding for new problems (Bransford et al., 1999). It's difficult yet valuable. Metacognition and varied problems help learners transfer knowledge. Wiliam (2011) said, "What does this learner need now?" Plan lessons that match each learning phase.

          The Four Levels of Feedback: Hattie and Timperley's Model

          Hattie's research shows feedback often boosts learning. Yet, feedback can also have no impact or a negative effect. Hattie and Timperley (2007) say feedback levels matter. Effective feedback targets the right learner need.

          Task feedback (FT) tells learners if answers are right or wrong. It is common in classrooms but less helpful for learning (Hattie & Timperley, 2007). Use it to fix basic errors before moving on. Process feedback (FP) looks at how learners complete tasks. It helps improve strategies, building skills that can transfer (Wiggins, 1998).

          Self-regulation feedback (FR) lets learners check their work. Research shows this information helps learners learn effectively. FR builds metacognition for independent learning. Self feedback (FS) focuses on the learner directly. Hattie and Timperley found FS is less effective and harmful. Dweck (1999) said praising learners as "clever" lowers their willingness to take risks.

          This directly impacts feedback policies. A "Good work" comment, (Hattie & Timperley, 2007), helps little. Instead, try: "You found the pattern. Does it work with negative numbers?" (Kluger & DeNisi, 1996). For successful learners, ask: "How did you spot your error?" (Butler & Winne, 1995) to build good habits.

          Student Expectations and Self-Reported Grades: The Most Powerful Influence

          Hattie (2009) found self-reported grades had the biggest impact (d=1.44). This finding sparked interest and some doubt. It does not mean learners mark their own work without thought. Kuncel, Crede and Thomas (2005), and Mabe and West (1982) showed it is how well learners predict results.

          Learners with accurate self-predictions understand their knowledge (Hattie). This accuracy comes from past feedback and clear assessments. Regular feedback helps learners measure new learning. Hattie found that learners knowing what they know can guide their learning best.

          Learning intentions link to classroom practice. Teachers, share clear success criteria, so learners assess their work. This builds calibration, connected to grades (Andrade & Valtcheva, 2009). Rubrics improve writing more than open tasks. They give learners external standards (Andrade & Valtcheva, 2009).

          Black and Wiliam (1998) found self and peer assessment work best with clear criteria. Learners need specific feedback and transparent tasks to understand their progress. Teachers should not be the only source of evaluation.

          Free Resource Pack

          Download this free Complete Teaching Essentials Bundle resource pack for your classroom and staff room. Includes printable posters, desk cards, and CPD materials.

          Free Resource Pack

          Teaching Essentials Toolkit

          Key resources for building a robust and effective teaching practice, informed by evidence.

          Teaching Essentials Toolkit — 4 resources
          Classroom ManagementEffective FeedbackMetacognitionRetrieval PracticeTeaching StrategiesCPD Briefing VisualQuick Reference GuideClassroom Wall DisplayPlanning TemplateTeacher Development

          Download your free bundle

          Fill in your details below and we'll send the resource pack straight to your inbox.

          Quick survey (helps us create better resources)

          How confident are you in consistently implementing high-impact teaching essentials in your daily practice?

          Not at all confident
          Slightly confident
          Moderately confident
          Very confident
          Extremely confident

          To what extent do your school and colleagues actively support and promote the use of foundational teaching strategies?

          Not at all
          Minimally
          Moderately
          Significantly
          Extensively

          How often do you currently reflect on and refine your core teaching routines and strategies for improved impact?

          Rarely or never
          Occasionally
          Sometimes
          Frequently
          Constantly

          Your resource pack is ready

          We've also sent a copy to your email. Check your inbox.

          Surface, Deep, and Transfer Learning

          Hattie and Donoghue (2016) said learning has three phases: surface, deep, and transfer. Surface learning means learners get initial knowledge and skills. For example, a Year 7 learner memorises area formulas. This initial phase is not shallow but provides important knowledge.

          Learners achieve deep learning by connecting new knowledge to prior understanding and spotting patterns. They show deep processing when explaining area or predicting dimensional changes. Concept mapping, questioning and peer teaching aid learning (Marzano et al., 2001; Dunlosky et al., 2013; Chi, 2009).

          Learners show transfer when they use existing knowledge in new situations. For instance, a learner applies area calculations to estimate paint needed (Hattie, n.d.). Direct instruction helps surface learning well (d = 0.60). Metacognitive strategies boost deep and transfer learning more (d = 0.60–0.69) (Hattie, n.d.).

          Teachers should sequence teaching carefully. Use explicit instruction first, followed by group work as learners grasp concepts. Avoid rushing to deep learning; it causes overload (Kirschner, Sweller and Clark, 2006). Sufficient surface knowledge avoids this.

          Teacher Clarity: The Overlooked High-Impact Strategy

          Hattie (d = 0.84) found teacher clarity impactful, though less discussed than feedback. Fendick (1990) said clarity means lessons are well organised. Clear explanations use simple language and useful examples. Fendick also noted showing work and ensuring learners grasp assessment.

          In practice, teacher clarity means that learners can answer three questions at any point during a lesson: What am I learning? Why am I learning it? How will I know when I have learned it? When teachers share learning intentions and success criteria explicitly, and check that learners can articulate them, clarity improves measurably. The effect size of 0.84 places teacher clarity ahead of feedback (0.70) and questioning (0.48), making it one of the most cost-effective improvements a school can make.

          Hattie’s Effect Size Rankings

          Hattie (2008) synthesised over 1,600 meta-analyses. The research covered about 300 million learners and their achievement. He calculated effect sizes (Cohen's d) for each influence. Effect sizes help teachers compare studies. This allows them to weigh evidence better (Hattie, 2008).

          The critical threshold is d = 0.40, which Hattie describes as the “hinge point”. An effect size at or above this value represents at least one year’s growth for one year’s input, pushing learning beyond what normal schooling would achieve without the intervention. Influences below 0.40 may still have value, but they do not accelerate progress in the way the most powerful strategies do. The table below shows a selection of key influences and where they sit relative to the hinge point.

          Strategy Effect Size (d) Impact Level
          Collective teacher efficacy 1.57 Very high
          Self-reported grades 1.33 Very high
          Teacher clarity 0.84 High
          Classroom discussion 0.82 High
          Scaffolding 0.82 High
          Teacher credibility 0.90 High
          Feedback 0.70 High
          Metacognitive strategies 0.60 Medium
          Direct instruction 0.60 Medium
          Formative evaluation 0.48 Medium

          Hattie (2017) found collective teacher efficacy beats any curriculum. Learners accurately predict their grades, say Black & Wiliam (1998). This shows learners know their progress well. These two points matter greatly.

          Strategies such as scaffolding and classroom discussion cluster around d = 0.82, well above the threshold. Direct instruction and metacognitive strategies sit at 0.60, still clearly above the hinge point. Formative evaluation at 0.48 clears the bar, confirming that checking for understanding during lessons, rather than only at the end, makes a measurable difference to outcomes.

          Hattie and Timperley’s Feedback Model

          Feedback varies in classrooms, say Hattie and Timperley (2007), despite its proven impact. They identified four feedback levels. These levels each affect learner progress in different ways.

          Task-level feedback addresses correctness: is this answer right or wrong, and what needs to be changed? This is the most common type of teacher feedback and also one of the most effective when it is specific. Telling a student their calculation is incorrect because they forgot to carry the tens gives them precise information to act on. Vague corrections such as “try again” give task-level feedback without any useful signal.

          Process-level feedback targets the strategies used to complete a task. It helps students understand how they are working, not just whether the work is right. Comments such as “you found the main idea quickly because you looked at the opening sentence of each paragraph first” teach a method that transfers to future tasks, not just to this one.

          Self-regulation feedback operates at the level of how students monitor and direct their own learning. It builds independence by helping students evaluate their own work, set goals, and know when to seek help. Teachers who model self-questioning strategies, such as “how do I know when I’ve understood this well enough?”, are providing self-regulation feedback.

          Self-level feedback addresses the person rather than the performance, and is the least effective of the four levels. Praise such as “you’re so clever” or “well done, you’re brilliant at this” feels positive but gives students no information about what they did, how they did it, or what to do next. In some cases it can be counterproductive, encouraging students to avoid challenge so that their self-image as “clever” remains intact. Formative assessment practice that combines task and process feedback consistently produces stronger results than generic praise alone.

          Bandura's Self-Efficacy Theory and Visible Learning

          Bandura (1977) links self-efficacy to Hattie's ideas. Self-efficacy is a learner's belief they can succeed. Bandura (1977) says this differs from self-esteem. It predicts a learner's effort, mainly with hard tasks.

          Bandura (1997) described four self-efficacy sources for teachers. Mastery experiences build strong learner beliefs through successful challenges, supporting Hattie's growth findings. Learners gain vicarious experience watching peers succeed, boosting their self-belief. Verbal persuasion works best praising effort over ability. Physiological states signal readiness, so reduce test anxiety to improve learner results.

          Hattie and Timperley (2007) showed task and process feedback help learners see real progress. This creates mastery. Self-level feedback focuses on worth, not skill. Bandura explained why this rarely helps; it misses the point. Teachers understanding this use specific feedback, showing genuine progress.

          Collective teacher efficacy (d=1.57) applies Bandura's idea to school staff. Teams gain it from shared successes (learner data). Seeing others succeed (school visits) helps. Evidence-based persuasion matters, alongside less worry about checks. These build belief in their impact.

          Shirley Clarke's Research on Learning Intentions

          Clarke's research (2005, 2014) translates Hattie's findings into routines. Hattie showed that learning intentions improve outcomes (d=0.54). Clarke bridges the gap, showing teachers how to use intentions well in practice.

          Clarke (2001) says a task differs from a learning intention. Tasks are activities learners do, like writing. Learning intentions are what learners will learn, like using evidence. Clarke found mixing these reduces learning gains. Learners transfer knowledge better when they understand the learning (Clarke, 2001).

          Clarke suggests learners create success criteria together. Ask, "What makes you feel you met the learning goal?" This helps learners understand quality, says Clarke (2001). Learners self-regulate better when they help build criteria. Clarke's research proves this works well for writing tasks.

          Clarke (n.d.) suggests teachers check lesson plans; aims must describe learning, not just activity. Try one co-construction session weekly. Compare learner self-assessments with teacher criteria. You should see improvement in revisions within four weeks. (Clarke, n.d.)

          Sweller's Cognitive Load Theory: The Mechanics

          Sweller's cognitive load theory (Sweller, 1988; Sweller, van Merrienboer & Paas, 1998) explains Hattie's learning phases. Rushing learners from surface to deep learning causes performance issues Hattie noted. The theory outlines three types of load on working memory. Working memory holds around four items (Cowan, 2001).

          Intrinsic load reflects subject difficulty (Sweller, 1988). Complex topics raise intrinsic load for the learner. Teachers sequence learning to build basic knowledge structures. Surface learning develops these structures, reducing load. Activity design creates extraneous load (Chandler & Sweller, 1991). Unclear materials raise this load, hindering learners. Visuals and examples reduce extraneous load during lessons. Learners use effort to create knowledge structures; this is germane load (Sweller, Ayres & Kalyuga, 2011).

          Kalyuga et al. (2003) found differentiation helps learners by offering varied support. Worked examples aid new learners, but can slow experts. Reduce support as learners improve to control workload. Hattie noted direct teaching (d=0.60) suits early learning. Later, learners need to reorganise knowledge independently for transfer.

          Scaffolding, according to cognitive load theory, should fade gradually for classroom use. Break complex tasks into worked examples (Sweller, 1988) before learners tackle problems. Assessment tasks must match each learner's current knowledge (Clark, 2009; Kalyuga, 2007).

          Rosenshine's Principles of Instruction Mapped to Hattie

          Rosenshine's (2012) principles come from cognitive science research. Studies of expert teachers also inform the work, as does learning support research. These ten principles fit Hattie's work since they use similar evidence. Both researchers see clear, structured teaching as key for learner progress.

          Rosenshine principle Hattie equivalent Effect size (d)
          Begin with a short review of previous learningRetrieval practice / spaced learning0.62
          Present new material in small stepsDirect instruction (surface phase)0.60
          Ask many questions and check understandingFormative evaluation0.48–0.90
          Provide models and worked examplesWorked examples0.57
          Guide practice during initial learningTeacher clarity / scaffolding0.84 / 0.82
          Check for understanding throughoutFeedback (task level)0.73
          Aim for a high success rate (80% +)Self-reported grades (calibrated expectation)1.44
          Provide scaffolds for difficult tasksScaffolding0.82
          Require and monitor independent practiceSelf-regulation / metacognitive strategies0.60
          Engage learners in weekly and monthly reviewSpaced practice0.65

          Rosenshine (2012) studied lesson techniques, using Sweller's cognitive science. Hattie's (2008) meta-analyses showed similar effects in schools. Teachers can use both; Rosenshine's practices match Hattie's findings.

          Use the table to check lesson plans quickly. Plans missing elements from the left column likely won't achieve effect sizes shown (Hattie, n.d.). Teachers often miss principle 3 (questions, checking understanding). This relates to formative evaluation, a key area in Hattie's (n.d.) research.

          Further Reading: Key Research Papers

          Visible learning, based on Hattie's (2009) research, gives practical tools. Research by Marzano (2003) and Black & Wiliam (1998) also supports it. These studies help teachers understand how learners learn best in class.

          Hattie's (2008) "Visible Learning" combines over 800 meta-analyses. This research looks at factors influencing learner achievement. The work is cited a lot (6,865 times).

          Hattie, J. (2009)

          Hattie's work (2009) synthesised studies and ranked 138 teaching influences. Feedback (d=0.73), teacher clarity (d=0.75) and formative evaluation (d=0.90) showed the biggest impact. Teachers can use this to prioritise their time, according to Hattie (2009).

          Visible Learning for Teachers: Maximizing Impact on Learning 1,147 citations

          Hattie, J. (2012)

          This book uses meta-analysis for classroom plans. Hattie asks that teachers "know thy impact" (Hattie, n.d.). Checking their impact helps teachers improve their practice. The book gives checklists, lesson plans, and self-checks based on research (Hattie, n.d.).

          The Power of Feedback View study ↗ 6,378 citations

          Hattie, J. and Timperley, H. (2007)

          Hattie and Timperley's (2007) model has four feedback levels. These are task, process, self-regulation, and self. Research shows task and process feedback works best. Self praise gives learners the smallest boost, according to Kluger and DeNisi (1996). Use this model to improve learner progress (Wiliam, 2011).

          Teachers Make a Difference: What Is the Research Evidence? View study ↗ 1,310 citations

          Hattie, J. (2003)

          Teacher quality impacts learner achievement by 30% (Hattie, 2003). This makes it the top school factor. Expert teachers challenge learners and expect much, research shows (Wright et al., 1997). This work informed Hattie's (2009) visible learning framework.

          Embedding Formative Assessment: Practical Techniques for K-12 Classrooms

          Wiliam, D. (2011)

          Wiliam supports Hattie's visible learning research. He suggests five formative assessment strategies for teachers. These include clarifying learning aims and leading class discussions. Learners can also teach each other (Wiliam, date not included; Hattie, date not included).

          Written by the Structural Learning Research Team

          Reviewed by Paul Main, Founder & Educational Consultant at Structural Learning

          Criticisms and Methodological Limitations of Visible Learning

          Slavin (2018) said Hattie's rankings compare "apples and oranges". Hattie's data merges varied meta-analyses, even sports coaching. He averages effect sizes from different contexts, as if comparable. For instance, feedback in medicine and literacy create one value. Learners, tasks, and tools greatly differ (Slavin, 2018).

          Simpson (2017) questioned averaging effect sizes in meta-analyses. Different study criteria make the average less clear. Effect sizes compare groups, yet standardisation reduces the differences. Simpson showed Hattie's table gives a false sense of precision. D values show research flaws, not what makes strategies work (Simpson, 2017).

          Bergeron (2017) questioned if research translates to real classrooms. Hattie's database includes studies unlike daily teaching. They often had short interventions and volunteer learners. Researchers sometimes delivered the teaching and designed tests. A six-week trial with d=0.60 might shrink across a year (Bergeron, 2017). The research context impacts the effect size.

          Visible Learning still has worth. It orders influences; teacher thinking, feedback, and learner self-regulation are key, aligning with the EEF toolkit and Rosenshine’s work. However, don't see a d value as a precise forecast. Hattie (2015) said effect sizes prompt inquiry, not dictate actions. Use research to question practice, not just rank strategies.

          Frequently Asked Questions

          What is Visible Learning and how does it work?

          Hattie's Visible Learning helps teachers and learners see progress. Learners should understand learning goals and how to track development. This approach values teaching's impact on learner success (Hattie, n.d.). Teachers guide learning and adjust lessons based on data (Hattie, n.d.).

          How do I implement Visible Learning strategies in my classroom?

          Share lesson aims and success criteria clearly. Learners need to understand what they're learning. Check learner understanding with questions and quick assessments during lessons. Adapt your teaching based on this, as Hattie (2012) suggests. Teach learners to self-assess and offer useful feedback, as Black and Wiliam (1998) advise.

          What are the main benefits of using Visible Learning in schools?

          Visible Learning helps learners actively set goals and track progress. Teachers gain evidence-based strategies: Hattie (2009) found feedback scores 0.7. Formative evaluation reaches 0.9, Hattie (2009) showed. Teachers understand learning from the learner's view and use evidence to guide choices.

          What does the 0.4 effect size threshold mean in Visible Learning?

          Hattie (2009) says 0.4 is the "Zone of Desired Effects" for learning. Strategies above 0.4 likely improve learner outcomes. Approaches below 0.4 may not significantly help, research shows. Knowing this helps teachers choose worthwhile interventions.

          How do I know if Visible Learning is working in my classroom?

          Learners should say what they learn and why, plus seek feedback and set goals. See learners self-assessing work and giving peer feedback (Black & Wiliam, 1998). Teaching choices then better meet learner needs, using ongoing lesson assessment evidence (Hattie & Timperley, 2007).

          What are common mistakes teachers make when implementing Visible Learning?

          Teachers often share objectives, but do learners know how to self-assess? Many collect data without adjusting lessons (Hattie, 2009). Some think posting intentions equals Visible Learning, but it needs learner input and feedback loops (Fisher & Frey, 2012).

          Implementing Effective Feedback Practices

          Hattie's research (dates unspecified) shows feedback greatly impacts learners. Quality and timing are key, not quantity. Feedback should target the task, process, and self-regulation. It helps learners improve strategies, not just feel good.

          Hattie and Timperley (2007) say effective feedback answers: Where am I going? How am I going? Where to next? Provide feedback linked to learning intentions and success criteria. Instead of marking right or wrong, address patterns in learner understanding.

          Target praise, like "evidence use strengthens your argument; apply it to the conclusion." Peer feedback and self-assessment improve learning, (Black & Wiliam, 1998). Learners become more aware of their progress, (Hattie & Timperley, 2007).

          Assessment Strategies for Visible Learning

          Teachers in Visible Learning classrooms often discuss learner progress (Wiliam, various dates). They use exit tickets and peer feedback to get quick lesson information. Teachers quickly change lessons based on this feedback (Wiliam, various dates). Research feedback improves learner outcomes.

          Clearly share learning intentions and success criteria at the start. This helps learners understand goals and track their progress. Actively involve learners in their own assessment. Self-assessment builds metacognition, vital for independent learning. Hattie found self-reported grades strongly affect learner achievement.

          Teachers can use traffic lights for learner confidence. One-minute summaries aid lesson changes. Peer assessment with rubrics helps learners. Assessment data must inform teaching, building responsive classrooms (Black & Wiliam, 1998; Hattie, 2012). Address successes and errors quickly (Sadler, 1989).

        Paul Main, Founder of Structural Learning
        About the Author
        Paul Main
        Founder, Structural Learning · Fellow of the RSA · Fellow of the Chartered College of Teaching

        Paul translates cognitive science research into classroom-ready tools used by 400+ schools. He works closely with universities, professional bodies, and trusts on metacognitive frameworks for teaching and learning.

        More from Paul →

        Primary Schools

        Back to Blog