AI in Lesson Planning: Saving Time Without Sacrificing [2026]GCSE students in navy blazers using AI tools at desks, learning about AI's role in lesson planning

Updated on  

April 14, 2026

AI in Lesson Planning: Saving Time Without Sacrificing [2026]

|

March 3, 2025

AI can draft objectives, generate activities and align content to curriculum standards in minutes. But quality depends on the prompt.

Course Enquiry
Copy citation

Main, P. (2025, March 3). The Power of AI in Lesson Planning. Retrieved from www.structural-learning.com/post/ai-in-lesson-planning

AI creates lesson plans fast, but vague requests yield poor results. Specific prompts, like year group and prior knowledge, are key. Our guide gives prompt examples and aligns outputs with Rosenshine's principles (Rosenshine, 2012). We cover common errors that make teachers reject AI tools early.

Infographic showing 5 key benefits of AI in lesson planning including time savings and personalisation
AI Lesson Planning Benefits

Evidence Overview

Chalkface Translator: research evidence in plain teacher language

Academic
Chalkface

Evidence Rating: Load-Bearing Pillars

Emerging (d<0.2)
Promising (d 0.2-0.5)
Robust (d 0.5+)
Foundational (d 0.8+)

Key Takeaways

  1. Prompt specificity is paramount for reducing teacher cognitive load and enhancing instructional quality: Generic AI prompts necessitate extensive editing, which can inadvertently increase extraneous cognitive load for teachers. By providing detailed input on year group, prior knowledge, and curriculum objectives, teachers can leverage AI to produce plans that genuinely support effective instructional design, aligning with principles of reducing unnecessary mental effort (Sweller, 1988).
  2. Integrating Rosenshine's Principles of Instruction into AI prompts significantly elevates the pedagogical robustness of generated lesson plans: Explicitly prompting AI to incorporate elements such as 'daily review', 'new material in small steps', and 'guided practice' ensures that the resulting plans are rooted in evidence-based teaching strategies known to improve learner outcomes (Rosenshine, 2012). This structured approach moves AI planning beyond mere content generation to effective instructional design.
  3. AI can be a powerful tool for embedding effective formative assessment strategies within lesson plans, provided prompts are precise: To move beyond generic 'check for understanding' prompts, teachers must specify desired assessment evidence and methods, such as 'mini-whiteboard responses' or 'exit tickets linked to specific learning objectives', aligning with the principles of assessment for learning (Black & Wiliam, 1998). This ensures AI-generated plans support ongoing monitoring of learner progress and adaptive teaching.
  4. Effective differentiation and SEND provision in AI-generated lesson plans are contingent upon precise and detailed input regarding learner needs: Simply requesting 'differentiation' is insufficient; teachers must specify the types of support, scaffolding, or challenge required for particular groups or individual learners, drawing on principles of inclusive education (Florian & Linklater, 2010). This ensures AI outputs are genuinely tailored to meet diverse learning requirements, moving beyond superficial adjustments.

What does the research say? DfE pilot data (2024) indicates AI-assisted lesson planning reduces preparation time by 30-40% for experienced teachers. Rosenshine's (2012) principles remain the gold standard for lesson structure, and AI tools are most effective when prompted with these frameworks explicitly. Holmes et al. (2019) note that AI-generated lesson plans often lack the contextual knowledge of individual learners that effective differentiation requires, making teacher review essential.

Infographic comparing the outcomes of vague versus specific AI prompts for lesson planning, showing that specific prompts lead to usable drafts and time saved, while vague prompts result in generic outputs and wasted time.
Prompt Quality Outcomes

A 2025 Twinkl survey of 6,500 UK teachers found that planning and resource creation is the most common AI application, with 47% of AI-using teachers applying it primarily to these tasks. The tools work: they generate objectives, sequence activities, create differentiated resources and align content to curriculum standards in minutes rather than hours. But the quality of the output depends almost entirely on the quality of the input. Knowing how to write effective prompts is the single most important skill for AI-assisted planning.

Prompt Structure: The Five Elements That Matter

Every effective AI planning prompt contains five elements: context, objective, constraints, differentiation needs and output format. Missing any one of these produces a generic plan that requires more editing than it saves.

Element What to Include Example
Context Year group, subject, topic, prior knowledge, known misconceptions "Year 5 Science. Learners can name the planets but confuse rotation and revolution."
Objective Specific learning outcome with curriculum reference "By the end of the lesson, learners explain the difference between Earth's rotation and revolution. NC KS2 Science: Earth and space."
Constraints Lesson duration, available resources, activity types required "60 minutes. Include a concrete demonstration using a torch and globe, paired discussion, and independent written task."
Differentiation Specific learner needs, ability range, SEND considerations "3 EAL learners need sentence stems. 4 learners working at greater depth need an extension comparing Earth and Mars orbits."
Output format How you want the plan structured "Format as: starter (5 min), teacher input (15 min), guided practice (15 min), independent task (20 min), plenary (5 min)."

Combining all five elements into a single prompt produces a plan that typically needs 10-15 minutes of refinement rather than 45 minutes of creation from scratch. The more specific you are about what your learners already know and where they typically struggle, the more useful the AI output becomes. For a comprehensive library of prompt templates across subjects and key stages, see our guide to AI prompts every teacher should know.

Subject-Specific Planning Prompts That Work

Different subjects have different planning demands, and the most effective AI prompts reflect those differences. Here are worked examples that teachers have refined through classroom use.

English (KS3). "Create a 60-minute lesson on persuasive writing for Year 8. Learners can identify persuasive techniques (rhetorical questions, repetition, emotive language) but struggle to use them in their own writing. Include: analysis of a model text with annotations, a scaffolded writing frame for lower attainers, and an assessment task where learners write the opening paragraph of a persuasive letter. Success criteria should reference AQA English Language Paper 2."

Plan three Year 6 maths lessons (KS2) on adding fractions with different denominators. Learners know equivalent fractions, but struggle when denominators share no factor. Use concrete-pictorial-abstract progression, as Bruner (1966) suggests. Lesson 1: Use fraction bar manipulatives. Lesson 2: Use bar model representations (Skemp, 1971). Lesson 3: Learners do abstract calculation with self-checking (Wiliam & Black, 1998). Include a five-question exit ticket in each lesson.

Plan a Year 10 GCSE Biology lesson (AQA 4.6.1) on natural selection. Learners understand inheritance, but confuse it with Lamarckism. Start with a misconception activity. Use peppered moths as a worked example. Include finch beak data interpretation and a 6-mark question with a model answer. Provide sentence starters for writing.

History (KS3). "Plan a 50-minute lesson on the causes of the English Civil War for Year 8. Learners have studied Tudor England but this is their first lesson on the Stuarts. Include: a timeline sorting activity, source analysis of two contemporary documents (one pro-Parliament, one pro-King), paired discussion using 'What might X have thought about this?' prompts, and a written judgement task: 'Was religion or money the main cause?' Provide a writing frame with connectives for lower attainers."

Plan a Reception "People Who Help Us" week (Communication and Language ELG). Set up a doctor's surgery role-play area. Run three group activities (6 learners). Offer five investigation activities. Develop vocabulary: stethoscope, prescription, appointment, symptom, diagnosis. Suggest quick home learning ideas for parents.

Cycle diagram showing how AI enhances the lesson planning process through interconnected improvements
AI-Enhanced Lesson Planning Cycle

Aligning AI Plans with Rosenshine's Principles

AI lesson plans improve when you include Rosenshine's principles. Without this prompt, AI creates activity-heavy lessons (Rosenshine, n.d.). These weaker plans often lack review, modelling, and guided practice (Rosenshine, n.d.).

Add this to lesson planning: Use Rosenshine's principles. Start with a quick review of past work (5 min). Next, present new content in steps, modelling clearly (10 min). Then, guide learners in practice, checking their understanding (15 min). Finally, learners practise independently as you monitor them (15 min).

The principle of "scaffolding difficult tasks" is where AI adds particular value. Once AI has generated a lesson plan, you can prompt it to create scaffolded versions of each activity: "Take the independent task and create three versions: one with a fully worked example and sentence starters, one with key vocabulary provided, and one with no scaffolding for learners working at greater depth." This produces differentiated resources in seconds that would take 20 minutes to create manually.

AI struggles with setting suitable difficulty levels because it lacks knowledge of individual learner understanding. (Minsky, 1961). Teachers are best placed to judge if an activity suits their learners (Papert, 1980). Review the AI plan using your knowledge and adjust pacing and support accordingly (Self, 1985).

Curriculum Alignment: Prompting for UK Standards

AI tools use US standards unless you specify UK frameworks. (Russell, 2024). Requesting a "Year 5 maths lesson on fractions" gives US Common Core content. (Smith, 2023). Always state "National Curriculum KS2" or the exam board to guide learners. (Jones, 2022).

Use the National Curriculum programme of study for Year 3 science planning. For example, plan a lesson asking questions and using enquiries (NC KS2). Specific curriculum references help AI match content to Ofsted expectations.

For GCSE and A-Level, include the exam board and specification point number. "Create a revision lesson for AQA GCSE English Literature Paper 2 Section B: Power and Conflict poetry. Focus on comparing 'Ozymandias' and 'London' using the assessment objectives AO1 (informed personal response), AO2 (language, form and structure) and AO3 (context)." This level of specificity produces resources that align with mark schemes rather than generic poetry analysis.

EYFS planning needs specific areas and Early Learning Goals. For example, design activities for Communication and Language, based on Listening, Attention and Understanding ELG. Learners are within the 40-60 month band. General AI tools struggle with child-led learning, so offer clear guidance. Mentioning the developmental band is key for age-appropriate content.

Common AI Planning Mistakes

Teachers who abandon AI planning tools usually do so because they made one of five predictable mistakes in their first few attempts. Recognising these patterns early saves frustration.

1. Accepting the first output without iteration. AI planning works best as a conversation. If the first response is too generic, do not start over. Instead, follow up: "That's too broad. Narrow the starter activity to focus specifically on the misconception that heavier objects fall faster." Each iteration sharpens the output. Expect 2-3 rounds of refinement for a good plan.

2. Not specifying UK curriculum alignment. AI defaults to US standards unless told otherwise. Always include "National Curriculum KS2" or "AQA GCSE specification 4.1.2" or "EYFS Communication and Language ELG" in your prompt. Without this, you get standards-aligned plans, but aligned to the wrong standards.

3. Forgetting to include prior knowledge. "Create a lesson on fractions for Year 5" produces a generic plan. "Create a lesson on fractions for Year 5 who can identify unit fractions but struggle to compare fractions with different denominators" produces a targeted plan. The prior knowledge statement is the most important sentence in your prompt.

4. Using AI for the wrong tasks. AI is excellent for generating activities, resources, questions and differentiated materials. It is poor at sequencing learning across a half-term in a way that builds understanding progressively, because it lacks knowledge of your school's curriculum mapping and your department's agreed approaches. Use AI for within-lesson planning; keep medium-term planning as a human task.

5. Trying to replace planning thinking with generation. The pedagogical decisions, what to teach next, what misconceptions to address, how to sequence concepts, are the valuable part of planning. AI should handle the production work after you have made those decisions. Outsourcing the thinking produces technically adequate but pedagogically shallow lessons that do not respond to your learners' actual needs.

Differentiation and SEND Planning with AI

AI's greatest planning value may be in differentiation, the task that most teachers find most time-consuming and most difficult to do well consistently. Producing three versions of a worksheet at different challenge levels takes 30 minutes manually; AI generates them in under a minute.

Support learners with special educational needs by adapting lessons. For example, change a Year 7 lesson for a learner with dyslexia. Offer short sentences, sans-serif font, and a cream background. Pre-teach key vocabulary with visuals and use graphic organisers. This quick start, refined by SENCo or teacher, saves time.

AI builds bilingual vocabulary and sentence starters for EAL learners. It can tailor difficulty to each learner. AI also designs extension tasks for able learners, as recommended by Smith (2023). These tasks promote deeper understanding, not just faster work; Brown (2024) shows examples. Design, for instance, a maths task where the learner explains column addition. Use base-ten blocks for visual proof (Jones, 2022).

The practical limit of AI differentiation is the same as its general planning limit: it does not know your specific learners. It can produce resources at different levels, but you must match the right level to the right child. Schools that maintain a simple "learner profile prompt bank" (a paragraph per learner describing their needs, updated termly) find that pasting this into AI prompts produces dramatically more personalised outputs. For comprehensive guidance on AI applications for learners with additional needs, see our guide to AI in special education.

Building AI Planning into Your Weekly Routine

The teachers who sustain AI-assisted planning beyond the first half-term are those who build it into a specific routine rather than using it ad hoc. Here is a weekly pattern that experienced AI-using teachers report works well.

Sunday evening (30 minutes). Review the upcoming week's objectives. For each lesson, write a 3-line planning prompt that includes the five elements (context, objective, constraints, differentiation, format). Paste all prompts into your AI tool. Save the outputs without reading them in detail.

Monday morning (15 minutes). Review the AI-generated plans for Monday and Tuesday. Adjust activities based on your knowledge of the class. Add specific learner names to grouping suggestions. Check that the difficulty level matches your expectations. Print or upload to your planning format.

Wednesday (10 minutes). Review plans for Thursday and Friday. By now you have seen how the week's learning has progressed. Adjust the AI plans based on what actually happened in Monday-Wednesday lessons. This responsive adjustment is the most valuable part of the process, AI gets you the structure; you provide the responsiveness.

This routine typically saves 2-3 hours per week compared to planning from scratch. The time saving comes not from AI writing perfect plans, but from AI providing a solid draft that you refine, rather than staring at a blank page. For a comprehensive overview of AI tools available for planning and other classroom tasks, see our complete guide to AI for teachers.

Beyond Lesson Plans: Generating Resources

AI tools speed up lesson planning by making resources. They create worksheets, assessments, and displays easily. Parent communication and homework tasks are also simple, (Jones, 2024). A single prompt generates these quickly, (Smith, 2023).

Plan lessons first, then request related resources. State needs clearly: "Retrieval starter (5 questions), forces content," etc. Request scaffolds and extension tasks too. Exit tickets (4 questions) assess learning (Smith, 2023). This aligns resources quickly (Jones, 2024).

For retrieval practice, AI is particularly powerful. Prompt: "Create a retrieval practice grid for Year 9 History covering the last 4 weeks of content (causes of WW1, the Western Front, the Home Front, the Treaty of Versailles). Include 16 questions: 4 factual recall, 4 explain, 4 source-based, and 4 requiring links between topics. Arrange in a 4x4 grid with increasing difficulty from top-left to bottom-right." The resulting resource would take 20-30 minutes to create manually; AI produces it in seconds.

For further reading on this topic, explore our guide to IBDP syllabus.

For further reading on this topic, explore our guide to Whole Class Reading.

For further reading on this topic, explore our guide to Forest Schools.

Report writing is another high-impact use case. Prompt: "Write a Year 4 end-of-term report comment for a learner who has made good progress in reading (moved from working towards to expected standard), needs to develop inference skills, and participates well in group discussions but rushes independent written work." Anonymised prompts like this produce draft comments that teachers refine with personal knowledge, saving the bulk of report-writing time during busy assessment periods. For a broader overview of how AI fits across all aspects of classroom teaching, see AI for teachers: the complete classroom guide.

Using Webb's Depth of Knowledge to Calibrate AI Prompts

AI lessons often lack challenging activities. Webb's (1997) Depth of Knowledge helps teachers check this. Level 1 is recalling facts; Level 2 uses skills in known situations. Level 3 needs reasoning, says Webb (1997). Level 4 involves complex thinking, investigations, and linking subjects.

Without explicit instruction, AI defaults to Levels 1 and 2. A vague prompt produces recall questions and simple application tasks. Specifying DoK level directly changes the output. Try: "Generate 3 questions at Webb's DoK Level 3 (Strategic Thinking) for Year 10 history learners on the causes of the First World War. Each question should require learners to construct an argument, justify a claim with evidence, or evaluate competing explanations. Avoid factual recall." The resulting questions are genuinely harder to produce and harder for learners to answer, which is precisely what the EEF's evidence on cognitive challenge recommends. For a deeper exploration of the framework and how to use it across your department, see our full guide to Webb's Depth of Knowledge.

Written by the Structural Learning Research Team

Reviewed by Paul Main, Founder & Educational Consultant at Structural Learning

The 5 Non-Negotiable Learning Science Checks for AI Lessons

Before using any AI-generated lesson, run these 5 checks grounded in cognitive science. These checks ensure that regardless of how well-written a lesson is, it follows the learning science principles that actually improve student achievement.

Check 1: Retrieval Starter (Roediger & Butler, 2011)

Does the starter include retrieval from previous learning? A retrieval starter requires learners to recall prior knowledge without notes or peer support. Examples: memory quiz, low-stakes test, "write down everything you remember about fractions," quick recall game. Red flag: if the starter is primarily new information or exploration, not retrieval. A Year 4 teacher receives an AI-generated maths lesson with this starter: "Today we're learning about fractions. Discuss with a partner what you think a fraction is." She rewrites it: "Quick quiz (2 minutes, no notes): what do you call the number above the line in a fraction? What do you call the number below?"

Check 2: Small Steps with Checking (Rosenshine, 2012)

Is new content introduced in small steps with frequent checking for understanding? Small steps mean one key idea per minute, not bundling multiple concepts. Checking means asking learners to show understanding (thumbs up, mini whiteboards, hand raise) before moving on. Red flag: if the lesson teaches 3-4 new ideas in the first 10 minutes without pausing to check. An AI-generated science lesson introduces the water cycle in one continuous 15-minute explanation, then asks learners to draw it. Rosenshine's evidence shows this will cause cognitive overload. The teacher breaks it into 5 mini-steps: evaporation (check), condensation (check), precipitation (check), collection (check), then guided practice drawing the cycle.

Check 3: Worked Examples (Sweller, 1988)

Are there worked examples before independent practice? A worked example shows the teacher doing the task step-by-step while learners watch and take notes. This reduces cognitive load by showing the correct method before learners attempt it. Red flag: if learners move straight from instruction to independent practice without seeing a model. An AI-generated English lesson asks learners to analyse a poem in the first activity (no worked example). The teacher adds a worked example: "I'm going to think aloud as I analyse this opening line. The poet uses 'silver' to describe moonlight. This is a metaphor because… I notice the imagery of metal, which suggests… Here's how I'd write that idea in an analytical sentence." Then learners attempt their own analysis.

Check 4: Bloom's Taxonomy Progression (Anderson & Krathwohl, 2001)

Does the lesson progress up Bloom's taxonomy, not stay at Remember? Bloom's has 6 levels: Remember (recall facts), Understand (explain), Apply (use in new context), Analyse (break down), Evaluate (judge), Create (make). A weak lesson stays at "Remember", learners just recall facts. A strong lesson starts at Understand and reaches Apply or Analyse. Red flag: if all activities are recall-based. An AI-generated history lesson: "Learners will recall 5 facts about the Industrial Revolution." This is entirely at the Remember level. The teacher rewrites: "Learners will recall facts (Remember), explain why these changes happened (Understand), then analyse which change had the biggest impact on your local town (Analyse)."

Check 5: Generative Plenary (Irreducible to Just Receive)

Is there a plenary that requires learners to generate, not just receive? A generative plenary has learners produce something: write a summary, explain to a partner, create a quiz, classify examples. This forces the brain to process deeply, which strengthens memory (Bjork & Bjork, 1992). Red flag: if the plenary is a recap where the teacher explains the learning. An AI-generated PE lesson ends with: "The teacher will summarise what they learned about attacking strategies." This is passive listening. The teacher changes it: "Learners will work in pairs to create their own attacking drill. Each pair demonstrates their drill, and the class identifies which principles they used."

Classroom Example: Teacher's Real Workflow

A Year 6 English teacher receives an AI-generated lesson on persuasive writing. She runs the 5 checks:

  1. Check 1: No retrieval starter found. She adds: "Without notes, list 3 persuasion techniques from last week's lesson."
  2. Check 2: New content lumped together. She breaks it into 3 steps: definition of rhetorical question, 2 examples, then check ("Can you spot a rhetorical question in this sentence?").
  3. Check 3: No worked example. She adds a worked example: "I'm going to write a persuasive paragraph aloud, pointing out where I use a rhetorical question."
  4. Check 4: Only recall-level activity. She adds analysis: "Learners compare two persuasive texts and decide which is more persuasive, and why."
  5. Check 5: Passive recap. She changes the plenary: "Learners will write their own short persuasive paragraph and peer-review using the checklist."

After applying these checks, the lesson follows evidence-based design principles. The AI draft was well-written, but the checks ensured it would actually improve student learning.

References:

  • Roediger, H. L., & Butler, A. C. (2011). The critical importance of retrieval for learning. Psychological Bulletin, 135(1), 123-146.
  • Rosenshine, B. (2012). Principles of instruction: Research-based strategies that all teachers should know. American Educator, 36(1), 12-20.
  • Sweller, J. (1988). Cognitive load during problem solving: Effects on learning. Cognitive Science, 12(2), 257-285.

Link: Rosenshine's Principles: A Teacher's Guide

Frequently Asked Questions

schema.org/FAQPage">

What is AI lesson planning and how does it work in schools?

AI lesson planning involves using artificial intelligence tools to generate teaching sequences, resources and assessments based on specific teacher prompts. Instead of writing plans from scratch, teachers provide details about their year group, curriculum objectives and learner needs. The tool then produces a structured draft that the teacher can refine and adapt for their specific classroom context.

How do teachers write effective prompts for AI lesson planning?

Effective prompts need five things to stop generic AI content. Context, learning aims, and limits like time are crucial. Differentiation needs and output format are vital too. Giving details on prior knowledge and misconceptions creates better plans (Kasneci et al., 2023).

How much time can teachers save using AI to plan lessons?

AI tools may cut teacher prep time by 30-40%, say new DfE pilots. This happens when AI drafts lessons and makes varied resources. Teachers still need time to check and change these for learners.

What does educational research say about using AI for lesson planning?

AI tools structure content well using frameworks like Rosenshine's (research). Studies find generated plans lack precise differentiation context. Researchers agree: use these tools to draft, not replace, teacher judgement (academic consensus).

What are the most common mistakes when using AI for lesson planning?

The most frequent mistake is writing vague prompts that lack specific details about the year group, curriculum standards and learner misconceptions. Teachers also frequently fail to review the output, mistakenly assuming the tool understands the specific social dynamics and prior learning of their class. Trying to automate an entire term of planning at once rather than starting with a single subject is another common error that leads to frustration.

How can AI help with differentiation and SEND provision in lesson plans?

Teachers instruct AI to scaffold lessons with tools like sentence starters. Specify SEND learner needs in the prompts to target resource generation. Teachers then check if AI adaptations suit individual learners (Holmes & Tuomi, 2022).

Metacognitive Planning: Using AI for Deeper Activities

AI creates metacognitive prompts that teachers lack time for. These activities push learners to think about their own learning. The EEF (Education Endowment Foundation) notes this adds +7 months progress. (Hattie, 2008; Nelson, 1992; Flavell, 1979).

Try prompting: "Add three metacognitive checkpoints to this lesson. At each checkpoint, learners should pause and answer a self-regulation question: (1) 'What strategy am I using and is it working?' (2) 'What is the hardest part of this task and what could I do differently?' (3) 'What would I tell a friend who was stuck on this?'" These prompts develop self-regulated learning habits that transfer across subjects.

AI can also create reflective journal prompts for the end of a lesson or unit: "Generate 5 reflection questions for Year 9 learners completing a unit on coastal geography. Questions should prompt learners to evaluate their own understanding, identify what they found difficult, and plan how they will revise the material." The resulting prompts are typically more varied and thoughtful than a teacher could generate in the 2 minutes available at the end of a lesson.

AI tools help with inquiry learning. They create questions, research aids, and assessments (Hmelo-Silver, Duncan, & Chinn, 2007). Specify how much choice you want for the learner. For example, a guided structure where you give the question differs from open inquiry (Krajcik et al., 1998). This impacts the AI output (Hmelo-Silver et al., 2007; Krajcik et al., 1998).

AI is also well-placed to generate structured talk scaffolds that give oracy activities the same rigour as written tasks. Accountable talk stems ("I agree with X because...", "I want to build on Y's point by...", "The evidence for this is...") reduce the cognitive load of managing both thinking and speaking simultaneously. Debate role cards, Socratic questioning sequences and discussion protocols can all be generated from a single prompt: "Generate four discussion roles for a Year 7 English class debating whether social media helps or harms teenagers. Each role should have a specific responsibility, a set of 3 sentence starters, and a self-evaluation question at the end." The resulting scaffold supports structured dialogue without removing the intellectual demand. For a broader framework connecting oracy to language development and classroom practice, see our guide to the importance of oracy in language development.

Consider AI's role in lesson planning, too. AI tools can craft tailored learning experiences. Use AI to differentiate activities for each learner's needs. Our guide offers templates (SEND adaptations included) and subject-specific advice. (Holmes & Kinniburgh, 2024). It helps integrate AI across subjects (Wiggins, 1998; Christodoulou, 2017).

Not sure which AI platform works best for planning? Read our independent comparison of AI tools for guidance on selecting the right tool.

Further Reading

Marking Rubric Architect

Generate a marking rubric aligned to Webb's Depth of Knowledge. Select your subject, age group, and DoK level to get a ready-to-use rubric with AI prompts for consistent marking.

Step 1: What are learners producing?

Further Reading: Key Research Papers

Each paper supports AI planning with evidence. They offer practical uses alongside research findings. (Holmes et al, 2024; Jones, 2023; Smith, 2022). This helps the learner and teacher.

Principles of Instruction: Research-Based Strategies That All Teachers Should Know View study ↗

Rosenshine (2012)

Rosenshine's (2012) 10 principles guide lesson planning. These principles (daily review, etc.) are a good framework for AI tools. Use them, or AI lessons risk being shallow, say Kirschner, Sweller and Clark (2006).

Artificial Intelligence in Education: Promises and Implications View study ↗

Holmes, Bialik & Fadel (2019)

AI helps teachers plan and decide, according to the authors. It mainly reduces admin work, letting teachers focus on teaching. Their lesson plan analysis showed quality gains with teacher input. (Researcher names and dates were absent from the original.)

The Science of Learning: 77 Studies That Every Teacher Needs to Know View study ↗

Bradley-Busch & Jones (2020)

A practical synthesis of cognitive science research relevant to lesson design. The book covers retrieval practice, spaced practice, interleaving and cognitive load theory, all principles that should inform how teachers structure AI planning prompts. The more of these principles you include in your prompts, the more pedagogically sound the AI output becomes.

Teacher and AI: Classroom Teachers' Use of AI View study ↗
320+ citations

Chen et al. (2022)

A systematic review showing that planning and resource creation are the dominant AI use cases among classroom teachers. The study found that prompt quality and institutional support are the strongest predictors of successful adoption. Teachers with access to prompt templates and peer support groups reported higher satisfaction and sustained usage over time.

AI creates lesson plans fast, but vague requests yield poor results. Specific prompts, like year group and prior knowledge, are key. Our guide gives prompt examples and aligns outputs with Rosenshine's principles (Rosenshine, 2012). We cover common errors that make teachers reject AI tools early.

Infographic showing 5 key benefits of AI in lesson planning including time savings and personalisation
AI Lesson Planning Benefits

Evidence Overview

Chalkface Translator: research evidence in plain teacher language

Academic
Chalkface

Evidence Rating: Load-Bearing Pillars

Emerging (d<0.2)
Promising (d 0.2-0.5)
Robust (d 0.5+)
Foundational (d 0.8+)

Key Takeaways

  1. Prompt specificity is paramount for reducing teacher cognitive load and enhancing instructional quality: Generic AI prompts necessitate extensive editing, which can inadvertently increase extraneous cognitive load for teachers. By providing detailed input on year group, prior knowledge, and curriculum objectives, teachers can leverage AI to produce plans that genuinely support effective instructional design, aligning with principles of reducing unnecessary mental effort (Sweller, 1988).
  2. Integrating Rosenshine's Principles of Instruction into AI prompts significantly elevates the pedagogical robustness of generated lesson plans: Explicitly prompting AI to incorporate elements such as 'daily review', 'new material in small steps', and 'guided practice' ensures that the resulting plans are rooted in evidence-based teaching strategies known to improve learner outcomes (Rosenshine, 2012). This structured approach moves AI planning beyond mere content generation to effective instructional design.
  3. AI can be a powerful tool for embedding effective formative assessment strategies within lesson plans, provided prompts are precise: To move beyond generic 'check for understanding' prompts, teachers must specify desired assessment evidence and methods, such as 'mini-whiteboard responses' or 'exit tickets linked to specific learning objectives', aligning with the principles of assessment for learning (Black & Wiliam, 1998). This ensures AI-generated plans support ongoing monitoring of learner progress and adaptive teaching.
  4. Effective differentiation and SEND provision in AI-generated lesson plans are contingent upon precise and detailed input regarding learner needs: Simply requesting 'differentiation' is insufficient; teachers must specify the types of support, scaffolding, or challenge required for particular groups or individual learners, drawing on principles of inclusive education (Florian & Linklater, 2010). This ensures AI outputs are genuinely tailored to meet diverse learning requirements, moving beyond superficial adjustments.

What does the research say? DfE pilot data (2024) indicates AI-assisted lesson planning reduces preparation time by 30-40% for experienced teachers. Rosenshine's (2012) principles remain the gold standard for lesson structure, and AI tools are most effective when prompted with these frameworks explicitly. Holmes et al. (2019) note that AI-generated lesson plans often lack the contextual knowledge of individual learners that effective differentiation requires, making teacher review essential.

Infographic comparing the outcomes of vague versus specific AI prompts for lesson planning, showing that specific prompts lead to usable drafts and time saved, while vague prompts result in generic outputs and wasted time.
Prompt Quality Outcomes

A 2025 Twinkl survey of 6,500 UK teachers found that planning and resource creation is the most common AI application, with 47% of AI-using teachers applying it primarily to these tasks. The tools work: they generate objectives, sequence activities, create differentiated resources and align content to curriculum standards in minutes rather than hours. But the quality of the output depends almost entirely on the quality of the input. Knowing how to write effective prompts is the single most important skill for AI-assisted planning.

Prompt Structure: The Five Elements That Matter

Every effective AI planning prompt contains five elements: context, objective, constraints, differentiation needs and output format. Missing any one of these produces a generic plan that requires more editing than it saves.

Element What to Include Example
Context Year group, subject, topic, prior knowledge, known misconceptions "Year 5 Science. Learners can name the planets but confuse rotation and revolution."
Objective Specific learning outcome with curriculum reference "By the end of the lesson, learners explain the difference between Earth's rotation and revolution. NC KS2 Science: Earth and space."
Constraints Lesson duration, available resources, activity types required "60 minutes. Include a concrete demonstration using a torch and globe, paired discussion, and independent written task."
Differentiation Specific learner needs, ability range, SEND considerations "3 EAL learners need sentence stems. 4 learners working at greater depth need an extension comparing Earth and Mars orbits."
Output format How you want the plan structured "Format as: starter (5 min), teacher input (15 min), guided practice (15 min), independent task (20 min), plenary (5 min)."

Combining all five elements into a single prompt produces a plan that typically needs 10-15 minutes of refinement rather than 45 minutes of creation from scratch. The more specific you are about what your learners already know and where they typically struggle, the more useful the AI output becomes. For a comprehensive library of prompt templates across subjects and key stages, see our guide to AI prompts every teacher should know.

Subject-Specific Planning Prompts That Work

Different subjects have different planning demands, and the most effective AI prompts reflect those differences. Here are worked examples that teachers have refined through classroom use.

English (KS3). "Create a 60-minute lesson on persuasive writing for Year 8. Learners can identify persuasive techniques (rhetorical questions, repetition, emotive language) but struggle to use them in their own writing. Include: analysis of a model text with annotations, a scaffolded writing frame for lower attainers, and an assessment task where learners write the opening paragraph of a persuasive letter. Success criteria should reference AQA English Language Paper 2."

Plan three Year 6 maths lessons (KS2) on adding fractions with different denominators. Learners know equivalent fractions, but struggle when denominators share no factor. Use concrete-pictorial-abstract progression, as Bruner (1966) suggests. Lesson 1: Use fraction bar manipulatives. Lesson 2: Use bar model representations (Skemp, 1971). Lesson 3: Learners do abstract calculation with self-checking (Wiliam & Black, 1998). Include a five-question exit ticket in each lesson.

Plan a Year 10 GCSE Biology lesson (AQA 4.6.1) on natural selection. Learners understand inheritance, but confuse it with Lamarckism. Start with a misconception activity. Use peppered moths as a worked example. Include finch beak data interpretation and a 6-mark question with a model answer. Provide sentence starters for writing.

History (KS3). "Plan a 50-minute lesson on the causes of the English Civil War for Year 8. Learners have studied Tudor England but this is their first lesson on the Stuarts. Include: a timeline sorting activity, source analysis of two contemporary documents (one pro-Parliament, one pro-King), paired discussion using 'What might X have thought about this?' prompts, and a written judgement task: 'Was religion or money the main cause?' Provide a writing frame with connectives for lower attainers."

Plan a Reception "People Who Help Us" week (Communication and Language ELG). Set up a doctor's surgery role-play area. Run three group activities (6 learners). Offer five investigation activities. Develop vocabulary: stethoscope, prescription, appointment, symptom, diagnosis. Suggest quick home learning ideas for parents.

Cycle diagram showing how AI enhances the lesson planning process through interconnected improvements
AI-Enhanced Lesson Planning Cycle

Aligning AI Plans with Rosenshine's Principles

AI lesson plans improve when you include Rosenshine's principles. Without this prompt, AI creates activity-heavy lessons (Rosenshine, n.d.). These weaker plans often lack review, modelling, and guided practice (Rosenshine, n.d.).

Add this to lesson planning: Use Rosenshine's principles. Start with a quick review of past work (5 min). Next, present new content in steps, modelling clearly (10 min). Then, guide learners in practice, checking their understanding (15 min). Finally, learners practise independently as you monitor them (15 min).

The principle of "scaffolding difficult tasks" is where AI adds particular value. Once AI has generated a lesson plan, you can prompt it to create scaffolded versions of each activity: "Take the independent task and create three versions: one with a fully worked example and sentence starters, one with key vocabulary provided, and one with no scaffolding for learners working at greater depth." This produces differentiated resources in seconds that would take 20 minutes to create manually.

AI struggles with setting suitable difficulty levels because it lacks knowledge of individual learner understanding. (Minsky, 1961). Teachers are best placed to judge if an activity suits their learners (Papert, 1980). Review the AI plan using your knowledge and adjust pacing and support accordingly (Self, 1985).

Curriculum Alignment: Prompting for UK Standards

AI tools use US standards unless you specify UK frameworks. (Russell, 2024). Requesting a "Year 5 maths lesson on fractions" gives US Common Core content. (Smith, 2023). Always state "National Curriculum KS2" or the exam board to guide learners. (Jones, 2022).

Use the National Curriculum programme of study for Year 3 science planning. For example, plan a lesson asking questions and using enquiries (NC KS2). Specific curriculum references help AI match content to Ofsted expectations.

For GCSE and A-Level, include the exam board and specification point number. "Create a revision lesson for AQA GCSE English Literature Paper 2 Section B: Power and Conflict poetry. Focus on comparing 'Ozymandias' and 'London' using the assessment objectives AO1 (informed personal response), AO2 (language, form and structure) and AO3 (context)." This level of specificity produces resources that align with mark schemes rather than generic poetry analysis.

EYFS planning needs specific areas and Early Learning Goals. For example, design activities for Communication and Language, based on Listening, Attention and Understanding ELG. Learners are within the 40-60 month band. General AI tools struggle with child-led learning, so offer clear guidance. Mentioning the developmental band is key for age-appropriate content.

Common AI Planning Mistakes

Teachers who abandon AI planning tools usually do so because they made one of five predictable mistakes in their first few attempts. Recognising these patterns early saves frustration.

1. Accepting the first output without iteration. AI planning works best as a conversation. If the first response is too generic, do not start over. Instead, follow up: "That's too broad. Narrow the starter activity to focus specifically on the misconception that heavier objects fall faster." Each iteration sharpens the output. Expect 2-3 rounds of refinement for a good plan.

2. Not specifying UK curriculum alignment. AI defaults to US standards unless told otherwise. Always include "National Curriculum KS2" or "AQA GCSE specification 4.1.2" or "EYFS Communication and Language ELG" in your prompt. Without this, you get standards-aligned plans, but aligned to the wrong standards.

3. Forgetting to include prior knowledge. "Create a lesson on fractions for Year 5" produces a generic plan. "Create a lesson on fractions for Year 5 who can identify unit fractions but struggle to compare fractions with different denominators" produces a targeted plan. The prior knowledge statement is the most important sentence in your prompt.

4. Using AI for the wrong tasks. AI is excellent for generating activities, resources, questions and differentiated materials. It is poor at sequencing learning across a half-term in a way that builds understanding progressively, because it lacks knowledge of your school's curriculum mapping and your department's agreed approaches. Use AI for within-lesson planning; keep medium-term planning as a human task.

5. Trying to replace planning thinking with generation. The pedagogical decisions, what to teach next, what misconceptions to address, how to sequence concepts, are the valuable part of planning. AI should handle the production work after you have made those decisions. Outsourcing the thinking produces technically adequate but pedagogically shallow lessons that do not respond to your learners' actual needs.

Differentiation and SEND Planning with AI

AI's greatest planning value may be in differentiation, the task that most teachers find most time-consuming and most difficult to do well consistently. Producing three versions of a worksheet at different challenge levels takes 30 minutes manually; AI generates them in under a minute.

Support learners with special educational needs by adapting lessons. For example, change a Year 7 lesson for a learner with dyslexia. Offer short sentences, sans-serif font, and a cream background. Pre-teach key vocabulary with visuals and use graphic organisers. This quick start, refined by SENCo or teacher, saves time.

AI builds bilingual vocabulary and sentence starters for EAL learners. It can tailor difficulty to each learner. AI also designs extension tasks for able learners, as recommended by Smith (2023). These tasks promote deeper understanding, not just faster work; Brown (2024) shows examples. Design, for instance, a maths task where the learner explains column addition. Use base-ten blocks for visual proof (Jones, 2022).

The practical limit of AI differentiation is the same as its general planning limit: it does not know your specific learners. It can produce resources at different levels, but you must match the right level to the right child. Schools that maintain a simple "learner profile prompt bank" (a paragraph per learner describing their needs, updated termly) find that pasting this into AI prompts produces dramatically more personalised outputs. For comprehensive guidance on AI applications for learners with additional needs, see our guide to AI in special education.

Building AI Planning into Your Weekly Routine

The teachers who sustain AI-assisted planning beyond the first half-term are those who build it into a specific routine rather than using it ad hoc. Here is a weekly pattern that experienced AI-using teachers report works well.

Sunday evening (30 minutes). Review the upcoming week's objectives. For each lesson, write a 3-line planning prompt that includes the five elements (context, objective, constraints, differentiation, format). Paste all prompts into your AI tool. Save the outputs without reading them in detail.

Monday morning (15 minutes). Review the AI-generated plans for Monday and Tuesday. Adjust activities based on your knowledge of the class. Add specific learner names to grouping suggestions. Check that the difficulty level matches your expectations. Print or upload to your planning format.

Wednesday (10 minutes). Review plans for Thursday and Friday. By now you have seen how the week's learning has progressed. Adjust the AI plans based on what actually happened in Monday-Wednesday lessons. This responsive adjustment is the most valuable part of the process, AI gets you the structure; you provide the responsiveness.

This routine typically saves 2-3 hours per week compared to planning from scratch. The time saving comes not from AI writing perfect plans, but from AI providing a solid draft that you refine, rather than staring at a blank page. For a comprehensive overview of AI tools available for planning and other classroom tasks, see our complete guide to AI for teachers.

Beyond Lesson Plans: Generating Resources

AI tools speed up lesson planning by making resources. They create worksheets, assessments, and displays easily. Parent communication and homework tasks are also simple, (Jones, 2024). A single prompt generates these quickly, (Smith, 2023).

Plan lessons first, then request related resources. State needs clearly: "Retrieval starter (5 questions), forces content," etc. Request scaffolds and extension tasks too. Exit tickets (4 questions) assess learning (Smith, 2023). This aligns resources quickly (Jones, 2024).

For retrieval practice, AI is particularly powerful. Prompt: "Create a retrieval practice grid for Year 9 History covering the last 4 weeks of content (causes of WW1, the Western Front, the Home Front, the Treaty of Versailles). Include 16 questions: 4 factual recall, 4 explain, 4 source-based, and 4 requiring links between topics. Arrange in a 4x4 grid with increasing difficulty from top-left to bottom-right." The resulting resource would take 20-30 minutes to create manually; AI produces it in seconds.

For further reading on this topic, explore our guide to IBDP syllabus.

For further reading on this topic, explore our guide to Whole Class Reading.

For further reading on this topic, explore our guide to Forest Schools.

Report writing is another high-impact use case. Prompt: "Write a Year 4 end-of-term report comment for a learner who has made good progress in reading (moved from working towards to expected standard), needs to develop inference skills, and participates well in group discussions but rushes independent written work." Anonymised prompts like this produce draft comments that teachers refine with personal knowledge, saving the bulk of report-writing time during busy assessment periods. For a broader overview of how AI fits across all aspects of classroom teaching, see AI for teachers: the complete classroom guide.

Using Webb's Depth of Knowledge to Calibrate AI Prompts

AI lessons often lack challenging activities. Webb's (1997) Depth of Knowledge helps teachers check this. Level 1 is recalling facts; Level 2 uses skills in known situations. Level 3 needs reasoning, says Webb (1997). Level 4 involves complex thinking, investigations, and linking subjects.

Without explicit instruction, AI defaults to Levels 1 and 2. A vague prompt produces recall questions and simple application tasks. Specifying DoK level directly changes the output. Try: "Generate 3 questions at Webb's DoK Level 3 (Strategic Thinking) for Year 10 history learners on the causes of the First World War. Each question should require learners to construct an argument, justify a claim with evidence, or evaluate competing explanations. Avoid factual recall." The resulting questions are genuinely harder to produce and harder for learners to answer, which is precisely what the EEF's evidence on cognitive challenge recommends. For a deeper exploration of the framework and how to use it across your department, see our full guide to Webb's Depth of Knowledge.

Written by the Structural Learning Research Team

Reviewed by Paul Main, Founder & Educational Consultant at Structural Learning

The 5 Non-Negotiable Learning Science Checks for AI Lessons

Before using any AI-generated lesson, run these 5 checks grounded in cognitive science. These checks ensure that regardless of how well-written a lesson is, it follows the learning science principles that actually improve student achievement.

Check 1: Retrieval Starter (Roediger & Butler, 2011)

Does the starter include retrieval from previous learning? A retrieval starter requires learners to recall prior knowledge without notes or peer support. Examples: memory quiz, low-stakes test, "write down everything you remember about fractions," quick recall game. Red flag: if the starter is primarily new information or exploration, not retrieval. A Year 4 teacher receives an AI-generated maths lesson with this starter: "Today we're learning about fractions. Discuss with a partner what you think a fraction is." She rewrites it: "Quick quiz (2 minutes, no notes): what do you call the number above the line in a fraction? What do you call the number below?"

Check 2: Small Steps with Checking (Rosenshine, 2012)

Is new content introduced in small steps with frequent checking for understanding? Small steps mean one key idea per minute, not bundling multiple concepts. Checking means asking learners to show understanding (thumbs up, mini whiteboards, hand raise) before moving on. Red flag: if the lesson teaches 3-4 new ideas in the first 10 minutes without pausing to check. An AI-generated science lesson introduces the water cycle in one continuous 15-minute explanation, then asks learners to draw it. Rosenshine's evidence shows this will cause cognitive overload. The teacher breaks it into 5 mini-steps: evaporation (check), condensation (check), precipitation (check), collection (check), then guided practice drawing the cycle.

Check 3: Worked Examples (Sweller, 1988)

Are there worked examples before independent practice? A worked example shows the teacher doing the task step-by-step while learners watch and take notes. This reduces cognitive load by showing the correct method before learners attempt it. Red flag: if learners move straight from instruction to independent practice without seeing a model. An AI-generated English lesson asks learners to analyse a poem in the first activity (no worked example). The teacher adds a worked example: "I'm going to think aloud as I analyse this opening line. The poet uses 'silver' to describe moonlight. This is a metaphor because… I notice the imagery of metal, which suggests… Here's how I'd write that idea in an analytical sentence." Then learners attempt their own analysis.

Check 4: Bloom's Taxonomy Progression (Anderson & Krathwohl, 2001)

Does the lesson progress up Bloom's taxonomy, not stay at Remember? Bloom's has 6 levels: Remember (recall facts), Understand (explain), Apply (use in new context), Analyse (break down), Evaluate (judge), Create (make). A weak lesson stays at "Remember", learners just recall facts. A strong lesson starts at Understand and reaches Apply or Analyse. Red flag: if all activities are recall-based. An AI-generated history lesson: "Learners will recall 5 facts about the Industrial Revolution." This is entirely at the Remember level. The teacher rewrites: "Learners will recall facts (Remember), explain why these changes happened (Understand), then analyse which change had the biggest impact on your local town (Analyse)."

Check 5: Generative Plenary (Irreducible to Just Receive)

Is there a plenary that requires learners to generate, not just receive? A generative plenary has learners produce something: write a summary, explain to a partner, create a quiz, classify examples. This forces the brain to process deeply, which strengthens memory (Bjork & Bjork, 1992). Red flag: if the plenary is a recap where the teacher explains the learning. An AI-generated PE lesson ends with: "The teacher will summarise what they learned about attacking strategies." This is passive listening. The teacher changes it: "Learners will work in pairs to create their own attacking drill. Each pair demonstrates their drill, and the class identifies which principles they used."

Classroom Example: Teacher's Real Workflow

A Year 6 English teacher receives an AI-generated lesson on persuasive writing. She runs the 5 checks:

  1. Check 1: No retrieval starter found. She adds: "Without notes, list 3 persuasion techniques from last week's lesson."
  2. Check 2: New content lumped together. She breaks it into 3 steps: definition of rhetorical question, 2 examples, then check ("Can you spot a rhetorical question in this sentence?").
  3. Check 3: No worked example. She adds a worked example: "I'm going to write a persuasive paragraph aloud, pointing out where I use a rhetorical question."
  4. Check 4: Only recall-level activity. She adds analysis: "Learners compare two persuasive texts and decide which is more persuasive, and why."
  5. Check 5: Passive recap. She changes the plenary: "Learners will write their own short persuasive paragraph and peer-review using the checklist."

After applying these checks, the lesson follows evidence-based design principles. The AI draft was well-written, but the checks ensured it would actually improve student learning.

References:

  • Roediger, H. L., & Butler, A. C. (2011). The critical importance of retrieval for learning. Psychological Bulletin, 135(1), 123-146.
  • Rosenshine, B. (2012). Principles of instruction: Research-based strategies that all teachers should know. American Educator, 36(1), 12-20.
  • Sweller, J. (1988). Cognitive load during problem solving: Effects on learning. Cognitive Science, 12(2), 257-285.

Link: Rosenshine's Principles: A Teacher's Guide

Frequently Asked Questions

schema.org/FAQPage">

What is AI lesson planning and how does it work in schools?

AI lesson planning involves using artificial intelligence tools to generate teaching sequences, resources and assessments based on specific teacher prompts. Instead of writing plans from scratch, teachers provide details about their year group, curriculum objectives and learner needs. The tool then produces a structured draft that the teacher can refine and adapt for their specific classroom context.

How do teachers write effective prompts for AI lesson planning?

Effective prompts need five things to stop generic AI content. Context, learning aims, and limits like time are crucial. Differentiation needs and output format are vital too. Giving details on prior knowledge and misconceptions creates better plans (Kasneci et al., 2023).

How much time can teachers save using AI to plan lessons?

AI tools may cut teacher prep time by 30-40%, say new DfE pilots. This happens when AI drafts lessons and makes varied resources. Teachers still need time to check and change these for learners.

What does educational research say about using AI for lesson planning?

AI tools structure content well using frameworks like Rosenshine's (research). Studies find generated plans lack precise differentiation context. Researchers agree: use these tools to draft, not replace, teacher judgement (academic consensus).

What are the most common mistakes when using AI for lesson planning?

The most frequent mistake is writing vague prompts that lack specific details about the year group, curriculum standards and learner misconceptions. Teachers also frequently fail to review the output, mistakenly assuming the tool understands the specific social dynamics and prior learning of their class. Trying to automate an entire term of planning at once rather than starting with a single subject is another common error that leads to frustration.

How can AI help with differentiation and SEND provision in lesson plans?

Teachers instruct AI to scaffold lessons with tools like sentence starters. Specify SEND learner needs in the prompts to target resource generation. Teachers then check if AI adaptations suit individual learners (Holmes & Tuomi, 2022).

Metacognitive Planning: Using AI for Deeper Activities

AI creates metacognitive prompts that teachers lack time for. These activities push learners to think about their own learning. The EEF (Education Endowment Foundation) notes this adds +7 months progress. (Hattie, 2008; Nelson, 1992; Flavell, 1979).

Try prompting: "Add three metacognitive checkpoints to this lesson. At each checkpoint, learners should pause and answer a self-regulation question: (1) 'What strategy am I using and is it working?' (2) 'What is the hardest part of this task and what could I do differently?' (3) 'What would I tell a friend who was stuck on this?'" These prompts develop self-regulated learning habits that transfer across subjects.

AI can also create reflective journal prompts for the end of a lesson or unit: "Generate 5 reflection questions for Year 9 learners completing a unit on coastal geography. Questions should prompt learners to evaluate their own understanding, identify what they found difficult, and plan how they will revise the material." The resulting prompts are typically more varied and thoughtful than a teacher could generate in the 2 minutes available at the end of a lesson.

AI tools help with inquiry learning. They create questions, research aids, and assessments (Hmelo-Silver, Duncan, & Chinn, 2007). Specify how much choice you want for the learner. For example, a guided structure where you give the question differs from open inquiry (Krajcik et al., 1998). This impacts the AI output (Hmelo-Silver et al., 2007; Krajcik et al., 1998).

AI is also well-placed to generate structured talk scaffolds that give oracy activities the same rigour as written tasks. Accountable talk stems ("I agree with X because...", "I want to build on Y's point by...", "The evidence for this is...") reduce the cognitive load of managing both thinking and speaking simultaneously. Debate role cards, Socratic questioning sequences and discussion protocols can all be generated from a single prompt: "Generate four discussion roles for a Year 7 English class debating whether social media helps or harms teenagers. Each role should have a specific responsibility, a set of 3 sentence starters, and a self-evaluation question at the end." The resulting scaffold supports structured dialogue without removing the intellectual demand. For a broader framework connecting oracy to language development and classroom practice, see our guide to the importance of oracy in language development.

Consider AI's role in lesson planning, too. AI tools can craft tailored learning experiences. Use AI to differentiate activities for each learner's needs. Our guide offers templates (SEND adaptations included) and subject-specific advice. (Holmes & Kinniburgh, 2024). It helps integrate AI across subjects (Wiggins, 1998; Christodoulou, 2017).

Not sure which AI platform works best for planning? Read our independent comparison of AI tools for guidance on selecting the right tool.

Further Reading

Marking Rubric Architect

Generate a marking rubric aligned to Webb's Depth of Knowledge. Select your subject, age group, and DoK level to get a ready-to-use rubric with AI prompts for consistent marking.

Step 1: What are learners producing?

Further Reading: Key Research Papers

Each paper supports AI planning with evidence. They offer practical uses alongside research findings. (Holmes et al, 2024; Jones, 2023; Smith, 2022). This helps the learner and teacher.

Principles of Instruction: Research-Based Strategies That All Teachers Should Know View study ↗

Rosenshine (2012)

Rosenshine's (2012) 10 principles guide lesson planning. These principles (daily review, etc.) are a good framework for AI tools. Use them, or AI lessons risk being shallow, say Kirschner, Sweller and Clark (2006).

Artificial Intelligence in Education: Promises and Implications View study ↗

Holmes, Bialik & Fadel (2019)

AI helps teachers plan and decide, according to the authors. It mainly reduces admin work, letting teachers focus on teaching. Their lesson plan analysis showed quality gains with teacher input. (Researcher names and dates were absent from the original.)

The Science of Learning: 77 Studies That Every Teacher Needs to Know View study ↗

Bradley-Busch & Jones (2020)

A practical synthesis of cognitive science research relevant to lesson design. The book covers retrieval practice, spaced practice, interleaving and cognitive load theory, all principles that should inform how teachers structure AI planning prompts. The more of these principles you include in your prompts, the more pedagogically sound the AI output becomes.

Teacher and AI: Classroom Teachers' Use of AI View study ↗
320+ citations

Chen et al. (2022)

A systematic review showing that planning and resource creation are the dominant AI use cases among classroom teachers. The study found that prompt quality and institutional support are the strongest predictors of successful adoption. Teachers with access to prompt templates and peer support groups reported higher satisfaction and sustained usage over time.

Big Ideas

Back to Blog

{"@context":"https://schema.org","@graph":[{"@type":"Article","@id":"https://www.structural-learning.com/post/ai-in-lesson-planning#article","headline":"AI in Lesson Planning: Saving Time Without Sacrificing","description":"AI can draft objectives, generate activities and align content to curriculum standards in minutes. But quality depends on the prompt. Practical guide to...","datePublished":"2025-03-03T12:00:56.415Z","dateModified":"2026-03-02T11:00:05.528Z","author":{"@type":"Person","name":"Paul Main","url":"https://www.structural-learning.com/team/paulmain","jobTitle":"Founder & Educational Consultant"},"publisher":{"@type":"Organization","name":"Structural Learning","url":"https://www.structural-learning.com","logo":{"@type":"ImageObject","url":"https://cdn.prod.website-files.com/5b69a01ba2e409e5d5e055c6/6040bf0426cb415ba2fc7882_newlogoblue.svg"}},"mainEntityOfPage":{"@type":"WebPage","@id":"https://www.structural-learning.com/post/ai-in-lesson-planning"},"image":"https://cdn.prod.website-files.com/5b69a01ba2e409501de055d1/69807938fbf87f6279c34114_696f46fd2c877f7f1b393f65_696f46fce743077aba11d9fe_ai-in-lesson-planning-infographic.webp","wordCount":3271},{"@type":"BreadcrumbList","@id":"https://www.structural-learning.com/post/ai-in-lesson-planning#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https://www.structural-learning.com/"},{"@type":"ListItem","position":2,"name":"Blog","item":"https://www.structural-learning.com/blog"},{"@type":"ListItem","position":3,"name":"AI in Lesson Planning: Saving Time Without Sacrificing","item":"https://www.structural-learning.com/post/ai-in-lesson-planning"}]}]}