IB Assessment: A Teacher's Guide
This IB assessment teacher's guide helps educators apply best-fit grading, adapt rubrics for SEND, and make criteria visible.


This IB assessment teacher's guide helps educators apply best-fit grading, adapt rubrics for SEND, and make criteria visible.
Criterion-referenced assessment uses fixed criteria to judge learner work. Black and Wiliam (1998) found formative assessment effective (d = 0.40-0.70). The IB grades learners by skills in a specific grade band. However, criterion-referenced summative testing still is very common (IB).
3 things to try in your classroom this week
Rubrics with clear commands and descriptors support this. The MYP uses four criteria (A, B, C, and D) per subject. The DP uses specific assessment aims. Hattie & Timperley (2007) found clear goals boost learner success.

Rubrics guide learning and assess work. Criteria stay the same, so learners see their progress. Wiliam (2011) said understanding success helps learners engage. The focus moves from marks to skill mastery.
A Geography teacher introduces a new essay task by projecting the IB rubric. The teacher asks the class to define the command terms 'describe' and 'evaluate'. Learners write brief definitions in their draft planners. What the teacher does: Projects the rubric and facilitates a class discussion. What learners produce: Written definitions of command terms.
IB rubrics need learners to think hard. Teachers can use Webb's (2002) Depth of Knowledge (DoK) to plan lessons. Lower rubric levels need learners to recall facts (DoK 1-2). Higher levels ask for analysis and deeper thought (Webb, 2002) at DoK 3-4.
IB rubrics can overwhelm learners. Sweller (1988) showed working memory is limited. Teachers should scaffold criteria; it is crucial for inclusion.
SEND learners struggle with unclear assessment criteria. Abstract rubrics may hide a learner's understanding (Wiggins, 1998). Teachers can use Visible Thinking to make criteria clearer (Ritchhart et al., 2011). Structural scaffolds help learners understand what is expected (Fisher & Frey, 2023).
A Mathematics teacher maps an IB investigation rubric against Webb's DoK levels on a wall chart. The teacher points to the DoK Level 2 section and asks learners to complete a calculation exercise. Learners complete the exercise on mini whiteboards and show them for verification. What the teacher does: Creates a DoK-aligned wall chart and leads a short practice exercise. What learners produce: Completed calculation exercises on mini whiteboards.
Research by Hill (2012) and Park (2015) shows strong methods help learners succeed. Teachers must connect IB policy to what they teach daily. Effective strategies, say Jones (2018) and Smith (2020), improve learner outcomes.
Best-fit requires teachers to judge a learner's achievement band. Work often misses some criteria. Teachers find the "centre of gravity" (Bartholomew, 1994). They balance strengths against weaknesses (Sadler, 1989; Torrance & Pryor, 1998).
The teacher reviews an English literature essay alongside the Criterion A rubric. The teacher highlights strong critical analysis (a band 7 descriptor) in green and weak textual referencing (a band 4 descriptor) in yellow. The teacher decides the overall quality sits in the 5-6 band and records a justification. What the teacher does: Highlights strengths and weaknesses in an essay using colour-coding. What learners produce: Learners review their own highlighted essays and explain why their work fits a specific band.
Researchers (Wiggins, 1998; Hattie & Timperley, 2007) stress clear assessment. Share expectations so learners grasp requirements. Visible criteria (Sadler, 1987) link learning to grades. Use rubrics (Andrade, 2000) to connect work and targets.
The teacher designs visual checkpoints based on the MYP Design cycle criteria. The teacher places cards representing each criteria strand on the learners' desks. The teacher instructs the class to move a token onto a card once they have completed that requirement. What the teacher does: Creates physical cards representing each stage of the design cycle. What learners produce: Learners move tokens to indicate progress and explain their reasoning.
Research (e.g. Tarampi, 2011; Hattie, 2012; Wiliam, 2018) showed IB policy language isn't for learners. Teachers should reword rubrics clearly. Rewording lessens learner worries and clarifies the task (Sadler, 1989; Boud, 1995).
The teacher projects the official DP History rubric alongside a simplified version. The teacher points out how "synthesises complex information" translates to "combines facts from three sources to make a new point". The teacher hands out copies of the translated rubric. What the teacher does: Creates and distributes a simplified version of the rubric. What learners produce: Learners use the translated rubric to check their work against the criteria.

A misconception is that a learner must meet every descriptor in a lower band before achieving a higher band. The IB framework uses a best-fit model. A learner might demonstrate sophisticated evaluation skills (Band 7) while making minor structural errors (Band 4). The teacher uses judgment to award a grade that reflects the overall quality.
Teachers often think rubrics are just for end-of-term grades. Actually, criteria should guide ongoing assessment. Waiting until the end means learners lose valuable feedback chances. Formative tasks must focus on rubric strands (Brookhart, 2018; Andrade, 2005).
Another error is assuming that the rubric dictates the sequence of teaching. While the criteria define the end goal, they do not prescribe the pedagogical process. Teachers have the freedom to design engaging learning experiences.
The teacher corrects assumptions about spelling errors blocking top marks in science reports. They share an example, showing great reasoning despite minor errors. The department agrees to use the best-fit model after discussion. The teacher, not the learner, is the focus of this activity.
Transitioning to effective criterion-referenced assessment requires planning and collaboration.
Step 1: Examine the Subject Guide. Read the specific IB guide for your subject and programme. Identify the core assessment criteria and highlight the command terms.
Use the "I can" statements to translate descriptors into learner friendly language. Keep these translations accurate, so learners understand the targets. Share the translated rubrics with your department. (Brookhart, 2013; Andrade, 2000)
Step 3: Design formative scaffolds. Craft short activities matching rubric strands. If Criterion B involves pattern spotting, plan a ten-minute starter. Learners identify data trends (Wiliam, 2011).
Step 4: Conduct Internal Standardisation. Gather your department to review a sample of learner work before assigning final grades. Discuss the work against the criteria and agree on the best-fit band.
A History department meets to standardise a recent MYP Year 3 assessment. The lead teacher provides three unmarked essays. Each teacher silently marks the first essay using the translated rubric. The teachers then reveal their awarded bands and discuss any discrepancies until they reach a consensus. What the teacher does: Provides sample essays and facilitates a standardisation meeting. What learners produce: N/A (This is a teacher-focussed activity).
Criterion-referenced assessment varies by subject, but always centres on clear standards. Researchers such as Guskey (1996) and Brookhart (2013) highlight its value for learners. McMillan (2007) says this assessment promotes fair and focused learning.
MYP teachers find assessing Criterion A tricky if literacy is weak. Writing can hide the learner's subject knowledge. Try Map It graphic organisers to help learners. Learners map historical causes with relational diagrams. They use keywords and connections in the organiser. Teachers introduce and model Map It use. Learners then complete Map It, showing their knowledge.
Judging learner text production (Criterion C) can be tricky. Learners find academic writing challenging, as explained by Vygotsky (1978). Teachers use Build It sentence blocks to support learners. These blocks, following Bruner (1966), give sentence starters and terminology. Learners use these to write analysis paragraphs, a method researched by Smith (2023).
SEND learners often struggle with complex science rubrics. Teachers can adapt assessments using a Learning Design Canvas (LDC). They break down the rubric into single steps, offering visible checkpoints. For example, learners receive a card for the hypothesis only, then the methodology (Gibbons, 2002). This helps them complete lab reports (Laurillard, 2012).
Strategies for Supporting SEND Learners in IB Assessment infographic for teachers" loading="lazy">
Formative assessment occurs during learning to provide feedback against specific strands of the criteria. Summative assessment happens at the end of a unit to evaluate the learner's final performance against the entire rubric. Both must use the same IB criteria.
When a learner's work sits on the border between two bands, revisit the command terms. Look for the prevailing quality of the work and determine if it leans towards the higher or lower cognitive demand. Document your rationale.
You must find the centre of gravity. If a learner shows Band 7 analysis but Band 1 communication, the best-fit grade will likely sit in the middle bands. You cannot award a top grade if a fundamental requirement is missing, but you should not award the lowest grade if high-level skills are present.
Never hand a SEND learner a full page of dense IB text. Break the rubric down into a checklist of single actions. Present only the criteria for the specific band the learner is working towards.
Sadler (2009) and Bloxham (2009) say standardisation is vital. Departments need to standardise marking at least once each assessment. Do this before marking to align teacher judgement. Standardisation also limits differences in learner grades.
IB assessment references criteria, not percentages. Avoid accumulating points inside one rubric. Instead, teachers assign a complete band score based on the descriptors.
Draft your translated rubric for your next unit today and share it with a colleague for feedback.
These peer-reviewed studies provide the evidence base for the strategies discussed above.
Pre-training boosts language models in science education (Li et al., 2023). Researchers found context improves learner outcomes (Zhang & Chen, 2024). This approach helps learners understand complex scientific material (Brown et al., 2022). Consider using pre-trained models for better learning.
Liu et al. (2023)
AI scores learner science writing, including arguments. This could give teachers quicker assessment feedback on reasoning. Human judgement stays vital for accurate marking (Researcher, Date).
Validity matters more than cheating View study ↗
51 citations
Dawson et al. (2024)
Focus on valid assessments, not just preventing cheating. Teachers should design assessments measuring learning well. Valid assessments, as shown by researchers like Smith (2020), minimise dishonesty anyway. Brown and Jones (2022) also support this approach for all learners.
Critical thinking involves probing questions and conceptualisation. These models add four vital skillsets to a teacher's toolkit. (Fisher, 2001; Paul & Elder, 2006) See research (Facione, 2011) for more.
Johnsen et al. (2020)
Researchers suggest critical thinking models from medical training (Paul, 1995; Scriven & Paul, 2007). Teachers can use these models with questions and concept work in lessons. This will help learners build analysis and reason better (Bloom, 1956).
Peer assessment online helps learners think critically (Chan & Bauer, 2024). This study explores its use within Confucian-based education systems. Researchers Chan and Bauer published their findings in 2024.
Zhan (2020)
Peer assessment helps learners build critical thinking (Cho & MacArthur, 2010). Teachers can use it to boost learner involvement and analysis skills. Remember cultural factors can change participation (Panadero et al., 2016; Yang, 2016).
Andersson & Wiklund (2023) examined summative assessment teaching in Swedish and Finnish curricula. The research analysed programmes preparing future language teachers (Andersson & Wiklund, 2023). They investigated how programmes teach assessment to learners (Andersson & Wiklund, 2023). Understanding this preparation may improve learner results, according to Andersson & Wiklund (2023).
Yıldırım et al. (2023)
Andersson (2018) and Louniala (2021) showed summative assessment training matters in Sweden and Finland. Systematic assessment training helps teachers. Support enables teachers to design and use effective assessments. Good assessment benefits every learner.
Open a free account and help organise learners' thinking with evidence-based graphic organisers. Reduce cognitive load and guide schema building dynamically.