Why facilitation changes assessment

You can't use a recall quiz
to assess reasoning.

Formative assessment in a facilitated classroom looks different from a lecture-based one — not because the assessment instruments are different per se, but because the learning objectives are different. A lecture-based lesson often has a recall objective. A facilitated lesson typically has a reasoning objective: students should be able to apply, analyse, or evaluate content.

A multiple-choice quiz that tests recall of the day's content produces accurate data about whether students can remember what was covered. It produces no data about whether they can reason with it. For a facilitated lesson that aimed at reasoning, this is the wrong data source — a student can score 100% on recall and still be unable to construct an argument, apply the concept to a new case, or identify the assumptions in a position.

💡Backwards design applied to assessment
The formative assessment instrument for a facilitated lesson must require the same cognitive operation as the learning objective. If the objective was to construct an argument, the formative check must ask for an argument. This is backwards design (C2/A1) applied to assessment design — the assessment is calibrated to the objective, not to the content covered.
Four formative instruments for facilitated classrooms

Each assesses reasoning,
not recall.

1
The reasoning-visible exit question
Requires visible reasoning, not a conclusion

Not 'What was the most significant cause of WWI?' (which can be answered with a memorised conclusion) but 'Choose one cause of WWI and explain why you think it was more significant than the other two we discussed today. Address the strongest counterargument.' This question cannot be answered correctly without visible reasoning — the structure of the argument is assessable, not just the conclusion.

What to look for
Can the student make a claim, support it with specific evidence, and address an alternative? The quality of the reasoning is visible in the structure of the response — not in whether the student picked the 'right' cause.
2
The misconception reveal
Shows whether the student has the correct conceptual model

'A student argues that X. What is wrong with their reasoning?' This format requires students to identify an error in a presented position — which reveals whether they have the correct mental model or are themselves holding the same misconception. It is diagnostic in a way that a correct-answer question cannot be.

Why this is more diagnostic
A student who answers 'The student is wrong because X' correctly may have the right answer for the wrong reason. A student who can explain precisely where the reasoning in the presented argument fails has demonstrated a deeper conceptual grasp — they understand not just the correct position but why the incorrect one fails.
3
The transfer question
Novel context application

Present a context not covered in the lesson that requires the same concept. A facilitated lesson on economic principles of monopoly: 'Apply today's analysis to the following market situation [novel case].' Students who understood the principle can do this. Students who memorised the lesson's examples cannot.

Design requirement
The novel context must be genuinely new — not a variant of a lesson example with different numbers. If students can answer by substituting figures into a memorised structure, the question is testing recall dressed as application.
4
The self-assessment reasoning check
Student-identified quality of their own reasoning

'Look at your contribution to today's discussion. What was the strongest part of your reasoning? What assumption in your argument were you least certain about?' This produces metacognitive data — students who can identify the weakest link in their own argument are developing the critical thinking capacity that facilitation aims for.

What this reveals
A student who identifies a specific weakness in their own argument ('I assumed that market share implies pricing power, but I didn't check whether that's always true') is demonstrating more sophisticated reasoning than one who says 'I think I explained it well.' The self-assessment is itself a reasoning task.
Acting on facilitation formative data

Reasoning gaps require different responses
from recall gaps.

When an exit ticket reveals that most students cannot recall a fact, the next lesson begins with a retrieval practice activity. When an exit ticket reveals that most students can make a claim but cannot support it with specific evidence, the next lesson begins with a worked example of evidence-based argumentation — not content re-teaching.

The reasoning gap is not filled by re-explaining the content; it is filled by modelling the reasoning process more explicitly, then creating another opportunity to practise it. This is the agile teaching loop (P6) applied to reasoning objectives: identify the specific reasoning failure, design an activity that addresses that specific failure, run it, check again.