Why Choosing the Next Problem Matters More Than the Perfect Explanation
Penn study insight: the next problem often drives more learning than a better explanation.
If you have ever watched a student nod through a polished explanation and then freeze on the very next question, you have seen the core problem this guide addresses. In teaching practice, the difference between a memorable lesson and a meaningful learning gain is often not the elegance of the explanation; it is whether the next problem is chosen well. That is the central lesson to draw from the Penn study on AI tutoring: when practice stays inside a student’s zone of proximal development, learning improves more reliably than when every student receives the same clean, linear sequence. For tutors, teachers, and AI tutor designers, this shifts the question from “How do I explain it better?” to “What should the student attempt right now?”
This matters because students usually cannot accurately diagnose their own gaps. As the Penn researchers noted, a student may feel that a chatbot or tutor is “personal” simply because it answers their question, but personalization goes deeper than response style. It requires adaptive sequencing—the deliberate selection of practice problems that are neither too easy nor too hard. For practical guidance on how technology and instruction can work together, see our piece on the teacher’s roadmap to AI and the broader discussion of agentic AI systems that act with human-quality judgment. The key insight is simple: explanations teach ideas, but problem selection builds capability.
1) What the Penn Study Suggests About Learning Gains
1.1 The study design in plain English
In the Penn experiment summarized by the Hechinger Report, close to 800 high school students in Taiwan learned Python with the same AI tutor. The critical difference was not the tutor’s voice, tone, or sophistication of explanation. Half the students got a fixed problem sequence, while the other half received a personalized sequence that changed based on performance and interaction data. That distinction is important because it isolates sequencing as the key variable. In other words, the study asked whether what comes next can matter more than how well the tutor explains the current topic.
The result was striking: the personalized group performed better on the final exam. The article reported the gains as equivalent to several months of additional schooling, though the researchers themselves cautioned that any conversion of statistical effects into “months” is approximate. Even with that caution, the message is robust enough for teaching practice: better calibration of difficulty can produce significant learning gains without changing the core content. For a related look at how educational systems evaluate impact, compare this with our guide on what metrics matter when moving from AI pilots to outcomes.
1.2 Why this is more than an AI story
It would be a mistake to treat the Penn study as proof that AI tutors are magical. In fact, the broader evidence on chatbots is mixed, and some studies suggest students can become overly dependent on them or get spoon-fed answers. The bigger lesson is not “use AI” but “optimize practice selection.” That principle applies equally to human tutoring, classroom instruction, homework design, and digital learning platforms. A strong teacher who assigns the wrong next problem can blunt progress just as surely as a weak tutor can. For more on evaluating instruction quality, see our rubric for hiring great instructors.
That is why adaptive sequencing deserves more attention than explanation polish. A beautifully explained concept can still fail if the student has not yet practiced enough near the edge of current competence. Conversely, a moderately concise explanation paired with the right sequence of practice can generate durable understanding. This is especially relevant in subjects like physics, math, and programming, where conceptual knowledge and problem-solving fluency must develop together. If you want another example of learning through structure and progression, see our piece on solving puzzles like a pro—the mechanics differ, but the learning logic is similar.
1.3 The zone of proximal development is the engine
The study’s adaptive design works because it keeps learners in the zone of proximal development—the sweet spot where a task is challenging but still reachable with effort, hints, or partial scaffolding. If the next problem is too easy, the student rehearses what they already know and gets little growth. If it is too hard, the student experiences failure without enough productive struggle. The best next problem stretches, but does not shatter, current understanding. This is the instructional equivalent of progressive overload in training: enough resistance to force adaptation, not so much that performance collapses.
For tutors, this means the real skill is not delivering a flawless explanation once; it is continuously adjusting the task set based on evidence. That can be a student’s error pattern, response time, hint usage, confidence rating, or the kinds of steps they skip. Many modern systems already collect these signals, but too often they are used only to explain the last mistake. Better tutoring uses those signals to choose the next question. For an adjacent example of structured adaptation, see the gaming-to-real-world pipeline, where practice environments teach by sequencing challenges, not just by narrating rules.
2) Why Explanations Alone Often Plateau Learning
2.1 Explanation is necessary, but not sufficient
Students need explanations. Without a clear model, they are forced to guess at meaning, memorize procedures blindly, or copy examples without comprehension. But explanation has a ceiling: after the student understands the idea at a basic level, additional polish produces diminishing returns. At that point, learning depends more on retrieval, application, and transfer than on listening. This is why a tutor can spend ten minutes giving a brilliant lecture and still fail to move the needle if the student never practices the exact decision point that matters.
In physics and other quantitative subjects, this shows up constantly. A student may understand Newton’s laws when they are discussed in the abstract, but still fail to decide which force diagram to draw or which variable to solve for. The issue is not missing prose; it is missing decision practice. For more on turning explanations into usable methods, visit our guide on instructional quality signals and our practical article on AI adoption in the classroom.
2.2 Students overestimate what they know
One of the Penn study’s most useful insights is that students often cannot tell what they do not know. This is a major reason sequencing beats explanation polish. Students may ask for help on the thing they already suspect is hard, while the actual obstacle sits one layer earlier or later in the skill chain. A student in algebra might request help with factoring when the real issue is identifying the correct equation structure. A physics student may ask for a formula, but the real bottleneck is representing the motion correctly. Personalized practice reveals these hidden gaps faster than conversation alone.
This pattern mirrors findings across education and other complex decision domains: people tend to ask for help at the moment of awareness, not at the root cause of difficulty. That is why adaptive systems need an external selector, not just a responsive explainer. If you are interested in how editors and teams systematize better selection decisions, our guide on building a citation-ready content library offers a useful analogy: good systems make the next action easier to choose.
2.3 The polished-explanation trap
There is a seductive trap in teaching practice: assuming that if a student still struggles, the explanation simply was not polished enough. Sometimes that is true, but often the issue is timing. The student may need a more targeted sequence of attempts, not a more elaborate lecture. In that sense, explanation is like opening a door, while sequencing is like choosing the hallway beyond it. If the hallway is wrong, the door’s craftsmanship does not matter much.
This is also why “coverage” is not the same as “competence.” A teacher can cover a topic beautifully and still leave students unable to execute on an exam. The most effective teaching strategy is usually a blend of concise explanation, then tightly calibrated practice, then feedback. For more on measuring educational impact rather than assuming it, see Measure What Matters and our analysis of moving from AI pilots to repeatable outcomes.
3) How Adaptive Sequencing Works in Practice
3.1 The feedback loop
Adaptive sequencing is a simple idea with a powerful loop: assign a problem, observe how the student handles it, infer the level of mastery, and choose the next task accordingly. That next task might be slightly easier to rebuild confidence, at the same level to reinforce a weak spot, or slightly harder to extend the skill. This is how a good tutor operates in real time, even without software. The major difference with an AI tutor is scale: the system can adjust after every response and do it consistently.
The Penn study’s core insight is that this loop can outperform a fixed path. A fixed sequence assumes all learners should progress in the same order and at the same pace, which is rarely true in real classrooms. Adaptive sequencing respects differences in prior knowledge, speed, and error patterns. If you are comparing instructional systems, this is similar to how achievement systems in games work: the next challenge is matched to readiness, not just calendar order.
3.2 What data you actually need
You do not need a sophisticated lab to practice adaptive sequencing. You need enough information to estimate readiness. The most useful signals are surprisingly ordinary: whether the student solved independently, whether they used a hint, whether they made a conceptual or procedural error, how long they took, and whether they can explain the answer afterward. A tutor watching a student solve two or three problems can often infer more than from a long conversation. The trick is to record the observations and act on them deliberately.
For schools and tutoring teams, this suggests a practical workflow. Use one problem to probe baseline understanding, one to confirm the likely barrier, and one to stretch. That small sequence often reveals more than a lengthy explanation. For a broader systems view of adapting to changing conditions, see how to build an internal news and signals dashboard and how to spot long-term topic opportunities.
3.3 What good sequencing feels like to students
Well-chosen sequencing feels less like being “tested” and more like being guided through a staircase. Students should sense that each problem is doable with effort and that the difficulty is rising for a reason. That emotional experience matters because it reduces avoidable frustration and boredom at the same time. When the next problem is correctly chosen, students often stay engaged longer and ask better questions because the task itself reveals what they need to know.
This is one reason personalized learning can outperform static homework sets. Students do not need endless variation; they need the right variation. That distinction is subtle but critical. For more on designing learning tools that respect human attention and momentum, see our article on autonomous assistants that respect standards.
4) A Simple Decision Framework Tutors Can Use
4.1 The 4-question framework
When assigning practice, tutors can use a simple decision framework that keeps sequencing focused on growth rather than perfection. Ask four questions: Can the student start? Can the student finish with effort? What kind of mistake is most likely? What should the next problem teach? If the answer to the first two is no, the task is too hard. If the answer to the first two is yes and the student finishes with no struggle, it is too easy. The ideal task usually produces a visible but solvable challenge.
The third question prevents generic grading. A student who misses a problem because they misread a word needs a different next step than one who lacks the underlying concept. The fourth question ensures the next item has a purpose beyond repetition. If you want a practical rubric for evaluating those teaching choices, our guide on what great instructors do differently is a helpful companion.
4.2 The three-level assignment rule
For everyday tutoring, use a three-level rule: one problem at current level, one slightly below, one slightly above. Start with the current level to diagnose independence. If the student succeeds quickly, move up. If they fail, step down just enough to rebuild the chain of reasoning. This keeps the student within the zone of proximal development while preserving momentum. The idea is not to avoid challenge but to shape challenge so it remains productive.
This rule works especially well in physics tutoring, where prerequisite skills often determine success more than the headline topic. If a student cannot manipulate units, a kinematics problem that “should” be easy may still be too hard. In that case, the best next problem is not a fancier explanation of displacement; it is a targeted units or graph-reading task. For more on structuring practice around skill ladders, see our coverage of problem-solving strategies in games.
4.3 A tutor’s decision tree
Use this compact decision tree during live sessions. If the student solved the last item independently, raise difficulty slightly or increase novelty. If they solved it with one hint, keep the same level and change the context. If they made the same error twice, move down one step and isolate the subskill. If they became discouraged, assign a quick win to restore confidence before returning to challenge. This approach is simple enough to use on paper and sophisticated enough to drive meaningful personalized learning.
Below is a quick comparison of common tutoring moves and their likely effects:
| Tutoring move | Best used when | Likely effect on learning | Risk |
|---|---|---|---|
| Longer explanation | Student lacks initial conceptual frame | Improves clarity and reduces confusion | Diminishing returns if overused |
| Slightly harder next problem | Student is comfortable and accurate | Expands transfer and stamina | Can overwhelm if jump is too large |
| Slightly easier next problem | Student is stuck or discouraged | Rebuilds confidence and momentum | Can become under-challenging |
| Same-level new context | Student knows the method but not the application | Strengthens flexible retrieval | May still hide the root issue |
| Hinted problem with reflection | Student needs scaffolded independence | Improves self-correction and metacognition | Dependency if hints are too generous |
For teams building systems around decision quality, our guide to not applicable is not used here; instead, think of this as an instructional version of choosing the right next move in a game or workflow. For broader context on AI-enabled assistance, see the AI operating model playbook and outcome-based AI.
5) Why This Matters for Physics Teaching in Particular
5.1 Physics is a sequencing subject
Physics is often taught as if success depends mainly on understanding the formula sheet. In reality, most difficulty comes from sequence decisions: which principle applies, which representation helps, and which simplification is legitimate. That makes physics an ideal subject for adaptive sequencing because students rarely fail for only one reason. They may need conceptual clarity, visual reasoning, algebra, or graph interpretation in different combinations. A single polished explanation cannot address all those prerequisites at once.
When tutoring physics, the next problem should test the next decision, not merely the next chapter heading. If a student can solve constant-acceleration numericals but struggles when the graph changes, the next assignment should focus on graph-to-equation translation. If they can calculate but cannot set up free-body diagrams, the next problem should emphasize representation. For additional support in making teaching more inclusive and rigorous, see our article on sensitivity and rigor in the classroom.
5.2 Worked examples still matter
This article is not an argument against worked examples. It is an argument against stopping at the explanation. A worked example is often the perfect way to introduce a new idea, but after one or two examples, the student needs a smart next problem to convert observation into performance. The best teachers use explanation to lower the entry barrier and sequencing to create growth. Both matter, but sequencing is usually the more leverage-rich decision once basic comprehension exists.
In a physics context, that might look like: explain Newton’s second law, walk through one force-diagram example, then assign a nearly identical problem with a twist. That twist could be friction, an incline, or a two-body system. The twist is where learning lives. For more on skill-building through structured progression, explore sim-based skill development and achievement-style progression.
5.3 Exam prep depends on readiness, not exposure
Students often mistake “I saw this before” for “I can do this under exam conditions.” Adaptive sequencing corrects that illusion by forcing retrieval in the right difficulty band. Good exam prep should therefore alternate among retrieval, transfer, and timed performance. A tutor who assigns the perfect explanation but the wrong practice set may leave a student feeling reassured without becoming more capable. By contrast, a slightly uncomfortable but well-chosen problem can produce the exact cognitive effort that strengthens recall and speed.
This is especially valuable for AP, A-level, and college courses where problem variation is the norm. Students need to learn not just the topic but the decision procedure for unknowns. If you are shaping a tutoring program, check our guide to hiring great instructors and our analysis of whole-class AI adoption for implementation ideas.
6) How to Turn This Into Better Tutoring Strategy
6.1 Diagnose before you explain
The fastest way to improve tutoring effectiveness is to diagnose with a problem before launching into a lecture. Give a short task that reveals the student’s current model, then respond to the evidence. This often shortens sessions because you avoid explaining material the student already understands. It also prevents the common mistake of over-teaching the wrong subskill. The best tutors are not those who talk the most; they are those who ask the most informative questions.
Diagnostic problems should be short and revealing. Ideally, they should expose whether the student understands concepts, procedures, representations, or vocabulary. For a related approach to collecting and acting on signals, our piece on signals dashboards shows how structured observation improves decisions.
6.2 Use a “struggle budget”
Students should struggle, but not for so long that the session becomes demoralizing. A practical way to manage this is to think in terms of a “struggle budget.” If a student hits multiple dead ends on one task, you can reduce complexity, give a hint, or switch to a precursor problem. The goal is to preserve productive friction while avoiding useless friction. This balance is the heart of adaptive sequencing.
In practice, that means not treating every error as proof the student needs more explanation. Sometimes they need a smaller step. Sometimes they need a different representation. Sometimes they need a problem that removes one variable and isolates the real obstacle. For additional insight into managing tradeoffs in systems, see our buyer’s guide on balancing battery life, portability, and power—a different domain, but the same logic of choosing the right constraint.
6.3 End each session with a calibrated exit ticket
A strong tutoring session ends with a final problem that is slightly easier than the hardest item attempted, but not so easy that it is meaningless. This “exit ticket” tests whether the student can now perform the target move with less support. It gives both tutor and student a concrete signal of readiness. If the exit ticket is too easy, you learn little. If it is too hard, you learn only that more scaffolding is needed.
The point is to leave the student with a sense of progress and a clear next step. That next step might be another problem at the same level, a review of a prerequisite, or a transfer task in a new context. For more on designing strategic next steps, see build your own dashboard and navigating a new market with good decision rules—again, different field, same principle.
7) Common Mistakes Tutors Make When Choosing the Next Problem
7.1 Mistake: Always moving on too quickly
Some tutors assume that if a student completed a problem, they are ready for a harder one. But correctness alone is not mastery. The student may have used a hint, guessed, or relied on recognition rather than retrieval. If you move on too quickly, you skip the consolidation step and create fragile knowledge. That is one reason students can seem strong in a session and weak on a test.
Instead, use success quality as the criterion. Did the student solve it independently? Did they explain the method? Could they transfer it to a new context? The answer determines the next problem better than right-or-wrong status alone. For a wider view on avoiding superficial metrics, see Measure What Matters.
7.2 Mistake: Overcorrecting with long explanations
When students make mistakes, tutors often jump into a longer and more detailed explanation. That can help, but it can also flood working memory and turn the session into passive listening. In many cases, a shorter clarification plus a targeted new attempt produces better retention. The student learns by doing the repaired version, not by hearing an extended monologue about the repair.
Think of explanation as a tool for re-entry, not a replacement for practice. You want the student back in action quickly so the feedback loop can continue. This is one reason adaptive systems can be powerful: they reduce the distance between diagnosis and the next chance to succeed. For more on designing intelligent support systems, see agentic assistants with editorial standards.
7.3 Mistake: Ignoring emotional state
Sequencing is not just cognitive; it is emotional. A student who feels defeated is less able to benefit from even the best next problem. Sometimes the right pedagogical move is to temporarily lower difficulty to restore momentum. Other times, it is to give a small but meaningful win, then re-raise the challenge. Good tutors watch for frustration, disengagement, and overconfidence, because these states affect the usefulness of the next task.
If you want a broader perspective on how human factors shape performance, see our articles on accessible trails and adaptive gear and wearables and home diagnostics, both of which show how design choices can either support or impede user success.
8) A Practical Workflow for Teachers and AI Tutor Designers
8.1 Build problem banks by difficulty bands
Instead of organizing practice only by topic, organize it by difficulty bands and subskills. For each concept, create a few easy checks, several medium “growth” items, and a few stretch problems. This allows tutors and AI systems to select the most appropriate next item based on current performance. It also makes personalization easier because the tutor is choosing among known options rather than inventing a new question from scratch.
This mirrors how good content systems work in other fields: when categories are clear, decision quality improves. For an instructive parallel, see citation-ready content libraries and internal signals dashboards. The structure does not replace judgment, but it makes judgment faster and more consistent.
8.2 Let the system adapt, but keep teacher control
AI tutors are most useful when they handle recommendation, not authority. The system can surface likely next problems based on performance data, but teachers should still define the learning goals, prerequisite map, and acceptable jump sizes. In other words, use AI to support sequencing, not to abdicate it. That balance preserves pedagogical intent while gaining the efficiency of personalization.
For institutions adopting these tools, the implementation question is not “Can the AI explain well?” but “Can it choose well?” That is the deeper lesson from the Penn study. A responsive tutor that chooses the wrong next step is still not enough. For a broader implementation mindset, see the AI operating model playbook and the teacher’s roadmap to AI.
8.3 Track learning gains, not just satisfaction
It is easy to overvalue student satisfaction because it is immediate and visible. But a student can feel pleased by a helpful explanation and still not improve much. If you care about learning gains, you need post-assessment evidence, transfer tasks, and delayed recall checks. The Penn study is compelling precisely because it looked beyond comfort to exam performance. That is the standard tutoring should strive for.
To measure whether sequencing is working, compare not only accuracy but also speed, independence, and retention over time. If you need a framework for impact evaluation, our guide on what matters in AI outcomes is a strong model.
Conclusion: The Best Explanation Is Often the Next Right Problem
The Penn study gives tutors and educators a valuable correction: the most powerful lever is not always the most polished explanation. Often, the bigger gains come from choosing the right next problem—one that keeps the student inside the zone of proximal development and forces productive effort. That is the essence of adaptive sequencing, and it is one of the most practical ideas in modern teaching practice. The student who is always given the perfect explanation may feel supported; the student who is given the perfect next problem becomes more capable.
For tutors, the takeaway is actionable. Diagnose first, classify the error, choose the nearest useful challenge, and adjust after each response. This is how personalized learning becomes more than a buzzword and starts producing real learning gains. Whether you are teaching physics, programming, or exam strategy, the best question to ask after every explanation is not “Was that clear?” but “What should they try next?”
If you want to keep building your instructional toolkit, explore our guides on evaluating instructors, adopting AI in teaching, and designing progression systems that keep learners moving forward.
Frequently Asked Questions
What does “adaptive sequencing” mean in tutoring?
Adaptive sequencing means choosing each next practice problem based on how the student just performed. Instead of giving everyone the same fixed set, the tutor adjusts difficulty, context, and subskill focus so the student stays in the zone of proximal development. The goal is to keep tasks challenging enough to promote growth without causing overwhelm. In practice, that means each problem is selected for its instructional value, not just its topic label.
Why can the next problem matter more than a better explanation?
A great explanation can improve understanding, but learning gains usually depend on what the student does after the explanation. If the next problem is too easy, too hard, or unrelated to the real gap, the student may not consolidate the skill. The Penn study suggests that careful problem selection can outperform a fixed sequence even when the tutor’s explanation quality stays the same. That is because practice converts understanding into usable ability.
How can a tutor tell whether a problem is in the student’s zone of proximal development?
Look for evidence that the student can start the problem, make progress with effort, and finish with limited help. If the student is completely stuck, the task is likely too hard; if they solve it instantly without strain, it may be too easy. Error type also matters: a conceptual error suggests a different next step than a careless one. A good fit usually produces productive struggle rather than confusion or boredom.
Can AI tutors really improve learning gains?
Yes, but not automatically. The evidence is mixed, and some AI tutoring tools can backfire if they give away too much or encourage passive dependence. The promising part of the Penn study was not that the tutor “explained better,” but that it adapted the sequence of practice problems. In other words, AI helps most when it supports strong instructional decisions, especially problem selection.
What is the simplest framework tutors can use right away?
Use a four-question check: Can the student start? Can they finish with effort? What kind of mistake are they making? What should the next problem teach? Then assign one problem at current level, one slightly easier if needed, or one slightly harder if readiness is clear. This keeps sessions responsive and prevents over-teaching or under-challenging. Over time, it improves both confidence and mastery.
Does this approach work only for STEM subjects?
No. It is especially visible in STEM because the skills are concrete and the feedback is immediate, but the same logic applies in writing, languages, history, and test prep. Any subject with prerequisites and transfer demands benefits from choosing the right next task. The details change, but the principle is the same: practice should match readiness. That is what makes learning gains more likely.
Related Reading
- What the AI Index Means for Creator Niches - Learn how to spot durable opportunities as AI shifts education and content discovery.
- The AI Operating Model Playbook - See how teams move from experiments to repeatable instructional outcomes.
- Beyond Test Scores: A Rubric to Hire Great Instructors for Test Prep - A practical lens for evaluating teaching quality beyond charisma.
- Measure What Matters - A metrics framework you can adapt to tutoring and student progress.
- The Teacher’s Roadmap to AI - A step-by-step look at adopting AI without losing instructional control.
Related Topics
Daniel Mercer
Senior SEO Content Strategist
Senior editor and content strategist. Writing about technology, design, and the future of digital media. Follow along for deep dives into the industry's moving parts.
Up Next
More stories handpicked for you
Parent Tutors: Turn Your Subject Knowledge into Flexible Income Without Burning Out
From Brick‑and‑Mortar to Platform Play: What New Oriental’s Product Mix Means for Test‑Prep Providers
The Physics Behind Demand: Solving for Price Dynamics in Agriculture
Understanding the Impact of Global Markets on US Agriculture: A Physics Perspective
Digital Minimalism: Apps That Boost Focus and Learning Efficiency
From Our Network
Trending stories across our publication group