Table of Contents
Teaching Kids to Evaluate AI Output: The Critical Thinking Skill Schools Skip
Kids can prompt AI. They can't reliably evaluate what it gives back. Here's what lateral reading is, why it matters, and how to teach it at home.
Your kid figured out how to prompt ChatGPT. They can coax it into writing a story in the voice of a pirate, summarize a chapter in bullet points, or explain photosynthesis three different ways until one clicks. That’s real. It’s also only half the skill. The other half — the one most schools are not teaching — is knowing what to do once the AI gives something back. How do you check it? How do you know when it’s wrong? How do you know when it’s subtly wrong, which is far more dangerous than obviously wrong? Teaching kids to evaluate AI output may be the most important critical thinking skill of the next decade, and it’s largely missing from K-12 curricula.
The Problem: Fluency Without Verification
Here is the gap. Children are learning to interact with AI tools at a pace that has surprised researchers, educators, and parents alike. Common Sense Media’s 2025 survey found that a majority of teenagers use AI tools at least weekly, and roughly a third use them for schoolwork on a regular basis. The tools are available, easy to use, and often genuinely helpful.
What isn’t keeping pace is the evaluative side. A 2021 study by Breakstone, McGrew, Smith, Ortega, and Wineburg published in PNAS tested 3,446 high school students across 14 U.S. states on their ability to evaluate online information. The results were striking: most students could not reliably distinguish between credible and unreliable sources. They evaluated websites based on visual appearance, they trusted official-looking logos, and they read sources from the top down rather than verifying them laterally against independent references.
That study was conducted before large language models became household tools. The problem it identified has only sharpened since. AI outputs have a specific characteristic that makes them harder to evaluate than ordinary web pages: they are fluent. They read well. They are structurally coherent and grammatically confident. They arrive without the visual signals — shaky formatting, unfamiliar domains, missing bylines — that sometimes alert a careful reader to distrust a webpage. An AI that is wrong doesn’t look wrong. It looks like everything else the AI generates, which is usually right.
This is what CISA (the Cybersecurity and Infrastructure Security Agency) flagged in its 2024 guidance on AI and misinformation: AI-generated content is becoming harder to distinguish from human-generated content, and that difficulty extends to factual claims. The risk isn’t limited to deep fakes and synthetic media. It includes the confident, incorrect summary a child submits as homework.
The ISTE 2024 AI literacy standards identify “evaluate AI output critically” as one of six core competencies for AI-literate students. The problem is that standards and curricula are two different things. A standard says what students should be able to do. A curriculum tells teachers how to teach it, and when, and through what activities. As of 2024, very few districts have a clear, structured curriculum for AI output evaluation. Parents are largely on their own.
What the Research Actually Says
The most useful body of research on this topic comes not from AI studies — those are still young — but from civic online reasoning research developed over the past decade at the Stanford History Education Group (SHEG), led by Sam Wineburg.
Wineburg and his colleagues spent years studying how different people evaluate online information. Their 2022 paper in Science, co-authored with Joel Breakstone and Mark Smith, introduced a framework that distinguished two approaches: vertical reading and lateral reading.
Vertical reading is what most students do. They open a source, read it carefully from top to bottom, and evaluate it based on what it says about itself — the “About” page, the credentials listed, the tone. This approach is intuitive and feels thorough. It is also unreliable. A source that says it is credible may not be. Internal evaluation of a source using only the source itself cannot detect bias, error, or fabrication that the source doesn’t acknowledge.
Lateral reading is what professional fact-checkers do. Instead of reading a source vertically, they immediately open new tabs and search for what others say about the source. They look for independent corroboration before they commit to reading the source in depth. They treat the source as a claim to be verified rather than a document to be understood.
Wineburg’s research found that professional fact-checkers were dramatically faster and more accurate than university professors and students at evaluating the reliability of sources — not because they knew more, but because they used lateral reading. The expertise being tested wasn’t subject-matter knowledge; it was epistemic method.
McGrew et al.’s 2018 study in Social Education applied this framework specifically to students and found that lateral reading could be taught. High school students who received explicit instruction in lateral reading outperformed control groups in source evaluation tasks, including on novel sources they hadn’t encountered in training.
The translation to AI output evaluation is direct but requires one additional step. With a webpage, lateral reading means checking who created the source. With AI output, the source is the model itself — and the model doesn’t know whether it’s wrong. So lateral reading of AI output means: identify the specific claims made, and then verify those claims against independent sources before treating them as facts.
This is different from the commonly-taught skill of “checking your sources.” It’s more specific and more active. It’s not about citing references — AI tools now generate references, often hallucinated ones. It’s about independently verifying the content of the claims, not trusting the fluency of the presentation.
The WEF’s 2024 AI literacy framework for education makes this point explicitly: AI literacy is not just the ability to use AI tools effectively; it includes the ability to recognize the limitations of those tools and the conditions under which their outputs should not be trusted. Verification is identified as a core literacy component, not an advanced or optional one.
| Evaluation Approach | What It Checks | Reliable for AI Output? | Time Required |
|---|---|---|---|
| Vertical reading | Internal consistency, writing quality, stated credentials | No — AI is fluent even when wrong | Low |
| Trusting citations | Whether references exist | No — AI hallucinates citations | Low |
| Spot-check search | Whether one other source agrees | Sometimes — depends on source chosen | Medium |
| Lateral reading | What multiple independent sources say about the claim | Yes — most reliable method | Medium–High |
| Domain expert review | Whether a knowledgeable person confirms the claim | Yes — most reliable overall | High |
What to Actually Do
The good news: lateral reading is a skill, and skills are teachable. It doesn’t require any special technology. It requires practice and a shift in default behavior — from “read and trust” to “read, then check.”
Start with the “one weird fact” test
Pick an AI response about any topic your child knows something about and look for one specific factual claim — a date, a name, a number, a causal relationship. Then ask: how would you check whether that’s true? Don’t look at the AI again. Open a search engine and find two sources that independently confirm or deny that specific claim. This exercise is not about distrusting AI. It’s about building a verification reflex. Do it once a week with any AI-generated content, until it becomes automatic.
Teach the difference between confident and correct
AI tools express uncertainty inconsistently. Sometimes they hedge (“This may vary depending on…”). Often they don’t. A claim delivered without hedging is not a verified claim — it’s a fluent output. Help your child notice this by pulling up an AI response and asking: how confident does the AI sound? Now how confident should we be? What would it take to find out for sure? The goal is to disconnect the perceived confidence of a statement from the child’s internal confidence in that statement.
Introduce the hallucinated-citation drill
This is uncomfortable but necessary. Ask an AI tool to provide references for a claim it made. Then — and this is the key step — look up those references. Verify that the journal, the author, the year, and the finding described all correspond to a real publication. Many will not. Children who do this exercise once tend to remember it permanently. The experience of finding a confident, well-formatted citation that doesn’t exist is more instructive than any lecture on AI limitations. For related context on how AI affects kids’ information habits, see the article on AI writing and kids’ brain learning.
Practice claim decomposition
Fluent AI text obscures how many distinct claims it contains. Train your child to read an AI paragraph and list every verifiable claim in it as a separate bullet point. A single paragraph might contain five or six distinct factual assertions. Treating each as individually needing verification is different from evaluating the paragraph as a whole, which reads coherently and therefore feels reliable. This is a writing and reading skill as much as an AI skill — it transfers to evaluating any complex source.
Build the lateral reading habit specifically
Teach the phrase: “What do other sources say about this?” Not “is there another source that agrees?” but “what do independent sources say?” The distinction matters. Finding agreement doesn’t verify a claim if both sources drew from the same original. True lateral reading means finding independent corroboration — sources that arrived at the same conclusion through different methods or evidence. Show your child what it looks like to open three tabs and check, rather than reading one source deeply. For more on how to build AI reasoning skills in kids, the article on teaching kids to use AI as a thinking partner covers the complementary skill.
Use structured skepticism, not blanket distrust
The goal is calibrated trust, not reflex suspicion. Some AI outputs should be trusted readily — the boiling point of water, the year a famous novel was published. Others require verification — political facts, scientific claims in contested areas, statistics, historical causation. Teach children to ask: what kind of claim is this, and how much independent verification does it need? Building this meta-cognitive habit is more durable than any list of AI weaknesses, because it adapts as AI tools change.
What to Watch for Over the Next 3 Months
Pay attention to how your child responds when you ask where they got a piece of information. If the answer is consistently “the AI said so” without any secondary verification, that’s the gap this piece is addressing. It’s not a moral failing — it’s the default behavior the tools encourage, because fluent, confident output doesn’t signal the need for checking.
Watch also for how schools describe their AI policies. Most current AI policies focus on whether students used AI, not on how they evaluated what it produced. A school that bans AI use entirely is addressing the easier problem. A school that teaches students to verify AI output is addressing the harder and more durable one.
The ISTE and WEF frameworks both anticipate that AI output evaluation will become a standard literacy benchmark — the way source citation became a standard research skill in the 1990s. The window in which parents can get ahead of this, before it’s tested or formally assessed, is probably the next two to three years. Children who have already built the lateral reading habit before it’s required will have a structural advantage.
Frequently Asked Questions
How is evaluating AI output different from regular source evaluation?
Standard source evaluation asks whether a source is credible. AI output evaluation asks whether the claims in the output are accurate — which requires lateral reading (checking claims against independent sources), not just assessing the credibility of the AI itself. The AI is always “credible” in the sense of being a known, widely-used tool. That tells you nothing about whether any specific output is accurate.
At what age can kids start learning lateral reading?
Breakstone and Wineburg’s research has been applied to students as young as middle school (grades 6–8) with strong results. The core concept — “look at what other sources say” — can be introduced even earlier, in elementary school, in simplified form: “let’s check if two other places say the same thing.”
What if my child’s school is actively promoting AI use without teaching verification?
This is the current state at most schools. You can raise the question directly with teachers or administrators — specifically asking whether the AI tools being used include instruction in output verification. In the meantime, the home practices described in this article are sufficient to build the habit without school support.
Should kids distrust AI entirely?
No. Distrust that is generalized and non-specific doesn’t build useful skills. Calibrated skepticism — knowing which claims to verify and how — is more useful than either blanket trust or blanket distrust. Many AI outputs are accurate. The goal is to be able to tell the difference.
How do I know if my child has actually learned this?
Test it. Give them an AI-generated paragraph and ask them to find one claim that might be wrong. Then watch their process. Do they search independently? Do they look for more than one confirming source? Do they report back with specific evidence rather than “I didn’t find anything wrong”? The process is the skill — the outcome of the search is secondary.
Is lateral reading a skill that transfers beyond AI?
Yes — this is one of the strongest findings in Wineburg’s research. Students who learn lateral reading as an AI-evaluation skill apply it to evaluating social media posts, news articles, and research summaries. It is not AI-specific; it is an epistemic habit that makes students better consumers of all information.
About the author
Ricky Flores is the founder of HiWave Makers and an electrical engineer with 15+ years of experience building consumer technology at Apple, Samsung, and Texas Instruments. He writes about how kids learn to build, think, and create in a tech-saturated world. Read more at hiwavemakers.com.
Sources
- Breakstone, J., McGrew, S., Smith, M., Ortega, T., & Wineburg, S. (2021). “Students’ civic online reasoning: A national portrait.” PNAS, 118(39). https://doi.org/10.1073/pnas.2108870118
- Wineburg, S., Breakstone, J., & Smith, M. (2022). “Lateral reading and the nature of expertise: Reading less and learning more when evaluating digital information.” Science, 374(6572). https://doi.org/10.1126/science.abm8613
- McGrew, S., Ortega, T., Breakstone, J., & Wineburg, S. (2018). “The challenge that’s bigger than fake news: Civic reasoning in a social media environment.” Social Education, 82(6), 316–322.
- ISTE. (2024). AI Literacy Standards for K-12 Students. International Society for Technology in Education.
- CISA. (2024). AI and Misinformation: What You Need to Know. Cybersecurity and Infrastructure Security Agency.
- World Economic Forum. (2024). AI Literacy Framework for Education. WEF Global Coalition for Digital Safety.
- Common Sense Media. (2025). AI Use Among Teens and Tweens: 2025 Survey Report.