Skip to main content
All Industries

Education

Khan Academy's Khanmigo and Duolingo Max with GPT-4 have put AI tutors in front of millions of students. Turnitin's AI detection is generating false positives that get innocent students flagged. The "homework is dead" debate has moved from think pieces to policy — some school districts banning ChatGPT, others mandating it. The deeper disruption is economic: if Khanmigo can teach calculus for free and Coursera is pivoting to AI-first credentials, the $50K/year university value proposition is under serious pressure. The institutions and EdTech builders that survive are those redesigning what "learning" means in an AI-native world.

Education industry
Overview

Education is the industry where AI capability created a crisis before the infrastructure for responsible deployment existed. Khanmigo and Duolingo Max are in production at scale. Turnitin is generating false positives. School districts are writing ChatGPT policies. The EdTech builders navigating this well are the ones who recognized that the homework completion problem is an assessment design problem — not a detection problem — and started redesigning what learning evidence looks like in an AI-native world.

···

AI Tutors: The Evidence Gap

The category is real. Khanmigo, Duolingo Max, and a dozen funded AI tutoring startups have millions of users. The product experience is genuinely better than nothing for many students. The harder question is durability: does AI tutoring produce knowledge that transfers and persists, or does it produce in-session correct answers without durable learning? The honest answer is that the evidence for knowledge-tracing-based adaptive systems (DreamBox, ALEKS) is stronger than for LLM-first tutors. Knowledge tracing models what the student knows; LLMs model what a helpful response sounds like. The distinction matters for learning outcomes.

The practical implication: EdTech products that combine LLM fluency with knowledge tracing rigor are building on defensible ground. Products that are GPT-4 with a system prompt that says "be a tutor" are not differentiated and will not produce outcomes that survive scrutiny as the evidence base develops.

Redesigning Assessment for an AI-Native World

The institutions handling the AI academic integrity challenge well are not those with the best detection tools. They are those that redesigned what assessment evidence looks like. If the task is completable by AI without student engagement, it is the wrong task. The shift is toward competency demonstration — work that requires a student to show their thinking, defend their reasoning, and apply knowledge in contexts that cannot be fully delegated to an AI that has not done the course.

Assessment Designs Resistant to AI Delegation
  • Oral defense components requiring students to explain reasoning and respond to follow-up questions in real time
  • Iterative work products where instructors review drafts and require explanation of revision decisions
  • In-class supervised problem-solving that cannot be delegated to an AI running outside the room
  • Portfolio assessments evaluated on growth trajectory and process, not final product quality alone
  • Competency demonstrations requiring application of course-specific knowledge to novel scenarios

Integrating AI Into Educational Workflows

01
LTI 1.3 Integration First

Any EdTech AI tool must integrate through LTI 1.3 to be deployable in Canvas, Blackboard, or Moodle without IT-level intervention. Build to the standard, not to individual LMS APIs — institutions will not accept tools that require custom IT work for each deployment.

02
FERPA and COPPA Data Classification

Classify every data element the AI system touches: directory information, education records, or non-FERPA data. In K-12, treat COPPA as the binding constraint — district-level data use agreements are required before any student data reaches an AI inference endpoint.

03
Learning Analytics via xAPI

Use xAPI to record learner activity in a Learning Record Store. This creates a portable, standards-compliant audit trail of learning interactions analyzable for outcome improvement without being tied to a single LMS.

04
Outcome Measurement Before Scale

Deploy AI tutoring tools with a measurement framework before scaling: pre/post knowledge assessments, retention testing at 30 and 90 days, comparison against control groups. EdTech that cannot demonstrate outcome improvement will face increasing institutional skepticism as the evidence base matures.

Domain Challenges
01

The AI cheating crisis has no clean technical solution — Turnitin's AI detection has documented false positive rates that have accused innocent students, and GPT detectors are unreliable enough that they cannot be used for disciplinary action without legal risk

02

FERPA (and COPPA for under-13 users) creates strict data privacy requirements for student records — AI systems that process student performance data require careful compliance architecture; the COPPA constraint in K-12 is particularly restrictive

03

The digital divide is being amplified by AI: Khanmigo, Duolingo Max, and AI writing tools are accessible to students with devices and internet; students without them fall further behind

04

LMS fragmentation: Canvas, Blackboard, Moodle, and Google Classroom have different APIs, data models, and integration capabilities — multi-platform EdTech requires substantial integration investment

05

Teacher adoption is the hardest problem in EdTech — tools that add to administrative burden rather than reducing it will not be used regardless of pedagogical merit, regardless of how impressive the demo is

06

The unbundling of universities accelerated by AI is a structural threat to institutions — free or cheap AI tutoring, AI-native skill credentials, and employer skepticism of degrees is compressing the perceived value of traditional higher education

What Sets It Apart

FERPA and COPPA create data privacy requirements specific to student data that general-purpose AI infrastructure does not address by default — the K-12 COPPA constraint alone rules out most standard AI API deployments without a district data use agreement

The cheating crisis is primarily an assessment design problem, not a detection problem — Turnitin false positives are already generating institutional liability; detection-first approaches are losing this battle

The digital divide argument is real and getting louder in K-12 policy — EdTech AI that assumes device and broadband access is building for the already-advantaged half of the student population

The unbundling threat to universities is structural — AI tutors, employer-recognized AI-native credentials, and free content from Coursera and edX are compressing the perceived ROI of a four-year degree

Teacher adoption requires reducing administrative burden first — lesson planning AI, IEP documentation AI, and grading assistance have faster adoption paths than student-facing learning AI

Domain Insights
01

AI Tutors Are Mainstream — the Question Is Whether They Produce Learning

Khanmigo has millions of users. Duolingo Max with GPT-4 is in production. The category is real and the adoption is happening. The harder question is whether AI tutors produce durable learning or just fluent-feeling sessions. The evidence for knowledge-tracing-based adaptive tutoring (DreamBox, ALEKS) is stronger than for general-purpose LLM tutors — because knowledge tracing systems model the student's actual knowledge state, while LLMs optimize for producing responses that feel helpful. The EdTech products that will hold up to scrutiny are those combining LLM fluency with knowledge tracing rigor. Products that are just GPT-4 with a "tutor" system prompt are not.

Education — AI Tutors Are Mainstream — the Question Is Whether They Produce Learning
02

Turnitin False Positives Are an Institutional Liability Problem

Turnitin's AI detection has generated documented false positives — students whose original work was flagged as AI-generated, leading to disciplinary proceedings. The false positive rate for AI detection tools is high enough that using them as evidence in academic misconduct proceedings creates legal exposure for institutions. The University of California system and others have issued guidance against using AI detection outputs as standalone evidence. The practical implication for EdTech builders: do not build on AI detection as a core product capability. The liability is moving toward the institutions and tools that make false positive-driven accusations.

Education — Turnitin False Positives Are an Institutional Liability Problem
03

Competency-Based Assessment Is the Durable Response to AI Completion

The institutions that are genuinely solving the AI academic integrity problem are not investing in better detection — they are redesigning assessment around competency demonstration rather than task completion. Oral defenses, in-class problem solving, portfolio assessments graded on process and revision history, and practical demonstrations of skill are all difficult to AI-complete in the same way that a take-home essay is not. EdTech tools that help educators design competency-based assessments, manage oral defense scheduling, and grade portfolios efficiently are building on the right side of this shift. Tools that rely on AI-generated content detection are building on sand.

Education — Competency-Based Assessment Is the Durable Response to AI Completion
Common Pitfalls
01

Building AI integrity detection into the product as a core feature — Turnitin false positive exposure and the fundamental unreliability of AI detection at disciplinary evidentiary standards makes this a liability, not a value-add

02

Collecting student data beyond what FERPA authorizes and using it for product improvement without proper consent — triggers enforcement and destroys institutional trust in ways that are very hard to repair

03

Building EdTech that does not integrate through LTI — requires IT involvement for every deployment and creates an adoption barrier that compounds at scale across a fragmented district landscape

04

AI tutoring systems optimized for session engagement rather than knowledge retention — time-on-platform and correct-answer-rate in-session are not valid proxies for learning; 30-day retention testing is

05

K-12 products that assume reliable broadband and modern devices without equity planning — exacerbates the digital divide that regulators and school boards are increasingly focused on

Our Approach

We build EdTech AI with FERPA and COPPA compliance at the infrastructure level — not bolted on after the fact. Student performance data is never used for advertising, is not shared across institutions, and is subject to retention limits aligned with applicable state law. Our tutoring AI implementations are designed to scaffold learning rather than provide answers — the pedagogical objective is building student capability, not maximizing session length. We integrate through LTI 1.3 and xAPI standards. We do not build AI integrity detection systems — that arms race is lost — we help institutions redesign assessments that make AI completion irrelevant.

Ready to build for Education?

We bring domain expertise, not just engineering hours.

Start a Conversation

Free 30-minute scoping call. No obligation.