AI in College Admissions Essays: 2026 Statistics, Detection Data, and Application Impact

Published:

Updated:

AI in College Admissions Essays: 2026 Statistics, Detection Data, and Application Impact

By Detection Drama Research Team
Published May 14, 2026
Read time 11 min
1 in 3
high school seniors used AI to help write their college admissions essays in 2023-24
Source: foundry10, Navigating College Applications with AI, 2024

Three years after ChatGPT became publicly available, generative AI has quietly become a fixture of the college application. Roughly a third of high school seniors now reach for a chatbot at some point while crafting their personal statements. Half use AI to brainstorm. One in five uses it to draft. And almost two-thirds of the colleges receiving those essays still have no formal policy telling applicants what is — or isn’t — allowed.

This report compiles the verified statistics on AI use in college admissions essays as of May 2026, drawing on the foundry10 white paper, the Kaplan 2025 admissions officer survey, a 2026 Cornell/Carnegie Mellon study of tens of thousands of real essays, and a BestColleges survey of 1,000 current undergraduates. The picture that emerges is messy: high adoption, low policy clarity, deep equity concerns, and detection tools that are still wildly unreliable on the kind of writing that actually appears in admissions files. For context on how the broader detection landscape evolved, see our 2026 AI Detection Industry Report.

Key takeaways

  • 50% of applicants use AI to brainstorm their admissions essay (foundry10)
  • 20% use AI to draft a first version of the essay (foundry10)
  • 6% rely on AI for the final draft they submit (foundry10)
  • 68% of colleges still have no formal AI policy on essays (Kaplan, 2025)
  • 30% of colleges officially ban AI in admissions essays (Kaplan, 2025)
  • 2% of colleges explicitly permit AI essay drafting (Kaplan, 2025)
  • 50% of admissions officers view AI essay use unfavorably (Kaplan, 2025)
  • 61% of TOEFL essays by non-native English speakers were misclassified as AI in a Stanford study

How Many Students Use AI to Write Their Admissions Essays

Adoption Data

About one-third of high school seniors who applied to college in the 2023-24 cycle used an AI tool somewhere in their admissions essay process, according to foundry10’s white paper of 1,000+ student responses. That figure has only climbed since.

foundry10’s Navigating College Applications with AI remains the most rigorous public dataset on student adoption. Among graduating seniors who applied to college in 2023-24, roughly 33% reported using a generative AI tool — most often ChatGPT — at some stage of their essay work. Sixty-five percent said they used AI rarely or not at all for school in general, but nearly a third of all surveyed students still reached for it specifically when working on college applications. That gap is telling: students who avoid AI for homework still treat the college essay as worth the help.

Cohort Used AI on essay Source
All 2023-24 applicants ~33% foundry10 (2024)
Current undergrads who say they would have used it 47% BestColleges (Oct 2023, n=1,000)
First-generation college students (would-have) 58% BestColleges (2023)
Non-first-generation students (would-have) 33% BestColleges (2023)
Avoided AI for fear of cheating accusation 39% foundry10 (2024)
47% of current college students surveyed by BestColleges in October 2023 said they would have used AI tools on their admissions essay if those tools had been available to them — meaning the demand was there before the supply caught up.

The 39% who avoided AI specifically because they worried it would count as cheating is the single most underreported finding in the foundry10 dataset. It explains why AI detection anxiety statistics spike sharply during application season — applicants are making suppression decisions on their own writing not because they did anything wrong, but because they fear being misread.

What Applicants Actually Use AI For

Use Cases

Among the third of applicants who used AI, the most common use was brainstorming (50%), followed by spelling and grammar checks (48%), outlining (47%), and drafting a first version (20%). Only 6% used AI to produce the final version they submitted.

The drop-off between “brainstorm” and “final draft” is the most important statistic in this entire report. It tells us that the dominant pattern is not students cheating with ChatGPT — it’s students using it the way they’d use a thoughtful older sibling: someone to talk through ideas with, get an outline from, and clean up grammar at the end. Whether colleges consider that cheating is a separate, unresolved question that the 68% policy void in the next section makes worse.

Applicant AI use cases (% of AI-using students)

Brainstorm ideas
50%
Spelling & grammar
48%
Create outline
47%
Draft first version
20%
Write final draft
6%
Source: foundry10, Navigating College Applications with AI, 2024 (among AI-using applicants).

Teachers, notably, mirror this same pattern in their own application-related work. About 31% of teachers reported using generative AI to help craft letters of recommendation in the same foundry10 cycle. The norms students are absorbing aren’t just emerging in dorm rooms — they’re being modeled by the adults writing the supporting materials. This dovetails with our analysis of professors using ChatGPT statistics, which shows even higher AI adoption rates among instructors than among students.

The 68% Policy Void: What Colleges Actually Say (or Don’t)

Policy Data

According to Kaplan’s 2025 survey of 220 admissions offices at top national, regional, and liberal arts colleges, 68% have no formal policy on AI use in admissions essays, 30% officially ban it, and just 2% explicitly allow it.

This is the dataset that makes the entire applicant experience confusing. Kaplan polled 220 of the U.S. News & World Report-ranked colleges between July and August 2025. The result is a policy landscape where the modal college simply hasn’t decided yet. For applicants this means there is no single rule to follow — the answer depends on which 50 to 100 schools an individual student is applying to, and most of those schools haven’t published guidance.

98% of surveyed colleges either ban AI essay writing outright or have no published policy at all. Applicants are functionally guessing in three out of four applications they submit.

When colleges do publish policies, the line is usually drawn between “ideation” (allowed) and “generation” (not). Brainstorming, outlining, and line-edit grammar are widely permitted. AI-generated paragraphs in the final submission are not. The Common App’s 2024 disclosure language reflects this consensus by asking applicants to certify the essay represents their own work. Schools that explicitly banned external AI assistance during essay drafting are a small but growing cluster — a different pattern from the universities we documented in our piece on universities that banned AI detectors after high false-positive rates began producing wrongful accusations.

What Admissions Officers Actually Think

Officer Attitudes

Half of admissions officers (50%) hold an unfavorable view of AI use in essays. Just 14% are favorable, and 26% are neutral, per Kaplan’s 2025 survey. That distribution drives a lot of the off-the-record skepticism that flagged essays receive.

The Kaplan attitude breakdown is the single best signal of how flagged or stylistically anomalous essays get read inside admissions offices. Half the room is already predisposed against AI-assisted writing. Only a small minority sees it positively, and most of those treat it as benign brainstorming, not generated prose. Combined with the policy void above, this means that an applicant whose essay reads AI-generated faces a coin flip: half the readers will treat it as a strike against authenticity, even where no formal rule has been violated.

Admissions officer attitudes toward AI in essays (Kaplan 2025)

Unfavorable
50%
Neutral
26%
Favorable
14%
Remaining ~10% expressed no opinion or weren’t asked. Source: Kaplan, 2025 (n=220 admissions officers).

Separately, an Inside Higher Ed survey from September 2023 found that ~80% of colleges expected to use AI somewhere in admissions during the 2024 cycle, with about 60% of officers reporting AI was being used to review personal essays. That doesn’t mean every essay runs through a detector — far from it — but it confirms that the technology is now on both sides of the application. For schools investing in detection software, see our breakdown of AI detection tool spending at universities in 2026.

Detection Accuracy: Why Flagged ≠ Cheated

Detection Data

AI detection tools are unreliable on admissions-essay-length text. Turnitin’s own published false-positive rate is around 4% per sentence, and a 2023 Stanford study found 61% of TOEFL essays from non-native English speakers were misclassified as AI. A flag is a signal, not proof.

This is the section every applicant should read. The detector tools that schools and counselors casually invoke are not calibrated for short, stylized personal essays written by 17-year-olds — they’re trained primarily on academic prose. Two well-documented failure modes appear repeatedly in published research: false positives on highly polished writing and systemic bias against non-native English speakers. For a complete breakdown of how Turnitin’s own thresholds work, see our piece on the Turnitin AI detection score threshold.

Detection failure mode Approximate rate Source
Turnitin false-positive rate (per sentence) ~4% Turnitin (vendor-reported)
Mislabeling rate on TOEFL essays (non-native English) 61% Stanford / Liang et al., 2023
Common detector false-flag rate on highly edited human prose 9–15% Multiple independent benchmarks
Applicants who avoided AI specifically to dodge a flag 39% foundry10 (2024)
61% of TOEFL essays — written entirely by humans — were classified as AI-generated by leading detectors in Stanford’s 2023 study. This is the single most important reason admissions officers should not lean on detector output alone.

Even within the broader Turnitin ecosystem, the company itself has acknowledged that essays under 300 words trigger more unreliable scoring. Most Common App supplements fall in that danger zone. Combined with the documented ESL bias in AI detection, this means applicants who write in second-language English or who are heavily edited by counselors face elevated false-positive risk. If you’ve already been flagged, our checklist on defending against Turnitin AI false positives walks through the evidence to assemble. And if the flag came after submission, the first 24 hours matter — see what to do in the next 24 hours after an AI accusation.

The Equity Problem: Who Gets Penalized

Equity Findings

A May 2026 Cornell/Carnegie Mellon study of tens of thousands of essays found that lower-income applicants were more likely to use AI and more likely to be rejected — even within the AI-using cohort. The research authors interpret this as a digital divide: higher-income students have counselors and premium AI tools that produce better essays; lower-income students often have only free-tier ChatGPT.

This is the finding that has begun reshaping how admissions researchers talk about AI. The Cornell/Carnegie Mellon paper (lead author Jinsook Lee, Cornell) analyzed tens of thousands of essays submitted to an unnamed selective institution over four years, beginning before ChatGPT’s late-2022 release. Two patterns emerged. First, fee-waived applicants — the proxy for lower-income status — increased their AI use faster than any other group. Second, AI-assisted writing was correlated with lower predicted admission probabilities, and that penalty fell hardest on the same lower-income applicants.

Inequality of returns: high-income applicants who use AI have counselors, essay coaches, and access to premium tools like Claude Pro. Low-income applicants using free ChatGPT produce more obviously AI-styled prose — and pay the price in the read.

The paper’s authors also documented a striking linguistic homogenization trend: essays at the selective college converged toward similar phrasings, openings, and rhetorical structures after AI tools went mainstream. The convergence was greatest among rejected applicants and among lower-income applicants. Read alongside the broader landscape of AI cheating consequences statistics, the picture is clear: AI penalties land unevenly, and the equity story is the part of the admissions-AI debate moving fastest in 2026.

Interactive: Estimate Your Essay’s Flag Risk

Interactive Tool

A back-of-envelope estimator. Answer the four questions and we’ll give you a rough flag-risk band based on the patterns documented in the studies above. This is not a guarantee — admissions readings are subjective.

Quick essay flag-risk estimator

1. How did you use AI on the essay?




2. Is English your first language?

3. Did you edit heavily after AI involvement (paraphrase, restructure, change voice)?


4. Does the essay’s voice match your short-answer responses and recommendations?


Methodology & Limitations

How we built this report

All statistics in this report are sourced from named, publicly available studies, surveys, or vendor-published documentation. Each finding is tagged with one of four verification levels: verified (single named source with primary data), cross-verified (two or more independent sources report the same finding), single-source (only one published source available), and commonly-claimed (vendor-reported figures we surface for context but cannot independently audit).

The foundry10 white paper sampled high school seniors who applied during the 2023-24 cycle. The Kaplan 2025 survey polled 220 admissions offices at U.S. News & World Report-ranked colleges between July and August 2025. The BestColleges survey polled 1,000 current undergraduate and graduate students in October 2023. The Cornell/Carnegie Mellon study (May 2026) analyzed tens of thousands of essays at one unnamed selective institution across four admissions cycles. None of these datasets represent the full U.S. applicant population, and findings should be read as directional rather than census-grade.

Frequently Asked Questions

How many students use AI to write their college admissions essays?
Roughly one in three high school seniors who applied in the 2023-24 cycle used an AI tool for some part of their admissions essays, according to foundry10. About 50% used it to brainstorm, 47% to outline, and 20% to draft. Only 6% relied on AI to write the final draft.
Do colleges check admissions essays for AI?
Most do not run essays through dedicated AI detectors as official policy. Kaplan’s 2025 survey found 68% of colleges have no formal AI policy, and admissions officers acknowledge detector tools are unreliable. However, 60% of officers in a separate 2023 survey said AI was being used somewhere in reviewing personal essays. Suspicious essays typically trigger a human comparison against other writing samples in the application.
Will my essay be flagged if I used ChatGPT to brainstorm?
Brainstorming with AI is the most common — and most widely accepted — use case. About half of all applicants do it, and most college policies that exist permit AI for ideation, outlining, or line-level grammar editing. Risk rises only when AI-generated text appears verbatim in the final submission. Even then, AI detectors return false positives at meaningful rates, so a flag is not proof.
Are AI detectors accurate on admissions essays?
No. Turnitin’s reported false-positive rate is around 4% per sentence, and a Stanford study found 61% of TOEFL essays written by non-native English speakers were mislabeled as AI. That bias matters: a single flagged paragraph in a 650-word Common App essay does not constitute reliable evidence of AI use.
Do students who use AI in their essays get admitted at lower rates?
Yes, according to a 2026 Cornell/Carnegie Mellon study analyzing tens of thousands of essays at a selective college. AI-assisted writing was associated with lower predicted admission probabilities, with the largest penalty falling on lower-income applicants who likely use free-tier AI tools rather than premium versions.
What is the safest way to use AI on a college essay?
Use AI for brainstorming topics, outlining structure, and line-edit grammar checks only — never to draft full paragraphs. Always submit content that sounds like your own writing voice and matches the style of your short-answer responses. Disclose AI use if the application asks. The Common App and most top-30 colleges permit ideation but draw the line at AI-generated prose.

Sources & references

  1. foundry10. Navigating College Applications with AI white paper. foundry10.org/resources/navigating-college-applications-with-ai
  2. Inside Higher Ed (Johanna Alonso). “Low-Income Students More Likely to Submit AI-Generated Admissions Essays.” May 8, 2026. insidehighered.com/news/admissions
  3. Kaplan. “Colleges Increasingly Clarify Rules on GenAI Use in Admissions Essays.” 2025 survey of 220 admissions offices. kaplan.com
  4. Lee, J., Alvero, AJ et al. Cornell/Carnegie Mellon working paper, May 2026. Coverage at Inside Higher Ed
  5. BestColleges. “Half of College Students Would Have Used AI on Admissions Essay.” October 2023 (n=1,000). bestcolleges.com/research
  6. Liang, W. et al. “GPT detectors are biased against non-native English writers.” Stanford HAI, 2023. hai.stanford.edu
  7. Pangram Labs. “Do College Admissions Offices Check for AI?” 2025. pangram.com/blog
  8. Hechinger Report. “Students who want to try AI in college admission essays should use it for brainstorming.” 2024. hechingerreport.org