Chancing Engines Landscape: Every Admissions Prediction Model vs. Our Simulation

chancing_engines_landscape.md


Chancing Engines Landscape: Every Admissions Prediction Model vs. Our Simulation

Comprehensive inventory of 50+ commercial tools and 15 academic models, compared side-by-side with our agent-based simulation. March 2026.


Table of Contents

  1. Executive Summary
  2. The Taxonomy
  3. Tier 1: Major Consumer Chancing Engines
  4. Tier 2: AI-Powered Newer Entrants
  5. Tier 3: Institutional / Counselor Platforms
  6. Tier 4: Academic Index Calculators
  7. Tier 5: Institution-Side Yield Prediction
  8. Tier 6: Academic Research Models
  9. The Master Comparison Matrix
  10. Feature Coverage Heat Map
  11. Where Our Simulation Sits
  12. Key Findings

1. Executive Summary

We identified 50+ admissions prediction tools across 6 tiers — from billion-user platforms (Naviance, CollegeVine) to single-developer GitHub repos. The landscape breaks into two fundamentally different categories:

Individual prediction tools (all commercial engines): "Given YOUR profile, what are your chances at School X?" These treat each student-college pair in isolation.

Market simulation models (our simulation + Reardon 2016 ABM): "Given ALL applicants, seat constraints, and round dynamics, what happens in the aggregate?" These model the system.

No commercial tool models hooks (legacy/athlete/donor), seat constraints, round sequencing, yield competition, or waitlist dynamics. Our simulation is the only system that combines a calibrated logistic admissions model with full market-clearing mechanics.

The closest analog is the Stanford CEPA agent-based model (Reardon et al. 2016) — same architecture, same multi-round matching — but that model uses stylized parameters while ours is calibrated to real CDS data for 55 named colleges.


2. The Taxonomy

INDIVIDUAL PREDICTION                           MARKET SIMULATION
(What are MY chances?)                          (How does the system work?)

┌─────────────────────────────┐                ┌──────────────────────────┐
│ TIER 1: Major Consumer      │                │ ACADEMIC ABMs            │
│   CollegeVine (75 factors)  │                │   Reardon 2016 (8K stu) │
│   Scoir AI (ML, round-aware)│                │   Assayed 2023 (Jordan) │
│   Niche (GPA+SAT scatter)   │                │                          │
│   PrepScholar (GPA+SAT)     │                │ OUR SIMULATION           │
│   Naviance (school scatter) │                │   55 colleges, 6 rounds  │
│   Parchment (766K outcomes) │                │   Logistic + market clear│
│   CollegeData (2K colleges) │                │   Hooks, DI, yield, WL   │
│   College Transitions       │                │   Monte Carlo validated  │
├─────────────────────────────┤                └──────────────────────────┘
│ TIER 2: AI Newcomers        │
│   20+ startups (2023-2026)  │                EMPIRICAL RESEARCH MODELS
│   GradGPT, Ultra, Chancify  │                (What factors matter?)
│   Admissionsly, DreamyUni   │                ┌──────────────────────────┐
├─────────────────────────────┤                │ Arcidiacono (SFFA trial) │
│ TIER 3: Institutional       │                │   150K apps, 370 vars    │
│   Scoir, Overgrad, Kollegio │                │   Pseudo R²=0.56         │
├─────────────────────────────┤                │ Chetty 2023 (Ivy-Plus)   │
│ TIER 4: Academic Index      │                │   ALDC decomposition     │
│   Top Tier, Ivy League Guru │                │ Espenshade (SAT equiv)   │
├─────────────────────────────┤                │   124K apps, 10 colleges │
│ TIER 5: Yield Prediction    │                │ Avery-Hoxby (revealed    │
│   ENROLL, Yield+, Slate     │                │   preference ranking)    │
│   (institution-facing only) │                └──────────────────────────┘
└─────────────────────────────┘

3. Tier 1: Major Consumer Chancing Engines

CollegeVine — The Market Leader

Dimension Detail
URL collegevine.com/admissions-calculator
Model ML (undisclosed algorithm, post-May 2021)
Inputs ~75 factors: GPA, SAT/ACT, AP count, EC tiers (4-tier/12-sub-tier), gender, state, intended major, race (reduced post-SFFA)
Excludes Essays, recommendations, legacy, athlete, donor, DI, yield protection, school context
Coverage 1,500+ colleges
Cost Free
Calibration Published: 50% predicted → 48.1% actual. Overpredicts at <20% acceptance rate schools
Round-aware ED added May 2022 (~1.6× average)

See collegevine.md and scoring_model_comparison.md for deep analysis.

Scoir Admission Intelligence — The Dark Horse

Dimension Detail
URL scoir.com/high-schools/admission-intelligence
Model ML trained on tens of millions of de-identified historical application records imported from high schools
Inputs GPA, test scores, first-gen, geographic location, race/ethnicity, sex, high school profile
Coverage Large (exact count unpublished)
Cost Institutional license only (students access through school)
Round-aware Yes — provides ED, EA, RD predictions separately
Calibration "A 60% chance means historically ~60 out of 100 students with similar profiles were admitted"

Scoir is the only commercial platform with round-level predictions, making it architecturally the most similar to our simulation among consumer-facing tools. However, it is institution-only (not public) and does not model hooks or market-clearing dynamics.

Niche.com — Scattergram Approach

Dimension Detail
Model k-nearest-neighbors visual comparison against self-reported user outcomes
Inputs GPA, SAT/ACT, intended major (optional)
Excludes ECs, essays, recs, demographics, hooks, round
Coverage Thousands of colleges
Cost Free
Revenue ~$110M/yr (colleges pay for recruitment ads)

PrepScholar — The Minimal Baseline

Dimension Detail
Model Percentile lookup against CDS data
Inputs GPA, SAT/ACT only
Excludes Everything else
Cost Free
Dimension Detail
Model Not predictive — displays scattergrams of past applicants from YOUR high school
Inputs GPA, SAT/ACT (historical, not user-entered)
Strength The only tool with school-specific historical context
Weakness Only GPA/SAT; no ECs, essays, or hooks. Requires school subscription
Research Mulhern 2021 found scattergrams deter high-achievers from selective schools by 50%

Parchment "My Chances"

Dimension Detail
Model Statistical predictions from 765,689 self-reported applications
Inputs GPA, test scores, self-reported academic qualifications
Cost Free

CollegeData "College Chances"

Dimension Detail
Model Algorithm comparing against enrolled freshmen profiles
Inputs GPA, test scores, honors courses, EC hours, leadership roles
Coverage 2,000+ colleges
Cost Free
Note One of the few free tools that includes any EC input beyond CollegeVine

College Transitions

Dimension Detail
Model Data-driven formula + decades of counseling experience
Inputs Grades, test scores, geography, ability to pay, high school attended, athlete recruitment, first-gen, legacy
Output Color-coded scale: Far Reach → Likely (ranges, not single %)
Note Only free tool besides our sim that includes legacy and recruited athlete

4. Tier 2: AI-Powered Newer Entrants (2023–2026)

A wave of 20+ AI-powered chancing startups emerged in 2023–2026. Most are thin wrappers around LLMs or simple ML models with unverified accuracy claims.

Tool Claimed Accuracy Model Detail Notable Feature
Ultra (YC-backed) Undisclosed Simulates AO evaluation process Founded by former AOs; closest to holistic review
GradGPT 90% ML + essay analysis Built by former AOs
Admissionsly 85–90% "Latest admissions data" 3,000+ colleges, no signup
stdnt.space 72% (basic) / 98.2% (pro) AI on real-world data Two-tier system
Chancify AI Undisclosed ML on real admission data Probability ranges, test-optional aware
AdmitYogi Undisclosed Proprietary on 6K data points Gap analysis with improvement recs
CollegeAI Undisclosed AI-powered Profile questionnaire
DreamyUni Undisclosed AI on "thousands of data points" SAT prep + essay help bundled
LumiSource Undisclosed AI on 10K+ accepted essays Continuous algorithm updates
EligioAI Undisclosed AI profile analysis Financial analysis included
Test Ninjas Undisclosed AI "admissions officer" Personal statement quality input
KapAdvisor (Kaplan) Undisclosed AI from senior advisors $199/yr premium; PDF transcript upload
Sups (acq. by USNews) N/A AI counselor Essay/activity optimization, not chancing
ChanceMe GPTs N/A Custom ChatGPT prompts Uses Harvard Reading Procedures as context

None of these publish independent calibration data. The 98.2% and 90% claims are self-reported and should be treated skeptically. None model hooks, rounds, or market dynamics.

Ultra (Y Combinator-backed) is the most interesting — it "simulates the process that elite colleges use to evaluate profiles" by running simulated AO evaluations, which is conceptually closer to our approach than a statistical lookup.


5. Tier 3: Institutional / Counselor Platforms

Tool Model Access Round-Aware Hooks
Scoir AI ML on millions of historical records School license Yes No
Overgrad Custom match algorithm, region-specific Institutional No No
Kollegio AI ML on historical data Free (basic) No No
Universily Data-driven with Common App compliance Free (basic) No No
Reach Best AI on demographics + stats Free (basic) No No

Scoir stands alone as the only platform with round-level predictions (ED vs. EA vs. RD).


6. Tier 4: Academic Index Calculators

These compute the Ivy League Academic Index — a pure academic metric, not an admission probability.

Tool Formula Coverage
Top Tier Admissions SAT EBRW + SAT Math + GPA conversion (CGS). Bands A/B/C/D above minimum 171 Ivy League
Ivy League Guru ~2/3 test scores + 1/3 rank/GPA Ivy League
Academic Index AI Modernized AI from PDF uploads (Princeton/Yale students) Elite colleges

Our simulation uses a similar Academic Index: AI = gpaToCSG(gpa) + section_weighted_SAT, range ~100–240. Our formula is inspired by the Ivy AI but adapted for a broader college universe.


7. Tier 5: Institution-Side Yield Prediction (Not Student-Facing)

These tools predict which admitted students will enroll, not who gets admitted. They serve colleges, not students.

Tool Maker Purpose
ENROLL Capture Higher Ed Yield prediction from behavioral + academic data
Yield+ Encoura (ACT subsidiary) Admit-to-enroll scoring
Slate Technolutions Full CRM with predictive analytics (2,000+ institutions)
CollegeVine AI Recruiter CollegeVine AI outreach to prospective students (250+ universities)

Our simulation models yield from the student sidestudentFinalDecisions() uses Chetty yield data by income bracket. These tools model yield from the institution side — predicting which students in an admitted class will deposit.


8. Tier 6: Academic Research Models

Arcidiacono et al. — The SFFA Harvard Model (The Gold Standard)

The most detailed empirical admissions model with publicly known coefficients.

Dimension Detail
Data 150,000 applicants to Harvard, Classes of 2014–2019
Model Binary logistic regression
Variables 370+ controls including Harvard's internal ratings (academic, EC, athletic, personal), Academic Index, race, gender, ALDC status, intended major, geographic indicators
Fit Pseudo R² = 0.56 (excellent for logit)

Key Odds Ratios:

Variable Odds Ratio Our Simulation Equivalent
Recruited Athlete ~5,075× HOOK_MULT_BY_TIER.athlete[1] = 4.5×
Legacy ~8.5× HOOK_MULT_BY_TIER.legacy[1] = 5.7×
Dean's Interest List >7× HOOK_MULT_BY_TIER.donor[1] = 12.0×
Asian American 0.63× SFFA_MULT_BY_TIER.asian_american[1] = 1.35× (post-SFFA reversal)
African American Large positive SFFA_MULT_BY_TIER.urm[1] = 0.85× (post-SFFA reversal)

Critical comparison: Arcidiacono's athlete odds ratio (~5,075×) is far larger than our 4.5× because his model includes Harvard's internal athletic rating as a separate variable. Our model combines the athletic rating effect into the single recruited_athlete hook multiplier. Also, our multipliers operate in logit space on a pre-threshold score, while Arcidiacono's are direct odds ratios from a fully specified model.

Chetty, Deming, Friedman (2023) — "Diversifying Society's Leaders?"

Dimension Detail
Data Ivy-Plus colleges linked to federal tax records and College Board files. Millions of students
Model Regression decomposition + instrumental variables

Decomposition of top-1% family admissions advantage:

Source Share
Legacy preferences 46%
Athletic recruitment 24%
Higher non-academic ratings 30%

Our simulation captures all three channels: legacy via HOOK_MULT_BY_TIER.legacy, athletes via HOOK_MULT_BY_TIER.athlete, and non-academic ratings via ec_quality and essay_base (which correlate with structural advantage through archetype generation).

Key finding: Legacy applicants from top 1% families are 5× more likely to be admitted; recruited athletes are admitted "with near certainty." Both match our model's calibration.

Espenshade, Chung, Walling (2004) — SAT-Point Equivalents

Dimension Detail
Data 124,374 applications to 10 selective colleges
Model Binary logistic regression with microsimulation

SAT-point equivalents of admission advantages (1600 scale):

Hook SAT-Point Equivalent Our Logit Equivalent
African American +230 points log(SFFA_MULT_BY_TIER.urm[1]) = -0.16 (post-SFFA: now a penalty)
Hispanic +185 points included in URM
Recruited Athlete +200 points log(4.5) = +1.50 logit units
Legacy +160 points log(5.7) = +1.74 logit units

These Espenshade figures are pre-SFFA. Post-SFFA, the race advantages are eliminated or reversed in our model — a key update that no commercial chancing engine has fully incorporated.

Reardon et al. (2016) — The Agent-Based Model

The closest published analog to our simulation.

Dimension Reardon 2016 Our Simulation
Platform NetLogo Vanilla JS (browser)
Students 8,000 agents ~4,000 from 20–485 high schools
Colleges 40 (generic, 150 seats each) 55 (named, real CDS data, real class sizes)
Student attributes Resources (SES), Caliber (SAT-like) 6 behavioral × 4 structural archetypes, GPA, SAT (section scores), ECs, essays, hooks, income, state
Application model Expected utility maximization with imperfect information Utility = prestige + fit + λ·log(P_admit), stratified allocation
Admission model Perceived caliber ranking → admit top N Logistic P(admit) with ~30 factors → Bernoulli trial
Rounds Single round ED → EA/REA → EDII → RD → Decisions → Waitlist
Yield model Utility-based enrollment choice Chetty income-bracket yield + financial aid
Hooks None (SES proxy only) Legacy, athlete, donor, first-gen, Pell, URM, Asian, gender, geography, DI
Phantom applicants None (full population modeled) MODEL_SCALE = 0.013, phantom admits consume seats
Calibration Validated against IPEDS 2010–2011 Monte Carlo validated against CDS for 55 named colleges
Key finding Resource-caliber correlation (r=0.3) drives most stratification Hook multipliers (ALDC) are the dominant force at T1–T2

Parameter comparison:

Parameter Reardon Ours
R-C correlation 0.3 (fixed) Implicit via feeder tier × archetype generation
Applications 4 + 0.5 × resources Lognormal(archetype mean × structural mult), range 3–20
College quality update 0.9 × Q + 0.1 × mean_caliber Static (CDS data, no dynamic quality)
Information accuracy Varies by SES Perfect school data; holistic noise ≈ Normal(0, 0.7) in logit
Seat capacity 150 per college Real class sizes (400–6,000+)

Avery, Glickman, Hoxby, Metrick (2004/2013) — Revealed Preference Ranking

Dimension Detail
Model Bradley-Terry paired comparison (logistic tournament model)
Data 3,240 high-achieving students, all admission/enrollment decisions
Innovation Each student holds a "tournament" among colleges that admitted them; enrolled college "wins"
Output Desirability parameter (δ) per college

Our studentFinalDecisions() function is conceptually similar — students choose among acceptances based on utility (prestige, aid, fit). Avery-Hoxby formalizes this as a maximum likelihood estimation of college desirability parameters.

Avery and Levin (2010) — Early Admissions Game Theory

Dimension Detail
Model Game-theoretic signaling model
Key finding ED credibly signals enthusiasm → colleges rationally apply lower thresholds
Prediction Lower-ranked schools benefit most from binding ED (can "capture" uncertain strong students)

Our ED multipliers (1.2–8.0×) are the empirical realization of this theoretical prediction. Schools like Dartmouth (3.5×) and Columbia (3.4×) have massive ED advantages precisely because of this signaling equilibrium.

Hurwitz (2011) — Legacy at 30 Elite Institutions

Dimension Detail
Data 294,457 applicants, 30 institutions, Fall 2007
Model Conditional logistic regression with college fixed effects
Key finding Legacy odds ratio = 3.13×; primary legacy = +45.1 percentage points

Our HOOK_MULT_BY_TIER.legacy values (5.7× at T1, 4.0× at T2, 3.0× at T3) bracket this finding — they're calibrated higher at T1 (HYPSM) and lower at T3+.

Dale and Krueger (2002) — Does Selectivity Matter?

Dimension Detail
Model OLS with self-revelation selection correction
Key innovation The set of schools a student applies to reveals unobserved ability
Key finding After controlling for application behavior, the selectivity premium disappears — except for low-income students

Not directly an admissions model, but validates our structural approach: modeling the application portfolio (list building) is as important as modeling admission probability.

UC Irvine Deep Learning (2024)

Dimension Detail
Data 4,442 CS applications to UCI, Fall 2021
Model Feed-forward + Input Convex Neural Networks with LIME interpretability
Accuracy 80.6% (3pp above classical ML baselines)
Top features Weighted GPA > Unweighted GPA > AP total > AP CS A score
Excluded Gender, ethnicity (California Prop 209)

Cornell Learned Admission-Prediction (2023, Best Paper Award)

Dimension Detail
Data 13,248 applications to a selective US institution
Model Gradient Boosting Decision Trees (scikit-learn)
Use case Replacing SAT-based heuristics in subset generation for holistic review
Finding ML model matched demographic composition of last admitted class better than SAT-based heuristic

Lira et al. (2023) — AI for Personal Qualities (Science Advances)

Dimension Detail
Data 3,131 training; 309,594 validation
Model Fine-tuned pretrained language models
Innovation Assesses 7 personal qualities from application essays: intrinsic motivation, prosocial purpose, etc.
Finding Incremental validity for predicting 6-year graduation beyond GPA/SAT

This is the frontier: NLP models that can actually evaluate essay quality. Our simulation approximates this with essay_base (archetype-generated 1–10 score), but Lira's approach could eventually replace the proxy.


9. The Master Comparison Matrix

Tool Model Type Factors Colleges Hooks Rounds Market Calibration Access
Our Simulation Logistic ABM ~30 55 Yes (6 types) 6 rounds Full MC validated Research
Reardon ABM Utility ABM 2 (SES, caliber) 40 No 1 round Full IPEDS validated Academic
Arcidiacono Logistic 370+ 1 (Harvard) Yes (ALDC) No No R²=0.56 Court record
CollegeVine ML (secret) ~75 1,500+ No ED only No Published ±3pp Free
Scoir AI ML Many Large No Yes No Claims calibrated School license
Niche kNN scatter 2–3 1000s No No No None Free
PrepScholar Lookup 2 Wide No No No None Free
Naviance Historical scatter 2 Per-school No No No Mulhern 2021 School license
Parchment Statistical ~3 Wide No No No None Free
CollegeData Algorithm 5–6 2,000+ No No No None Free
College Transitions Data + expertise 8+ Selective Partial No No None Free
Ultra AO simulation Full profile Elite Unknown Unknown No None Free
Espenshade Logistic + microsim 10+ 10 Yes No No N/A Academic
Chetty 2023 Regression Many Ivy-Plus Yes No No N/A Academic
Avery-Hoxby Paired logistic δ per college 100+ No No Partial Theory Academic

10. Feature Coverage Heat Map

What each model includes (filled) vs. excludes (empty):

                        CV  Scoir Niche Prep Nav  Parch CData CTrans Our  Rear  Arc  Esp  Chet
GPA                     ●    ●     ●     ●    ●    ●     ●     ●     ●    ●     ●    ●    ●
SAT/ACT                 ●    ●     ●     ●    ●    ●     ●     ●     ●    ●     ●    ●    ●
SAT section scores      ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ●    ○    ○
Course rigor (APs)      ●    ○     ○     ○    ○    ○     ●     ○     ○*   ○     ●    ○    ○
Extracurriculars        ●    ○     ○     ○    ○    ○     ●     ○     ●    ○     ●    ○    ○
Essay quality           ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ●    ○    ○
Recommendations         ○    ○     ○     ○    ○    ○     ○     ○     ○    ○     ●    ○    ○
Gender                  ●    ●     ○     ○    ○    ○     ○     ○     ●    ○     ●    ●    ○
Race/ethnicity          ◐    ●     ○     ○    ○    ○     ○     ○     ●    ○     ●    ●    ○
State/geography         ●    ●     ○     ○    ○    ○     ○     ●     ●    ○     ●    ○    ○
Intended major          ●    ○     ●     ○    ○    ○     ○     ○     ●    ○     ●    ○    ○
Legacy                  ○    ○     ○     ○    ○    ○     ○     ●     ●    ○     ●    ●    ●
Recruited athlete       ○    ○     ○     ○    ○    ○     ○     ●     ●    ○     ●    ●    ●
Donor/development       ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ●    ○    ●
First-gen               ○    ●     ○     ○    ○    ○     ○     ●     ●    ○     ●    ○    ○
Income effects          ○    ○     ○     ○    ○    ○     ○     ●     ●    ●     ○    ○    ●
Feeder school           ○    ○     ○     ○    ●**  ○     ○     ●     ●    ○     ○    ○    ○
ED/EA round             ◐    ●     ○     ○    ○    ○     ○     ○     ●    ○     ○    ○    ○
Demonstrated interest   ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ○    ○    ○
Yield protection        ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ○    ○    ○
Seat constraints        ○    ○     ○     ○    ○    ○     ○     ○     ●    ●     ○    ○    ○
Multi-round processing  ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ○    ○    ○
Yield/enrollment model  ○    ○     ○     ○    ○    ○     ○     ○     ●    ●     ○    ○    ●
Waitlist                ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ○    ○    ○
Summer melt             ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ○    ○    ○
Phantom applicants      ○    ○     ○     ○    ○    ○     ○     ○     ●    ○     ○    ○    ○
Holistic noise          ○    ○     ○     ○    ○    ○     ○     ○     ●    ●     ○    ○    ○

● = modeled   ◐ = partially   ○ = not modeled
*  AP count generated per archetype but not directly in admission score
** Naviance uses YOUR school's historical data (implicit feeder effect)

CV=CollegeVine, Prep=PrepScholar, Nav=Naviance, Parch=Parchment,
CData=CollegeData, CTrans=College Transitions, Rear=Reardon 2016,
Arc=Arcidiacono, Esp=Espenshade, Chet=Chetty 2023

11. Where Our Simulation Sits

What No One Else Does

  1. Full market clearing: ED fills seats → fewer for RD. Student choices affect yield. Waitlist fills gaps. Summer melt creates final attrition. No other tool models these cascading effects.

  2. Phantom applicant correction: At MODEL_SCALE = 0.013, we represent 1.3% of real applicants. Phantom applicants consume seats proportionally. No other model handles this.

  3. Hook stacking in logit space: A legacy athlete donor doesn't get P > 1 because hooks are additive in log-odds. The Arcidiacono model does this too, but for a single institution; we do it across 55 colleges with tier-dependent multipliers.

  4. Demonstrated interest with round interaction: DI matters most in RD (full boost), less in EA (0.5×), minimal in ED (0.2×). No tool models this.

  5. Yield protection: Overqualified unhooked students get penalized at schools like Tufts and Northeastern. Only our simulation and College Transitions even acknowledge this exists.

  6. Monte Carlo batch mode: 100-run headless simulations producing per-college mean ± SE for acceptance rate, yield, enrolled count, and avg SAT. No consumer tool offers this.

What Others Do Better

  1. College coverage: CollegeVine covers 1,500+ schools; we cover 55. For a student considering mid-selectivity schools, CollegeVine is more useful.

  2. Individual precision: CollegeVine takes YOUR actual GPA and activities. We generate synthetic students from archetype distributions. For personal prediction, CollegeVine wins.

  3. School-specific history: Naviance shows where students FROM YOUR SCHOOL were admitted. We use feeder tier as a proxy but lack school-specific granularity.

  4. Training data scale: Scoir has "tens of millions" of historical records; Parchment has 766K; we calibrate against CDS aggregates. Their n > our n.

Our Unique Position

                    INDIVIDUAL ←————————→ SYSTEMIC
                    PREDICTION              SIMULATION

MOST                CollegeVine
FACTORS             Scoir AI
                    Arcidiacono (370 vars)
                    CollegeData
                    College Transitions
                                            ╔════════════════╗
                                            ║ OUR SIMULATION ║
                                            ║ 30 factors +   ║
                                            ║ market clearing ║
                                            ╚════════════════╝
                                            Reardon 2016 ABM
FEWEST              PrepScholar
FACTORS             Niche

We sit at the intersection of rich factor modeling (comparable to CollegeVine's factor count, with hooks that CollegeVine can't model) and system simulation (comparable to Reardon's ABM, but calibrated to real data). No other tool occupies this position.


12. Key Findings

1. Most chancing tools are remarkably primitive

PrepScholar, Niche, CampusReel, U.S. News, and Parchment are essentially GPA + SAT lookup tables. They provide no more insight than checking the CDS yourself.

2. CollegeVine is the clear consumer leader but has structural blind spots

75 factors and 1,500+ schools, but it cannot distinguish a legacy athlete from a first-gen student with identical stats. At Harvard, the legacy athlete's admission probability is ~30× higher. This isn't a minor omission — it's the dominant factor at selective schools.

3. Scoir is the only commercial platform with round-level predictions

This makes it architecturally the most similar to our approach among commercial tools. But it's institutional-only, doesn't model hooks, and doesn't simulate market dynamics.

4. The AI startup wave is mostly noise

20+ new "AI chancing" tools launched in 2023–2026. Most are thin wrappers with unverified accuracy claims (98.2%, 90%, 85–90%). None publish calibration data. None model hooks or rounds.

5. Academic models provide the best coefficients

Arcidiacono's 370-variable Harvard model (Pseudo R² = 0.56), Espenshade's SAT-point equivalents, Chetty's ALDC decomposition, and Hurwitz's legacy odds ratios are the gold standard for calibrating hook multipliers — and they're what we used.

6. Reardon 2016 is our closest ancestor

Same ABM architecture, same multi-round matching, same focus on structural inequality. But Reardon uses stylized colleges with 150 generic seats; we use 55 named colleges with real CDS data. Reardon uses 2 student attributes (SES, caliber); we use ~30.

7. No tool models the full system

The admissions market is a multi-round, constrained matching market with information asymmetries, strategic behavior (ED signaling), supply constraints (class sizes), and cascading effects (one student's enrollment affects another school's yield). Our simulation is the only system that models all of these simultaneously.


Sources

Commercial Tools

Academic Research