Install our app for a better experience!
AI Language Testing

AI-Powered Language Test System

Practice IELTS, TOEFL iBT, GRE, GMAT, SAT and 9 more standardized tests with AI-generated content that is 96% indistinguishable from official exams. Instant AI scoring for Speaking and Writing — calibrated to official band descriptors.

12+ Test Types 60+ Question Formats 120B AI Model 23 TTS Voices
Book a Demo Start Free
96%
Match to Official Exam Quality
120B parameter generation model
12+
Standardized Test Types
IELTS, TOEFL, GRE, GMAT, SAT & more
60+
Distinct Question Formats
Reading, Listening, Writing, Speaking
99.9%
TTS Audio Reliability
23 voices, multi-accent coverage
How It Works

Three Steps to a Scored Practice Test

From test selection to detailed AI feedback — the entire loop runs in one session.

1
Choose Your Test

Select from 12+ standardized test types or build a fully custom test. Set your section preferences, timing mode (timed or untimed), and difficulty target.

2
Take the Test

Work through sections with authentic timing, AI-generated audio for Listening, and a clean test-day interface. Section-by-section progression mirrors the real exam flow.

3
Get Scored Instantly

Receive official-scale band scores, section breakdowns, per-criterion AI feedback for Speaking and Writing, multi-perspective feedback from 3 AI personas (Examiner, Study Coach, Fellow Student), model answers, and concrete improvement suggestions — all immediately after submission.

Supported Tests

12+ Standardized Test Types — All in One Platform

Every test follows official timing, section structure, question formats, and scoring scales. Updated for 2023 international standards.

IELTS
165 min · 0–9 band
TOEFL iBT
116 min · 0–120
GRE
Adaptive · 4-check verdict verifier
GMAT
4-check verdict · Cross-model verbal
SAT
Adaptive · Desmos calc · 4-check verdict
Duolingo
English Test
PTE Academic Aug 2025
22 task types · 10–90 scale · partial & negative marking
PTE Core
Immigration track · 10–90 scale
OET 2025
Healthcare English · 12 professions
CELPIP
CLB 0–12
Adaptive Language
IRT-based
Custom
Your content
Unique Differentiator

Custom Test Builder

One-click IELTS / TOEFL / PTE / OET / CELPIP replicas — real-exam section counts and durations auto-fill on page load. Pick a result format that fits your use case: numeric points, CEFR placement (A1–C2), or native band (IELTS Band 6.5, PTE 70/90, OET Grade B, TOEFL, CLB). Or go Fully Custom with your own passages, transcripts, and prompts.

IELTS/TOEFL/PTE replica auto-fill
10 languages · A1–C2 difficulty
Output: Points · CEFR · Native Scale
Topic pools per test type
Assigned-only placement workflow
Your passages, transcripts, prompts
Numeric Points

Classic numeric/percentage score (e.g. 72 / 100). Best for corporate assessments, internal practice, and ungraded check-ins.

CEFR Level (A1–C2)

Student gets a CEFR placement: "CEFR B2 — Upper Intermediate", per-skill CEFR badges, A1–C2 band axis, and a reference table. Recommended for placement testing at Intermediate difficulty.

Native Scale

Real-exam scale: Band 6.5 (IELTS), 70 / 90 (PTE), Grade B (OET), 87 / 120 (TOEFL), CLB 8 (CELPIP). Per-skill native scores. Ideal for exam-prep cohorts.

Reading Section

MCQ · True/False/Not Given · Fill-blank · Summary completion · Short-answer. Topics auto-pulled from base-style pool (IELTS academic disciplines, OET healthcare, CELPIP everyday-Canada) for natural per-passage variety.

Listening Section

MCQ · Form completion · Matching · Fill-blank questions with AI-generated audio from your transcript using 23 TTS voices in the target language.

Writing Section

AI-evaluated tasks using your prompt or pool-pulled topics. Rubric calibrated to official scoring (IELTS, TOEFL, OET, PTE, CELPIP).

Speaking Section

AI-evaluated spoken response to your topic. Scored on fluency, pronunciation, grammar, vocabulary, and (for OET) clinical communication.

Built-in Placement-Test Workflow

Running a single-test cohort (e.g. CEFR placement on intake)? Set max_visible_tests=0 on the student's enrollment and they see only the test you assign them — your full Custom Test library stays hidden. Open it up later by removing the limit. No more "student sees five practice tests when you wanted them to take one placement."

Result page + PDF include an honest calibration disclosure: "internal placement — expect ~1 level lower on a real proctored sitting." No over-claiming.

AI Evaluation Engine

Calibrated AI Scoring — Speaking & Writing

Our 17B parameter evaluation model uses official band descriptors for each test type. No generic rubrics — every score mirrors what a certified human examiner would give.

Speaking Evaluation

Scored per criterion on official scale. Example below shows an IELTS response.

Fluency & Coherence
7.0
Pronunciation
6.0
Grammatical Range
7.5
Lexical Resource
6.5
Scale applies per test:  IELTS 0–9 TOEFL 0–4 PTE per-task CELPIP CLB
Writing Evaluation

Rubric calibrated per test type. Example below shows a TOEFL Integrated task.

Task Achievement
4/5
Coherence & Cohesion
3/5
Lexical Resource
4/5
Grammatical Range
3/5
Scale applies per test:  IELTS 0–9 TOEFL 0–5 GRE 0–6 CELPIP CLB 0–12 PTE per-task

What You See on the Result Page

Instant results — no waiting. Example IELTS full test result.

Reading
7.5
Listening
6.5
Writing
6.0
Speaking
6.5
Overall Band
6.5

Result includes: AI feedback per section · Model answers · Multi-perspective feedback · Improvement suggestions · Downloadable report

New

Multi-Perspective Feedback Panel

Writing and Speaking results include feedback from 3 distinct AI personas — each referencing your actual text and full test context.

Examiner

Strict, rubric-focused

Identifies exactly which band descriptor you missed and how it affects your overall score. Quotes your actual response.

Study Coach

Supportive, actionable

Connects strengths from other sections to this one. Gives one before/after rewrite using your actual text.

Fellow Student

Relatable, practical

Shows how fixing one area could raise your overall band. Shares a concrete study technique that works.

Available for all test types: IELTS, TOEFL, PTE, OET, CELPIP, GRE, Duolingo, Adaptive Language, and Custom tests.

OET 2025 — The Deepest Healthcare-English Track

Not a generic language test with a medical coat of paint. Real OET 2025 format, real 2025 rubrics, real examiner-grade feedback.

Exact 2025 Format

Listening 24/6/12 split (2 consultations × 12 gaps + 6 workplace MCQs + 2 presentations × 6 MCQs). Reading 20/6/16 with matching, sentence-completion, and short-answer across Part A's 4 texts. Writing = 1 profession-specific letter. Speaking = warm-up + 2 role-plays on canonical three-section cards.

Transcript-Grounded Listening

Every Part A gap answer must appear verbatim in the generated transcript — a post-parse filter drops any gap whose answer isn't actually spoken. The "question asked for a date but the audio never mentioned one" bug can't exist.

Real 2025 Rubrics

Writing scored on Purpose (0-3) + Content / Conciseness & Clarity / Genre & Style / Organisation & Layout / Language (0-7 each) — max 38, anchored at raw 27 = Grade B floor. Speaking scored on 4 Linguistic criteria (0-6) + 5 Clinical Communication criteria (0-3) with a 60/40 weighted formula.

Examiner-Grade Writing Feedback

Every criterion score is backed by a verbatim quotation from the candidate's letter. The lowest-scoring paragraph gets a band-7 rewrite side-by-side with the original. 12-item examiner-gotcha checklist flags things like copy-paste from case notes, bullet points in the body, missing salutation, purpose buried past paragraph 1.

Patient-ICE-Aware Speaking Scoring

Every role-play card ships with a hidden patient brief (Ideas / Concerns / Expectations + verbal cues). The evaluator uses it to score whether the candidate actually elicited and incorporated the ICE framework — not just fluency. Moment-level feedback with timestamps shows exactly where empathy cues were missed or signposting broke down.

Multi-Session Band Projection

After a handful of sittings, students see their current grade, projected grade in 4 weeks at the current pace, weakest section with point-gap to the next grade, and an estimate of days to reach Grade B. Trend-based — no fake optimism when performance is flat or declining.

Harder-Than-Real Difficulty Tiers

Intermediate matches real-OET rigor 1:1. Advanced is measurably harder (C1 medical vocabulary, near-miss distractors, co-morbid scenarios). Expert is two tiers above real (C2+ academic medical language, multi-system vignettes). Scoring stays real-exam calibrated — a Grade B at Advanced means a Grade A on the real test.

12 Healthcare Professions

Medicine, nursing, pharmacy, dentistry, dietetics, occupational therapy, optometry, physiotherapy, podiatry, radiography, speech pathology, veterinary science — each with dedicated scenarios, workplace extracts, presentation topics, and role-play situations. Thousands of distinct mock tests before anything visibly repeats.

360-Combo Standardised Pools

When institutes create a 30-test standardised OET pool, each test #N lands on a deterministic (profession × topic) pair. 12 professions × 30 themes = 360 distinct combinations before any repeat — students never see the same test twice.

For Institutes & Test Prep Centers

Teacher Review Gate — Hold, Comment, Release

Give teachers a 24-hour window to review every result before scores publish — or let it auto-publish if they don't act. Per-student opt-in. Never blocks the student indefinitely.

Per-Student Opt-In

Org admin flips a "Teacher Review" switch on the student's enrollment in the Institute Student Management page. Some students get the hold (e.g. exam-bound IELTS students whose final mock needs teacher sign-off); others don't (e.g. self-paced learners on daily reading drills). Works on any language test the student takes — org-scoped or pool tests like IELTS / TOEFL — because the gate keys off the student, not the test.

24-Hour Hard Ceiling

When a held student submits a test, the score is computed normally but visibility is gated. Teachers get an in-app notification + email. They can release early or let the result auto-publish at the 24-hour mark via an hourly background sweep. The student never waits longer than 24 hours — even if the institute goes silent for a week.

Live Countdown for the Student

The "Awaiting teacher review" page shows the test name, submission time, deadline, and a live ticking countdown ("Auto-release in 17h 28m 04s"). The page polls the deadline and reloads itself the moment auto-release fires — no manual refresh. An email lands the same second.

Teacher Comments

Teachers can save a free-text comment without releasing (mid-review pause), or attach it to the release action — either way the comment is shown to the student on the result page. Stored on the session itself, audited with reviewed_by and reviewed_at.

Inline Annotations + Replies

Teachers can highlight any text in a student response and attach a threaded comment ("Strong topic sentence — but the supporting example doesn't connect"). Once released, students can reply on the same thread — turning a one-shot score into an actual coaching dialogue. Teachers can mark threads resolved, audit-stamped with resolved_by / resolved_at.

Teacher Review Dashboard

A dedicated dashboard at /language-tests/reviews/ with two tabs: Pending (sorted by deadline, with per-row hours-left countdown) and Released (filterable by 30 / 90 / 365 days / all-time, showing who released and whether it was teacher-released or auto-released after 24h). Text search across student name, email, and test title.

Multi-Channel Teacher Alerts

When a held result lands, every active org admin and examiner gets two simultaneous nudges: an in-app bell-icon notification (with deep-link to the result page) and an email ("Result pending your review — Student Name / Test Title"). De-duplicated per session so retries never re-spam teachers.

Strict Org-Level Authorization

Only active org admins and examiners on the test's organization (or the student's enrollment institute) can review. Site-wide super-admins are intentionally excluded from the teacher-review action — keeping reviews local to the institute that owns the student. Django staff/superuser retain inspection access via the standard authorization path.

Full Audit Trail

Every reviewed session stamps review_status, review_deadline, teacher_comment, reviewed_by, reviewed_at, released_at, and released_auto (true when the 24h sweep released it, false when a teacher released early). Institutes can audit which sessions slipped past teacher review for staffing and SLA reporting.

Why this works for a "practice platform that must always deliver results"

Most AI test platforms either drop raw AI scores on the student instantly with no teacher in the loop, or they make the teacher a hard bottleneck for every result. This feature gives institutes the best of both: a real teacher review window with comments and inline annotations, but never at the cost of the student waiting indefinitely. The 24-hour ceiling is a hard guarantee — auto-release is unconditional and runs hourly. It's the difference between AI that replaces teachers and AI that gives teachers their time back.

PTE Academic — Aug 2025 Format-Authentic

Format-authentic UI, real partial credit, and the new SGD task — all 22 task types from the official Aug 2025 spec, end-to-end.

All 22 Task Types — 4 sections

Reading (5): R&W FIB dropdown, Reading FIB drag-bank, Re-order Paragraphs, MC Single, MC Multi. Listening (8): SST, FIB inline, HCS, SMW, HIW, WFD, MC Single, MC Multi. Speaking (8 incl. Personal Intro): Read Aloud, Repeat Sentence, Describe Image, Re-tell Lecture, Answer Short Question, Summarize Group Discussion, Respond to a Situation. Writing (2): Summarize Written Text, Write Essay.

Real-exam UI mechanics

R&W FIB renders one passage with inline dropdowns per blank — not separate one-blank-per-question fragments. Reading FIB drag-bank uses one shared word bank (more words than blanks) to drag into inline drop-targets. Re-order Paragraphs is two-pane drag (source left → target right). The passage itself is the task for FIB — no separate read-only panel that would reveal the answers.

Real partial & negative marking

Reorder uses adjacent-pair scoring (1 point per correctly placed consecutive pair, max n−1). MC Multi and Highlight Incorrect Words apply +1 correct / −1 wrong, floored at 0. Multi-blank FIB and Listening FIB give per-blank credit. Write from Dictation scores per-word. None of this is binary all-or-nothing.

Summarize Group Discussion (NEW Aug 2025)

3-speaker discussion synthesised through our multi-voice TTS pipeline (10-voice US/UK pool, deterministic per-speaker assignment). Counts toward both Listening and Speaking. AI eval explicitly penalises memorised templates — content score depends on capturing each speaker's distinct viewpoint.

Audio plays once · Real Exam Mode

Per-session opt-in. When on: audio replay disabled, custom player with no scrubbing, per-task hard timers, no pause. Practice mode is forgiving (auto-submit on timer expiry, soft timing); real-exam mode mirrors Pearson's strict conditions for serious mock attempts. Toggle on the test-start screen.

No leaked answer markers

Pre-save sanitiser strips leaked A) /B) letter prefixes and (correct)/(incorrect) tags from option text — both at write time and via a one-shot backfill on existing items. The validator also tolerates JSON-encoded full-text answers that legacy data may produce, so students aren't penalised for AI artifacts.

10–90 score with explainer

Each section converts net partial-credit points to the official 10–90 scale; overall = average of 4 sections. The result page shows per-question slot ratios (e.g. "2 / 3 blanks correct" plus the 0–1 normalised score) and a collapsible "How is PTE Academic scored?" explainer covering every task type's scoring rule. Honest about being indicative, not Pearson-calibrated.

Save queue · pause & resume

Every answer routes through a resilient save queue with localStorage persistence and exponential-backoff retry. Multi-blank dropdowns, drag-bank chips, two-pane reorder, inline FIB inputs, clickable HIW transcripts — all restore correctly on resume. Mid-test pauses preserve progress on every question type.

Tested end-to-end

Every question type has automated coverage from the JS wire format → submit_response → validator → DB → complete_test → 10–90 conversion → result-page render. Includes regression tests for JSON-string MC submissions, leaked answer markers, partial-credit display, and per-blank scoring across 13 distinct question types.

Who It's For

Built for Students, Educators, and Institutions

Individual Test-Takers

Preparing for an exam abroad? Practice with tests indistinguishable from the real thing and get targeted feedback to fix your weakest areas.

  • Unlimited practice tests on your schedule
  • Instant band scores with section breakdown
  • AI feedback on every Speaking & Writing response
  • Track progress across multiple attempts
  • Adaptive GRE and SAT for accurate ability estimate
Language Schools & Test Prep Centers

Run mock exams, track cohort progress, and deploy custom tests built from your own course materials — all from one platform.

  • Batch student enrollment and management
  • Custom tests from your proprietary content
  • Examiner role with review and override access
  • 24h teacher review hold with comments and inline annotations
  • Cohort-level analytics and reporting
  • White-label under your school's brand
Universities & Corporate Programs

Deploy standardized English proficiency screening or professional language certification at scale — self-hosted under your institution's domain.

  • SSO and organization-level access control
  • OET for healthcare hiring and credentialing
  • CELPIP for immigration language assessment
  • Custom writing prompts for domain-specific tasks
  • Full API access for LMS integration
FAQ

Frequently Asked Questions

Our 120B parameter AI generation model produces practice tests that are 96% indistinguishable from official exams. The model is trained on official test formats, authentic question styles, timing structures, and scoring rubrics for every supported test type — including IELTS (165 min, 0-9 band), TOEFL iBT (116 min, 0-120 scale, July 2023 format), GRE, GMAT, SAT, PTE, CELPIP, OET, and more.

Speaking responses are evaluated by our 17B parameter language model across four dimensions: fluency, pronunciation, grammar, and vocabulary — each scored on the official 0-9 band scale for IELTS or the test-specific scale for other exams. Writing is evaluated on task achievement, coherence and cohesion, lexical resource, and grammatical range — calibrated per test type (IELTS 0-9, TOEFL 0-5, GRE 0-6, CELPIP CLB 0-12, PTE per-task criteria). You receive detailed AI feedback, model answers, and improvement suggestions immediately after submission.

The Custom Test Builder lets you create fully personalized language tests using your own content. Provide your own reading passages, listening transcripts, writing prompts, or speaking topics — the AI can automatically generate questions from your content, or you can use it 100% as-is. All four sections are supported: Reading (MCQ, True/False, Fill-blank), Listening (MCQ, Fill-blank with AI-generated audio), Writing (AI-evaluated essay), and Speaking (AI-evaluated response). Ideal for language schools, universities, and corporate training programs.

The platform supports 12+ test types: IELTS, TOEFL iBT (July 2023 format), GRE (adaptive), GMAT, SAT (adaptive), Duolingo English Test, PTE Academic (Aug 2025 format with all 22 task types incl. Summarize Group Discussion), PTE Core, OET, CELPIP, Adaptive Language, and Custom. Each test type follows official timing, section structure, question formats, and scoring scales. IELTS uses 165 minutes with a 0-9 band scale; TOEFL iBT uses 116 minutes with a 0-120 scale; PTE Academic uses ~2 hours with a 10-90 scale and real partial/negative marking.

Full PTE Academic Aug 2025 spec: all 22 task types across Reading (5), Listening (8), Speaking (8 incl. Personal Intro), and Writing (2). Real-exam UI mechanics — multi-blank R&W FIB with inline dropdowns, Reading FIB drag-bank with one shared word bank (more words than blanks), two-pane Re-order Paragraphs, Summarize Group Discussion with multi-voice 3-speaker audio (NEW Aug 2025). Real partial & negative marking: Re-order uses adjacent-pair scoring, MC Multi and Highlight Incorrect Words apply +1/−1 floored at 0, Write from Dictation scores per-word, FIB types give per-blank credit. Optional Real Exam Mode locks audio to play-once with no replay/scrubbing/pause. Score conversion to the official 10–90 scale per skill plus an in-page "How is PTE Academic scored?" explainer covering every task type's rule. Practice scores are indicative — Pearson's psychometric IRT calibration isn't replicated.

All listening audio is generated using professional text-to-speech technology with 23 available voices — powered by UnrealSpeech V8 with Groq Orpheus as a fallback, delivering 99.9% reliability. Voices span multiple accents and speaker profiles to replicate the variety found in official IELTS, TOEFL, and PTE listening sections. Audio is generated fresh for each test instance, so no two tests sound the same.

Yes. The language test system is fully white-label ready. Institutions — universities, language schools, test prep centers, and corporate training programs — can deploy the entire platform under their own branding. The system supports organization-level configurations, custom test libraries, student management, and role-based access for administrators, examiners, and students.

Yes. The platform has an opt-in Teacher Review Gate per student. When the org admin enables "Teacher Review" on a student's enrollment, every language test that student submits is held for up to 24 hours of teacher review before the result is released. Teachers (org admins and examiners) get an in-app notification + email the moment a result is pending, and they can act from a dedicated Teacher Review Dashboard at /language-tests/reviews/. They can save comments without releasing, add inline annotations on any selection in the student's response (with threaded student replies once released), or release immediately with an optional final comment. If the teacher does nothing, results auto-publish at the 24-hour mark via an hourly background sweep — so students never wait longer than 24 hours, even if the teacher is silent. Every action is audited with reviewed_by, reviewed_at, released_at, and released_auto.

A clean "Awaiting teacher review" page with the test title, submission timestamp, the deadline ("Results released by Apr 30, 2026 11:14 PM"), and a live ticking countdown to auto-release. The page polls the deadline and reloads itself the moment the teacher releases or the 24-hour auto-release fires — students don't have to refresh. An email also lands in their inbox the moment the result is available. Scores are never lost — they're computed and stored on submission; only visibility is gated. Once released, the standard result page renders with the teacher's comment block at the top and any inline annotations the teacher attached to specific parts of the student's response (which the student can reply to in-thread).

Start Your AI Language Test Preparation Today

12+ test types · 60+ question formats · Instant AI speaking and writing scores

Book a Demo Sign Up Free