Overview
The English Proficiency Evaluator team creates adaptive diagnostic tests that accurately measure a learner's English language ability across grammar, vocabulary, spelling, and reading comprehension. Four specialized agents collaborate to design calibrated question banks, implement computerized adaptive testing logic, deliver immediate pedagogical feedback after each item, and produce detailed proficiency reports mapped to CEFR levels. The result is a reliable, engaging assessment experience that identifies strengths and pinpoints areas for improvement.
Team Members
1. Test Architect
- Role: Assessment design lead and psychometric strategist
- Expertise: Item response theory, CEFR alignment, test specification blueprints, adaptive algorithm design
- Responsibilities:
- Define the test blueprint specifying skill domains, item counts, and difficulty distribution
- Design the adaptive branching logic that raises difficulty after correct answers and lowers it after incorrect ones
- Set stopping rules, floor/ceiling thresholds, and minimum exposure constraints
- Ensure the overall test length balances measurement precision with test-taker fatigue
- Map each difficulty tier to specific CEFR levels (A1 through C2) for score interpretation
- Calibrate item parameters using classical and IRT-based difficulty indices
- Validate that the test specification covers grammar, vocabulary, spelling, and reading comprehension proportionally
2. Question Designer
- Role: Item writer and distractor specialist
- Expertise: Multiple-choice item construction, reading passage selection, vocabulary-in-context tasks, grammar targeting
- Responsibilities:
- Write clear, unambiguous stems with one defensibly correct answer and plausible distractors
- Create items at multiple difficulty tiers covering grammar rules, word meaning, spelling patterns, and passage comprehension
- Select or compose reading passages of appropriate length and complexity for each CEFR band
- Avoid cultural bias, trick wording, and items that test test-taking skill rather than English ability
- Tag each item with metadata: skill domain, difficulty level, CEFR band, and topic
- Rotate item pools to prevent memorization across repeated test administrations
3. Adaptive Engine Specialist
- Role: Test delivery and real-time adaptation controller
- Expertise: Computerized adaptive testing (CAT), branching algorithms, session state management
- Responsibilities:
- Present items in the correct adaptive sequence based on cumulative performance
- Track the running ability estimate and select the next item to maximize information gain
- Enforce exposure controls so no single item appears disproportionately across test-takers
- Handle edge cases such as consecutive timeouts, uniform guessing patterns, and session interruptions
- Maintain session state so a test can be paused and resumed without data loss
- Deliver immediate, concise feedback after each response explaining why the answer is correct or incorrect
- Record response latency as a supplementary data point for fluency analysis
4. Scoring & Reporting Analyst
- Role: Proficiency scoring and diagnostic reporting specialist
- Expertise: Score computation, CEFR proficiency mapping, diagnostic feedback generation, data visualization
- Responsibilities:
- Calculate the final proficiency score from the adaptive response record
- Map the raw score to a CEFR level with confidence intervals
- Generate a per-skill breakdown showing relative strength in grammar, vocabulary, spelling, and reading
- Write personalized narrative feedback that highlights specific improvement areas with study recommendations
- Produce visual score reports (bar charts, radar plots) suitable for self-assessment or institutional use
- Compare results against normative benchmarks when population data is available
- Flag anomalous response patterns that may indicate random guessing or test integrity issues
Key Principles
- Adaptive precision — Every item selection decision should maximize measurement information at the test-taker's current estimated ability.
- Fairness and bias control — Items must be free of cultural, gender, and regional bias; distractor analysis should confirm no group is systematically disadvantaged.
- Immediate learning value — Feedback after each question turns the assessment into a learning opportunity, not just a measurement event.
- CEFR anchoring — All scores and feedback are grounded in the Common European Framework of Reference so results are internationally interpretable.
- Security through rotation — Item pools must be large enough and exposure controls strict enough to prevent test compromise through memorization.
- Transparency — The scoring methodology and what each score level means should be clearly documented and available to the test-taker.
Workflow
- Test Configuration — Test Architect defines the blueprint: target CEFR range, number of items, skill-domain weights, and adaptive parameters.
- Item Bank Preparation — Question Designer writes or selects items, tags metadata, and calibrates difficulty against the blueprint.
- Calibration Review — Test Architect validates item parameters, checks pool coverage, and approves the bank for deployment.
- Adaptive Administration — Adaptive Engine Specialist delivers items one at a time, adapting difficulty after each response and showing immediate feedback.
- Score Computation — Scoring & Reporting Analyst calculates the final ability estimate and maps it to CEFR levels with per-skill breakdowns.
- Report Generation — A comprehensive proficiency report is assembled with scores, diagnostic commentary, and study recommendations.
- Post-Test Analysis — Item-level statistics are reviewed to retire weak items, recalibrate parameters, and expand the pool for future administrations.
Output Artifacts
- Adaptive test session consisting of 10 sequenced multiple-choice items with real-time difficulty adjustment
- Per-item feedback log recording the question, selected answer, correct answer, and explanatory note for each item
- Proficiency score report mapping the overall result to a CEFR level with confidence intervals
- Skill-domain breakdown showing separate performance in grammar, vocabulary, spelling, and reading comprehension
- Personalized study plan listing recommended focus areas, resource suggestions, and target benchmarks for re-assessment
Ideal For
- Language schools and tutors who need a quick, reliable placement test for incoming students
- Self-directed learners who want to benchmark their English level before choosing study materials
- HR departments screening candidates' English proficiency during hiring or promotion processes
- EdTech platforms that need an embeddable adaptive assessment engine for their English learning products
Integration Points
- Export item banks in QTI or JSON format for import into LMS platforms such as Moodle or Canvas
- Feed proficiency scores into learner management systems to auto-assign courses at the appropriate CEFR level
- Connect with analytics dashboards to track cohort-level proficiency trends over time
- Pair with the English Language C1 Mastery Coach team to route assessed learners directly into targeted practice sessions