Quiz System - XPs, Classification, and Analytics

You can also view the article in Markdown

Launch a high-retention practice system that is simple to use, fair to learners, and useful to creators—while keeping taxonomy light, analytics strong, and recommendations reliable.

Product Principles

System Architecture

Learner → Quiz Player → XP Engine → XP Ledger
                ↘︎ events       ↘︎ dashboards

Creator → Builder/AI → Classification → Questions Store
                          ↘︎ Item Health
                          ↘︎ Coverage Analytics

Telemetry → Analytics Layer → Recommender
              (Mastery/Health/Spacing)

Admin/QA → Reliability Suite → Classifier Tuning

High-Level System Overview

XP & Gamification

XP Rules

XP is immediate and predictable, focused on engagement rather than difficulty

Base Rewards

  • Correct Answer: +10 XP
  • Incorrect Answer: +2 XP (participation reward)

Bonus System

Daily soft cap: After 300 XP/day, payouts at 50% rate to prevent grinding

Gamification Features

1

Seasons & Leagues

Monthly seasons with XP-based leagues (Bronze/Silver/Gold). Leaderboards reset each season.

2

Cell-Aligned Badges

Earn badges like "RISK–Analyze Adept" for 30 verified corrects in that cell. Tiers at 10/30/60.

3

Session Goals

Auto-set daily goals (e.g., "Earn 60 XP today") with progress tracking.

Anti-Exploit & Accessibility

No-Speed Mode: Swaps Speed Bonus for Review Bonus to accommodate learners who need more time or have device limitations

Classification Model

4×3 Matrix (12 Cells)

Domains

DomainFocusExample Applications
ANALYSISReading/interpreting market dataPattern recognition, signal validation
STRATEGYChoosing approach and playbookPlan selection, setup comparison
RISKPosition sizing and capital managementStop placement, size calculation
EXECUTIONOrder mechanics and timing policyOrder type selection, venue choice

Bloom's Levels

Classification Precedence

Precedence order: EXECUTION > RISK > STRATEGY > ANALYSIS

Hybrid Classifier

1

Rules Layer

Keyword markers + precedence resolution

2

LLM Challenge

Request domain & Bloom with "why-not" reasoning for other domains

3

Decision Logic

Rules == LLM → high confidence; else apply precedence or review

4

Reliability Target

κ & AC1 ≥ 0.75 on balanced 12-cell set

Content Creation

Doc-Driven AI Pipeline

{
  "domain": "RISK",
  "bloom": "Analyze",
  "constraints": ["policy-level execution only"],
  "grounding": ["doc://ch12#para3", "doc://ch12#fig2"]
}

AI Content Generation Flow

Manual Builder Features

  • Coverage-aware prompts ("Low on EXECUTION–Analyze")
  • Creator chooses domain + bloom or accepts auto-suggestion
  • Parametric templates offered (not required)
  • Same validator ensures consistency

Analytics Layer

Item Health Metrics

MetricDescriptionUse Case
Success RateCorrect answers per itemBasic performance
DiscriminationΔ success between top/bottom quartilesItem quality
Abandon RateQuits/timeouts per itemDifficulty indicator
Time Z-scoreDeviation from medianAmbiguity detection

Health status progression: new → stable → needs_edit → quarantined

Mastery Tracking

Beta model per cell: α = 1 + correct, β = 1 + incorrect Mastery = α/(α+β)

Spacing & Review

  • If days_since_last_correct > 7 and mastery < 0.70 → inject review
  • Light-touch spacing without heavy scheduler

Dashboard Types

Parametric Templates

Why Parametric Templates?

Template Structure

Each template includes:

  • Inputs/Slots: {ASSET}, {TIMEFRAME}, {ENTRY_RULE}
  • Constraints: Value ranges, mutually exclusive combos
  • Answer Rule: How to compute correct choice
  • Distractors: Principled wrong answers
  • Validation: Classification guardrails

Example Template: Position Size Calculation

Prompt: Account {EQUITY}. Risk {RISK_PCT}% per trade. 
        Entry {ENTRY}, stop {STOP}. What size?

Answer Rule: size = (EQUITY × RISK_PCT) / |ENTRY–STOP|

Distractors:
- Swap TP for stop
- Percent-of-equity share count
- Decimal errors

Implementation Timeline

1

Week 1

XP engine with base rewards, bonuses, streaks, session goals

2

Week 2

4-domain/3-Bloom classifier, precedence rules, builder UI

3

Weeks 3-4

Performance tracking, item health, dashboards v0

4

Month 2

LLM validator, confidence scoring, reduce manual review to ~20%

5

Month 3

Beta mastery, recommender v1, spaced review, badges & leagues

6

Month 6

Consider info_complexity if analytics justify (analytics-only)

Success Metrics

Primary Metrics

  • κ & AC1 ≥ 0.75 (balanced 12-cell set)
  • D7 retention ≥ 40%; D30 ≥ 20%

Secondary Metrics

  • +8–12% improvement in cell-level mastery within 7 days
  • Creator efficiency: < 90s median to confirm labels
  • Content health: < 5% traffic to needs_edit/quarantined items

Alert Thresholds

  • Speed bonus payout > 40% → tighten normative threshold
  • Any domain < 15% inventory for 2 weeks → trigger creator challenges

Risk Mitigation

RiskMitigation
Speed farmingAccuracy gate + normative pace + daily cap
Label driftPrecedence rule + balanced validation
Supply imbalanceCoverage-aware prompts + creator challenges
AccessibilityNo-Speed Mode with Review Bonus
Data sparsityMinimum attempts for quarantine

Future Enhancements

Phase A (Months 1-2)

  • Creator impact tracking
  • Basic personalization
  • Friendly competitions

Phase B (Months 3-6)

  • Creator payments based on learning impact
  • Study groups
  • Content difficulty ranking (analytics-only)

Phase C (Months 6-12)

  • Advanced predictions
  • Seasonal narratives
  • Mentorship programs

All future features would be for analytics only (no XP impact) and added carefully to avoid complexity bloat

    Quiz System - XPs, Classification, and Analytics - Wawe Docs