Every claim is verifiable

NeuroFrame is built on psychometrics, neurophysiology, and machine learning. We publish our metrics, validate our models, and hold ourselves to the highest scientific standards in the assessment industry.

How NeuroFrame works

Four scientific pillars that make NeuroFrame fundamentally different from questionnaires, interviews, and traditional assessment centers.

Game-Based Assessment (GBA)

Instead of questionnaires, candidates play an adaptive Tower Defense game. The game engine captures 4,000+ micro-behavioral signals per session: click timing, strategy changes, resource allocation patterns, error correction speed, risk-taking under pressure. These signals are impossible to fake through test preparation.

  • Adaptive difficulty that responds to player behavior in real time
  • Anti-cheat detection via behavioral pattern analysis
  • Cross-platform: iOS / Android / Web — identical signal capture
  • 30-minute session replaces 3–5 hours of traditional testing

Behavioral Signal Extraction

Raw game data is processed through a proprietary signal extraction pipeline. We don't look at "did they win" — we analyze how they played. The pipeline converts raw telemetry into 120+ behavioral features organized into cognitive and personality domains.

  • Temporal patterns: reaction times, decision intervals, micro-pauses
  • Strategic patterns: planning depth, adaptation speed, resource efficiency
  • Emotional patterns: stress response, error recovery, persistence after failure
  • Meta-cognitive patterns: self-monitoring, strategy evaluation, learning curves

Machine Learning Models

Competency scores are generated by an ensemble of ML models trained on validated performance data. Each competency has its own calibrated model with published metrics. The system is not a black box — every prediction has an explainability layer showing which behavioral signals contributed to the score.

  • Ensemble approach: Gradient Boosting + Neural Networks + Logistic Regression
  • Cross-validated on held-out samples (k=10 fold)
  • Calibrated probability outputs (not just scores)
  • Continuous retraining with new performance data quarterly

Psychometric Validation

Every scale in the NeuroFrame model is validated using Confirmatory Factor Analysis (CFA), reliability analysis, and criterion validity studies. We don't publish metrics for marketing — we publish them because it's scientific standard.

  • Structural validity: CFA with CFI > 0.95, RMSEA < 0.06
  • Internal consistency: Cronbach α > 0.70 for all scales
  • Criterion validity: predictive of job performance (KPIs, 360° ratings)
  • Convergent validity: correlates with validated instruments (NEO-PI-R, WAIS)

What's behind the numbers

Every metric is published. Every claim is verifiable. Here's what NeuroFrame's psychometric properties actually mean — in plain English.

0.96

CFI (Confirmatory Fit Index)

What is it?

CFI shows how well the assessment model fits real data. It's a structural validity check — does the test actually measure what it's supposed to? A score of 1.0 is perfect; above 0.95 is excellent by all global standards.

Comparison

MBTI: 0.80–0.89. Most personality questionnaires sit at 0.85–0.92. NeuroFrame: 0.96 — our model is rigorously validated against real behavioral data, not self-report.

0.74

Cronbach α (Internal Consistency)

What is it?

Cronbach's alpha measures whether the signals within each competency scale are coherent — are we measuring one construct, or noise? Above 0.70 is the accepted scientific threshold. 0.74 confirms NeuroFrame's scales are internally consistent and reliable across demographics.

Comparison

Many commercial tests don't publish α at all. Those that do often report 0.60–0.70 for behavioral scales. NeuroFrame exceeds the threshold — your data is reliable, not random.

0.46

R² (Predictive Power)

What is it?

R² shows what share of job performance NeuroFrame can predict. 0.46 means our model explains 46% of performance variance — which sounds modest in everyday terms, but in psychometrics it's among the highest published figures for any single assessment tool.

Comparison

Unstructured interviews: R² = 0.05–0.10. Assessment centers: R² ≈ 0.15. Big Five questionnaires: R² ≈ 0.10. NeuroFrame: R² = 0.46 — 5–9× more accurate than interviews, 3× more than assessment centers.

0.77

ROC AUC (Classification Accuracy)

What is it?

ROC AUC measures how well the algorithm separates "good fit" from "bad fit" for a given role. A score of 0.5 is random; 1.0 is perfect. NeuroFrame's 0.77 means the system correctly classifies candidates in 77% of cases — which is exceptional for a 30-minute game.

Comparison

Resume screening: ~0.55 (barely above random). Structured interviews: ~0.60. NeuroFrame: 0.77 — reliable, data-driven separation of top performers vs. the rest.

0.83+

Test-Retest Reliability

What is it?

Test-retest reliability measures stability: if the same person is assessed twice (weeks apart), do they get the same result? Above 0.80 is "excellent" in psychometric standards. NeuroFrame reports 0.83+ — meaning the assessment captures stable traits, not mood or luck.

Comparison

Traditional interviews: near-zero test-retest reliability (different interviewer = different result). NeuroFrame: 0.83+ — the result is reproducible and consistent.

5,000+

Validation Sample Size

What is it?

Our model was trained and validated on a dataset of over 5,000 real professionals from 500+ companies across industries and geographies. This is not an academic prototype — it's a production system built on real-world performance data.

Comparison

Many competing tools report sample sizes of 200–500. Academic benchmarks typically require N > 300. NeuroFrame: N > 5,000 — statistically robust and generalizable.

What the algorithm sees

NeuroFrame measures 8 competencies across two domains: cognitive (how you think) and personality (how you behave). Each competency is mapped from specific in-game behaviors to real business impact.

Learning Agility

How quickly a person grasps new requirements and adapts in changing environments. Measured through tactical adaptation speed across game levels with increasing complexity.

In Game

Rate of improvement across levels. How fast the player abandons failing strategies and adopts new ones when the environment changes.

In Business

Speed of onboarding, adaptation to organizational change, ability to operate in ambiguity.

Mental Efficiency

Ability to find optimal solutions under time and resource constraints. Reflects planning capability, prioritization, and cognitive bandwidth.

In Game

Resource utilization efficiency — how much output per unit of input. Correlation between planning time and execution quality.

In Business

Task prioritization, resource allocation, strategic decision-making quality under deadlines.

Progress Monitoring

Ability to track key indicators, notice deviations, and self-correct toward goals. Reflects meta-cognitive awareness and feedback sensitivity.

In Game

Frequency of status checks, response time to declining KPIs, proactive vs. reactive adjustments.

In Business

KPI tracking, project management discipline, ability to self-correct before problems escalate.

The 8-competency profile

Every candidate and employee receives a detailed individual report with scores across all 8 competencies, presented on an octagonal radar chart. The report includes:

  • Individual scores (percentile-ranked against benchmark population)
  • Strengths and development areas
  • Fit score against target role profile
  • Hire / Don't Hire recommendation (for Test)
  • Team compatibility analysis (for Team)
  • Leadership potential index (for HiPo)
  • Explainability layer — which behavioral signals drove each score
NeuroFrame Report — 8-competency radar chart

NeuroFrame vs. alternatives

Side-by-side comparison with traditional assessment tools and interviews — on metrics that matter.

FeatureNeuroFrameTraditional AssessmentInterviews
Assessment Duration30 minutes3–5 hours45–60 min / round
Predictive Validity (R²)0.460.10–0.150.05–0.10
Test-Retest Reliability0.83+0.70–0.80~0.20
Fakability / Social DesirabilityMinimal (behavioral data)High (self-report)Very High
Scalability5,000+ per cycle50–200 per cycle10–30 per recruiter
Bias RiskLow (blind assessment)Medium (demographic patterns)High (unconscious bias)
Candidate ExperienceEngaging (game)Tedious (questionnaires)Variable (interviewer-dependent)
Cost per AssessmentFrom $45$200–$500$300–$1,000+
Time to ResultsInstant (real-time ML)2–5 business days1–3 weeks
Anti-CheatingBuilt-in (behavioral anomaly detection)NoneNone

Peer-reviewed research

Our scientific team publishes in international peer-reviewed journals. We believe in open science — if you can't verify it, you shouldn't trust it.

2024Peer-reviewed

Game-Based Assessment of Cognitive Competencies: Validation Study

Journal of Applied Psychology

2024Peer-reviewed

Behavioral Micro-Signals in Digital Games as Predictors of Job Performance

Personality and Individual Differences

2023Peer-reviewed

Psychometric Properties of a Game-Based Personality Assessment: CFA and Reliability Analysis

European Journal of Psychological Assessment

2023Peer-reviewed

Adaptive Game Environments for Non-Invasive Psychological Profiling

Computers in Human Behavior

2025Conference

Anti-Cheating Detection in GBA: Behavioral Anomaly Framework

International Conference on AI in HR

2025Industry

Cross-Cultural Validity of Game-Based Cognitive Assessment

Assessment & Development Matters

Scientists & Engineers

Behind NeuroFrame is a multidisciplinary team from science, ML, and the HR industry — united by the belief that hiring decisions should be based on evidence, not intuition.

Founders

Neuroscience + AI/ML + HR-Tech. Three domains in one founding team — each with 10+ years of domain experience.

Scientific Board

5 PhDs spanning psychometrics, neurophysiology, organizational psychology, and computational cognitive science.

ML Engineers

Machine learning engineers specializing in behavioral signal processing, time-series modeling, and adaptive algorithms.

Full-Stack Engineers

Game developers and platform engineers building cross-platform adaptive assessment systems.

12+Scientific Publications
3Domains of Expertise
25K+Profiles Processed
5PhDs on the Team
500+Companies Served
4Years of R&D

Want to see the data?

Request a validation report, schedule a scientific briefing, or explore a pilot assessment for your team.