Rigorous, independent certification for AI skills across every platform. Three pillars. One standard.
Every skill submitted to Scientia Ex Machina undergoes a weighted three-pillar evaluation. The overall score is a composite of functional correctness, security posture, and runtime performance.
Does the skill do what it claims?
Is it safe to run?
Does it perform under load?
SXM certification is deliberately hard to earn. If it were easy, it would be worthless.
Most skills fail on their first submission. That is by design.
Every score is deterministic and reproducible. Here is exactly how each pillar is assessed.
We generate test scenarios from the skill's declared manifest (inputs, outputs, failure modes). Each scenario is run multiple times to verify consistency.
We run an adversarial test suite that evolves weekly based on published CVEs, research papers, and real-world incidents.
We benchmark under simulated production load. Skills must perform consistently, not just on a single quiet run.
The overall score is a weighted composite: overall = (functional * 0.40) + (security * 0.35) + (performance * 0.25)
A skill must score 90+ overall and 85+ on the security pillar independently. There is no trading off security for performance. Learn more about scoring.
Our evaluation criteria are not static. The threat landscape changes every week, and our evaluator changes with it.
Every week, our evaluator ingests new data from three sources:
Certification is not permanent. As our evaluator evolves:
GET /api/evolution/history). You can see exactly what changed and when.A certification from January that does not account for a vulnerability discovered in February is worthless. Static certification creates a false sense of security.
SXM certifications are living credentials. They reflect the current state of the threat landscape, not a snapshot from the day the skill was submitted.
This is modelled on how financial audit standards evolve: the bar rises over time as the industry matures.
Our certification provides evidence toward NIST AI RMF, ISO/IEC 42001, EU AI Act, and Colorado AI Act alignment. We do not certify compliance. We help you demonstrate it.
View Standards MappingTask automation and tool skills
Anthropic model capabilities
IDE and code generation
Model Context Protocol
Any AI platform