Verification Standard
Every artifact in the Soulmetric corpus carries two things: a verification tier (who checked it) and a Scharp Scale score (how well it held up). Here's how both work.
Tiers answer the question: who verified this?
Human-verified: multiple experts.
At least two independent domain experts have reviewed the working paper's claims, methodology, citations, and conclusions. Disagreements between reviewers are noted and resolved or flagged.
Human-verified: single expert.
One qualified domain expert has reviewed the working paper and confirmed its core claims, structure, and sources. The reviewer's domain and credentials are recorded.
External AI adjudication.
External AI adjudication by independent AI systems (not the generating model). This is machine verification, not human certification. It catches structural errors, citation failures, and logical inconsistencies but cannot substitute for human domain expertise.
Unverified generated output.
Raw output from the generation pipeline. Has not passed through any verification step. May contain errors, hallucinations, unsupported claims, or fabricated citations.
The verification tier and the Scharp Scale measure different things:
A working paper can be Bronze-tier (AI-adjudicated only) but score an 8 on the Scharp Scale—meaning the AI verification was thorough and the paper held up well. Conversely, a Silver-tier working paper might score a 4 if the human reviewer found significant issues that required correction.
Both are shown on every artifact. You always see WHO verified it and HOW it scored.
A verification quality score from Negative through 10.
Developed by Kevin Scharp. The scale is proprietary and has not been externally validated. We publish it for transparency, not as an industry standard.
AI agents produce working papers using specified methodologies. Each agent specializes in different research tasks: literature synthesis, data analysis, argument construction, cross-referencing.
Hallucination detection systems check every factual claim. Citation verification confirms sources exist and say what's attributed. Internal consistency checks flag contradictions.
Independent AI systems (different from the generating model) review each working paper against source material, known literature, and domain-specific standards. This produces the Bronze tier.
Domain experts review selected working papers for accuracy, methodology, and insight. Single-expert review yields Silver; multi-expert independent review yields Gold.
Each verified working paper receives a Scharp Scale score based on how well it survived scrutiny. The score reflects the artifact's quality, not the process used to check it.
Working papers are published with their tier, Scharp Score, reviewer information, and any flagged issues or corrections. Full provenance is always visible.
The Soulmetric verification pipeline uses multiple independent AI systems to avoid single-model bias:
The generating model never adjudicates its own output. Verification always uses a different system than generation.
Selection. Expert validators are selected based on demonstrated domain expertise—academic credentials, publication history, industry experience, or recognized practitioner status in the relevant field.
Compensation. Expert validators are compensated for their time. Compensation is flat-rate per review and does not vary based on the score they assign or whether they approve or reject the working paper.
Conflicts of interest. Validators disclose any conflicts of interest (financial, institutional, personal) related to the working paper's subject matter. Disclosures are recorded and available on request.
Independence. For Gold-tier verification, reviewers work independently and do not see each other's assessments until both are submitted. Disagreements are flagged and documented, not silently resolved.
Verification is not final. Any reader can challenge any claim in any working paper.
To submit a correction or challenge: kevin@soulmetric.com