Quality Controls for Preprints

Putting trust in preprints

With 10,000+ preprints posted weekly, there's no quality filter. AI-generated papers, recycled content, and methodologically flawed studies sit alongside groundbreaking research.

We're building a first-pass filter using open source research integrity tools — helping researchers, journalists, and LLMs distinguish signal from noise.

This is an experiment. All assessments are machine-generated indicators requiring human expert review.

Try: 10.1101/2025.01.15.633214 or paste any bioRxiv URL

The A5–E1 Grading Matrix

Every paper receives a grade combining Strength of Evidence (A–E) and Significance of Findings (1–5)

EvidenceSignificance → 5Landmark 4Fundamental 3Important 2Valuable 1Useful
ACompelling A5A4A3A2A1
BConvincing B5B4B3B2B1
CSolid C5C4C3C2C1
DIncomplete D5D4D3D2D1
EInadequate E5E4E3E2E1
A: Compelling evidence
B: Convincing evidence
C: Solid — major revision needed
D: Incomplete — not submission-ready
E: Inadequate — do not submit

Significance (1–5) reflects the importance of the research question — not publication tier. A D5 paper asks a landmark-scope question but lacks the evidence to support its claims.

The Problem

Preprint quality is impossible to assess at scale

2–5%
Paper Mill Output

Estimated percentage of publications from paper mills

50%
Statistical Errors

Papers contain at least one statistical inconsistency

10K+
Weekly Preprints

Posted to bioRxiv and medRxiv alone

Integrated Tools

We combine the best open-source research integrity tools

Paper Mill Detection Layer 1

8,000+ tortured phrases, SCIgen fingerprints, and fabrication indicators

Statistical Verification statcheck

P-value recalculation, GRIM tests, and impossible-result detection

Open Data Checks Layer 1

Verify data availability statements and detect datasets mentioned but not shared

Retraction Watch External

Cross-reference citations against retracted paper database

9-Agent AI Panel Phase 2

Methodologist, Statistician, Ethics, Domain Expert, and 5 specialist reviewers — with adversarial verification

eLife-Style Deliberation Phase 4

3 specialist agents (Claude, GPT-4o, Gemini) independently review, consult, and reconcile on borderline papers

REST API for Developers

Embed quality signals into any workflow — a single GET request, zero infrastructure

GET /v2/score/{doi} v2 REST API
# One line. Any DOI, arXiv ID, or bioRxiv URL.
curl -H "X-API-Key: pak_your_key" \
     "https://preprints.ai/v2/score/10.1101/2025.01.15.633214"

# Response
{
  "doi": "10.1101/2025.01.15.633214",
  "grade": "B4",
  "integrity": { "grade": "B", "score": 0.82, "label": "Convincing" },
  "novelty":    { "grade": "4", "score": 0.76, "label": "Fundamental" },
  "confidence": 0.85,
  "badge_url": "https://preprints.ai/badge/10.1101/...",
  "report_url": "https://preprints.ai/report/10.1101/..."
}
Full API Docs → View Pricing

Infrastructure for Publishers

From preprint servers to journals — integrate quality signals at submission, review, or post-publication

🔍

Submission Screening

Flag potential integrity concerns before peer review begins. Statcheck, ODDPub, and paper mill detection run automatically on every submission.

🏷️

Embeddable Badges

One line of HTML. Grade badges update automatically as assessments are revised. Colour-coded for immediate comprehension.

Webhook Callbacks

Publisher tier: receive an HMAC-signed POST to your endpoint the moment an assessment completes. No polling required.

📦

Bulk Processing

Backfill your entire catalogue. Submit up to 500 DOIs per batch job with progress tracking and email notification on completion.

See the full API →

Built on Open Science Infrastructure

Every finding is traceable to a peer-reviewed tool. No proprietary black boxes.