Grille v1

Scoring modes

How Publi-Score calculates the score depending on available data. Empirical data on 18 articles.

Comparison table

⚑ QuickπŸ”¬ Partial AI (abstract)πŸ“ Full manualπŸ€– Full AI (PDF)
InputPMID/DOIPMID onlyPMID + PDFPMID + PDF
Criteria coverage~53/100~85/100100/100100/100
Integrity coverage~80%~80%100%100%
Duration~2–5 sec~30–60 sec~10–30 min~30–60 sec
Objectivityβœ… Autoβœ… LLM⚠️ Human biasβœ… LLM
Published in catalogueβœ…βœ…βŒβœ…
Account requiredNoYes (free)NoYes (free)
QuotaUnlimitedShared with Full AI (same quota)Unlimited5/month (free)

What the quick mode covers (and doesn't cover)

βœ“ What it covers

  • β€’ Β§2.3 Bibliometric impact β€” 100% (citations, h-index)
  • β€’ Β§2.6 Freshness β€” ~69% (publication date)
  • β€’ Β§2.1 Level of evidence β€” ~57% (study type, randomisation)
  • β€’ Retractions β€” 100% (PubMed API + Retraction Watch)
  • β€’ Alert signals β€” 100% (predatory journals, EoC)

Total: ~53% of criteria Β· ~80% integrity

⚠ What quick mode doesn't cover

  • β€’ Real ITT analysis (intention to treat)
  • β€’ Compliance with pre-registered protocol
  • β€’ Raw data sharing
  • β€’ Clinical benefit/risk ratio
  • β€’ Β§2.7 Reporting quality β€” 0% (requires PDF)

~47% of criteria not evaluable without PDF

The Partial AI mode without PDF

Partial AI (abstract) mode uses only the abstract and article metadata β€” without PDF. AI evaluates all criteria accessible from these sources, covering ~85% of the total score.

~85%

criteria covered

+3 pts

βˆ† average vs. full with PDF: 3 pts (measured on 18 corpus articles)

1/12

Only 1 tier change in 12 articles (TOGETHER: B→A)

0/3 pts

criterion not evaluable without PDF

Why ~85% and not 100%?

  • β€’ Β§2.7 Reporting quality (3 pts) β€” evaluates the clarity of results, tables and figures. Inaccessible without the full PDF: always 0 pt.
  • β€’ Β§2.4 Reproducibility & transparency β€” some sub-criteria (code sharing, raw data) are partially inferable from the abstract, but without certainty. AI scores them conservatively.

This mode produces a detailed score with per-criterion justifications β€” it is partial, not degraded. It remains significantly more reliable than quick mode (~53%) and activates automatically when the PDF is not open access.

Activation: automatic fallback if OA PDF unavailable. Measured on 18 articles, COVID-19 and Vaccination clusters β€” Publi-Score calibration corpus.

Empirical data

Measured on 18 articles, COVID-19 and Vaccination clusters β€” Publi-Score calibration corpus.

4 critical overestimation cases in quick mode

These 4 articles are among the most viewed in the corpus. Quick mode assigns them tier A or B, while Full AI mode reveals tier D.

Article⚑ QuickπŸ”¬ Partial AI (abstract)πŸ€– Full AI (PDF)Gap Qβ†’FMain reason
Polack/Pfizer β€” NEJM 2020AEDβˆ’51 ptsMajor industrial COI + short editorial delay not captured
Voysey/AZ β€” Lancet 2021BDDβˆ’38 ptsAstraZeneca COI + adaptive design + data not shared
Hammond/Paxlovid β€” NEJM 2022BEDβˆ’37 ptsIndustry-only trial + raw data unavailable
Molnupiravir β€” NEJM 2022BDDβˆ’33 ptsMerck/Ridgeback trial β€” non-public data

Why quick mode overestimates: it normalises on criteria accessible via APIs. Β§2.3 (bibliometric impact) weighs ~14 pts and is maximal for NEJM/Lancet articles β€” which are often industrial trials (Pfizer, AZ, Merck) whose strong COI is only visible on in-depth analysis.

Which mode to choose?

ContextRecommended mode
First exploration, monitoring⚑ Quick
Clinical decision, citation, teachingπŸ€– Full AI (PDF)
PDF unavailable (NEJM, Lancet…)πŸ”¬ Partial AI (abstract)
Personal learning, methodological explorationπŸ“ Full manual