How Many Databases To Search In A Systematic Review? | Practical Number Guide

Most methods advise searching two to four databases for a systematic review, chosen by topic, plus trial registers and gray sources.

You came here for a clear number, not a shrug. There is no single magic count, because coverage varies by field and question. Still, you can pin a usable range. Most teams land on a core of two to four bibliographic databases, then add trial registers and gray sources that match the topic. That mix balances recall, time, and duplicate handling.

What A Thorough Search Means For Database Coverage

Thorough search work is about reducing missed studies, not chasing every index on earth. Core health reviews often pull from MEDLINE and Embase, while education leans on ERIC and social science draws from APA PsycInfo or Sociological Abstracts. STEM topics may point to Web of Science or Scopus. Multidisciplinary projects blend a few of these to widen reach without burning days on marginal gains.

Database Best Fit Add When
MEDLINE/PubMed Biomedicine, clinical Any health review
Embase Drugs, devices, European journals Interventions, adverse events
Cochrane CENTRAL Randomized trials Intervention questions
Web of Science Core Multidisciplinary citation reach Cross-field topics
Scopus Broad science and social science When Web of Science is missing
APA PsycInfo Behavioral and mental health Psych outcomes or exposures
CINAHL Nursing and allied health Practice, workforce, settings
ERIC Education School, pedagogy, training
EconLit Economics Costs, incentives
AGRIS/AGRICOLA Food and farming Nutrition, food systems
LILACS Latin American literature Regional breadth
Global Health Public health, LMIC scope Global programs

How Many Databases To Search For A Systematic Review: Real Ranges

Start with two core databases that index your field well, then judge the gap. Add a third or fourth source if pilot searches show poor recall, if outcomes sit in a niche database, or if stakeholders demand high sensitivity. Past audits found strong recall from sets like MEDLINE plus Embase, bolstered by Web of Science or Scopus. Some projects add CENTRAL for trials or APA PsycInfo for mental health angles.

Transparent reporting beats a fixed count. Methods checklists ask you to name every source, show full strategies, and give dates. That makes the work reproducible and lets readers see trade-offs. If you used a multi-database platform, still list each database it queried.

What The Methods Literature Says About Counts

Major handbooks stress multiple sources and full reporting, without locking you to a number. The Cochrane Handbook lists bibliographic databases, trials registers, and other routes, with advice on sensitivity versus precision in search strings. PRISMA-S lays out how to report information sources and strategies step by step. Method papers that tracked yield show that two to four well chosen databases capture the bulk of records for many topics, with niche files adding new items in some fields.

For wording you can cite, see the Cochrane Handbook search guidance and the PRISMA-S checklist and explanation. Both outline sources to search, strategy design, and full reporting of every database and register.

In short: use more than one, tailor to the question, and document every choice. That mix gives coverage without unmanageable screening loads.

Build Your Set Step By Step

Define The Field And Study Types

List the main concepts, populations, and designs. Trials, cohorts, and qualitative work live in different corners of the literature. If randomized trials drive your question, CENTRAL and trials registers help. If policy or social outcomes lead, add databases that index those streams.

Pilot Two Databases, Then Check Gaps

Draft a broad search in the top two databases. Screen the first few hundred titles and note misses. If a stack of seed papers are not retrieved, find out where they sit and add that database. If yield is thin, expand subject headings or text words and add a third database that covers adjacent fields.

Add Trials Registers And Gray Routes

Search ClinicalTrials.gov or the WHO ICTRP for ongoing and completed trials. Pull regulatory documents when relevant. Scan conference abstracts, theses repositories, and society sites. These routes catch time-lag bias and non-indexed work.

Decide When To Stop

Stop when another database adds little beyond duplicates and noise. Record the cut-off date and save full strategies for every source. If the topic moves fast, set alerts and plan an update window.

Search Quality Moves The Needle More Than Raw Count

Poor strings across six databases waste hours. Sharp strings in three hit more. Use both controlled vocabulary and free text. Map synonyms, acronyms, and spellings. Combine with tested filters for study design only when needed and only after checking loss of recall in pilot sets.

Use Peer Review For Search Strings

Have an information specialist or trained peer check your strings. Peer review catches missed terms and logic errors. Small fixes often add missed studies without adding a new database.

Document Everything Cleanly

Save the exact strategies, with line numbers and counts, for each database and date. Report platform names, years indexed, and any limits. If you used a multi-search tool that queries several databases at once, still list each database by name.

De-Duplication And Screening Volume

Expect high overlap between broad databases. Use a reference manager or screening tool with strong de-duplication. Keep a log of numbers at each step so the PRISMA flow is easy to build. Set daily screening goals to keep progress steady.

Reporting And Transparency That Reviewers Expect

Make reporting as strong as the search. In your methods, list every database by name and platform, the years indexed, the last search date, and any language or document limits. Provide at least one full strategy in an appendix with field tags, operators, and line numbers. If you adapt strings across databases, show the changes so readers can trace logic.

Spell out gray routes as well: registers searched, regulators checked, and sites scanned, with dates. If you reran searches before submission, give the refresh date and counts. These details match what PRISMA-S and major handbooks ask for and they save rounds of peer review back-and-forth.

Gray Literature And Trial Registers Count As Sources

Your database count does not tell the whole story. Trial registers, regulatory files, conference sets, and theses fill in gaps and cut bias from non-publication. Plan time for these routes and keep a log with links and dates.

When Fewer Databases Are Reasonable

Scoping reviews, rapid reviews, and updates may cap the search set to a tight core. Pick the two best databases for the field, add a third if pilot screening shows holes, and offset with stronger gray routes. State the limits up front and explain the trade-off in precision and recall.

When You Need More Than Four

Some topics sprawl across fields. Device safety, rare diseases, or method questions might need five databases plus registers. You will see this in pilot checks: many papers sit outside your first picks. Add the niche files that hold them and plan a careful de-duplication pass.

Time, Tools, And Team

Plan the workflow before you start. Pick a reference manager and a de-duplication method. Decide who builds strings, who runs them, and who screens. Set batch sizes so screening stays steady. Good process trims wasted effort more than any single database choice.

Recommended Core Sets By Topic And Aim

The right mix depends on the question. Use the table below as a starting point, then adjust with pilot data. The core sets here mirror common picks seen in protocols and method texts.

Review Aim Minimum Core Set Add If
Clinical intervention effects MEDLINE + Embase + CENTRAL Trials register, Web of Science or Scopus for cross-checking
Public health programs MEDLINE + Global Health Regional file (LILACS), trials register
Mental health outcomes MEDLINE + APA PsycInfo Embase, Web of Science, trials register
Nursing and practice MEDLINE + CINAHL Scopus or Web of Science
Education interventions ERIC + Scopus MEDLINE for health-ed cross-over
Economics or cost EconLit + MEDLINE Web of Science for citations
Food and farming AGRICOLA/AGRIS + MEDLINE Global Health for programs
Cross-disciplinary topics MEDLINE + Embase + Web of Science Scopus or a niche file

Worked Example: From Two Databases To Four

Say your review tests a drug-device combo for outpatient care. You start with MEDLINE and Embase using subject headings and text words. After screening, you notice missing trial reports cited in a major guideline. You add CENTRAL for trial records and run a trials register search. A quick check in Web of Science pulls several conference papers that fill gaps. The final set uses four databases plus two registers, with full strategies in an appendix.

Search Cost, Time, And Yield

Each extra database adds run time, export time, and de-duplication time. The payoff is new records. Track yield per database during pilot runs. If a source adds near-zero new items after de-duplication, drop it and document why. Spend saved time tuning strings and searching gray routes where fresh yield often sits.

Quick Decision Guide

  • Two core databases span the field; run pilots and tally misses.
  • Add a third or fourth when pilots show gaps or niche outcomes.
  • Always search at least one major trials register and list dates.
  • Stop adding databases once new records are near-zero and mostly duplicates.
  • Log yields per source to guide updates and later planned replications.

Bottom Line For Your Protocol

State the planned sources and why you chose them. Most protocols name two to four databases keyed to the field, plus at least one trials register and a short list of gray routes. Add a note on peer review, dates, and updates. That plan sets up a transparent, reproducible search that meets handbook and journal expectations.