Why systematic review methods sections break – and the fast way to fix them before peer review

Most systematic review methods sections fail not because the work wasn’t done, but because the logic isn’t reproducible or linear. This guide shows the most common “methods collapse” patterns reviewers flag, and how to fix them quickly


The uncomfortable truth

When systematic reviews get desk-rejected or heavily criticized, the problem is often not the results.

It’s the Methods section.

Not because Methods need to be long.
Because Methods need to be reproducible.

If an editor or reviewer can’t quickly answer:

  • What did you search?

  • What did you include/exclude and why?

  • Do the PRISMA numbers add up?

  • How were decisions made?

…then the review feels fragile, even if the underlying work is solid.

This is why many reviewers “miss” things that are technically in the document. The structure prevents them from seeing it.


What a Methods section must do (in one sentence)

A systematic review Methods section is a linear audit trail that allows a stranger to replicate your search and understand how your final sample was produced.

Not a narrative.
Not a dump of details.
A reproducible sequence.


The five failure modes that break Methods sections

1) Non-linear structure (jumping between steps)

A common problem is mixing components that belong in different phases:

  • research question appears inside Methods

  • quality assessment appears before data extraction

  • lineage of included studies is described instead of the search itself

  • “nice to have” details crowd out the essential trail

Fix: Make the Methods read like “baking a cake.”
Search → duplicates → screening → eligibility → inclusion → extraction → synthesis → (optional) quality assessment.

If the order is wrong, reviewers assume the underlying process was wrong.


2) PRISMA numbers don’t match (or can’t be reconstructed)

This is one of the fastest ways to lose credibility, because reviewers love catching inconsistencies.

Common causes:

  • duplicates removed but not reported cleanly

  • exclusions reported vaguely (“not relevant”)

  • counts scattered across text and figure

  • missing “reports not retrieved” numbers

Fix: Write the PRISMA narrative first, then build the diagram to match it.

A clean PRISMA narrative looks like this:

  • We identified X records…

  • After removing Y duplicates, Z remained for screening…

  • After screening, A full texts were assessed…

  • B were excluded for reasons…

  • Final included sample: N studies.

Then point the reader to Figure 1.


3) Inclusion/exclusion criteria are arbitrary (or appear arbitrary)

Reviewers don’t mind narrow criteria. They mind criteria that look like they were chosen to “make the project manageable.”

Red flag phrasing:

  • “We excluded genetic studies because there were too many.”

  • “We excluded a subgroup because it was overwhelming.”

Even if that’s emotionally true, it’s not defensible scientifically.

Fix: Convert “too many papers” into a scientific justification.

For example:

  • focus on routinely available clinical data (not specialised genetic markers)

  • focus on settings relevant to the intended application (e.g., real-world practice)

  • exclude study types that are not comparable to your research question

The principle: exclusions must be justified by scope, applicability, or reproducibility, not workload.


4) Quality assessment is used as a gate in a way reviewers won’t accept

Many students incorrectly treat risk-of-bias tools as “exclusion machines.”

In most reviews, quality assessment is used to:

  • describe the evidence quality

  • interpret findings

  • qualify confidence

—not to silently remove inconvenient studies.

Fix: In most cases:

  • include studies based on your inclusion criteria

  • report risk-of-bias at the end of Results (briefly)

  • use it to contextualize evidence strength

If you do exclude based on quality, state this explicitly as a criterion and justify it carefully.


5) The Methods include lots of “extra” information that distracts from what matters

Methods sections often bloat because authors include:

  • long explanations of frameworks everyone already knows

  • repeated statements that add no reproducibility

  • detailed tables described again in text

  • unnecessary lineage of “where included papers came from”

This creates a paradox:

  • the section becomes longer

  • but less clear

Fix: Keep it lean:

  • databases + date

  • keywords strategy (and where full strings are)

  • deduplication tool

  • inclusion/exclusion

  • screening process

  • PRISMA narrative

  • extraction approach

  • synthesis approach

  • optional quality assessment summary

Everything else is either:

  • Results

  • Appendix

  • or noise


A quick Methods “reproducibility test” (2 minutes)

If someone skim-reads your Methods, can they answer:

  1. Which databases did you search, and when?

  2. What were the main keywords / search strings (and where are full strings)?

  3. How many records did you find, and how many duplicates were removed?

  4. What were the inclusion and exclusion criteria?

  5. How did you screen and how does PRISMA match the text?

  6. What did you extract and how did you synthesize?

If any answer is unclear, the Methods need tightening