Recent from talks
Nothing was collected or created yet.
Systematic review
View on Wikipedia

| Part of a series on |
| Research |
|---|
| Philosophy portal |
A systematic review is a scholarly synthesis of the evidence on a clearly presented topic using critical methods to identify, define and assess research on the topic.[1] A systematic review extracts and interprets data from published studies on the topic (in the scientific literature), then analyzes, describes, critically appraises and summarizes interpretations into a refined evidence-based conclusion.[1][2] For example, a systematic review of randomized controlled trials is a way of summarizing and implementing evidence-based medicine.[3] Systematic reviews, sometimes along with meta-analyses, are generally considered the highest level of evidence in medical research.[4][5]
While a systematic review may be applied in the biomedical or health care context, it may also be used where an assessment of a precisely defined subject can advance understanding in a field of research.[6] A systematic review may examine clinical tests, public health interventions, environmental interventions,[7] social interventions, adverse effects, qualitative evidence syntheses, methodological reviews, policy reviews, and economic evaluations.[8][9]
Systematic reviews are closely related to meta-analyses, and often the same instance will combine both (being published with a subtitle of "a systematic review and meta-analysis"). The distinction between the two is that a meta-analysis uses statistical methods to induce a single number from the pooled data set (such as an effect size), whereas the strict definition of a systematic review excludes that step. However, in practice, when one is mentioned, the other may often be involved, as it takes a systematic review to assemble the information that a meta-analysis analyzes, and people sometimes refer to an instance as a systematic review, even if it includes the meta-analytical component.
An understanding of systematic reviews and how to implement them in practice is common for professionals in health care, public health, and public policy.[1]
Systematic reviews contrast with a type of review often called a narrative review. Systematic reviews and narrative reviews both review the literature (the scientific literature), but the term literature review without further specification refers to a narrative review.
Characteristics
[edit]A systematic review can be designed to provide a thorough summary of current literature relevant to a research question.[1] A systematic review uses a rigorous and transparent approach for research synthesis, with the aim of assessing and, where possible, minimizing bias in the findings. While many systematic reviews are based on an explicit quantitative meta-analysis of available data, there are also qualitative reviews and other types of mixed-methods reviews that adhere to standards for gathering, analyzing, and reporting evidence.[10]
Systematic reviews of quantitative data or mixed-method reviews sometimes use statistical techniques (meta-analysis) to combine results of eligible studies. Scoring levels are sometimes used to rate the quality of the evidence depending on the methodology used, although this is discouraged by the Cochrane Library.[11] As evidence rating can be subjective, multiple people may be consulted to resolve any scoring differences between how evidence is rated.[12][13][14]
The EPPI-Centre, Cochrane, and the Joanna Briggs Institute have been influential in developing methods for combining both qualitative and quantitative research in systematic reviews.[15][16][17] Several reporting guidelines exist to standardise reporting about how systematic reviews are conducted. Such reporting guidelines are not quality assessment or appraisal tools. The Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement[18] suggests a standardized way to ensure a transparent and complete reporting of systematic reviews, and is now required for this kind of research by more than 170 medical journals worldwide.[19] The latest version of this commonly used statement corresponds to PRISMA 2020 (the respective article was published in 2021).[20] Several specialized PRISMA guideline extensions have been developed to support particular types of studies or aspects of the review process, including PRISMA-P for review protocols and PRISMA-ScR for scoping reviews.[19] A list of PRISMA guideline extensions is hosted by the EQUATOR (Enhancing the QUAlity and Transparency Of health Research) Network.[21] However, the PRISMA guidelines have been found to be limited to intervention research and the guidelines have to be changed in order to fit non-intervention research. As a result, Non-Interventional, Reproducible, and Open (NIRO) Systematic Reviews was created to counter this limitation.[22]
For qualitative reviews, reporting guidelines include ENTREQ (Enhancing transparency in reporting the synthesis of qualitative research) for qualitative evidence syntheses; RAMESES (Realist And MEta-narrative Evidence Syntheses: Evolving Standards) for meta-narrative and realist reviews;[23][24] and eMERGe (Improving reporting of Meta-Ethnography) for meta-ethnograph.[15]
Developments in systematic reviews during the 21st century included realist reviews and the meta-narrative approach, both of which addressed problems of variation in methods and heterogeneity existing on some subjects.[25][26]
Types
[edit]There are over 30 types of systematic review and Table 1 below non-exhaustingly summarises some of these.[19][18] There is not always consensus on the boundaries and distinctions between the approaches described below.
| Review type | Summary |
|---|---|
| Mapping review/systematic map | A mapping review maps existing literature and categorizes data. The method characterizes the quantity and quality of literature, including by study design and other features. Mapping reviews can be used to identify the need for primary or secondary research.[19] |
| Meta-analysis | A meta-analysis is a statistical analysis that combines the results of multiple quantitative studies. Using statistical methods, results are combined to provide evidence from multiple studies. The two types of data generally used for meta-analysis in health research are individual participant data and aggregate data (such as odds ratios or relative risks). |
| Mixed studies review/mixed methods review | Refers to any combination of methods where one significant stage is a literature review (often systematic). It can also refer to a combination of review approaches, such as combining quantitative with qualitative research.[19] |
| Qualitative systematic review/qualitative evidence synthesis | This method integrates or compares findings from qualitative studies. The method can include 'coding' the data and looking for 'themes' or 'constructs' across studies. Multiple authors may improve the 'validity' of the data by potentially reducing individual bias.[19] |
| Rapid review | An assessment of what is already known about a policy or practice issue, which uses systematic review methods to search for and critically appraise existing research. Rapid reviews are still a systematic review, however parts of the process may be simplified or omitted in order to increase rapidity.[27] Rapid reviews were used during the COVID-19 pandemic.[28] |
| Systematic review | A systematic search for data using a repeatable method. It includes appraising the data (for example, the quality of the data) and a synthesis of research data. |
| Systematic search and review | Combines methods from a 'critical review' with a comprehensive search process. This review type is usually used to address broad questions to produce the most appropriate evidence synthesis. This method may or may not include quality assessment of data sources.[19] |
| Systematized review | Include elements of the systematic review process, but searching is often not as comprehensive as a systematic review and may not include quality assessments of data sources. |
Scoping reviews
[edit]Scoping reviews are distinct from systematic reviews in several ways. A scoping review is an attempt to search for concepts by mapping the language and data which surrounds those concepts and adjusting the search method iteratively to synthesize evidence and assess the scope of an area of inquiry.[25][26] This can mean that the concept search and method (including data extraction, organisation and analysis) are refined throughout the process, sometimes requiring deviations from any protocol or original research plan.[29][30] A scoping review may often be a preliminary stage before a systematic review, which 'scopes' out an area of inquiry and maps the language and key concepts to determine if a systematic review is possible or appropriate, or to lay the groundwork for a full systematic review. The goal can be to assess how much data or evidence is available regarding a certain area of interest.[29][31] This process is further complicated if it is mapping concepts across multiple languages or cultures.
As a scoping review should be systematically conducted and reported (with a transparent and repeatable method), some academic publishers categorize them as a kind of 'systematic review', which may cause confusion. Scoping reviews are helpful when it is not possible to carry out a systematic synthesis of research findings, for example, when there are no published clinical trials in the area of inquiry. Scoping reviews are helpful when determining if it is possible or appropriate to carry out a systematic review, and are a useful method when an area of inquiry is very broad,[32] for example, exploring how the public are involved in all stages systematic reviews.[33]
There is still a lack of clarity when defining the exact method of a scoping review as it is both an iterative process and is still relatively new.[34] There have been several attempts to improve the standardisation of the method,[30][29][31][35] for example via a Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guideline extension for scoping reviews (PRISMA-ScR).[36] PROSPERO (the International Prospective Register of Systematic Reviews) does not permit the submission of protocols of scoping reviews,[37] although some journals will publish protocols for scoping reviews.[33]
Stages
[edit]While there are multiple kinds of systematic review methods, the main stages of a review can be summarised as follows:
Defining the research question
[edit]Some reported that the 'best practices' involve 'defining an answerable question' and publishing the protocol of the review before initiating it to reduce the risk of unplanned research duplication and to enable transparency and consistency between methodology and protocol.[38][39] Clinical reviews of quantitative data are often structured using the mnemonic PICO, which stands for 'Population or Problem', 'Intervention or Exposure', 'Comparison', and 'Outcome', with other variations existing for other kinds of research. For qualitative reviews, PICo is 'Population or Problem', 'Interest', and 'Context'.
Searching for sources
[edit]Relevant criteria can include selecting research that is of good quality and answers the defined question.[38] The search strategy should be designed to retrieve literature that matches the protocol's specified inclusion and exclusion criteria. The methodology section of a systematic review should list all of the databases and citation indices that were searched. The titles and abstracts of identified articles can be checked against predetermined criteria for eligibility and relevance. Each included study may be assigned an objective assessment of methodological quality, preferably by using methods conforming to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement,[21] or the standards of Cochrane.[40]
Common information sources used in searches include scholarly databases of peer-reviewed articles such as MEDLINE, Web of Science, Embase, and PubMed, as well as sources of unpublished literature such as clinical trial registries and grey literature collections. Key references can also be yielded through additional methods such as citation searching, reference list checking (related to a search method called 'pearl growing'), manually searching information sources not indexed in the major electronic databases (sometimes called 'hand-searching'),[41] and directly contacting experts in the field.[42]
To be systematic, searchers must use a combination of search skills and tools such as database subject headings, keyword searching, Boolean operators, and proximity searching while attempting to balance sensitivity (systematicity) and precision (accuracy). Inviting and involving an experienced information professional or librarian can improve the quality of systematic review search strategies and reporting.[43][44][45][46][47]
'Extraction' of relevant data
[edit]
Relevant data are 'extracted' from the data sources according to the review method. The data extraction method is specific to the kind of data, and data extracted on 'outcomes' is only relevant to certain types of reviews. For example, a systematic review of clinical trials might extract data about how the research was done (often called the method or 'intervention'), who participated in the research (including how many people), how it was paid for (for example, funding sources) and what happened (the outcomes).[38] Relevant data are being extracted and 'combined' in an intervention effect review, where a meta-analysis is possible.[48]
Assess the eligibility of the data
[edit]This stage involves assessing the eligibility of data for inclusion in the review by judging it against criteria identified at the first stage.[38] This can include assessing if a data source meets the eligibility criteria and recording why decisions about inclusion or exclusion in the review were made. Software programmes can be used to support the selection process, including text mining tools and machine learning, which can automate aspects of the process.[49] The 'Systematic Review Toolbox' is a community-driven, web-based catalog of tools, to help reviewers chose appropriate tools for reviews.[50]
Analyse and combine the data
[edit]Analysing and combining data can provide an overall result from all the data. Because this combined result may use qualitative or quantitative data from all eligible sources of data, it is considered more reliable as it provides better evidence, as the more data included in reviews, the more confident we can be of conclusions. When appropriate, some systematic reviews include a meta-analysis, which uses statistical methods to combine data from multiple sources. A review might use quantitative data, or might employ a qualitative meta-synthesis, which synthesises data from qualitative studies. A review may also bring together the findings from quantitative and qualitative studies in a mixed methods or overarching synthesis.[51] The combination of data from a meta-analysis can sometimes be visualised. One method uses a forest plot (also called a blobbogram).[38] In an intervention effect review, the diamond in the 'forest plot' represents the combined results of all the data included.[38] An example of a 'forest plot' is the Cochrane Collaboration logo.[38] The logo is a forest plot of one of the first reviews which showed that corticosteroids given to women who are about to give birth prematurely can save the life of the newborn child.[52]
Recent visualisation innovations include the albatross plot, which plots p-values against sample sizes, with approximate effect-size contours superimposed to facilitate analysis.[53] The contours can be used to infer effect sizes from studies that have been analysed and reported in diverse ways. Such visualisations may have advantages over other types when reviewing complex interventions.
Communication and dissemination
[edit]Once these stages are complete, the review may be published, disseminated, and translated into practice after being adopted as evidence. The UK National Institute for Health Research (NIHR) defines dissemination as "getting the findings of research to the people who can make use of them to maximise the benefit of the research without delay".[54]
Some users do not have time to invest in reading large and complex documents and/or may lack awareness or be unable to access newly published research. Researchers are, therefore, developing skills to use creative communication methods such as illustrations, blogs, infographics, and board games to share the findings of systematic reviews.[55]
Automation
[edit]Living systematic reviews are a newer kind of semi-automated, up-to-date online summaries of research that are updated as new research becomes available.[56] The difference between a living systematic review and a conventional systematic review is the publication format. Living systematic reviews are "dynamic, persistent, online-only evidence summaries, which are updated rapidly and frequently".[57]
The automation or semi-automation of the systematic process itself is increasingly being explored. While little evidence exists to demonstrate it is as accurate or involves less manual effort, efforts that promote training and using artificial intelligence for the process are increasing.[58][56] In particular, since 2023, there has been a growing emergence of tools powered by large language models designed to support, automate, or even generate literature reviews. [59]
Research fields
[edit]Health and medicine
[edit]Current use of systematic reviews in medicine
[edit]Many organisations around the world use systematic reviews, with the methodology depending on the guidelines being followed. Organisations which use systematic reviews in medicine and human health include the National Institute for Health and Care Excellence (NICE, UK), the Agency for Healthcare Research and Quality (AHRQ, US), and the World Health Organization. Most notable among international organisations is Cochrane, a group of over 37,000 specialists in healthcare who systematically review randomised trials of the effects of prevention, treatments, and rehabilitation as well as health systems interventions. They sometimes also include the results of other types of research. Cochrane Reviews are published in The Cochrane Database of Systematic Reviews section of the Cochrane Library. The 2015 impact factor for The Cochrane Database of Systematic Reviews was 6.103, and it was ranked 12th in the Medicine, General & Internal category.[60]
There are several types of systematic reviews, including:[61][62][63][64]
- Intervention reviews assess the benefits and harms of interventions used in healthcare and health policy.
- Diagnostic test accuracy reviews assess how well a diagnostic test performs in diagnosing and detecting a particular disease. For conducting diagnostic test accuracy reviews, free software such as MetaDTA and CAST-HSROC in the graphical user interface is available.[65][66]
- Methodology reviews address issues relevant to how systematic reviews and clinical trials are conducted and reported.
- Qualitative reviews synthesize qualitative evidence to address questions on aspects other than effectiveness.
- Prognosis reviews address the probable course or future outcome(s) of people with a health problem.
- Overviews of Systematic Reviews (OoRs) compile multiple pieces of evidence from systematic reviews into a single accessible document, sometimes referred to as umbrella reviews.
- Living systematic reviews are continually updated, incorporating relevant new evidence as it becomes available.[67]
- Rapid reviews are a form of knowledge synthesis that "accelerates the process of conducting a traditional systematic review through streamlining or omitting specific methods to produce evidence for stakeholders in a resource-efficient manner".[68]
- Reviews of complex health interventions in complex systems are to improve evidence synthesis and guideline development.[69]
Patient and public involvement in systematic reviews
[edit]There are various ways patients and the public can be involved in producing systematic reviews and other outputs. Tasks for public members can be organised as 'entry level' or higher. Tasks include:
- Joining a collaborative volunteer effort to help categorise and summarise healthcare evidence[70]
- Data extraction and risk of bias assessment
- Translation of reviews into other languages
A systematic review of how people were involved in systematic reviews aimed to document the evidence-base relating to stakeholder involvement in systematic reviews and to use this evidence to describe how stakeholders have been involved in systematic reviews.[71] Thirty percent involved patients and/or carers. The ACTIVE framework provides a way to describe how people are involved in systematic review and may be used as a way to support systematic review authors in planning people's involvement.[72] Standardised Data on Initiatives (STARDIT) is another proposed way of reporting who has been involved in which tasks during research, including systematic reviews.[73][74]
There has been some criticism of how Cochrane prioritises systematic reviews.[75] Cochrane has a project that involved people in helping identify research priorities to inform Cochrane Reviews.[76][77] In 2014, the Cochrane–Wikipedia partnership was formalised.[78]
Environmental health and toxicology
[edit]Systematic reviews are a relatively recent innovation in the field of environmental health and toxicology. Although mooted in the mid-2000s, the first full frameworks for conduct of systematic reviews of environmental health evidence were published in 2014 by the US National Toxicology Program's Office of Health Assessment and Translation[79] and the Navigation Guide at the University of California San Francisco's Program on Reproductive Health and the Environment.[80] Uptake has since been rapid, with the estimated number of systematic reviews in the field doubling since 2016 and the first consensus recommendations on best practice, as a precursor to a more general standard, being published in 2020.[81]
Social, behavioural, and educational
[edit]In 1959, social scientist and social work educator Barbara Wootton published one of the first contemporary systematic reviews of literature on anti-social behavior as part of her work, Social Science and Social Pathology.[82][83]
Several organisations use systematic reviews in social, behavioural, and educational areas of evidence-based policy, including the National Institute for Health and Care Excellence (NICE, UK), Social Care Institute for Excellence (SCIE, UK), the Agency for Healthcare Research and Quality (AHRQ, US), the World Health Organization, the International Initiative for Impact Evaluation (3ie), the Joanna Briggs Institute, and the Campbell Collaboration. The quasi-standard for systematic review in the social sciences is based on the procedures proposed by the Campbell Collaboration, which is one of several groups promoting evidence-based policy in the social sciences.[84]
Others
[edit]Some attempts to transfer the procedures from medicine to business research have been made,[85] including a step-by-step approach,[86][87] and developing a standard procedure for conducting systematic literature reviews in business and economics.
Systematic reviews are increasingly prevalent in other fields, such as international development research.[88] Subsequently, several donors (including the UK Department for International Development (DFID) and AusAid) are focusing more on testing the appropriateness of systematic reviews in assessing the impacts of development and humanitarian interventions.[88]
The Collaboration for Environmental Evidence (CEE) has a journal titled Environmental Evidence, which publishes systematic reviews, review protocols, and systematic maps on the impacts of human activity and the effectiveness of management interventions.[89]
Review tools
[edit]A 2022 publication identified 24 systematic review tools and ranked them by inclusion of 30 features deemed most important when performing a systematic review in accordance with best practices. The top six software tools (with at least 21/30 key features) are all proprietary paid platforms, typically web-based, and include:[90]
- Giotto Compliance
- DistillerSR
- Nested Knowledge
- EPPI-Reviewer Web
- LitStream
- JBI SUMARI
The Cochrane Collaboration provides a handbook for systematic reviewers of interventions which "provides guidance to authors for the preparation of Cochrane Intervention reviews."[40] The Cochrane Handbook also outlines steps for preparing a systematic review[40] and forms the basis of two sets of standards for the conduct and reporting of Cochrane Intervention Reviews (MECIR; Methodological Expectations of Cochrane Intervention Reviews).[91] It also contains guidance on integrating patient-reported outcomes into reviews.
Limitations
[edit]Out-dated or risk of bias
[edit]While systematic reviews are regarded as the strongest form of evidence, a 2003 review of 300 studies found that not all systematic reviews were equally reliable, and that their reporting can be improved by a universally agreed upon set of standards and guidelines.[92] A further study by the same group found that of 100 systematic reviews monitored, 7% needed updating at the time of publication, another 4% within a year, and another 11% within 2 years; this figure was higher in rapidly changing fields of medicine, especially cardiovascular medicine.[93] A 2003 study suggested that extending searches beyond major databases, perhaps into grey literature, would increase the effectiveness of reviews.[94]
Some authors have highlighted problems with systematic reviews, particularly those conducted by Cochrane, noting that published reviews are often biased, out of date, and excessively long.[95] Cochrane reviews have been criticized as not being sufficiently critical in the selection of trials and including too many of low quality. They proposed several solutions, including limiting studies in meta-analyses and reviews to registered clinical trials, requiring that original data be made available for statistical checking, paying greater attention to sample size estimates, and eliminating dependence on only published data. Some of these difficulties were noted as early as 1994:
much poor research arises because researchers feel compelled for career reasons to carry out research that they are ill-equipped to perform, and nobody stops them.
— Altman DG, 1994[96]
Methodological limitations of meta-analysis have also been noted.[97] Another concern is that the methods used to conduct a systematic review are sometimes changed once researchers see the available trials they are going to include.[98] Some websites have described retractions of systematic reviews and published reports of studies included in published systematic reviews.[99][100][101] Eligibility criteria that is arbitrary may affect the perceived quality of the review.[102][103]
Limited reporting of data from human studies
[edit]The AllTrials campaign report that around half of clinical trials have never reported results and works to improve reporting.[104] 'Positive' trials were twice as likely to be published as those with 'negative' results.[105]
As of 2016, it is legal for-profit companies to conduct clinical trials and not publish the results.[106] For example, in the past 10 years, 8.7 million patients have taken part in trials that have not published results.[106] These factors mean that it is likely there is a significant publication bias, with only 'positive' or perceived favourable results being published. A recent systematic review of industry sponsorship and research outcomes concluded that "sponsorship of drug and device studies by the manufacturing company leads to more favorable efficacy results and conclusions than sponsorship by other sources" and that the existence of an industry bias that cannot be explained by standard 'risk of bias' assessments.[107]
Poor compliance with review reporting guidelines
[edit]The rapid growth of systematic reviews in recent years has been accompanied by the attendant issue of poor compliance with guidelines, particularly in areas such as declaration of registered study protocols, funding source declaration, risk of bias data, issues resulting from data abstraction, and description of clear study objectives.[108][109][110][111][112] A host of studies have identified weaknesses in the rigour and reproducibility of search strategies in systematic reviews.[113][114][115][116][117][118] To remedy this issue, a new PRISMA guideline extension called PRISMA-S is being developed.[119] Furthermore, tools and checklists for peer-reviewing search strategies have been created, such as the Peer Review of Electronic Search Strategies (PRESS) guidelines.[120]
A key challenge for using systematic reviews in clinical practice and healthcare policy is assessing the quality of a given review. Consequently, a range of appraisal tools to evaluate systematic reviews have been designed. The two most popular measurement instruments and scoring tools for systematic review quality assessment are AMSTAR 2 (a measurement tool to assess the methodological quality of systematic reviews)[121][122][123] and ROBIS (Risk Of Bias In Systematic reviews); however, these are not appropriate for all systematic review types.[124] Some recent peer-reviewed articles have carried out comparisons between AMSTAR 2 and ROBIS tools.[125][126]
History
[edit]The first publication that is now recognized as equivalent to a modern systematic review was a 1753 paper by James Lind, which reviewed all of the previous publications about scurvy.[127] Systematic reviews appeared only sporadically until the 1980s, and became common after 2000.[127] More than 10,000 systematic reviews are published each year.[127]
History in medicine
[edit]A 1904 British Medical Journal paper by Karl Pearson collated data from several studies in the UK, India and South Africa of typhoid inoculation. He used a meta-analytic approach to aggregate the outcomes of multiple clinical studies.[128] In 1972, Archie Cochrane wrote: "It is surely a great criticism of our profession that we have not organised a critical summary, by specialty or subspecialty, adapted periodically, of all relevant randomised controlled trials".[129] Critical appraisal and synthesis of research findings in a systematic way emerged in 1975 under the term 'meta analysis'.[130][131] Early syntheses were conducted in broad areas of public policy and social interventions, with systematic research synthesis applied to medicine and health.[132] Inspired by his own personal experiences as a senior medical officer in prisoner of war camps, Archie Cochrane worked to improve the scientific method in medical evidence.[133] His call for the increased use of randomised controlled trials and systematic reviews led to the creation of The Cochrane Collaboration,[134] which was founded in 1993 and named after him, building on the work by Iain Chalmers and colleagues in the area of pregnancy and childbirth.[135][129]
See also
[edit]References
[edit]
This article was submitted to WikiJournal of Medicine for external academic peer review in 2019 (reviewer reports). The updated content was reintegrated into the Wikipedia page under a CC-BY-SA-3.0 license (2020). The version of record as reviewed is:
Jack Nunn; Steven Chang; et al. (9 November 2020). "What are Systematic Reviews?" (PDF). WikiJournal of Medicine. 7 (1): 5. doi:10.15347/WJM/2020.005. ISSN 2002-4436. Wikidata Q99440266.
STARDIT report Q101116128.
- ^ a b c d "What is a systematic review?". Temple University Libraries. 6 June 2022. Retrieved 15 June 2022.
- ^ Armstrong R, Hall BJ, Doyle J, Waters E (March 2011). "Cochrane Update. 'Scoping the scope' of a cochrane review". Journal of Public Health. 33 (1): 147–150. doi:10.1093/pubmed/fdr015. PMID 21345890.
- ^ "What is EBM?". Centre for Evidence Based Medicine. 20 November 2009. Archived from the original on 6 April 2011. Retrieved 17 June 2011.
- ^ Wallace, Sowdhamini S.; Barak, Gal; Truong, Grace; Parker, Michelle W. (1 August 2022). "Hierarchy of Evidence Within the Medical Literature". Hospital Pediatrics. 12 (8): 745. doi:10.1542/hpeds.2022-006690. PMID 35909178.
The quality of evidence from medical research is partially deemed by the hierarchy of study designs. On the lowest level, the hierarchy of study designs begins with animal and translational studies and expert opinion, and then ascends to descriptive case reports or case series, followed by analytic observational designs such as cohort studies, then randomized controlled trials, and finally systematic reviews and meta-analyses as the highest quality evidence.
- ^ Murad, M Hassan; Asi, Noor; Alsawas, Mouaz; Alahdab, Fares (August 2016). "New evidence pyramid". Evidence Based Medicine. 21 (4): 125–127. doi:10.1136/ebmed-2016-110401. PMC 4975798. PMID 27339128.
A pyramid has expressed the idea of hierarchy of medical evidence for so long, that not all evidence is the same. Systematic reviews and meta-analyses have been placed at the top of this pyramid for several good reasons.
- ^ Ader HJ, Mellenbergh GJ, Hand DJ (2008). "Methodological quality". Advising on Research Methods: A consultant's companion. Johannes van Kessel Publishing. ISBN 978-90-79418-02-2.
- ^ Bilotta GS, Milner AM, Boyd I (2014). "On the use of systematic reviews to inform environmental policies". Environmental Science & Policy. 42: 67–77. Bibcode:2014ESPol..42...67B. doi:10.1016/j.envsci.2014.05.010.
- ^ Systematic reviews: CRD's guidance for undertaking reviews in health care (PDF). York: University of York, Centre for Reviews and Dissemination. 2008. ISBN 978-1-900640-47-3. Retrieved 17 June 2011.
- ^ Petticrew M, Roberts H (2006). Systematic reviews in the social sciences (PDF). Wiley Blackwell. ISBN 978-1-4051-2110-1. Archived from the original (PDF) on 16 June 2015.
- ^ Bearman M, Dawson P (March 2013). "Qualitative synthesis and systematic review in health professions education". Medical Education. 47 (3): 252–260. doi:10.1111/medu.12092. PMID 23398011. S2CID 11042748.
- ^ Higgins JP, Thomas J, Chandler J, Cumpston M, Li T, Page MJ, Welch VA, eds. (20 September 2019). Cochrane Handbook for Systematic Reviews of Interventions. version 6.1. pp. section 4.6. Retrieved 14 September 2020.
- ^ Siemieniuk R, Guyatt G. "What is GRADE?". BMJ Best Practice. Retrieved 26 August 2020.
- ^ Adèr HJ (2008). Advising on Research Methods: A Consultant's Companion. Johannes van Kessel Publishing. ISBN 978-90-79418-01-5.
- ^ Schünemann H, Brożek J, Guyatt G, Oxman A, eds. (2013). GRADE Handbook. Retrieved 26 August 2020.
- ^ a b Flemming K, Booth A, Hannes K, Cargo M, Noyes J (May 2018). "Cochrane Qualitative and Implementation Methods Group guidance series-paper 6: reporting guidelines for qualitative, implementation, and process evaluation evidence syntheses" (PDF). Journal of Clinical Epidemiology. 97: 79–85. doi:10.1016/j.jclinepi.2017.10.022. PMID 29222060.
- ^ Harden A, Thomas J, Cargo M, Harris J, Pantoja T, Flemming K, et al. (May 2018). "Cochrane Qualitative and Implementation Methods Group guidance series-paper 5: methods for integrating qualitative and implementation evidence within intervention effectiveness reviews" (PDF). Journal of Clinical Epidemiology. 97: 70–78. doi:10.1016/j.jclinepi.2017.11.029. PMID 29242095.
- ^ "EPPI-Centre Home". eppi.ioe.ac.uk. Retrieved 29 June 2020.
- ^ a b Booth A, Noyes J, Flemming K, Gerhardus A, Wahlster P, Van Der Wilt GJ, et al. (2016). Guidance on choosing qualitative evidence synthesis methods for use in health technology assessments of complex interventions. p. 32. OCLC 944453327.
- ^ a b c d e f g Grant MJ, Booth A (June 2009). "A typology of reviews: an analysis of 14 review types and associated methodologies". Health Information and Libraries Journal. 26 (2): 91–108. doi:10.1111/j.1471-1842.2009.00848.x. PMID 19490148.
- ^ Page, Matthew J.; McKenzie, Joanne E.; Bossuyt, Patrick M.; Boutron, Isabelle; Hoffmann, Tammy C.; Mulrow, Cynthia D.; Shamseer, Larissa; Tetzlaff, Jennifer M.; Akl, Elie A.; Brennan, Sue E.; Chou, Roger; Glanville, Julie; Grimshaw, Jeremy M.; Hróbjartsson, Asbjørn; Lalu, Manoj M. (29 March 2021). "The PRISMA 2020 statement: an updated guideline for reporting systematic reviews". BMJ (Clinical Research Ed.). 372: n71. doi:10.1136/bmj.n71. ISSN 1756-1833. PMC 8005924. PMID 33782057.
- ^ a b "Preferred Reporting Items for Systematic Reviews and Meta-Analyses: The PRISMA Statement". www.equator-network.org. Retrieved 29 June 2020.
- ^ Topor, Marta; Pickering, Jade S.; Mendes, Ana Barbosa; Bishop, Dorothy V. M.; Büttner, Fionn; Elsherif, Mahmoud M.; Evans, Thomas R.; Henderson, Emma L.; Kalandadze, Tamara; Nitschke, Faye T.; Staaks, Janneke P. C.; Akker, Olmo R. van den; Yeung, Siu Kit; Zaneva, Mirela; Lam, Alison (10 July 2023). "An integrative framework for planning and conducting Non-Intervention, Reproducible, and Open Systematic Reviews (NIRO-SR)". Meta-Psychology. 7. doi:10.15626/MP.2021.2840. hdl:11250/3106491. ISSN 2003-2714.
- ^ Wong G, Greenhalgh T, Westhorp G, Buckingham J, Pawson R (May 2013). "RAMESES publication standards: meta-narrative reviews". Journal of Advanced Nursing. 69 (5): 987–1004. doi:10.1111/jan.12092. PMC 3558334. PMID 23356699.
- ^ Wong G, Greenhalgh T, Westhorp G, Buckingham J, Pawson R (May 2013). "RAMESES publication standards: realist syntheses". Journal of Advanced Nursing. 69 (5): 1005–1022. doi:10.1111/jan.12095. PMC 3558331. PMID 23356726.
- ^ a b Arksey H, O'Malley L (2005). "Scoping studies: Towards a methodological framework" (PDF). International Journal of Social Research Methodology. 8: 19–32. doi:10.1080/1364557032000119616. hdl:11250/2477738. S2CID 12719181.
- ^ a b "PRISMA". www.prisma-statement.org. Archived from the original on 25 March 2024. Retrieved 1 July 2020.
- ^ "What is a rapid review? Systematic Review Library Guides at CQ University". library.cqu.edu.au. Archived from the original on 16 September 2020. Retrieved 16 September 2020.
- ^ "Home". covidrapidreviews.cochrane.org. Archived from the original on 1 July 2020. Retrieved 1 July 2020.
- ^ a b c Peters MD, Godfrey CM, Khalil H, McInerney P, Parker D, Soares CB (September 2015). "Guidance for conducting systematic scoping reviews". International Journal of Evidence-Based Healthcare. 13 (3): 141–146. doi:10.1097/XEB.0000000000000050. PMID 26134548. S2CID 8860037.
- ^ a b Levac D, Colquhoun H, O'Brien KK (September 2010). "Scoping studies: advancing the methodology". Implementation Science. 5 (1) 69. doi:10.1186/1748-5908-5-69. PMC 2954944. PMID 20854677.
- ^ a b Colquhoun HL, Levac D, O'Brien KK, Straus S, Tricco AC, Perrier L, et al. (December 2014). "Scoping reviews: time for clarity in definition, methods, and reporting". Journal of Clinical Epidemiology. 67 (12): 1291–1294. doi:10.1016/j.jclinepi.2014.03.013. hdl:1807/73365. PMID 25034198. S2CID 205843946.
- ^ Nunn JS, Tiller J, Fransquet P, Lacaze P (2019). "Public Involvement in Global Genomics Research: A Scoping Review". Frontiers in Public Health. 7: 79. doi:10.3389/fpubh.2019.00079. PMC 6467093. PMID 31024880.
- ^ a b Pollock A, Campbell P, Struthers C, Synnot A, Nunn J, Hill S, et al. (21 April 2017). "Stakeholder involvement in systematic reviews: a protocol for a systematic review of methods, outcomes and effects". Research Involvement and Engagement. 3 (1) 9. doi:10.1186/s40900-017-0060-4. PMC 5611627. PMID 29062534.
- ^ Munn Z, Peters MD, Stern C, Tufanaru C, McArthur A, Aromataris E (November 2018). "Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach". BMC Medical Research Methodology. 18 (1) 143. doi:10.1186/s12874-018-0611-x. PMC 6245623. PMID 30453902.
- ^ Arksey H, O'Malley L (1 February 2005). "Scoping studies: towards a methodological framework" (PDF). International Journal of Social Research Methodology. 8 (1): 19–32. doi:10.1080/1364557032000119616. hdl:11250/2477738. S2CID 12719181.
- ^ Tricco AC, Lillie E, Zarin W, O'Brien KK, Colquhoun H, Levac D, et al. (October 2018). "PRISMA Extension for Scoping Reviews (PRISMA-ScR): Checklist and Explanation" (PDF). Annals of Internal Medicine. 169 (7): 467–473. doi:10.7326/M18-0850. PMID 30178033. S2CID 52150954.
- ^ "PROSPERO". Centre for Reviews and Dissemination. University of York. Retrieved 24 February 2019.
- ^ a b c d e f g "Animated Storyboard: What Are Systematic Reviews?". cccrg.cochrane.org. Cochrane Consumers and Communication. Retrieved 1 June 2016.
- ^ Moher D, Shamseer L, Clarke M, Ghersi D, Liberati A, Petticrew M, et al. (January 2015). "Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement". Systematic Reviews. 4 (1) 1. doi:10.1186/2046-4053-4-1. PMC 4320440. PMID 25554246.
- ^ a b c Higgins JP, Green S (eds.). "Cochrane handbook for systematic reviews of interventions, version 5.1.0 (updated March 2011)". The Cochrane Collaboration. Retrieved 2 June 2016.
- ^ "5. Handsearching". training.cochrane.org. Retrieved 14 September 2020.
- ^ Papaioannou D, Sutton A, Carroll C, Booth A, Wong R (June 2010). "Literature searching for social science systematic reviews: consideration of a range of search techniques". Health Information and Libraries Journal. 27 (2): 114–122. doi:10.1111/j.1471-1842.2009.00863.x. PMID 20565552.
- ^ Rethlefsen ML, Farrell AM, Osterhaus Trzasko LC, Brigham TJ (June 2015). "Librarian co-authors correlated with higher quality reported search strategies in general internal medicine systematic reviews". Journal of Clinical Epidemiology. 68 (6): 617–626. doi:10.1016/j.jclinepi.2014.11.025. PMID 25766056.
- ^ Koffel JB (4 May 2015). "Use of recommended search strategies in systematic reviews and the impact of librarian involvement: a cross-sectional survey of recent authors". PLOS ONE. 10 (5) e0125931. Bibcode:2015PLoSO..1025931K. doi:10.1371/journal.pone.0125931. PMC 4418838. PMID 25938454.
- ^ Meert D, Torabi N, Costella J (October 2016). "Impact of librarians on reporting of the literature searching component of pediatric systematic reviews". Journal of the Medical Library Association. 104 (4): 267–277. doi:10.5195/jmla.2016.139. PMC 5079487. PMID 27822147.
- ^ Li L, Tian J, Tian H, Moher D, Liang F, Jiang T, et al. (September 2014). "Network meta-analyses could be improved by searching more sources and by involving a librarian". Journal of Clinical Epidemiology. 67 (9): 1001–1007. doi:10.1016/j.jclinepi.2014.04.003. PMID 24841794.
- ^ Rethlefsen ML, Murad MH, Livingston EH (September 2014). "Engaging medical librarians to improve the quality of review articles". JAMA. 312 (10): 999–1000. doi:10.1001/jama.2014.9263. PMID 25203078.
- ^ a b "Animated Storyboard: What Are Systematic Reviews?". cccrg.cochrane.org. Retrieved 24 November 2020.
- ^ Higgins JP, Thomas J, Chandler J, Cumpston M, Li T, Page MJ, Welch VA, eds. (20 September 2019). "Chapter 4: Searching for and selecting studies". Cochrane Handbook for Systematic Reviews of Interventions. version 6.1. section 4.6. Retrieved 14 September 2020.
{{cite book}}:|website=ignored (help) - ^ Marshall C, Brereton P (27 April 2015). "Systematic review toolbox". Proceedings of the 19th International Conference on Evaluation and Assessment in Software Engineering. EASE '15. Nanjing, China: Association for Computing Machinery. pp. 1–6. doi:10.1145/2745802.2745824. ISBN 978-1-4503-3350-4. S2CID 6679820.
- ^ Thompson Coon J, Gwernan-Jones R, Garside R, Nunns M, Shaw L, Melendez-Torres GJ, Moore D (July 2020). "Developing methods for the overarching synthesis of quantitative and qualitative evidence: The interweave synthesis approach". Research Synthesis Methods. 11 (4): 507–521. doi:10.1002/jrsm.1383. PMC 7383598. PMID 31725951.
- ^ "The difference we make". www.cochrane.org. Retrieved 8 March 2019.
- ^ Higgins JP, López-López JA, Becker BJ, Davies SR, Dawson S, Grimshaw JM, et al. (1 January 2019). "Synthesising quantitative evidence in systematic reviews of complex health interventions". BMJ Global Health. 4 (Suppl 1) e000858. doi:10.1136/bmjgh-2018-000858. PMC 6350707. PMID 30775014.
- ^ "How to disseminate your research". www.nihr.ac.uk. Archived from the original on 24 May 2022. Retrieved 20 May 2022.
- ^ Coon JT, Orr N, Shaw L, Hunt H, Garside R, Nunns M, et al. (April 2022). "Bursting out of our bubble: using creative techniques to communicate within the systematic review process and beyond". Systematic Reviews. 11 (1) 56. doi:10.1186/s13643-022-01935-2. PMC 8977563. PMID 35379331.
- ^ a b Tsafnat G, Glasziou P, Choong MK, Dunn A, Galgani F, Coiera E (July 2014). "Systematic review automation technologies". Systematic Reviews. 3 (1) 74. doi:10.1186/2046-4053-3-74. PMC 4100748. PMID 25005128.
- ^ Elliott JH, Turner T, Clavisi O, Thomas J, Higgins JP, Mavergames C, Gruen RL (February 2014). "Living systematic reviews: an emerging opportunity to narrow the evidence-practice gap". PLOS Medicine. 11 (2) e1001603. doi:10.1371/journal.pmed.1001603. PMC 3928029. PMID 24558353.
- ^ Reddy SM, Patel S, Weyrich M, Fenton J, Viswanathan M (October 2020). "Comparison of a traditional systematic review approach with review-of-reviews and semi-automation as strategies to update the evidence". Systematic Reviews. 9 (1) 243. doi:10.1186/s13643-020-01450-2. PMC 7574591. PMID 33076975.
- ^ Bolanos, Francisco; Salatino, Angelo; Osborne, Francesco; Motta, Enrico (2024). "Artificial intelligence for literature reviews: Opportunities and challenges". Artificial Intelligence Review. 57 (10). Springer: 259. doi:10.1007/s10462-024-10902-3.
- ^ The Cochrane Library. 2015 impact factor. Cochrane Database of Systematic Reviews (CDSR) Archived 2 July 2016 at the Wayback Machine Retrieved 2016-07-20.
- ^ Review Manager (RevMan) [Computer program]. Version 5.2. Copenhagen: The Nordic Cochrane Centre, The Cochrane Collaboration, 2012.
- ^ "Main page". Cochrane Library.
- ^ Silva V, Grande AJ, Carvalho AP, Martimbianco AL, Riera R (2015). "Overview of systematic reviews - a new type of study. Part II". Sao Paulo Medical Journal = Revista Paulista de Medicina. 133 (3): 206–217. doi:10.1590/1516-3180.2013.8150015. PMC 10876375. PMID 25388685.
- ^ Silva V, Grande AJ, Martimbianco AL, Riera R, Carvalho AP (2012). "Overview of systematic reviews - a new type of study: part I: why and for whom?". Sao Paulo Medical Journal = Revista Paulista de Medicina. 130 (6): 398–404. doi:10.1590/S1516-31802012000600007. PMC 10522321. PMID 23338737.
- ^ Freeman SC, Kerby CR, Patel A, Cooper NJ, Quinn T, Sutton AJ (April 2019). "Development of an interactive web-based tool to conduct and interrogate meta-analysis of diagnostic test accuracy studies: MetaDTA". BMC Medical Research Methodology. 19 (1) 81. doi:10.1186/s12874-019-0724-x. PMC 6471890. PMID 30999861.
- ^ Banno M, Tsujimoto Y, Luo Y, Miyakoshi C, Kataoka Y (February 2021). "CAST-HSROC: A Web Application for Calculating the Summary Points of Diagnostic Test Accuracy From the Hierarchical Summary Receiver Operating Characteristic Model". Cureus. 13 (2) e13257. doi:10.7759/cureus.13257. PMC 7953362. PMID 33717762.
- ^ "Living systematic reviews". community.cochrane.org. Retrieved 1 July 2020.
- ^ "Cochrane's work on Rapid Reviews in response to COVID-19". www.cochrane.org. Retrieved 1 July 2020.
- ^ Norris SL, Rehfuess EA, Smith H, Tunçalp Ö, Grimshaw JM, Ford NP, Portela A (1 January 2019). "Complex health interventions in complex systems: improving the process and methods for evidence-informed health decisions". BMJ Global Health. 4 (Suppl 1) e000963. doi:10.1136/bmjgh-2018-000963. PMC 6350736. PMID 30775018.
- ^ "Cochrane crowd". crowd.cochrane.org. Archived from the original on 21 October 2017. Retrieved 14 February 2019.
- ^ Pollock A, Campbell P, Struthers C, Synnot A, Nunn J, Hill S, et al. (November 2018). "Stakeholder involvement in systematic reviews: a scoping review". Systematic Reviews. 7 (1) 208. doi:10.1186/s13643-018-0852-0. PMC 6260873. PMID 30474560.
- ^ Pollock A, Campbell P, Struthers C, Synnot A, Nunn J, Hill S, et al. (October 2019). "Development of the ACTIVE framework to describe stakeholder involvement in systematic reviews" (PDF). Journal of Health Services Research & Policy. 24 (4): 245–255. doi:10.1177/1355819619841647. PMID 30997859. S2CID 121348214.
- ^ [Pre-print] Nunn J, Shafee T, Chang S, Stephens R, Elliott J, Oliver S, John D, Smith M, Orr N. "Standardised Data on Initiatives - STARDIT: Alpha Version". osf.io. doi:10.31219/osf.io/5q47h. S2CID 242815262. Retrieved 20 August 2020.
- ^ Nunn JS, Shafee T, Chang S, Stephens R, Elliott J, Oliver S, et al. (July 2022). "Standardised data on initiatives-STARDIT: Beta version". Research Involvement and Engagement. 8 (1) 31. doi:10.1186/s40900-022-00363-9. PMC 9294764. PMID 35854364.
- ^ Newman, Melanie (3 January 2019). "Has Cochrane lost its way?". BMJ. 364 k5302. doi:10.1136/bmj.k5302. PMID 30606713. S2CID 58623482.
- ^ Synnot AJ, Tong A, Bragge P, Lowe D, Nunn JS, O'Sullivan M, et al. (April 2019). "Selecting, refining and identifying priority Cochrane Reviews in health communication and participation in partnership with consumers and other stakeholders". Health Research Policy and Systems. 17 (1) 45. doi:10.1186/s12961-019-0444-z. PMC 6489310. PMID 31036016.
- ^ Synnot A, Bragge P, Lowe D, Nunn JS, O'Sullivan M, Horvat L, et al. (May 2018). "Research priorities in health communication and participation: international survey of consumers and other stakeholders". BMJ Open. 8 (5) e019481. doi:10.1136/bmjopen-2017-019481. PMC 5942413. PMID 29739780.
- ^ "The Cochrane-Wikipedia partnership in 2016". Cochrane. Retrieved 24 February 2019.
- ^ Rooney AA, Boyles AL, Wolfe MS, Bucher JR, Thayer KA (July 2014). "Systematic review and evidence integration for literature-based environmental health science assessments". Environmental Health Perspectives. 122 (7): 711–718. Bibcode:2014EnvHP.122..711R. doi:10.1289/ehp.1307972. PMC 4080517. PMID 24755067.
- ^ Woodruff TJ, Sutton P (October 2014). "The Navigation Guide systematic review methodology: a rigorous and transparent method for translating environmental health science into better health outcomes". Environmental Health Perspectives. 122 (10): 1007–1014. Bibcode:2014EnvHP.122.1007W. doi:10.1289/ehp.1307175. PMC 4181919. PMID 24968373.
- ^ Whaley P, Aiassa E, Beausoleil C, Beronius A, Bilotta G, Boobis A, et al. (October 2020). "Recommendations for the conduct of systematic reviews in toxicology and environmental health research (COSTER)". Environment International. 143 105926. Bibcode:2020EnInt.14305926W. doi:10.1016/j.envint.2020.105926. hdl:2066/229303. PMID 32653802. S2CID 220502683.
- ^ Gough D, Oliver S, Thomas J (2017). An Introduction to Systematic Reviews (2nd ed.). London: Sage. p. XIII.
- ^ Oakley A (2011). A critical woman: Barbara Wootton, social science and public policy in the twentieth century. London: Bloomsbury Academic. ISBN 978-1-84966-470-7.
- ^ "History of the Campbell Collaboration". Campbell Collaboration. Retrieved 15 April 2021.
- ^ Tranfield D, Denyer D, Smart P (2003). "Towards a methodology for developing evidence-informed management knowledge by means of systematic review". British Journal of Management. 14 (3): 207–222. CiteSeerX 10.1.1.622.895. doi:10.1111/1467-8551.00375.
- ^ Durach CF, Kembro J, Wieland A (2017). "A New Paradigm for Systematic Literature Reviews in Supply Chain Management". Journal of Supply Chain Management. 53 (4): 67–85. doi:10.1111/jscm.12145.
- ^ Linnenluecke, Martina K; Marrone, Mauricio; Singh, Abhay K (May 2020). "Conducting systematic literature reviews and bibliometric analyses". Australian Journal of Management. 45 (2): 175–194. doi:10.1177/0312896219877678. ISSN 0312-8962.
- ^ a b Hagen-Zanker J, Duvendack M, Mallett R, Slater R, Carpenter S, Tromme M (January 2012). "Making systematic reviews work for international development research". Overseas Development Institute. Archived from the original on 23 September 2012. Retrieved 19 January 2012.
- ^ "Environmental Evidence: Reliable evidence, informed decisions, better environment". www.environmentalevidence.org. Retrieved 1 July 2020.
- ^ Cowie K, Rahmatullah A, Hardy N, Holub K, Kallmes K (2022). "Web-Based Software Tools for Systematic Literature Review in Medicine: Systematic Search and Feature Analysis". JMIR. 10 (5) e33219. doi:10.2196/33219. PMC 9112080. PMID 35499859.
- ^ "Methodological Expectations of Cochrane Intervention Reviews (MECIR)". Cochrane. Archived from the original on 9 October 2014. Retrieved 6 October 2014.
- ^ Moher D, Tetzlaff J, Tricco AC, Sampson M, Altman DG (March 2007). "Epidemiology and reporting characteristics of systematic reviews". PLOS Medicine. 4 (3) e78. doi:10.1371/journal.pmed.0040078. PMC 1831728. PMID 17388659.
- ^ Shojania KG, Sampson M, Ansari MT, Ji J, Doucette S, Moher D (August 2007). "How quickly do systematic reviews go out of date? A survival analysis". Annals of Internal Medicine. 147 (4): 224–233. doi:10.7326/0003-4819-147-4-200708210-00179. PMID 17638714. S2CID 3988259.
- ^ Savoie I, Helmer D, Green CJ, Kazanjian A (2003). "Beyond Medline: reducing bias through extended systematic review search". International Journal of Technology Assessment in Health Care. 19 (1): 168–178. doi:10.1017/S0266462303000163. PMID 12701949. S2CID 42494895.
- ^ Roberts I, Ker K, Edwards P, Beecher D, Manno D, Sydenham E (June 2015). "The knowledge system underpinning healthcare is not fit for purpose and must change" (PDF). BMJ. 350 h2463. doi:10.1136/bmj.h2463. PMID 26041754. S2CID 33523612.
- ^ Altman DG (January 1994). "The scandal of poor medical research". BMJ. 308 (6924): 283–284. doi:10.1136/bmj.308.6924.283. PMC 2539276. PMID 8124111.
- ^ Shapiro S (November 1994). "Meta-analysis/Shmeta-analysis". American Journal of Epidemiology. 140 (9): 771–778. doi:10.1093/oxfordjournals.aje.a117324. PMID 7977286.
- ^ Page MJ, McKenzie JE, Kirkham J, Dwan K, Kramer S, Green S, Forbes A (October 2014). "Bias due to selective inclusion and reporting of outcomes and analyses in systematic reviews of randomised trials of healthcare interventions". The Cochrane Database of Systematic Reviews. 2015 (10): MR000035. doi:10.1002/14651858.MR000035.pub2. PMC 8191366. PMID 25271098.
- ^ Roberts I (12 June 2015). "Retraction Of Scientific Papers For Fraud Or Bias Is Just The Tip Of The Iceberg". IFL Science!. Retrieved 29 June 2015.
- ^ Ferguson C (2 April 2015). "Retraction and republication for Lancet Resp Med tracheostomy paper". Retraction Watch. Retrieved 29 June 2015.
- ^ Ferguson C (26 March 2015). "BioMed Central retracting 43 papers for fake peer review". Retraction Watch.
- ^ Flinders University Library. "Search Smart: Systematic Reviews: Methodology overview". flinders.libguides.com. Archived from the original on 28 October 2020. Retrieved 16 September 2020.
- ^ Pursell E, McCrae N (2020). How to Perform a Systematic Literature Review: a guide for healthcare researchers, practitioners and students. Springer Nature. ISBN 978-3-030-49672-2. OCLC 1182880684.
- ^ "Half of all clinical trials have never reported results". AllTrials. 20 August 2015. Retrieved 8 March 2019.
- ^ Song F, Parekh S, Hooper L, Loke YK, Ryder J, Sutton AJ, et al. (February 2010). "Dissemination and publication of research findings: an updated review of related biases". Health Technology Assessment. 14 (8): iii, ix–xi, 1–193. doi:10.3310/hta14080. PMID 20181324.
- ^ a b Iacobucci G (November 2016). "Nearly half of all trials run by major sponsors in past decade are unpublished". BMJ. 355 i5955. doi:10.1136/bmj.i5955. PMID 27815253. S2CID 43604202.
- ^ Lundh A, Lexchin J, Mintzes B, Schroll JB, Bero L, et al. (Cochrane Methodology Review Group) (February 2017). "Industry sponsorship and research outcome". The Cochrane Database of Systematic Reviews. 2017 (2): MR000033. doi:10.1002/14651858.MR000033.pub3. PMC 8132492. PMID 28207928.
- ^ Pidgeon TE, Wellstead G, Sagoo H, Jafree DJ, Fowler AJ, Agha RA (October 2016). "An assessment of the compliance of systematic review articles published in craniofacial surgery with the PRISMA statement guidelines: A systematic review". Journal of Cranio-Maxillo-Facial Surgery. 44 (10): 1522–1530. doi:10.1016/j.jcms.2016.07.018. PMID 27575881.
- ^ Lee SY, Sagoo H, Whitehurst K, Wellstead G, Fowler AJ, Agha RA, Orgill D (1 March 2016). "Compliance of Systematic Reviews in Plastic Surgery With the PRISMA Statement". JAMA Facial Plastic Surgery. 18 (2): 101–105. doi:10.1001/jamafacial.2015.1726. PMID 26719993.
- ^ Akhigbe T, Zolnourian A, Bulters D (May 2017). "Compliance of systematic reviews articles in brain arteriovenous malformation with PRISMA statement guidelines: Review of literature". Journal of Clinical Neuroscience. 39: 45–48. doi:10.1016/j.jocn.2017.02.016. PMID 28246008. S2CID 27738264.
- ^ Lee SY, Sagoo H, Farwana R, Whitehurst K, Fowler A, Agha R (December 2017). "Compliance of systematic reviews in ophthalmology with the PRISMA statement". BMC Medical Research Methodology. 17 (1) 178. doi:10.1186/s12874-017-0450-1. PMC 5745614. PMID 29281981.
- ^ E JY, Saldanha IJ, Canner J, Schmid CH, Le JT, Li T (May 2020). "Adjudication rather than experience of data abstraction matters more in reducing errors in abstracting data in systematic reviews". Research Synthesis Methods. 11 (3): 354–362. doi:10.1002/jrsm.1396. PMID 31955502. S2CID 210829764.
- ^ Koffel JB, Rethlefsen ML (26 September 2016). Thombs BD (ed.). "Reproducibility of Search Strategies Is Poor in Systematic Reviews Published in High-Impact Pediatrics, Cardiology and Surgery Journals: A Cross-Sectional Study". PLOS ONE. 11 (9) e0163309. Bibcode:2016PLoSO..1163309K. doi:10.1371/journal.pone.0163309. PMC 5036875. PMID 27669416.
- ^ Yoshii A, Plaut DA, McGraw KA, Anderson MJ, Wellik KE (January 2009). "Analysis of the reporting of search strategies in Cochrane systematic reviews". Journal of the Medical Library Association. 97 (1): 21–29. doi:10.3163/1536-5050.97.1.004. PMC 2605027. PMID 19158999.
- ^ Toews LC (July 2017). "Compliance of systematic reviews in veterinary journals with Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) literature search reporting guidelines". Journal of the Medical Library Association. 105 (3): 233–239. doi:10.5195/jmla.2017.246. PMC 5490700. PMID 28670210.
- ^ Mullins MM, DeLuca JB, Crepaz N, Lyles CM (June 2014). "Reporting quality of search methods in systematic reviews of HIV behavioral interventions (2000-2010): are the searches clearly explained, systematic and reproducible?". Research Synthesis Methods. 5 (2): 116–130. doi:10.1002/jrsm.1098. PMC 5861495. PMID 26052651.
- ^ Briscoe S (March 2018). "A review of the reporting of web searching to identify studies for Cochrane systematic reviews". Research Synthesis Methods. 9 (1): 89–99. doi:10.1002/jrsm.1275. PMID 29065246. S2CID 206155997.
- ^ Golder S, Loke Y, McIntosh HM (May 2008). "Poor reporting and inadequate searches were apparent in systematic reviews of adverse effects". Journal of Clinical Epidemiology. 61 (5): 440–448. doi:10.1016/j.jclinepi.2007.06.005. PMID 18394536.
- ^ Rethlefsen, Melissa L.; et al. (2021). "PRISMA-S: An extension to the PRISMA Statement for Reporting Literature Searches in Systematic Reviews". Systematic Reviews. 10 (1): 39. doi:10.1186/s13643-020-01542-z. PMC 7839230. PMID 33499930.
- ^ McGowan J, Sampson M, Salzwedel DM, Cogo E, Foerster V, Lefebvre C (July 2016). "PRESS Peer Review of Electronic Search Strategies: 2015 Guideline Statement". Journal of Clinical Epidemiology. 75: 40–46. doi:10.1016/j.jclinepi.2016.01.021. PMID 27005575.
- ^ Shea BJ, Reeves BC, Wells G, Thuku M, Hamel C, Moran J, et al. (September 2017). "AMSTAR 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both". BMJ. 358 j4008. doi:10.1136/bmj.j4008. PMC 5833365. PMID 28935701.
- ^ Shea BJ, Grimshaw JM, Wells GA, Boers M, Andersson N, Hamel C, et al. (February 2007). "Development of AMSTAR: a measurement tool to assess the methodological quality of systematic reviews". BMC Medical Research Methodology. 7 (1) 10. doi:10.1186/1471-2288-7-10. PMC 1810543. PMID 17302989.
- ^ Shea BJ, Hamel C, Wells GA, Bouter LM, Kristjansson E, Grimshaw J, et al. (October 2009). "AMSTAR is a reliable and valid measurement tool to assess the methodological quality of systematic reviews". Journal of Clinical Epidemiology. 62 (10): 1013–1020. doi:10.1016/j.jclinepi.2008.10.009. PMID 19230606. S2CID 5935312.
- ^ Whiting P, Savović J, Higgins JP, Caldwell DM, Reeves BC, Shea B, et al. (January 2016). "ROBIS: A new tool to assess risk of bias in systematic reviews was developed". Journal of Clinical Epidemiology. 69: 225–234. doi:10.1016/j.jclinepi.2015.06.005. PMC 4687950. PMID 26092286.
- ^ Perry, R.; Whitmarsh, A.; Leach, V.; Davies, P. (25 October 2021). "A comparison of two assessment tools used in overviews of systematic reviews: ROBIS versus AMSTAR-2". Systematic Reviews. 10 (1): 273. doi:10.1186/s13643-021-01819-x. ISSN 2046-4053. PMC 8543959. PMID 34696810.
- ^ Pieper, Dawid; Puljak, Livia; González-Lorenzo, Marien; Minozzi, Silvia (April 2019). "Minor differences were found between AMSTAR 2 and ROBIS in the assessment of systematic reviews including both randomized and nonrandomized studies". Journal of Clinical Epidemiology. 108: 26–33. doi:10.1016/j.jclinepi.2018.12.004. ISSN 1878-5921. PMID 30543911.
- ^ a b c Clarke, Mike; Chalmers, Iain (August 2018). "Reflections on the history of systematic reviews". BMJ Evidence-Based Medicine. 23 (4): 121–122. doi:10.1136/bmjebm-2018-110968. ISSN 2515-446X. PMID 29921711. S2CID 49311760.
- ^ "Report on Certain Enteric Fever Inoculation Statistics". British Medical Journal. 2 (2288). British Medical Journal Publishing Group: 1243–1246. November 1904. doi:10.1136/bmj.2.2288.1243. PMC 2355479. PMID 20761760. S2CID 30331745.
- ^ a b "1.1.2 A brief history of Cochrane". community.cochrane.org. Archived from the original on 17 January 2019. Retrieved 24 February 2019.
- ^ Glass GV, Smith ML, et al. (Far West Lab. for Educational Research and Development, San Francisco, CA) (1978). Meta-Analysis of Research on the Relationship of Class-Size and Achievement. The Class Size and Instruction Project. Washington, D.C.]: Distributed by ERIC Clearinghouse.
- ^ "History of Systematic Reviews". Evidence for Policy and Practice Information and Co-ordinating Centre (EPPI-Centre). Retrieved 24 February 2019.
- ^ Lau J, Antman EM, Jimenez-Silva J, Kupelnick B, Mosteller F, Chalmers TC (July 1992). "Cumulative meta-analysis of therapeutic trials for myocardial infarction". The New England Journal of Medicine. 327 (4): 248–254. doi:10.1056/NEJM199207233270406. PMID 1614465.
- ^ Cochrane AL (1972). Effectiveness and efficiency: random reflections on health services. [London]: Nuffield Provincial Hospitals Trust. ISBN 0-900574-17-8. OCLC 741462.
- ^ Shah HM, Chung KC (September 2009). "Archie Cochrane and his vision for evidence-based medicine". Plastic and Reconstructive Surgery. 124 (3): 982–988. doi:10.1097/PRS.0b013e3181b03928. PMC 2746659. PMID 19730323.
- ^ Evans I, Thornton H, Chalmers I, Glasziou P (2011). Testing treatments: better research for better healthcare (Second ed.). London: Pinter & Martin. ISBN 978-1-905177-48-6. OCLC 778837501.
External links
[edit]- Systematic Review Tools — Search and list of systematic review software tools
- Cochrane Collaboration
- MeSH: Review Literature—articles about the review process
- MeSH: Review [Publication Type] - limit search results to reviews
- Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) Statement Archived 27 July 2011 at the Wayback Machine, "an evidence-based minimum set of items for reporting in systematic reviews and meta-analyses"
- PRISMA Extension for Scoping Reviews (PRISMA-ScR): Checklist and explanation
- Animated Storyboard: What Are Systematic Reviews? - Cochrane Consumers and Communication Group
- Sysrev - a free platform with open access systematic reviews.
- STARDIT - an open access data-sharing system to standardise the way that information about initiatives is reported.
- Cao, Christian; Arora, Rohit; Cento, Paul; Manta, Katherine; Farahani, Elina; Cecere, Matthew; Selemon, Anabel; Sang, Jason; Gong, Ling Xi (19 June 2025), Automation of Systematic Reviews with Large Language Models, medRxiv, doi:10.1101/2025.06.13.25329541, retrieved 20 June 2025
Systematic review
View on GrokipediaDefinition and Characteristics
Core Principles
A systematic review is defined as a rigorous synthesis of empirical evidence on a clearly formulated research question, employing explicit, pre-specified eligibility criteria to identify, select, appraise, and synthesize all relevant studies while minimizing bias through structured methods. This approach ensures that the review process is methodical and comprehensive, aiming to reduce subjectivity and provide a reliable overview of the existing literature.[7] Central to systematic reviews are several core principles that uphold their scientific integrity. Comprehensiveness in literature searching requires exhaustive efforts to locate all pertinent studies across multiple databases and sources, avoiding selective inclusion that could skew results. Transparency in methods demands detailed documentation of every step, from search strategies to inclusion decisions, allowing readers to understand and scrutinize the process.[7] Reproducibility is achieved through the development and adherence to a pre-defined protocol, often registered in advance, which outlines the review's objectives, criteria, and analytical plans to enable replication by others.[8] Critical appraisal of study quality involves systematically evaluating the methodological strengths and limitations of included studies using validated tools, ensuring that only high-quality evidence informs the synthesis. These principles originated within the evidence-based medicine movement of the late 20th century, which sought to integrate the best available research evidence with clinical expertise and patient values to improve healthcare decisions.[5] By adhering to them, systematic reviews facilitate evidence-based decision-making, such as developing clinical guidelines and informing public health policies, by distilling complex evidence into actionable insights that reduce uncertainty and enhance reliability.[9]Distinctions from Other Reviews
Systematic reviews differ from narrative reviews primarily in their methodological rigor and approach to evidence synthesis. Narrative reviews, also known as traditional or non-systematic reviews, provide an expert-driven overview of a topic by selectively summarizing key literature, often relying on the authors' intuition and experience without predefined criteria for inclusion.[10] In contrast, systematic reviews employ explicit, reproducible protocols to identify, appraise, and synthesize all relevant evidence, minimizing selection bias through comprehensive searches across multiple databases and transparent eligibility criteria.[10] This structured process ensures higher confidence in the findings, as evidenced by studies showing narrative reviews can reach differing conclusions from the same body of evidence due to subjective selection.[10] Literature reviews, frequently synonymous with narrative reviews in practice, aim to contextualize a research area by thematically discussing existing studies but lack the systematic rigor required to control for bias or ensure completeness.[11] Unlike systematic reviews, which use prespecified protocols, exhaustive search strategies, and formal quality assessments to avoid cherry-picking studies, literature reviews often involve ad hoc searches and subjective synthesis, making them suitable for broad overviews but less reliable for informing policy or practice.[11] Umbrella reviews, sometimes called reviews of reviews, further distinguish themselves by synthesizing evidence from multiple existing systematic reviews and meta-analyses rather than primary studies, providing a high-level overview of broad topics such as intervention effects across interventions.[12] While systematic reviews focus on primary data synthesis with original eligibility criteria applied to individual studies, umbrella reviews assess the quality and consistency of those syntheses, reanalyzing data where needed to standardize methods and identify overarching patterns or gaps.[12] The following table enumerates key methodological differences across these review types:| Aspect | Systematic Review | Narrative/Literature Review | Umbrella Review |
|---|---|---|---|
| Search Strategy | Exhaustive, predefined across databases | Ad hoc, selective | Targets existing systematic reviews |
| Inclusion Criteria | Explicit, prespecified eligibility | Subjective, often implicit | Applied to systematic reviews/meta-analyses |
| Bias Control | Formal assessment and minimization | Informal, prone to author bias | Evaluates bias in included reviews |
| Synthesis Approach | Reproducible, may include meta-analysis | Thematic narrative summary | Overarching synthesis of review findings |
| Purpose | Answer specific question with high rigor | Provide context or overview | Broad evidence overview from reviews |
Types of Systematic Reviews
Scoping Reviews
Scoping reviews, also referred to as scoping studies, represent a type of knowledge synthesis designed to map the breadth of literature on a specific topic or field by identifying key concepts, main sources and types of evidence, and research gaps.[13] Unlike more focused reviews, they emphasize exploration over evaluation, providing an initial overview of the extent, range, and nature of available evidence without conducting in-depth critical appraisals of individual study quality.[14] This approach is particularly valuable for clarifying complex or undefined research areas, informing policy, practice, or subsequent studies by highlighting thematic patterns and areas needing further investigation.[15] The purpose of scoping reviews centers on delineating the landscape of existing research to guide decision-making, such as determining the feasibility of a full systematic review or identifying priorities for primary research.[16] They are especially suited to emerging topics where evidence is heterogeneous or evolving, allowing researchers to assess the volume and variety of studies before pursuing narrower syntheses. By focusing on descriptive mapping, scoping reviews help stakeholders understand what is known and unknown, thereby supporting evidence-based planning in fields like health policy and social services.[13] Methodologically, scoping reviews follow a structured yet flexible framework originally outlined by Arksey and O'Malley in 2005, comprising five core stages: formulating the research question to define the scope; searching for relevant studies across multiple databases and sources, including grey literature; selecting studies using predefined but broad inclusion criteria; charting data through extraction of key information into tables or charts for organization; and collating, summarizing, and reporting results via narrative or visual descriptions.[13] An optional sixth stage involves consulting stakeholders, such as practitioners or policymakers, to refine interpretations and enhance practical relevance.[13] This process employs wider inclusion parameters than traditional systematic reviews to capture diverse evidence types, prioritizing descriptive summaries over quantitative analysis or meta-synthesis.[15] Reporting standards, such as the PRISMA extension for scoping reviews (PRISMA-ScR), ensure transparency in these steps.[14] Scoping reviews are recommended when a topic is broad, complex, or preliminary, such as in emerging public health issues where a comprehensive synthesis would be premature or resource-intensive.[16] For example, they have been applied to map interventions for health equity in governmental public health practice, revealing implementation gaps and priority areas for targeted research.[17] In another case, a scoping review examined public health interventions delivered via beauty salons, identifying common strategies for disease prevention and underserved populations to inform scalable programs.[18] These applications demonstrate how scoping reviews serve as foundational tools for policy-oriented fields by providing actionable overviews without exhaustive depth.[19]Meta-Analyses
A meta-analysis represents a quantitative extension of systematic reviews, involving the statistical combination of results from multiple independent studies to produce an overall estimate of effect size, thereby enhancing precision and resolving inconsistencies among individual findings. Coined by Gene V. Glass in 1976 as "the statistical analysis of a large collection of analysis results from individual studies for the purpose of integrating the findings," this method is particularly valuable in fields like medicine and social sciences for synthesizing evidence on interventions or associations. Results from meta-analyses are commonly visualized through forest plots, which graphically display each study's effect estimate, confidence intervals, weights, and the pooled summary effect, facilitating intuitive assessment of consistency and overall impact.[20] Key methods in meta-analysis include the choice between fixed-effect and random-effects models, alongside rigorous assessment of heterogeneity. The fixed-effect model posits that all studies share a common true effect size, with observed differences attributable only to sampling variation, and assigns greater weight to larger studies; it is suitable when studies are sufficiently similar. Conversely, the random-effects model incorporates an additional source of variation to account for true differences in effect sizes across studies, providing more conservative estimates and broader confidence intervals, and is preferred when between-study heterogeneity is anticipated or observed. Heterogeneity is quantified using the I² statistic, which measures the proportion of total variability in study estimates that is due to heterogeneity rather than chance. This is computed via the formula: where is Cochran's heterogeneity statistic (a chi-squared test for variation) and is the degrees of freedom (typically the number of studies minus one). Values of I² range from 0% (no heterogeneity) to 100% (complete heterogeneity), with thresholds like 25% for low, 50% for moderate, and 75% for high often guiding interpretation, though context matters.[21] Conducting a meta-analysis necessitates selecting homogeneous studies with comparable outcomes, such as similar measures of effect (e.g., odds ratios or mean differences) and aligned PICO elements (population, intervention, comparison, outcomes), to ensure the validity of pooling and minimize confounding. When full homogeneity is absent, subgroup analyses allow exploration of potential moderators by dividing studies into categories—such as by age group, dosage, or geographic region—and performing separate meta-analyses within each, followed by tests for subgroup differences to identify sources of variation. Assessments of risk of bias in included studies, often using tools like the Cochrane Risk of Bias instrument, are essential to appropriately weight contributions and interpret results. Reporting standards for meta-analyses follow the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines, which mandate transparent documentation including a flow diagram of study selection, detailed descriptions of statistical methods (e.g., model choice and heterogeneity tests), forest plots, and sensitivity analyses to evaluate how results change with alternative assumptions or exclusions. These elements promote reproducibility and critical appraisal, with PRISMA emphasizing the reporting of effect sizes, confidence intervals, and any limitations like publication bias. The updated PRISMA 2020 statement further refines these requirements to reflect advances in synthesis methods.[22][23]Rapid Reviews
Rapid reviews are streamlined versions of systematic reviews that accelerate the evidence synthesis process to deliver timely information for urgent decision-making, particularly in policy, clinical, or crisis contexts such as public health emergencies.[24] They maintain the core aim of systematically identifying, selecting, and synthesizing relevant studies but abbreviate certain steps to shorten timelines, often responding to time-sensitive needs where full systematic reviews would be too slow.[25] This approach has gained prominence during events like the COVID-19 pandemic, where rapid evidence was essential for guiding interventions and resource allocation.[26] Key methodological adaptations distinguish rapid reviews from traditional systematic reviews, focusing on efficiency without entirely sacrificing rigor. These include limiting literature searches to a smaller number of databases (typically 2-3, such as MEDLINE, Embase, and the Cochrane Library), restricting inclusion to high-quality study designs or existing systematic reviews, and employing single-reviewer screening for titles and abstracts with optional dual verification on a subset (e.g., 20%).[27] Data extraction is often performed by one reviewer with verification, and full risk-of-bias assessments are abbreviated or omitted, prioritizing key outcomes.[25] The entire process typically spans 1 to 12 weeks, compared to 12-24 months for a standard systematic review, enabling quicker dissemination of findings.[26] Established frameworks guide the conduct of rapid reviews to ensure transparency and reproducibility. The Cochrane Rapid Reviews Methods Group provides interim guidance with 26 specific recommendations (updated to 24 in 2024), emphasizing stakeholder involvement in question formulation, protocol registration (e.g., on PROSPERO), narrative synthesis over meta-analysis unless feasible, and clear reporting of limitations.[27][26] These guidelines recommend focusing searches on English-language publications unless multilingual evidence is critical and using tools like GRADE for assessing evidence certainty when resources permit.[25] While rapid reviews offer significant advantages in speed and applicability to urgent scenarios, they involve inherent trade-offs that can affect comprehensiveness and reliability. Streamlined methods, such as single-reviewer processes, increase the risk of selection bias and may miss relevant studies, with estimates suggesting up to 5% fewer inclusions compared to dual-reviewer approaches.[25] Limited database coverage and abbreviated bias assessments can lead to less robust conclusions, potentially underestimating evidence gaps, though proponents argue the benefits of timely action outweigh these risks in high-stakes situations like health crises.[26] To mitigate issues, guidance stresses explicit documentation of methodological shortcuts and their implications for decision-makers.[27]Living Systematic Reviews
Living systematic reviews (LSRs) represent a dynamic approach to evidence synthesis, defined as systematic reviews that are continually updated to incorporate relevant new evidence as it becomes available.[28] Unlike traditional systematic reviews, LSRs involve ongoing surveillance of the literature through frequent searches, often utilizing automated alerts from bibliographic databases to identify emerging studies efficiently.[28] This methodology ensures that the review remains current, particularly in fields where evidence evolves rapidly, by integrating new data without restarting the entire process from scratch.[29] The core methodology of LSRs adheres to principles established by the Living Systematic Review Network in 2017, emphasizing streamlined workflows, technology integration for efficiency, and a focus on end-user utility while minimizing author burden.[28] Continuous screening occurs alongside regular literature searches, typically monthly for major databases like MEDLINE and Embase, with partial updates published every 2 to 6 months or sooner if new evidence significantly impacts findings.[29] These updates may include living network meta-analyses, which extend traditional network meta-analysis techniques to dynamically compare multiple interventions as new trials accrue, providing ongoing relative effect estimates.[30] Screening and incorporation follow standard systematic review protocols, often supported by software tools for automation and crowdsourcing to handle the influx of citations.[28] LSRs are particularly applied in fast-changing domains such as infectious diseases and clinical guideline development, where timely evidence is critical for decision-making.[28] A prominent example is the suite of Cochrane living systematic reviews on COVID-19 interventions, which addressed pharmacological treatments, diagnostics, and risk factors through monthly searches and iterative updates to inform global health responses during the pandemic.[31] These reviews demonstrated LSRs' value in synthesizing rapidly accumulating data from diverse study designs, including randomized trials and observational studies, to reduce research waste and support adaptive guidelines.[31] Despite their benefits, LSRs present notable challenges, primarily their resource-intensive nature, which demands sustained team commitment, frequent monitoring, and ongoing editorial support to maintain methodological rigor.[28] Version control poses additional difficulties, as managing multiple iterations requires clear documentation of changes, consistent use of tools like "What's New" tables, and standardized policies for digital object identifiers (DOIs) across publications to track evolution without confusion.[31] These demands can strain resources, particularly during high-volume evidence periods like pandemics, though automation and predefined stopping criteria help mitigate workload.[29]Conducting a Systematic Review
Formulating the Research Question
Formulating a clear and focused research question is the foundational step in conducting a systematic review, as it defines the scope, guides the selection of studies, and ensures the review addresses a specific knowledge gap. A well-defined question helps maintain methodological rigor, minimizes bias in the review process, and facilitates reproducibility by outlining the objectives upfront. According to the Cochrane Handbook for Systematic Reviews of Interventions, the research question should be structured to be answerable through synthesis of existing evidence, emphasizing precision to avoid ambiguity that could lead to inconsistent interpretations.[32] Common frameworks assist in developing research questions tailored to the review's purpose. For quantitative or clinical intervention studies, the PICO framework—standing for Population (or Patient), Intervention, Comparison, and Outcome—is widely used to break down the question into searchable components, such as "In adults with type 2 diabetes (P), does metformin (I) compared to lifestyle changes (C) improve glycemic control (O)?" This approach, originally proposed by Richardson et al. in 1995, promotes clarity and directs the literature search effectively.[33] For qualitative or mixed-methods reviews, alternatives like PICo (Population, phenomenon of Interest, and Context) or SPIDER (Sample, Phenomenon of Interest, Design, Evaluation, and Research type) are preferred to capture experiential or contextual elements, as PICO may be less sensitive for non-intervention queries. For instance, SPIDER can frame questions about patient experiences, such as "What are the experiences (E) of caregivers (S) regarding dementia care (PI) in community settings (D, RT)?" These frameworks, compared in Methley et al. (2014), enhance specificity in qualitative evidence synthesis.[34][35] Once formulated, the research question informs the development of a detailed protocol, which specifies the review's objectives, inclusion and exclusion criteria (e.g., study types, publication dates, and languages), and overall scope to prevent deviations during execution. Protocols are typically registered prospectively to promote transparency and reduce duplication; PROSPERO, maintained by the Centre for Reviews and Dissemination at the University of York, serves as an international registry for health-related systematic reviews, requiring submission of the question, methods, and rationale before data collection begins.[36] For non-health fields or scoping reviews ineligible for PROSPERO, the Open Science Framework (OSF) provides a platform for protocol preregistration, allowing timestamped archiving of plans and supporting open science practices. Involving stakeholders, such as clinicians or policymakers, during question formulation ensures relevance and practicality, as recommended in methodological guidelines.[37] A key benefit of rigorous question formulation is its role in preventing scope creep, where the review expands uncontrollably, leading to resource inefficiency and diluted focus; studies show that poorly defined questions correlate with higher rates of protocol amendments. Common pitfalls include crafting overly broad questions, such as "What treatments work for depression?" which yield unmanageable volumes of irrelevant studies, or neglecting to align the question with feasible evidence types, resulting in empty syntheses. To mitigate these, iterative refinement with frameworks and pilot testing of criteria is essential, as highlighted in analyses of review failures.[37][38]Literature Searching
Literature searching in systematic reviews involves a structured and exhaustive process to identify all potentially relevant studies on a predefined research question, minimizing the risk of bias due to incomplete retrieval.[39] This phase emphasizes sensitivity over precision to ensure comprehensive coverage, typically yielding thousands of records that are later screened for eligibility. Key sources for identifying studies include electronic databases such as MEDLINE (accessed via PubMed), Embase, and the Cochrane Central Register of Controlled Trials (CENTRAL) within the Cochrane Library, which collectively index millions of biomedical records from the mid-20th century onward.[39] Grey literature, encompassing unpublished or non-commercially published materials like theses, conference proceedings, clinical trial registries (e.g., ClinicalTrials.gov), and regulatory documents, is essential to capture studies not yet indexed in journals or those at risk of publication bias.[39] Hand-searching involves manually reviewing key journals, reference lists of included studies, and conference abstracts to uncover additional relevant citations.[39] Effective search strategies combine controlled vocabularies, such as Medical Subject Headings (MeSH) in MEDLINE, with free-text terms to enhance sensitivity, using Boolean operators like AND (to narrow intersections), OR (to broaden synonyms), and NOT (to exclude irrelevant concepts).[39] Citation tracking, including backward searching of references and forward searching via tools like Google Scholar or Web of Science, helps identify related studies by tracing connections from known relevant papers.[39] To ensure strategy quality, searches should undergo peer review by a librarian or information specialist using the PRESS (Peer Review of Electronic Search Strategies) checklist, which evaluates elements like translation of the research question, use of limits, and spelling. Comprehensiveness requires avoiding restrictions on language, date, or publication status unless justified by the review's scope, ideally including studies in any language to reduce bias, with translations sought as needed.[39] All search details—such as database names, full search strings, dates conducted, and any filters applied—must be documented meticulously for reproducibility and transparency, often appended to the review protocol or report.[40] The yield from these searches is reported via a flow diagram as per PRISMA 2020 guidelines, illustrating the total records identified (often in the thousands), duplicates removed, and progression to screening, thereby providing a visual audit trail of the process. While automation tools can assist in initial querying, manual oversight remains critical for strategy development and validation.[40]Screening and Selection
Screening and selection in a systematic review involve systematically applying predefined eligibility criteria to the pool of identified studies from the literature search, ensuring only relevant and appropriate studies proceed to further analysis. This process typically occurs in two sequential levels to minimize bias and maximize comprehensiveness: initial screening of titles and abstracts, followed by full-text assessment of potentially eligible studies.[39] At the title and abstract screening stage, at least two independent reviewers examine the records to determine initial relevance, prioritizing sensitivity to avoid excluding potentially eligible studies. This step filters out clearly irrelevant items, with reviewers working separately to reduce subjective bias. Software tools such as Rayyan, a web-based platform designed for collaborative screening, or Covidence, which supports team-based review management, facilitate this process by allowing simultaneous access, annotation, and prioritization features like machine learning-assisted suggestions.[39][41] Eligibility criteria, explicitly outlined in the review protocol, are applied rigorously during both screening levels, assessing aspects such as study design, population, intervention, and outcomes. Reviewers exclude studies that fail to meet any criterion, documenting reasons for exclusion at the full-text stage to ensure transparency and reproducibility; these records contribute to the PRISMA flow diagram illustrating the selection pathway. Disagreements between reviewers are resolved through discussion to reach consensus, or escalated to a third reviewer for arbitration if needed.[39] To evaluate the consistency of reviewer judgments, inter-rater reliability is often measured using Cohen's kappa statistic, which quantifies agreement beyond chance. The formula is given by: where represents the observed proportion of agreement between reviewers, and the expected proportion by chance. Kappa values above 0.8 indicate substantial agreement, helping to identify training needs or criterion ambiguities.[42]Data Extraction and Risk Assessment
Data extraction in systematic reviews involves systematically collecting relevant information from included studies to facilitate subsequent analysis. Reviewers typically use standardized data collection forms to record details on study characteristics, such as participant populations (e.g., demographics, eligibility criteria), interventions or exposures, comparators, outcomes (including measures, time points, and effect estimates), and methods (e.g., study design, randomization, blinding). These forms are developed in advance, often piloted for clarity and completeness, and tailored to the review's protocol to ensure consistency across extractors.[43] To minimize errors and subjectivity, data extraction is ideally performed independently by at least two reviewers, with discrepancies resolved through discussion or consultation with a third party; this dual extraction approach reduces transcription errors and interpretive biases, particularly for numerical outcome data. Extractors may also note any ambiguities in reporting, such as unclear denominators or missing standard deviations, which can be addressed by contacting study authors if necessary.[43] Risk of bias assessment evaluates the internal validity of included studies to determine how confidently results can be trusted. In a systematic review, quality criteria (often referred to as risk of bias or methodological quality assessment) are set by pre-specifying them in the review protocol. Reviewers select validated tools appropriate to the study designs included, such as Cochrane's RoB 2 tool for randomized trials. Signalling questions are used to guide judgments (low risk, some concerns, high risk), performed independently by at least two reviewers. Criteria are tailored to the review question and justified transparently.[44] For randomized controlled trials (RCTs), the Cochrane Risk of Bias 2 (RoB 2) tool is widely recommended, assessing bias across five domains: bias arising from the randomization process, due to deviations from intended interventions (performance bias), in measurement of the outcome (detection bias), due to missing outcome data (attrition bias), and in selection of the reported result (reporting bias). Each domain is judged as low risk, some concerns, or high risk based on responses to signaling questions, with an overall risk judgement derived algorithmically.[45] For non-randomized studies of interventions, the Risk Of Bias In Non-randomized Studies - of Interventions (ROBINS-I) tool is used, which evaluates bias in seven domains: bias due to confounding, selection of participants, classification of interventions, deviations from interventions, missing data, outcome measurement, and selection of reported results. Judgements are made as low, moderate, serious, critical, or no information, relative to a hypothetical ideal randomized trial, to inform the credibility of effect estimates.[46] Qualitative appraisal of evidence certainty complements risk of bias assessments by grading the overall quality of the body of evidence for specific outcomes. The Grading of Recommendations Assessment, Development and Evaluation (GRADE) approach rates certainty as high, moderate, low, or very low, starting from high for RCTs and low for observational studies, then upgrading or downgrading based on five domains: risk of bias (downgraded for limitations in study design or execution), inconsistency (unexplained heterogeneity in results), indirectness (evidence not directly applicable to the review question), imprecision (wide confidence intervals suggesting uncertain estimates), and publication bias (evidence of selective reporting). This structured evaluation helps reviewers communicate the strength of evidence supporting conclusions.[47] During data extraction, reviewers document variations in study characteristics—such as differences in population definitions, intervention intensities, or outcome measures—to inform later assessments of heterogeneity; these narrative notes highlight potential sources of clinical or methodological diversity without attempting statistical quantification at this stage.[43]Data Synthesis and Analysis
Data synthesis in systematic reviews involves integrating the extracted data from included studies to draw meaningful conclusions about the research question, often combining qualitative and quantitative approaches depending on the nature of the evidence. This process aims to assess the consistency, strength, and applicability of findings while addressing potential biases, such as those identified during risk assessment. When meta-analysis is feasible, quantitative methods provide a pooled estimate of effect; otherwise, narrative approaches organize and interpret the data thematically. The choice of method is guided by the heterogeneity of studies and the availability of comparable data, ensuring transparency in how conclusions are reached.[48] Narrative synthesis is employed when quantitative pooling via meta-analysis is inappropriate, such as in reviews with heterogeneous studies or non-numeric outcomes, relying primarily on words and text to summarize and explain findings from multiple studies. It typically involves four elements: developing preliminary synthesis, exploring relationships in the data, assessing robustness of the synthesis, and providing a textual summary that groups studies thematically or via tabulations. For instance, studies may be organized into tables by intervention type or outcome measures to highlight patterns, such as converging evidence on effectiveness across qualitative reports. This method promotes a structured textual approach to integration, avoiding ad hoc descriptions, and is particularly useful for synthesizing qualitative or mixed-methods evidence. Guidance emphasizes tabulating key features like study design and results to facilitate thematic grouping and identify gaps or contradictions.[49] Quantitative synthesis, often through meta-analysis, statistically combines effect estimates from comparable studies to produce an overall summary measure, such as odds ratios or mean differences, weighted by precision (e.g., inverse variance method). This is applicable when studies share similar populations, interventions, and outcomes with low clinical heterogeneity, using fixed-effect models for homogeneous data or random-effects for anticipated variation. Heterogeneity is quantified using the I² statistic, where values above 50% indicate moderate to substantial inconsistency, prompting exploration via subgroups rather than pooling. If meta-analysis is pursued, it provides a more precise estimate than individual studies but requires careful interpretation of forest plots displaying individual and pooled effects.[50][48] To detect publication bias in quantitative syntheses, funnel plots are constructed by plotting effect sizes against their standard errors, with asymmetry suggesting smaller studies with null or negative results may be missing. Egger's regression test formally assesses this asymmetry by regressing the standardized effect estimate against its precision, where a significant intercept (typically P < 0.10) indicates potential bias. This graphical and statistical approach is recommended for meta-analyses with at least 10 studies, helping to evaluate the robustness of pooled results, though it can be influenced by true heterogeneity or small-study effects.[51][50] Sensitivity analyses test the stability of synthesis findings by systematically varying methodological choices, such as excluding studies at high risk of bias, altering effect measures, or imputing missing data differently. Common applications include one-study-out analyses to identify influential studies or subgroup explorations to assess impact of factors like study quality. These analyses reveal how robust conclusions are to assumptions, with changes in direction or magnitude signaling the need for cautious interpretation; for example, excluding low-quality studies might strengthen evidence for an intervention's efficacy. Pre-specifying such analyses in the review protocol enhances transparency and minimizes selective reporting.[50] Certainty of evidence is evaluated using the GRADE (Grading of Recommendations Assessment, Development and Evaluation) framework, which rates the overall quality for each outcome as high, moderate, low, or very low based on five domains: risk of bias, inconsistency, indirectness, imprecision, and publication bias. Starting from high certainty for randomized trials or low for observational studies, ratings are downgraded (by 1 or 2 levels) for limitations like unexplained heterogeneity (I² > 50%) or wide confidence intervals failing to meet optimal information size thresholds, and upgraded for large effects or dose-response gradients. GRADE summaries, often presented in 'Summary of Findings' tables, provide a transparent judgment of confidence in effect estimates, informing the strength of review conclusions. This approach ensures systematic reviews communicate not just findings but their reliability across outcomes.[47][52]Reporting and Dissemination
Reporting systematic reviews requires adherence to established guidelines to ensure transparency, reproducibility, and completeness, allowing readers to assess the methods and findings critically. The Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) 2020 statement serves as the primary framework, consisting of a 27-item checklist for the main text and an eight-item checklist for abstracts, designed to guide the reporting of systematic reviews evaluating health interventions, while also applicable to other fields.[22][23] This update from the 2009 version incorporates advances in review methods, such as risk-of-bias assessments and certainty of evidence evaluations, and emphasizes the inclusion of a flow diagram to illustrate the study selection process.[53] The recommended structure for a systematic review report begins with a title and abstract that clearly identify the work as a systematic review and outline the objectives. The introduction provides rationale and objectives, followed by a detailed methods section covering eligibility criteria, information sources, search strategy, study selection, data collection, risk of bias assessment, and synthesis methods. Results sections present study selection (via the PRISMA flow diagram), characteristics of included studies, risk of bias, and synthesis outcomes, such as effect estimates from meta-analyses; the discussion interprets findings, discusses limitations, and outlines implications for practice or policy. Funding sources and competing interests are disclosed at the end.[23] For scoping reviews, the PRISMA extension for Scoping Reviews (PRISMA-ScR) provides a 20-item checklist tailored to mapping evidence and identifying knowledge gaps, without requiring the full meta-analytic detail of PRISMA 2020. This includes explicit reporting of rationale, protocol details, search strategy, selection of sources, data charting, and implications for future research, ensuring scoping reviews are distinguished from more evaluative systematic reviews.[14][54] Dissemination of systematic reviews typically occurs through publication in peer-reviewed journals, where adherence to PRISMA enhances acceptance and impact. Organizations like Cochrane mandate PRISMA compliance for their reviews, alongside plain language summaries to make findings accessible to non-experts, policymakers, and patients; these summaries highlight key results and implications without technical jargon.[55] Updates to existing reviews, such as those in the Cochrane Library, follow similar reporting standards to reflect new evidence. Open access publication broadens reach and accelerates uptake, with many journals offering gold open access options for systematic reviews to maximize dissemination. Sharing review materials, including protocols (often registered in repositories like PROSPERO), full datasets, and analytic code, promotes transparency and enables reuse; repositories such as Zenodo or Figshare facilitate this, aligning with growing expectations for data sharing in high-quality reviews.[56][57]Applications Across Disciplines
Health and Medicine
In health and medicine, systematic reviews form the cornerstone of evidence-based medicine by synthesizing rigorous evidence from multiple studies to guide clinical practice and policy decisions. They are integral to the development of clinical guidelines by organizations such as the National Institute for Health and Care Excellence (NICE) and the World Health Organization (WHO), where they ensure recommendations are based on comprehensive, unbiased assessments of intervention effectiveness, harms, and applicability.[58][59] The Cochrane Reviews, produced by the Cochrane Collaboration, represent the gold standard for such syntheses in health care, with over 11,000 published by 2023, spanning topics from pharmacological treatments to public health interventions.[60] Systematic reviews in this domain address specific applications, including evaluations of drug efficacy, comparisons of therapeutic interventions, and assessments of diagnostic test accuracy, often incorporating meta-analyses to quantify effects across studies. Patient involvement enhances their relevance through initiatives like the James Lind Alliance, which uses priority-setting partnerships to identify evidence gaps in collaboration with patients, caregivers, and clinicians, ensuring reviews align with real-world uncertainties in care.[61] These reviews have profoundly impacted medical practice by standardizing approaches and reducing variability in treatment decisions. A landmark example is the 1988 systematic review by the Antiplatelet Trialists' Collaboration, which analyzed 31 trials and demonstrated that prolonged antiplatelet therapy, primarily aspirin, reduced the odds of serious vascular events by 23% in patients with prior myocardial infarction, stroke, or transient ischemic attacks, prompting its broad adoption for secondary prevention of cardiovascular disease.[62] In subfields like environmental health and toxicology, systematic reviews synthesize evidence on exposure-related risks to inform regulatory and preventive measures. For example, they have evaluated the health impacts of pesticide exposure, identifying associations with endocrine disruption, neurotoxicity, and cancer, thereby supporting evidence-based policies on chemical safety and occupational limits.[63][64]Social and Behavioral Sciences
In the social and behavioral sciences, systematic reviews adapt traditional methodologies to address broader, non-clinical outcomes such as behavior change, educational attainment, and social equity, often incorporating diverse evidence from interventions in psychology, sociology, and education. Unlike reviews in health sciences that prioritize clinical efficacy, these adaptations emphasize contextual factors like community impacts and long-term societal effects, guided by standards from the Campbell Collaboration, which promotes rigorous evidence synthesis for policy-relevant questions in social welfare and international development. The Campbell Collaboration's Methodological Expectations of Campbell Collaboration Intervention Reviews (MECCIR) outline 35 items across seven sections, building on PRISMA-2020 to ensure transparency in searching, screening, and synthesizing studies with varied designs, including quasi-experimental and qualitative data.[65] Representative examples illustrate these applications. A Campbell systematic review and meta-analysis of guaranteed basic income interventions in high-income countries synthesized 27 studies, finding modest reductions in poverty-related outcomes like financial hardship and food insecurity, highlighting the role of such reviews in evaluating social safety nets. Similarly, in behavioral therapies for mental health, a Campbell review of Multisystemic Therapy examined 23 randomized trials, demonstrating moderate effects on reducing antisocial behaviors and improving family functioning among youth with social, emotional, and behavioral problems, underscoring the value of these reviews for community-based interventions.[66][67] Challenges in conducting systematic reviews in these fields arise from the heterogeneity of study designs, ranging from randomized trials to observational and qualitative research, complicating synthesis and risk-of-bias assessments. Integrating qualitative evidence, which captures nuanced social processes like participant experiences in educational programs, requires mixed-methods approaches to avoid losing contextual depth, as discussed in frameworks for combining diverse data types in reviews. Tools from the Evidence for Policy and Practice Information and Co-ordinating (EPPI-) Centre, such as EPPI-Reviewer software, facilitate this by supporting screening, coding, and thematic synthesis for education-focused reviews, enabling users to manage large datasets across review stages.[68][69] The use of systematic reviews in social and behavioral sciences has grown significantly since the 2010s, driven by demands for evidence-based policymaking. This expansion has informed policies on social interventions, such as anti-bullying programs in schools; for instance, a systematic review of 53 evaluations on school-based anti-bullying programs found that they reduced bullying perpetration by 20-23% and victimization by 17-20%, influencing guidelines from organizations like the U.S. Department of Education. These reviews align with reporting standards like those in broader dissemination guidelines to enhance accessibility for policymakers.[70]Environmental and Other Fields
Systematic reviews in environmental science play a crucial role in synthesizing evidence on complex ecological processes, such as the impacts of climate change on biodiversity. For instance, a systematic review examining climate change effects on biodiversity across multiple ecosystems identified shifts in species distributions and increased extinction risks, with analysis spanning genetic to biome levels.[71] These reviews often highlight how rising temperatures and altered precipitation patterns exacerbate habitat loss. The Collaboration for Environmental Evidence (CEE) provides specialized guidelines for conducting such reviews, emphasizing transparent protocols for searching heterogeneous environmental data sources, including gray literature from government reports and field observations.[72] CEE standards, updated in version 5.1 (2022), recommend pre-registering review protocols to minimize bias and ensure reproducibility, particularly when integrating qualitative evidence on ecosystem services.[73] A prominent example of environmental systematic reviews is the use of systematic maps to catalog evidence on conservation interventions. These maps identify and visualize the distribution of studies on topics like protected area effectiveness. For instance, a CEE-funded systematic map of nature conservation impacts on human well-being in developing countries collated over 1,000 articles (1,043 total), with over 25% examining linkages to economic well-being, including livelihood improvements, but highlighting sparse data on equity effects.[74] Unique to environmental applications, systematic reviews must address challenges in handling spatial data, such as integrating GIS-based analyses to model geographic range shifts, which requires standardized metadata for georeferenced studies to avoid aggregation biases.[75] Similarly, incorporating long-term studies—often spanning decades—poses issues like data incompleteness due to funding discontinuities, prompting CEE guidelines to advocate for sensitivity analyses on temporal trends.[76] Beyond ecology, systematic reviews have expanded into business, management, entrepreneurship, and engineering fields to evaluate policy effectiveness, management practices, and entrepreneurial phenomena. In business management, systematic reviews synthesize evidence on organizational strategies and contribute to theory development. Pioneering work by Tranfield, Denyer, and Smart (2003) adapted systematic review methods from health sciences to management, proposing a methodology to produce evidence-informed knowledge that accommodates heterogeneous study designs and emphasizes theoretical contributions.[77] More recently, Sauer and Seuring (2023) outlined a six-step process—defining the research question, determining characteristics of primary studies, retrieving literature, selecting pertinent studies, synthesizing the literature, and reporting results—with 14 associated decisions that allow for flexibility, iterative refinement, and focus on theory building (inductive, abductive, or deductive). These reviews commonly use databases such as Web of Science and Scopus, and may incorporate grey literature for emerging topics.[3] For example, a meta-analysis of 50 studies on agile practices demonstrated a 25% improvement in project success rates across industries. These reviews often employ content analysis, bibliometric techniques, or statistical synthesis to advance management theories and appraise diverse sources including case studies and surveys for replicability.[78] In entrepreneurship research, systematic reviews follow similar core steps but emphasize peer-reviewed journal articles from high-quality outlets, with quality appraisal often based on journal ranking systems such as VHB, ABS, or JCR. They are particularly valuable for addressing terminological inconsistencies and fragmentation in the field, synthesizing diverse perspectives to clarify concepts and support theory development. Kraus et al. (2020) proposed a three-stage approach—planning the review (including protocol development), conducting the review (identification, extraction, and synthesis), and reporting findings—with a focus on transparent methodologies, concept-centric analysis, and exclusion of grey literature to maintain rigor.[4] In engineering, systematic reviews assess intervention efficacy, for example, mapping evidence on sustainable infrastructure policies that reduced material waste by 15-30% in construction projects based on 40 empirical studies.[79] Policy-focused reviews in these areas often quantify effectiveness, such as evaluations of environmental regulations in business showing compliance costs offset by long-term efficiency gains in 60% of cases.[80] Emerging interdisciplinary areas, including One Health approaches that link environmental, animal, and human systems, have seen significant growth in systematic reviews since 2020, driven by global health crises. This surge underscores the value of systematic methods in cross-disciplinary policy, such as mapping interventions for antimicrobial resistance in agricultural settings. As of 2025, living systematic reviews have become more prominent in One Health to address evolving threats like zoonotic diseases.[81][82]Tools and Technologies
Software for Manual Processes
Software tools for managing manual processes in systematic reviews primarily support reference organization, screening, data extraction, and compliance with reporting standards, enabling researchers to handle large volumes of literature collaboratively and transparently. Reference management software like EndNote and Zotero plays a foundational role by allowing users to import, organize, and deduplicate citations from multiple databases, which is essential for preparing datasets for subsequent review stages. EndNote, a proprietary tool, facilitates the creation of customized libraries where references can be grouped by criteria such as study type or relevance, and its duplicate detection algorithm compares fields like title, author, year, and DOI to identify and merge overlaps, reducing manual effort in preprocessing search results.[83] Zotero, an open-source alternative, offers similar capabilities through its browser extension for seamless citation capture, folder-based organization, and automated deduplication, while also supporting PDF annotation and group libraries for team-based workflows in academic settings.[84] These tools ensure that initial literature imports—often exceeding thousands of records—are efficiently managed without data loss, as demonstrated in library protocols.[85] For the core manual tasks of screening and data extraction, specialized platforms such as Covidence and DistillerSR provide structured, collaborative interfaces tailored to systematic review protocols. Covidence enables teams to conduct title and abstract screening in parallel, with built-in conflict resolution tools and customizable forms for full-text assessment, streamlining the inclusion-exclusion decisions that form the basis of review eligibility criteria.[86] It supports data extraction by allowing reviewers to populate standardized templates for key study variables, such as participant demographics and outcomes, while tracking progress to minimize errors in multi-reviewer environments. DistillerSR complements this by offering configurable workflows for hierarchical screening—starting with titles and abstracts before advancing to full texts—and flexible forms for extracting both qualitative and quantitative data, accommodating projects with over 675,000 references per review.[87] Both tools emphasize human oversight, with DistillerSR's interface allowing supervisors to monitor reviewer assignments and ensure consistent application of inclusion criteria across distributed teams.[88] A 2022 evaluation highlights EPPI-Reviewer and Nested Knowledge as prominent options for comprehensive manual support, particularly in education and health sciences research.[89] EPPI-Reviewer, a web-based application, integrates reference import with Zotero for bulk handling, enables mobile-friendly screening and coding of studies against predefined frameworks, and includes audit trails via collaborative "wizards" that log all decisions and revisions for reproducibility.[69] Nested Knowledge similarly excels in dual-reviewer screening protocols, where teams tag and extract data using hierarchical structures to link concepts across studies, with traceable paths documenting each reference's journey from import to synthesis.[90] These platforms were ranked highly in comparative analyses for their density of manual features, such as customizable coding schemes and export options, outperforming generalist tools in handling complex, multi-stage reviews.[89] Compliance with reporting guidelines like PRISMA is embedded in these tools to facilitate transparent documentation and export of review outputs. DistillerSR generates PRISMA 2020 flow diagrams automatically from screening data, alongside reports on inter-rater reliability and full audit trails that capture 100% of changes to searches, references, and extractions, ensuring regulatory readiness in fields like pharmacovigilance.[88] Nested Knowledge produces PRISMA-compliant diagrams and supports exports to formats like Excel or RIS for integration with reporting software, while its real-time tracking maintains an immutable record of manual inputs.[90] EPPI-Reviewer aids compliance through Excel-compatible exports of screening results and synthesis summaries, aligning with Cochrane-endorsed standards for verifiable review processes.[69] Covidence similarly streamlines PRISMA adherence by visualizing screening flows and allowing direct export of extracted data, reducing the administrative burden of manual diagram creation.[86]Automation and AI Innovations
Automation in systematic reviews has increasingly incorporated machine learning (ML) techniques to streamline labor-intensive processes such as title and abstract screening. ASReview, an open-source tool employing active learning algorithms, enables reviewers to iteratively label records while the model predicts relevance for subsequent ones, thereby prioritizing potentially relevant studies.[91] This approach has demonstrated substantial workload reductions, with evaluations showing an average 83% decrease in screening time while identifying 95% of relevant records in biomedical reviews.[92] Large language models (LLMs) represent a more recent advancement, facilitating end-to-end automation of review workflows. For instance, otto-SR, introduced in 2025, leverages LLMs such as GPT-4o and o3-mini to handle tasks from literature searching and screening to data extraction and synthesis, completing workflows equivalent to 12 work-years of human effort in just two days for Cochrane-level reviews.[93] Similarly, Rayyan incorporates AI-driven prioritization features, using predictive algorithms to rank records by relevance during screening, which enhances efficiency in collaborative review environments.[41] Evidence PRIME's Laser AI tool further exemplifies this trend, applying natural language processing (NLP) to automate deduplication, screening, and quality assessment in living systematic reviews, minimizing human intervention while maintaining methodological rigor.[94] Integration of these AI innovations into established frameworks like Cochrane has accelerated since 2023, with guidelines now endorsing semi-automated tools for evidence synthesis to address rising review volumes. A 2025 analysis of 2,271 Cochrane, Campbell, and Environmental Evidence syntheses from 2017 to 2024 revealed a marked uptick in ML-based automation, particularly for screening and extraction, reflecting broader adoption in high-impact reviews.[95] NLP techniques have proven particularly effective for semi-automated data extraction, where models extract structured information from unstructured text with accuracies exceeding 80% for key variables like outcomes and interventions in clinical studies.[96] Despite these efficiencies, ethical considerations remain paramount, especially regarding transparency in AI decision-making to ensure reproducibility and mitigate biases. Studies emphasize the need for explainable AI models that disclose prediction rationales, as opaque "black box" processes could undermine trust in automated outputs for evidence-based policy.[97] A 2025 analysis indicates that AI tools are integrated into approximately 12% of evidence syntheses (via ML-enabled tools), with about 5% explicitly reporting ML use, driven by their ability to scale reviews amid growing literature volumes, though full automation is tempered by requirements for human oversight.[95]Limitations and Challenges
Methodological Biases and Risks
Publication bias, also known as the file-drawer problem, occurs when studies with statistically significant or positive results are more likely to be published than those with null or negative findings, leading to an overestimation of effect sizes in systematic reviews and meta-analyses.[98] This bias arises because researchers and journals preferentially disseminate favorable outcomes, leaving non-significant results unpublished and hidden in file drawers.[99] Selective reporting bias complements this issue by involving the selective inclusion or emphasis of certain outcomes within published studies, such as highlighting only statistically significant results while omitting others, which distorts the synthesized evidence.[100] These biases threaten the validity of systematic reviews by skewing the pooled estimates toward inflated effects.[101] Funnel plots serve as a primary visual tool for detecting publication bias, plotting study effect sizes against a measure of precision (such as standard error) to reveal asymmetry indicative of missing small studies with non-significant results.[102] In an unbiased meta-analysis, the plot resembles a symmetrical inverted funnel; asymmetry suggests bias.[103] At the review level, additional risks include outdated searches, as new evidence can emerge that alters conclusions, with update frequency depending on the topic's evolution and availability of recent studies rather than a fixed timeline, potentially rendering reviews misleading for decision-making.[104] Incomplete inclusion of grey literature—such as unpublished reports, conference proceedings, or government documents—further exacerbates this by underrepresenting non-commercial evidence, which often reports smaller or null effects.[105] To mitigate these biases, funnel plot asymmetry can be statistically tested using methods like Egger's regression test, which regresses the standardized effect estimate against its precision to detect significant deviation from symmetry (p < 0.05 indicating potential bias). The trim-and-fill method addresses identified bias by iteratively estimating and imputing "missing" studies on the less-populated side of the funnel plot, then recalculating the pooled effect to assess sensitivity.[106] For timeliness, living systematic reviews continuously update searches and incorporate new evidence as it emerges, reducing the risk of obsolescence in rapidly evolving fields.[107] Tools for assessing study-level biases, such as the RoB 2 tool, can complement these by evaluating individual trial risks during data extraction. Recent literature from 2024-2025 highlights emerging concerns with AI-induced biases in automated screening for systematic reviews, where machine learning models may perpetuate algorithmic unfairness—such as over- or under-retrieval of studies based on training data imbalances—potentially introducing new distortions in evidence selection.[108] For instance, stage-wise biases in AI pipelines can amplify disparities if not transparently reported, underscoring the need for human oversight and bias audits in hybrid workflows.[109]Resource and Reproducibility Issues
Conducting systematic reviews is highly resource-intensive, often requiring 12 to 24 months from initiation to completion, depending on the scope and complexity of the topic.[110] These timelines reflect the extensive stages involved, including protocol development, literature searching, screening, data extraction, and synthesis, which demand coordinated effort to maintain rigor. Typically, a team of 3 to 5 members is assembled for standard reviews, including subject experts, methodologists, librarians, and statisticians, though larger teams of up to 10 or more are common for comprehensive or multidisciplinary projects.[111] Financial costs can exceed $100,000 for large-scale reviews, encompassing personnel salaries, software licenses, database access fees, and publication expenses, with estimates ranging from $80,000 to $300,000 USD per review.[113] Reproducibility in systematic reviews is frequently compromised by poor adherence to established protocols and standards. Assessments using the AMSTAR 2 tool reveal widespread issues, with up to 90% of reviews rated as critically low quality due to deficiencies in protocol registration, search comprehensiveness, and risk of bias evaluation.[114] Additionally, access to proprietary or restricted data poses significant barriers, as unavailable datasets from industry-sponsored studies or paywalled sources hinder independent verification and replication of findings.[115] These challenges undermine the transparency essential for evidence-based decision-making, particularly in fields reliant on cumulative knowledge synthesis. Reporting gaps further exacerbate reproducibility concerns, with many systematic reviews featuring incomplete descriptions of methods, such as unclear inclusion criteria or unstated data extraction processes, and limited inclusion of human study data when applicable.[116] To address these, the NIRO-SR guidelines, introduced in 2023, provide a structured checklist for planning and conducting non-interventional, reproducible, and open systematic reviews, emphasizing pre-registration, open data sharing, and detailed methodological transparency.[117] Emerging solutions focus on mitigating these resource and reproducibility hurdles through open-source tools, enhanced training programs, and greater patient and public involvement. Open-source software, such as Rayyan for screening and R packages for meta-analysis, reduces costs and promotes shared methodologies accessible to diverse teams.[37] Specialized training initiatives, including workshops on protocol adherence and AMSTAR 2 application, build reviewer capacity and improve compliance rates.[118] Incorporating patient and public involvement from the protocol stage enhances review relevance, ensures ethical considerations, and fosters broader stakeholder buy-in for transparent dissemination.[119]History and Evolution
Early Origins
The origins of systematic reviews can be traced to early efforts in evidence synthesis within medicine, beginning in the 18th century. In 1747, Scottish naval surgeon James Lind conducted what is regarded as the first controlled clinical trial to evaluate treatments for scurvy, assigning 12 patients with the disease to six pairs receiving different interventions—such as cider, vinegar, or oranges and lemons—while controlling for diet and conditions.[120] His findings, published in 1753 in A Treatise of the Scurvy, included a critical and chronological review of prior literature on the disease, synthesizing historical accounts and proposed remedies to contextualize his trial results; this approach exemplified an early proto-systematic synthesis of evidence, though citrus treatments were not widely adopted until later.[121][122] In the early 20th century, statistical innovations laid groundwork for quantitative meta-analytic techniques, particularly in agriculture. During the 1920s and 1930s, Ronald A. Fisher, working at the Rothamsted Experimental Station, developed methods for analyzing data from multiple experiments, such as combining results on fertilizer effects across varying years and locations to assess consistency and variability.[123] In his 1935 book The Design of Experiments, Fisher illustrated how to integrate findings from several agricultural studies using techniques like the inverse sine transformation for proportions, promoting the idea of treating disparate datasets as part of a unified analysis to draw robust inferences.[123] The medical roots of systematic reviews emerged in the 1970s amid growing critiques of traditional narrative reviews, which often relied on selective or anecdotal evidence rather than rigorous evaluation. Archie Cochrane, a British epidemiologist, highlighted these shortcomings in his 1972 book Effectiveness and Efficiency: Random Reflections on Health Services, arguing that healthcare decisions should be based on systematic evaluations of randomized controlled trials to identify interventions that truly benefit patients.[124][125] Cochrane specifically called for compiling and analyzing all available randomized trial data on specific topics, such as the efficacy of treatments for conditions like lung cancer or back pain, to replace unreliable expert opinions with evidence-based assessments.[124] Initial frameworks for systematic reviews gained traction in the 1990s, revealing persistent delays in clinical adoption. A seminal 1992 study by Antman and colleagues examined cardiovascular therapies for myocardial infarction, finding that cumulative meta-analyses of randomized controlled trials demonstrated statistically significant mortality reductions for treatments like thrombolytics as early as 1973 (based on 10 trials with 2,544 patients), yet most expert recommendations and textbooks lagged by 10–13 years, continuing to endorse ineffective or harmful options like lidocaine.[126][127] This analysis underscored the need for systematic methods to accelerate the integration of trial evidence into practice, particularly in cardiology where narrative overviews had delayed uptake of proven interventions like beta-blockers and antiplatelet drugs.[127]Key Milestones and Modern Advances
The 1990s marked a pivotal era for systematic reviews, with the establishment of the Cochrane Collaboration in 1993, which aimed to organize production and maintenance of systematic reviews of healthcare interventions based on randomized controlled trials.[128] This initiative was followed by the publication of the first Cochrane Handbook in 1994, providing methodological guidance that standardized review processes and emphasized rigorous evidence synthesis.[129] Concurrently, meta-analyses gained prominence in leading medical journals, including the New England Journal of Medicine, where publications such as cumulative meta-analyses of therapeutic trials for myocardial infarction highlighted their utility in tracking evolving evidence and informing clinical strategies.[130][123] In the 2000s and 2010s, systematic reviews expanded beyond medicine into social sciences through the founding of the Campbell Collaboration in 2000, which focused on evidence synthesis for policy and practice in education, social welfare, and crime reduction.[131] A key methodological advancement came with the PRISMA statement in 2009, a 27-item checklist and flow diagram designed to enhance the transparency and completeness of reporting in systematic reviews and meta-analyses.[132] By the mid-2010s, the field had scaled significantly, with over 10,000 systematic reviews published annually, reflecting widespread adoption across disciplines and the growing volume of primary research.[133] The EQUATOR Network, launched in 2008, further supported global standardization by promoting reporting guidelines like PRISMA and fostering collaboration among researchers, journals, and funders to improve research reliability.[134] Recent developments from 2020 to 2025 have emphasized adaptability and technological integration. The PRISMA 2020 update refined the original guidelines to incorporate advances in search methods, risk-of-bias assessments, and synthesis approaches, addressing limitations in equity, certainty of evidence, and non-health research.[23] The COVID-19 pandemic accelerated the use of living systematic reviews, which continuously update with emerging evidence to guide rapid decision-making, as seen in initiatives like Cochrane's living reviews on treatments and vaccines.[31] AI integration advanced with tools like ASReview in 2019, an open-source machine learning framework that employs active learning to streamline title and abstract screening, reducing workload by up to 80% while maintaining accuracy.[135] By 2025, large language model (LLM) workflows have emerged as milestones in automation, enabling tasks such as automated data extraction and protocol generation, though human oversight remains essential for validity. For non-health fields, the NIRO-SR guidelines, published in 2023, provide a comprehensive checklist for non-interventional, reproducible, and open systematic reviews, filling gaps in protocol pre-registration and reporting outside clinical contexts.[136]References
- https://guides.lib.uw.edu/hsl/sr/[plan](/page/Plan)
