Skip to main content

Evaluating Your Own SLR Process

This section is distinct from Appraise Study Quality, which assesses the rigor of the primary studies you have included. This section asks a different question: how rigorously did you conduct the review itself?

Overview

Quality appraisal tools such as CASP, MMAT, and JBI look outward: they help you evaluate the studies in your dataset. The checklist and scoring rubric on this page look inward: they help you evaluate your own review process against recognized best practices.

The items below are adapted from Petersen, Vakkalanka, and Kuzniarz (2015), who derived them from a systematic mapping study of how SLRs and systematic mapping studies are conducted in practice. Use this rubric in two ways:

  • During planning: as a checklist of actions to build into your protocol.
  • Before submission: as a retrospective audit to identify gaps in your process and to disclose them transparently in your methods chapter.

This rubric was developed in the context of software engineering research. The core dimensions; motivating the review, search strategy, search evaluation, extraction/classification, and validity; apply equally to business and management SLRs. Items that refer to software-engineering-specific classification schemes may be skipped if they are not relevant to your discipline.


Part 1: Activities Checklist

The table below lists the 26 actions identified by Petersen et al. (2015) as relevant to a rigorous systematic review or mapping study. Work through each row and mark whether the action was taken (✓), partially taken (~), or not taken (✗). This produces a ratio score: count your ✓ marks and divide by 26 (or by the number of applicable items).

PhaseActionTaken?
Motivate the reviewMotivate the need and relevance of the review 
 Define objectives and research questions 
 Consult with the target audience (e.g., supervisor, domain expert) to refine questions 
Search strategyConduct a database search 
 Apply snowball sampling (backward and/or forward) 
 Conduct a manual search of key journals or conference proceedings 
Develop the searchUse a structured framework (PICO, SPIDER, or PCC) to derive keywords 
 Consult a librarian or domain expert during search design 
 Iteratively refine the search string to improve coverage 
 Derive additional keywords from known relevant papers 
 Use thesauri, encyclopedias, or controlled vocabularies (e.g., MeSH, EBSCO subject headings) 
Evaluate the searchTest the search against a set of known-relevant papers 
 Have an expert evaluate the search results 
 Check the web pages or profiles of key authors in the field 
 Conduct a test–retest to check consistency 
Inclusion and exclusionDefine objective, pre-specified criteria for inclusion and exclusion 
 Involve a second reviewer; resolve disagreements systematically 
 Define and apply explicit decision rules for borderline cases 
Data extractionDefine objective criteria for the extraction process 
 Blind or obscure information that could bias extraction 
 Involve a second reviewer; resolve disagreements in extraction 
 Conduct test–retest of extraction on a subset 
ClassificationClassify studies by research type (e.g., empirical, conceptual, review) 
 Classify studies by research method (e.g., case study, survey, experiment) 
 Classify studies by venue type (e.g., journal, conference, practitioner publication) 
ValidityDiscuss validity threats and limitations of the review process 

Part 2: Scoring Rubrics

After completing the checklist, use the rubrics below to assign a score to each of the five key dimensions. Record these scores in your methods chapter alongside a brief narrative.

Rubric 1 — Motivating the Review

ScoreLabelDescription
0Not describedThe review is not motivated and no objectives are stated
1PartialMotivations and research questions are provided
2FullMotivations and questions are provided and have been developed in dialogue with the target audience (supervisor, practitioners, or domain experts)

Rubric 2 — Search Strategy

ScoreLabelDescription
0Not describedOnly one type of search was conducted
1MinimalTwo search strategies were used
2FullAll three strategies were used: database search, snowball sampling, and manual search

Rubric 3 — Evaluating the Search

ScoreLabelDescription
0Not describedNo actions were taken to improve the reliability of the search or inclusion/exclusion process
1MinimalAt least one action was taken to improve either the reliability of the search or the inclusion/exclusion process
2PartialAt least one action was taken to improve both the search and the inclusion/exclusion process
3FullAll identified actions were taken

Rubric 4 — Extraction and Classification

ScoreLabelDescription
0Not describedNo actions were taken to improve extraction reliability or enable comparability between studies
1MinimalAt least one action was taken to increase extraction reliability
2PartialAt least one action to increase extraction reliability and studies were classified by research type and method
3FullAll identified actions were taken

Rubric 5 — Study Validity

ScoreLabelDescription
0Not describedNo threats or limitations are described
1FullThreats and limitations of the review process are described

Interpreting Your Scores

No minimum threshold is formally established in the literature for general SLRs; the rubric is a diagnostic tool, not a pass/fail gate. Use the results as follows:

  • In your methods chapter: Report your scores and briefly explain any dimension rated 0 or 1. A low score on a dimension is not automatically a fatal flaw, but it must be acknowledged as a limitation.
  • In your discussion: Dimensions scored 0 (especially search strategy and validity) should be discussed explicitly when qualifying the strength of your conclusions.
  • As a planning aid: If you are still in the protocol stage, any action not yet checked is a concrete item to build into your plan before searching begins.

For more detail on designing and evaluating your search strategy, see the Search Quality Self-Assessment Checklist (adapted from vom Brocke et al., 2015), which provides granular guidance on the search phase specifically.