Baethge, Christopher, Goldbeck-Wood, Sandra and Mertens, Stephan (2019). SANRA-a scale for the quality assessment of narrative review articles. Res. Integr. Peer Rev., 4 (1). LONDON: BMC. ISSN 2058-8615

Full text not available from this repository.

Abstract

BackgroundNarrative reviews are the commonest type of articles in the medical literature. However, unlike systematic reviews and randomized controlled trials (RCT) articles, for which formal instruments exist to evaluate quality, there is currently no instrument available to assess the quality of narrative reviews. In response to this gap, we developed SANRA, the Scale for the Assessment of Narrative Review Articles.MethodsA team of three experienced journal editors modified or deleted items in an earlier SANRA version based on face validity, item-total correlations, and reliability scores from previous tests. We deleted an item which addressed a manuscript's writing and accessibility due to poor inter-rater reliability. The six items which form the revised scale are rated from 0 (low standard) to 2 (high standard) and cover the following topics: explanation of (1) the importance and (2) the aims of the review, (3) literature search and (4) referencing and presentation of (5) evidence level and (6) relevant endpoint data. For all items, we developed anchor definitions and examples to guide users in filling out the form. The revised scale was tested by the same editors (blinded to each other's ratings) in a group of 30 consecutive non-systematic review manuscripts submitted to a general medical journal.ResultsRaters confirmed that completing the scale is feasible in everyday editorial work. The mean sum score across all 30 manuscripts was 6.0 out of 12 possible points (SD 2.6, range 1-12). Corrected item-total correlations ranged from 0.33 (item 3) to 0.58 (item 6), and Cronbach's alpha was 0.68 (internal consistency). The intra-class correlation coefficient (average measure) was 0.77 [95% CI 0.57, 0.88] (inter-rater reliability). Raters often disagreed on items 1 and 4.ConclusionsSANRA's feasibility, inter-rater reliability, homogeneity of items, and internal consistency are sufficient for a scale of six items. Further field testing, particularly of validity, is desirable. We recommend rater training based on the explanations and instructions document provided with SANRA. In editorial decision-making, SANRA may complement journal-specific evaluation of manuscripts-pertaining to, e.g., audience, originality or difficulty-and may contribute to improving the standard of non-systematic reviews.

Item Type: Journal Article
Creators:
CreatorsEmailORCIDORCID Put Code
Baethge, ChristopherUNSPECIFIEDUNSPECIFIEDUNSPECIFIED
Goldbeck-Wood, SandraUNSPECIFIEDUNSPECIFIEDUNSPECIFIED
Mertens, StephanUNSPECIFIEDUNSPECIFIEDUNSPECIFIED
URN: urn:nbn:de:hbz:38-153402
DOI: 10.1186/s41073-019-0064-8
Journal or Publication Title: Res. Integr. Peer Rev.
Volume: 4
Number: 1
Date: 2019
Publisher: BMC
Place of Publication: LONDON
ISSN: 2058-8615
Language: English
Faculty: Unspecified
Divisions: Unspecified
Subjects: no entry
Uncontrolled Keywords:
KeywordsLanguage
Ethics; History & Philosophy Of ScienceMultiple languages
Refereed: Yes
URI: http://kups.ub.uni-koeln.de/id/eprint/15340

Downloads

Downloads per month over past year

Altmetric

Export

Actions (login required)

View Item View Item