Evaluating Implementation of the Transparency and Openness Promotion Guidelines: Reliability of Instruments to Assess Journal Policies, Procedures, and Practices

dc.contributor.authorKianersi S.
dc.contributor.authorGrant S.P.
dc.contributor.authorNaaman K.
dc.contributor.authorHenschel B.
dc.contributor.authorMellor D.
dc.contributor.authorApte S.
dc.contributor.authorDeyoe J.E.
dc.contributor.authorEze P.
dc.contributor.authorHuo C.
dc.contributor.authorLavender B.L.
dc.contributor.authorTaschanchai N.
dc.contributor.authorZhang X.
dc.contributor.authorMayo-Wilson E.
dc.contributor.otherMahidol University
dc.date.accessioned2023-05-19T07:56:08Z
dc.date.available2023-05-19T07:56:08Z
dc.date.issued2023-01-01
dc.description.abstractThe Transparency and Openness Promotion (TOP) Guidelines describe modular standards that journals can adopt to promote open science. The TOP Factor quantifies the extent to which journals adopt TOP in their policies, but there is no validated instrument to assess TOP implementation. Moreover, raters might assess the same policies differently. Instruments with objective questions are needed to assess TOP implementation reliably. In this study, we examined the interrater reliability and agreement of three new instruments for assessing TOP implementation in journal policies (instructions to authors), procedures (manuscript-submission systems), and practices (journal articles). Independent raters used these instruments to assess 339 journals from the behavioral, social, and health sciences. We calculated interrater agreement (IRA) and interrater reliability (IRR) for each of 10 TOP standards and for each question in our instruments (13 policy questions, 26 procedure questions, 14 practice questions). IRA was high for each standard in TOP; however, IRA might have been high by chance because most standards were not implemented by most journals. No standard had “excellent” IRR. Three standards had “good,” one had “moderate,” and six had “poor” IRR. Likewise, IRA was high for most instrument questions, and IRR was moderate or worse for 62%, 54%, and 43% of policy, procedure, and practice questions, respectively. Although results might be explained by limitations in our process, instruments, and team, we are unaware of better methods for assessing TOP implementation. Clarifying distinctions among different levels of implementation for each TOP standard might improve its implementation and assessment (study protocol: https://doi.org/10.1186/s41073-021-00112-8).
dc.identifier.citationAdvances in Methods and Practices in Psychological Science Vol.6 No.1 (2023)
dc.identifier.doi10.1177/25152459221149735
dc.identifier.eissn25152467
dc.identifier.issn25152459
dc.identifier.scopus2-s2.0-85152087550
dc.identifier.urihttps://repository.li.mahidol.ac.th/handle/20.500.14594/82264
dc.rights.holderSCOPUS
dc.subjectPsychology
dc.titleEvaluating Implementation of the Transparency and Openness Promotion Guidelines: Reliability of Instruments to Assess Journal Policies, Procedures, and Practices
dc.typeArticle
mu.datasource.scopushttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85152087550&origin=inward
oaire.citation.issue1
oaire.citation.titleAdvances in Methods and Practices in Psychological Science
oaire.citation.volume6
oairecerif.author.affiliationIndiana University School of Education
oairecerif.author.affiliationThe University of North Carolina at Chapel Hill
oairecerif.author.affiliationIndiana University-Purdue University Indianapolis
oairecerif.author.affiliationIndiana University Bloomington
oairecerif.author.affiliationBrigham and Women's Hospital
oairecerif.author.affiliationFaculty of Medicine Ramathibodi Hospital, Mahidol University
oairecerif.author.affiliationUniversity of Oregon
oairecerif.author.affiliationIndiana University School of Informatics and Computing
oairecerif.author.affiliationCenter for Open Science

Files

Collections