Repeatability of manual coding of cancer reports in the South African National Cancer Registry, 2010

dc.contributor.authorDube, Nomathemba Michell
dc.contributor.authorGirdler-Brown, B.V. (Brendan)
dc.contributor.authorTint, Khin-San
dc.contributor.authorKellett, Patricia
dc.date.accessioned2014-04-22T12:21:08Z
dc.date.available2014-04-22T12:21:08Z
dc.date.issued2013
dc.descriptionThis study was carried out by Nomathemba Dube in partial fulfilment of the requirements for her Master’s degree in Public Health in the School of Health Systems and Public Health at the University of Pretoria.en_US
dc.description.abstractData validity is a very important aspect of cancer registries in ensuring data quality for research and interventions. This study focused on evaluating the repeatability of manual coding of cancer reports in the South African National Cancer Registry (NCR). This cross-sectional study used the Delphi technique to classify 48 generic tumour sites into sites that would be most likely (“difficult”) and least likely (“not difficult”) to give rise to discordant results among coders. Reports received from the Charlotte Maxeke Academic Hospital were manually recoded by five coders (2 301 reports, e.g. approximately 400 reports each) for intracoder agreement; and by four coders (400 reports) for inter-coder agreement. Unweighted kappa statistics were calculated and interpreted using Byrts’ criteria. After four rounds of the Delphi technique, consensus was reached on the classification of 91.7% (44/48) of the sites. The remaining four sites were classified according to modal expert opinion. The overall kappa was higher for intra-coder agreement (0.92) than for inter-coder agreement (0.89). “Not difficult” tumour sites reflected better agreement than “difficult” tumour sites. Ten sites (skin other, basal cell carcinoma of the skin, connective tissue, other specified, lung, colorectal, prostate, oesophagus, naso-oropharynx and primary site unknown) were among the top 80% misclassified sites. The repeatability of manual coding at the NCR was rated as “good” according to Byrts’ criteria. Misclassified sites should be prioritised for coder training and the strengthening of the quality assurance system.en_US
dc.description.librarianam2014en_US
dc.description.librarianay2014
dc.description.sponsorshipThe South African Field Epidemiology and Laboratory Training Programme (SAFELTP), funded by the Centers for Disease Control and Prevention (CDC)en_US
dc.description.urihttp://www.sajei.co.za/index.php/SAJEIen_US
dc.identifier.citationDube, N, Girdler-Brown, B, Tint, K & Kellett, P 2013, 'Repeatability of manual coding of cancer reports in the South African National Cancer Registry, 2010', Southern African Journal of Epidemiology and Infection, vol. 28, no. 3, pp. 157-165.en_US
dc.identifier.issn1015-8782 (print)
dc.identifier.issn2220-1084 (online)
dc.identifier.urihttp://hdl.handle.net/2263/39678
dc.language.isoenen_US
dc.publisherMedPharm Publicationsen_US
dc.rights© SAJEIen_US
dc.subjectCancer reportsen_US
dc.subjectManual codingen_US
dc.subjectRepeatabilityen_US
dc.subjectKappa scoreen_US
dc.subjectSouth African National Cancer Registry (NCR)en_US
dc.subject.lcshCancer -- Research -- South Africaen
dc.subject.lcshDelphi methoden
dc.titleRepeatability of manual coding of cancer reports in the South African National Cancer Registry, 2010en_US
dc.typeArticleen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Dube_Repeatability_2013.pdf
Size:
198.55 KB
Format:
Adobe Portable Document Format
Description:
Article

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: