Abstract.
Our objective was to evaluate interobserver agreement and to compare the performance score in quality control of screen-film mammography and computed radiography (CR) using a breast phantom. Eleven radiologists interpreted a breast phantom image (CIRS model X) by four viewing methods: (a) original screen-film; (b) soft-copy reading of the digitized film image; (c) hard-copy reading of CR using an imaging plate; and (d) soft-copy reading of CR. For the soft-copy reading, a 17-in. CRT monitor (1024×1536×8 bits) was used. The phantom image was evaluated using a scoring system outlined in the instruction manual, and observers judged each object using a three-point rating scale: (a) clearly seen; (b) barely seen; and (c) not seen. For statistical analysis, the kappa statistic was employed. For "mass" depiction, interobserver agreement using CR was significantly lower than when using screen-film (p<0.05). There was no significant difference in the kappa value for detecting "microcalcification"; however, the performance score of "microcalcification" on CR hard-copy was significantly lower than on the other three viewing methods (p<0.05). Viewing methods (film or CR, soft-copy or hard-copy) could affect how the phantom image is judged. Paying special attention to viewing conditions is recommended for quality control of CR mammograms.
Similar content being viewed by others
Author information
Authors and Affiliations
Additional information
Electronic Publication
Rights and permissions
About this article
Cite this article
Shimamoto, K., Ikeda, M., Satake, H. et al. Interobserver agreement and performance score comparison in quality control using a breast phantom: screen-film mammography vs computed radiography. Eur Radiol 12, 2192–2197 (2002). https://doi.org/10.1007/s00330-002-1357-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00330-002-1357-7