An open competition involving thousands of competitors failed to construct useful abstract classifiers for new diagnostic test accuracy systematic reviews

Yuki Kataoka, Shunsuke Taito, Norio Yamamoto, Ryuhei So, Yusuke Tsutsumi, Keisuke Anan, Masahiro Banno, Yasushi Tsujimoto, Yoshitaka Wada, Shintaro Sagami, Hiraku Tsujimoto, Takashi Nihashi, Motoki Takeuchi, Teruhiko Terasawa, Masahiro Iguchi, Junji Kumasawa, Takumi Ichikawa, Ryuki Furukawa, Jun Yamabe, Toshi A. Furukawa

研究成果: ジャーナルへの寄稿学術論文査読

抄録

There are currently no abstract classifiers, which can be used for new diagnostic test accuracy (DTA) systematic reviews to select primary DTA study abstracts from database searches. Our goal was to develop machine-learning-based abstract classifiers for new DTA systematic reviews through an open competition. We prepared a dataset of abstracts obtained through database searches from 11 reviews in different clinical areas. As the reference standard, we used the abstract lists that required manual full-text review. We randomly splitted the datasets into a train set, a public test set, and a private test set. Competition participants used the training set to develop classifiers and validated their classifiers using the public test set. The classifiers were refined based on the performance of the public test set. They could submit as many times as they wanted during the competition. Finally, we used the private test set to rank the submitted classifiers. To reduce false exclusions, we used the Fbeta measure with a beta set to seven for evaluating classifiers. After the competition, we conducted the external validation using a dataset from a cardiology DTA review. We received 13,774 submissions from 1429 teams or persons over 4 months. The top-honored classifier achieved a Fbeta score of 0.4036 and a recall of 0.2352 in the external validation. In conclusion, we were unable to develop an abstract classifier with sufficient recall for immediate application to new DTA systematic reviews. Further studies are needed to update and validate classifiers with datasets from other clinical areas.

本文言語英語
ページ(範囲)707-717
ページ数11
ジャーナルResearch synthesis methods
14
5
DOI
出版ステータス出版済み - 09-2023

All Science Journal Classification (ASJC) codes

  • 教育

フィンガープリント

「An open competition involving thousands of competitors failed to construct useful abstract classifiers for new diagnostic test accuracy systematic reviews」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル