바로가기 메뉴 본문 바로가기 주메뉴 바로가기
  • 06-3aHave you clearly established the data labeling standards and provided them to labelers?
    • Potential biases in healthcare AI can occur due to humans’ bias input and labeling.
    ✓ Bias that the incidence rate of diseases like cancer is higher in women than men, in other words, fixed ideas about gender, can lead to overdiagnosis for women.
    ✓ Several studies [28, 29] demonstrated the existence of a racial bias that led to the underdiagnosis of the cause of pain in Blacks as opposed to Whites. This diagnostic data is transmitted to the labeler as-is, resulting in bias.

    • Such biases occur due to unclear guidelines on data labeling which leads to dependency on individual judgment. To prevent this, it is necessary to pay close attention to establishing labeling standards in close collaboration with various clinical staff, as well as establishing standardized work standards by preparing detailed labeling guidelines. The procedure outlined below is an example of how to develop a guide and training program for labelers.
    ✓ Decision-making process mapping: Work closely with clinicians to establish standards and guidelines for data labeling.
    ✓ Utilize the proper labeling tools: Such tools help efficiently manage a large quantity of diverse data, including CT and MRI images, voice, and electrocardiogram data, and eliminate subjectivity in labeling by providing the same working environment for all labelers.
    ✓Create work guide and provide training: Crowdworkers are trained on the characteristics of the healthcare sector, diseases, and labeling targets, and healthcare professionals, such as specialists and nurses, are given work manuals and training that include clear explanations of tools and procedures.