바로가기 메뉴 본문 바로가기 주메뉴 바로가기
  • 06-2Have you analyzed features used in training and prepared selection criteria?
    Determine applicability: Consider this question if you are developing your own AI algorithm or model, or if you are using specific sensitive variables in addition, and determine if the requirement has been satisfied.

    • Because the collection process of medical data involves high costs and complex procedures, data on races or countries with relatively high profits were first constructed. Bias in the collected data can lead to low generalization performance and trustworthiness of AI systems trained using these data. Therefore, a meticulous recognition of the features used for training is necessary when collecting and processing medical data, and measures to remove potential biases must be implemented in an effective manner.

    • Bias can occur depending on where and how medical data were collected. Selection bias may occur if selectively collecting and acquiring data that can support the algorithm’s conclusion to correspond with the purpose of the model, whereas automation bias occurs when automatically collecting data using an algorithm or AI without a careful review. It is ideal to analyze features for training and establish the selection criteria to identify discriminatory elements from data in advance, as this is important in mitigating bias.

    • Discriminatory elements are characteristics that can cause social criticism and discrimination. Examples of sensitive attributes that international agencies or global companies mention are as follows. These are elements that have been selected as attributes that should not be reflected in data training to prevent bias.