We don’t think of everyday devices as biased against race or gender, but they can be. Electrical engineer and computer scienti

admin2022-02-15  44

问题   We don’t think of everyday devices as biased against race or gender, but they can be. Electrical engineer and computer scientist Achuta Kadambi is familiar with the problem both professionally and personally. "Being fairly dark-skinned myself," Kadambi says, he sometimes cannot activate no-touch faucets that detect light bouncing off skin. At one airport, he recalls, "I had to ask a lighter-skinned traveler to trigger a faucet for me".
  Medical devices, too, can be biased. In a recent article, Kadambi describes three ways that racial and gender bias can permeate medical devices and suggests a number of solutions.
  The first problem, Kadambi says, is physical bias, which is inherent in the mechanics of the device. Then there is computational bias, which lies in the software or in the data sets used to develop the gadget. Finally, there is interpretation bias, which resides not in the machine but in its user. It occurs when clinicians apply unequal, race-based standards to the readouts from medical devices and tests—an alarmingly common practice.
  Physical bias made news last December when a study at the University of Michigan found that pulse oximeters—which use light transmitted through skin and tissue to measure the oxygen in a person’s blood—are three times more likely to miss low oxygen levels in black patients than in white ones.
  Computational biases can creep into medical technology when it is tested primarily on a homogeneous group of subjects—typically white males. For instance, an artificial-intelligence system used to analyze chest x-rays and identify 14 different lung and chest diseases worked less well for women when trained on largely male scans. But training the system on a gender-balanced sample produced the best overall results, with no significant loss of accuracy for men.
  Stopping computational bias means making a much greater effort to recruit people from different populations to participate in the design and testing of medical devices. In addition to building diversity among researchers, Rachel Hardeman, a public health scientist at the University of Minnesota favors mandatory training of medical personnel, a step that might also help counter practices that lead to interpretation bias. California has moved in this direction, she notes, with a 2020 law requiring health-care providers treating pregnant women and their newborns to complete a curriculum aimed at closing racial gaps in maternal and infant mortality.
  Fairness, Kadambi argues, should be a criterion for evaluating new technology, along with effectiveness.
It can be inferred from Paragraph 6 that_________.

选项 A、there is a lack of capable researchers at universities
B、Bachel Hardeman takes charge of the medical training in California
C、California has taken measures to eliminate interpretation bias
D、the pregnant women have to take a mandatory training

答案C

解析 推断题。根据题干可定位至第六段。选项A为无中生有,故排除;选项B为无中生有,文中仅提到雷切尔.哈德曼支持对医务人员进行培训l,故排除;根据选项D可定位至第三句,由requiring health-care providers treating pregnant women and their newborns to complete a curriculum可知是医疗服务提供者参加课程,偷换概念,故排除。由第二句可知,进行强制性培训有助于遏制解释偏差,而加州已经朝这个方向前进了。可以推断出加州已经采取措施去消除解释偏差.故选项C正确。
转载请注明原文地址:https://jikaoti.com/ti/ykg7FFFM
0

最新回复(0)