In a recent study conducted by the US government, they have found that when it comes to facial recognition systems, they tend to have biases against people of color and gender, where they misidentify people of color more so than other Caucasian people.
According to the report, “For one-to-one matching, the team saw higher rates of false positives for Asian and African American faces relative to images of Caucasians. The differentials often ranged from a factor of 10 to 100 times, depending on the individual algorithm.” For example, the study found that Microsoft’s facial recognition tech has nearly 10 times more false positives for women of color than men of color.
Microsoft has since said that they are reviewing the report, and hopefully make tweaks to address it. This is not the first time that tech has been accused of being biased. For example, not too long ago, the algorithm behind the Apple Card was found to be offering men a higher credit limit compared to women.