Can machine-learning models overcome biased datasets?

Artificial intelligence systems may be able to complete tasks quickly, but that doesn’t mean they always do so fairly. If the datasets used to train machine-learning models contain biased data, it is likely the system could exhibit that same bias when it makes decisions in practice. For instance, if a dataset contains mostly images of white men, then a facial-recognition model trained with these data may be less accurate for women or people with different skin tones. A group of researchers at MIT, in collaboration with researchers at Harvard University…

This content is for Member members only.
Log In Register