
A team of researchers at Stanford and University of Massachusetts Amherst published a paper in Science outlining a new technique that translates a goal, such as to avoid gender or racial bias, into mathematical criteria to allow a machine-learning algorithm to train an AI application to avoid that behavior. “We want to advance AI that respects the values of its human users and justifies the trust we place in autonomous systems,” said Emma Brunskill, an assistant professor of computer science at Stanford and senior author of the paper.