Machine learning models increasingly make high-stakes decisions affecting people’s lives—determining who gets hired, who receives loans, who qualifies for parole, and who gets access to healthcare. Yet these systems can perpetuate and amplify societal biases, systematically disadvantaging certain groups whilst privileging others. The consequences are serious. Biased hiring algorithms exclude qualified candidates based on protected…