"We know of too many real-world examples of AI systems that have flawed or biased algorithms: automated processes used in hospitals that understate the health needs of Black patients; recruiting and hiring tools that discriminate against women and minority candidates; facial recognition systems with higher error rates among people with darker skin; and more. The Algorithmic Accountability Act would require that automated systems be assessed for biases, hold bad actors accountable, and ultimately help to create a safer AI future."