FTC, CFPB, DOJ, And EEOC Pledge Increased Focus On Discrimination In AI

KD
Kelley Drye & Warren LLP
Contributor
Kelley Drye & Warren LLP is an AmLaw 200, Chambers ranked, full-service law firm of more than 350 attorneys and other professionals. For more than 180 years, Kelley Drye has provided legal counsel carefully connected to our client’s business strategies and has measured success by the real value we create.
Artificial intelligence and algorithmic processes continue to remain at the top of federal law enforcement agencies' agendas. Yesterday, the FTC, CFPB, DOJ, and EEOC issued a joint statement pledging to use...
United States Employment and HR
To print this article, all you need is to be registered or login on Mondaq.com.

Artificial intelligence and algorithmic processes continue to remain at the top of federal law enforcement agencies' agendas. Yesterday, the FTC, CFPB, DOJ, and EEOC issued a joint statement pledging to use their respective tools to "protect the public from bias in automated systems and artificial intelligence."

While these agencies' general commitment to monitoring AI processes is not new (for example, see here, here, and here for recently-published guidance related to the use of AI in various contexts), the joint statement shows they are now making concerted efforts to approach AI enforcement in a methodical and coordinated manner. The statement summarizes the agencies' existing legal authority and prior work relating to AI issues, along with three main areas of concern:

  • Bad data: Datasets used to train algorithms may be unrepresentative, imbalanced, biased, or contain other errors that could result in discriminatory outcomes. Similar outcomes could occur if automated systems end up correlating data with protected classes.
  • Lack of transparency: Many algorithmic models are "black boxes" whose internal workings may not be clear even to their developers. The lack of transparency makes it difficult to evaluate whether the systems are acting fairly.
  • Unanticipated uses: Automated systems designed with one purpose in mind may be appropriated for other uses. In such cases, the repurposed algorithm may produce improper results because the system's design is based on flawed assumptions about its users, relevant context, or underlying practices or procedures it might replace.

In addition to discrimination harms, the joint statement also points to other possible AI-related harms, such as companies overstating AI capabilities and using improperly collected data to train AI systems.

The joint statement doesn't necessarily break new ground, but it communicates a level of urgency, prioritization, and cross-agency collaboration that should not be overlooked. Companies using automated systems to make decisions that could affect consumers should carefully monitor outcomes for discriminatory impact and take efforts to control for biases in training datasets.

The content of this article is intended to provide a general guide to the subject matter. Specialist advice should be sought about your specific circumstances.

We operate a free-to-view policy, asking only that you register in order to read all of our content. Please login or register to view the rest of this article.

FTC, CFPB, DOJ, And EEOC Pledge Increased Focus On Discrimination In AI

United States Employment and HR
Contributor
Kelley Drye & Warren LLP is an AmLaw 200, Chambers ranked, full-service law firm of more than 350 attorneys and other professionals. For more than 180 years, Kelley Drye has provided legal counsel carefully connected to our client’s business strategies and has measured success by the real value we create.
See More Popular Content From

Mondaq uses cookies on this website. By using our website you agree to our use of cookies as set out in our Privacy Policy.

Learn More