The term “artificial intelligence” covers a wide range of software or computers mimicking human cognition. But as the limits of AI get pushed, it seems like machines are doing some things that human could never do. And, because both criminals and cops are the ones testing the boundaries of AI, the legal limits of its use will become a hot topic.

Here’s a look at two recent cases in both crime and crimefighting where AI and similar technology could play a crucial role.

Can You Hear Me Now?

You may have heard about photoshopping one person’s head on another body. You may have even seen deepfakes, convincing faceswaps that use machine learning to create everything from fake celebrity pornography to viral videos mocking politicians. Now, criminals are using the same tech to mimic voices to rip off hundreds of thousands of dollars.

An unnamed company was tricked into transferring $243,000 into a secret account based on synthetic audio of the CEO’s voice. As Vice reported:

The thieves tried again, but the subordinate was suspicious this time, calling his boss. But the successful theft, along with earlier deepfake incidents, have individuals, businesses, and law enforcement wondering what’s next.

Can You See Me?

So, if the criminals are using facial technology, can the cops use it as well? Courts and communities are still trying to figure that out. While some cities, like San Francisco, have banned facial recognition technology, others, like Huntington Park just a few hours’ drive south, have “autonomous data machines” roaming the streets, collecting data on people, cars, and cellphones.

Courts have ruled that while law enforcement officers can follow you, they can’t put a GPS tracker on your car. And while courts have yet to address robotic collection of nearby cellphone data, the Supreme Court recently ruled that police need a warrant to access your phone’s location data from cell towers. So, the line between the policing that must be done by humans, and what can be done by artificial intelligence doesn’t seem exactly clear.

  • How Facial Recognition Is Used by Law Enforcement (FindLaw’s Technologist)
  • Robots Are Doing Pretrial Risk Assessments, and They’re Not Great at It (FindLaw Blotter)
  • Cops Track Suspects Using Spotify and Netflix Accounts (FindLaw Blotter)
  • Who Regulates How Police Use New Tech? (FindLaw Blotter)

You Don’t Have To Solve This on Your Own – Get a Lawyer’s Help

Civil Rights

Block on Trump’s Asylum Ban Upheld by Supreme Court

Criminal

Judges Can Release Secret Grand Jury Records

Politicians Can’t Block Voters on Facebook, Court Rules