Deepfakes: GenAI making phoney and real evidence harder to discern, says Maura Grossman

The organization C2PA is working on creating standards for marking AI-generated material

Deepfakes: GenAI making phoney and real evidence harder to discern, says Maura Grossman
Maura Grossman

While much of the legal profession has been focused on generative AI’s (GenAI) potential to hallucinate fictitious caselaw, Maura Grossman says she is more absorbed in the fact that we are entering a world in which it will be increasingly difficult to distinguish between genuine and AI-generated evidence.

The risk of deepfake evidence comes from the rising quality of AI-generated text, photos, video, and audio produced by free, open-source GenAI tools.

“That's going to be a really significant challenge for the court system,” she says.

Grossman is a research professor at the University of Waterloo’s School of Computer Science and an adjunct professor at Osgoode Hall Law School. She is also an affiliate faculty member of the Vector Institute and an eDiscovery attorney and consultant in Buffalo, New York.

Grossman will appear at the 2024 Canadian Lawyer LegalTech Summit, on the panel, “Best practices for addressing ethical and other concerns with Gen AI and Deepfakes.” The summit will take place June 12 at the Arcadian in Toronto.

Grossman began her career working at a law firm in New York before developing an interest in digital evidence – how to find the needle-in-the-haystack document in a hard drive full of them. That led her into the fields of machine learning and AI, and she left her legal practice to collaborate with a professor at Waterloo’s School of Computer Science, where she researched how machine-learning technologies could learn to find evidence better than lawyers.

Through her work, her interest in responsible and trustworthy AI grew. She now teaches computer scientists at Waterloo about the legal, ethical, and policy considerations relevant to their work. At Osgoode, she teaches law students about technology and its implications for legal rights, discrimination, and surveillance.

If courts become inundated with GenAI-produced evidence, that could mean that every trial will require a forensic expert to distinguish the real from the fake, she says, which will drive up litigation costs.

Typically, when a client gives their lawyer a piece of evidence – an audio recording of the client’s spouse verbally abusing them for a family law case, for example – there is no requirement for the lawyer to authenticate it.

“But I'm starting to be concerned that as we move forward, there's going to be more of an onus on lawyers to confirm that evidence given by their own client is likely to be correct, accurate, or true,” says Grossman. “That's very challenging.”

Grossman expects that there will be markers attached to identify AI-generated content that will enable determination of when and where the content originated.

“But it's going to take some years before we have a set of standards for every device to do this,” she says. In the meantime, it is possible to remove watermarks from AI-generated photos, for example, or to add them to genuine ones.

Grossman says organizations such as the Coalition for Content Provenance and Authenticity (C2PA) are working on creating a standard for AI-generated material that would allow for its simple identification.

“Eventually, we'll get there,” she says.

When it comes to AI-powered legal tech, Grossman says it is essential for lawyers and law students to become familiar with the tools because they will be prolific for evidence and in many other facets of the practice.  

“It behooves lawyers to get educated in this area and to make some decisions about what fruitfully can be used in practice and what is not ready for primetime yet. That includes learning how to vet the tools and how to evaluate them.”

“It's a buyer beware sort of atmosphere right now,” says Grossman. “But it's also very exciting. There’s going to be big changes, I think, in the way that law is practiced.”

Recent articles & video

Attorney General Arif Virani on how he works to expedite federal judicial nominations

Fasken M&A report ‘cautiously optimistic’ for market rebound in 2024

Voting is now open for Canadian Lawyer’s Top 25 Most Influential Lawyers

BC Court of Appeal upholds ‘competence-competence’ principle in Google and Apple lawsuit

Federal Court of Appeal retains plaintiffs’ counsel in a class action against tech firms

Naheed Bardai, Rochelle Wempe, Michael Morris appointed to Saskatchewan courts

Most Read Articles

BC Supreme Court orders father to pay fines for continuous breaches of conduct and parenting orders

BC Supreme Court revokes probate grant for failure to properly notify testator’s son in Mexico

NS Supreme Court clarifies disclosure standards in a divorce and property division case

Canadian lawyers need to replace resilience with real change