Anthony Bourdain’s documentary and the ethics of audio deepfakes – Quartz

Today, using machine learning to simulate a deceased person onscreen is an accepted Hollywood technique. Synthetic media, widely known as “deepfake” (a coat rack of “deep learning” and “fake”) Carrie Fisher in a Star Wars movie and Heath Ledger in “A Knight’s Tale”. In 2019, images of the actor Jimmy Fallon strangely transformed into Donald Trump demonstrated how advanced technology has become.

But no one was laughing when it was revealed that deepfake technology was used to simulate Anthony Bourdain’s voice in the new documentary. Roadrunner: a film about Anthony Bourdain. In an interview with GQ, director Morgan Neville, revealed he commissioned an AI model of the chef’s voice and TV personality and considered using it to narrate the entire movie. In the final cut, Neville tell the New Yorker that he used it for three lines in the two hour production. Among them is a poignant line from an email to artist David Choe, “My life is kind of shit now. You are successful, and I have succeeded, and I wonder: are you happy? “

Neville uses the AI-generated clip as an artistic touch to accentuate Choe’s pathos as he recounts the last email he received before Bourdain committed suicide in 2018. The audio is compelling, well a little flatter compared to the rest of Bourdain’s narration..

An emerging controversy over the making of Bourdain’s voice

Neville, a former reporter, didn’t see the problem of mixing AI-generated sound bytes with actual clips of Bourdain’s voice. “We can have a documentary ethics panel on this later,” he joked in the New Yorker interview.

Neville added that he obtained consent from Bourdain’s estate. “I checked, you know, with his widow and executor, just to make sure people were cool with that. And they were like, ‘Tony would have been cool with that.’ I didn’t put words in his mouth. I was just trying to bring them to life, ”he told GQ. Bourdain’s ex-wife Ottavia Busia-Bourdain, who appears extensively in the documentary, then disputed that she had never authorized an audio surrogate.

Meredith Broussard, professor of journalism at New York University and author of the book Artificial intelligence: how computers misunderstand the world, says it’s understandable that many find Bourdain’s audio clone deeply disturbing. “I’m not surprised his widow doesn’t feel like she gave her permission for this,” she said. “It’s such a new technology that no one really expects it to be used this way.”

The use of AI in journalism poses the biggest ethical dilemma, Broussard said. “People are more forgiving when we use this kind of technology in fiction as opposed to documentaries,” she explains. “In a documentary, people feel like it’s real and therefore they feel cheated.”

Simulated media and AI in journalism

Roadrunner, which is co-produced by CNN, is not the first instance where news outlets have relied on AI. Associated Press, for example, uses AI to automatically generate articles on quarterly results of companies since 2015. Each automatically generated AP article is accompanied by a note: “This story was generated by Automated Insights”, referring to the machine learning technology they use.

It is imperative to clearly disclose when AI is used, Broussard says. The fact that Neville discussed it after the fact is remarkable although it is questionable whether it satisfies his detractors. “It’s interesting that documentary filmmakers are going to have to think about the ethics of deepfakes,” Broussard says. “They’ve always thought about the ethics of storytelling, just like journalists, but here’s a whole new area where we’re going to have to develop ethical standards.”

The controversy rekindles a long-standing debate on how journalists quote their subjects. The famous writer Gay Talese, for example, reconstructs quotes as he remembers them, believing that the tape recorder is “the death knell of literary reporting”. In his book The journalist and the murderer, the New YorkerJanet Malcolm pointed out the problem of combining fragments of multiple interviews into one statement, as reporter Joe McGinniss did when covering the murder trial of former physician Jeffrey MacDonald. “The journalist can no more create his subjects than the analyst can create his patients. she wrote. The late writer herself has been embroiled in a decade-long legal battle over five quotes she used in a 1983 profile on Sigmund Freud’s archives. The libel case was dismissed in favor of Malcolm.

The ethics of deepfakes

Broussard is unsure of his position on Neville’s use of deepfake technology. “The problem with ethics is that it’s all about context,” she explains. “Three lines in a documentary film, it’s not the end of the world, but it’s important as a precedent. And it’s important to have a conversation about whether we think it’s the right thing to do.

Ultimately, Broussard says the Roadrunner controversy presents another argument for regulating the use of AI as a whole. “There is an emerging conversation in the field of machine learning about the need for ethics and machine learning,” she says, “I am grateful that this conversation has started because it is long overdue.”




Source link

About Ethel Nester

Ethel Nester

Check Also

The Best Podcasts in the Music Industry – Variety

With so many podcasts in the audiosphere and only a certain number of hours per …

Leave a Reply

Your email address will not be published. Required fields are marked *