Artwork

Innhold levert av Max Bodach and Foundation for American Innovation. Alt podcastinnhold, inkludert episoder, grafikk og podcastbeskrivelser, lastes opp og leveres direkte av Max Bodach and Foundation for American Innovation eller deres podcastplattformpartner. Hvis du tror at noen bruker det opphavsrettsbeskyttede verket ditt uten din tillatelse, kan du følge prosessen skissert her https://no.player.fm/legal.
Player FM - Podcast-app
Gå frakoblet med Player FM -appen!

What Should Be Done About Misinformation? w/Renée DiResta

1:14:06
 
Del
 

Manage episode 435597157 series 3530279
Innhold levert av Max Bodach and Foundation for American Innovation. Alt podcastinnhold, inkludert episoder, grafikk og podcastbeskrivelser, lastes opp og leveres direkte av Max Bodach and Foundation for American Innovation eller deres podcastplattformpartner. Hvis du tror at noen bruker det opphavsrettsbeskyttede verket ditt uten din tillatelse, kan du følge prosessen skissert her https://no.player.fm/legal.

The recent riots in the United Kingdom raise new questions about online free speech and misinformation. Following the murder of three children in Southport, England, false rumors spread across social media about the killer’s identity and religion, igniting simmering resentment over the British government’s handling of immigration in recent years. X, formerly Twitter, has come under fire for allowing the rumors to spread, and the company’s owner Elon Musk has publicly sparred with British politicians and European Union regulators over the issue.

The incident is the latest in an ongoing debate abroad and in the U.S. about free speech and the real-world impact of online misinformation. In the U.S., politicians have griped for years about the content policies of major platforms like YouTube and Facebook—generally with conservatives complaining the companies are too censorious and liberals bemoaning that they don’t take down enough misinformation and hate speech.

Where should the line be? Is it possible for platforms to respect free expression while removing “harmful content” and misinformation? Who gets to decide what is true and false, and what role, if any, should the government play? Evan is joined by Renee Diresta who studies and writes about adversarial abuse online. Previously, she was a research manager at the Stanford Internet Observatory where she researched and investigated online political speech and foreign influence campaigns. She is the author of Invisible Rulers: The People Who Turn Lies into Reality. Read her recent op-ed in the New York Times here.

  continue reading

82 episoder

Artwork
iconDel
 
Manage episode 435597157 series 3530279
Innhold levert av Max Bodach and Foundation for American Innovation. Alt podcastinnhold, inkludert episoder, grafikk og podcastbeskrivelser, lastes opp og leveres direkte av Max Bodach and Foundation for American Innovation eller deres podcastplattformpartner. Hvis du tror at noen bruker det opphavsrettsbeskyttede verket ditt uten din tillatelse, kan du følge prosessen skissert her https://no.player.fm/legal.

The recent riots in the United Kingdom raise new questions about online free speech and misinformation. Following the murder of three children in Southport, England, false rumors spread across social media about the killer’s identity and religion, igniting simmering resentment over the British government’s handling of immigration in recent years. X, formerly Twitter, has come under fire for allowing the rumors to spread, and the company’s owner Elon Musk has publicly sparred with British politicians and European Union regulators over the issue.

The incident is the latest in an ongoing debate abroad and in the U.S. about free speech and the real-world impact of online misinformation. In the U.S., politicians have griped for years about the content policies of major platforms like YouTube and Facebook—generally with conservatives complaining the companies are too censorious and liberals bemoaning that they don’t take down enough misinformation and hate speech.

Where should the line be? Is it possible for platforms to respect free expression while removing “harmful content” and misinformation? Who gets to decide what is true and false, and what role, if any, should the government play? Evan is joined by Renee Diresta who studies and writes about adversarial abuse online. Previously, she was a research manager at the Stanford Internet Observatory where she researched and investigated online political speech and foreign influence campaigns. She is the author of Invisible Rulers: The People Who Turn Lies into Reality. Read her recent op-ed in the New York Times here.

  continue reading

82 episoder

Alle episoder

×
 
Loading …

Velkommen til Player FM!

Player FM scanner netter for høykvalitets podcaster som du kan nyte nå. Det er den beste podcastappen og fungerer på Android, iPhone og internett. Registrer deg for å synkronisere abonnement på flere enheter.

 

Hurtigreferanseguide

Copyright 2024 | Sitemap | Personvern | Vilkår for bruk | | opphavsrett