Panel discussion: Content moderation beyond the ban: Reducing borderline, toxic, misleading, and low-quality content
Public debate about content moderation focuses almost exclusively on removal, such as what is deleted and who is suspended. But what about content that is identified as “borderline,” which almost—but not quite—violates the guidelines? Faced with an expanding sense of responsibility, many platform companies have started identifying this type of content, as well as content that may be toxic, misleading, or harmful in the aggregate. Rather than remove it, they can minimize its effect by taking some of the following approaches: reduce its visibility in recommendations, limit its discoverability in search, add labels or warnings, or provide fact-checks or additional context. Tarleton Gillespie (Senior Principal Researcher at Microsoft) and Zoe Darmé (Senior Manager of Search at Google) host a panel that includes Sarita Schoenebeck, (Associate Professor, School of Information, University of Michigan), Ryan Calo (Professor of Law, University of Washington), and Charlotte Willner (Founding Executive Director of the Trust & Safety Professional Association).
Join us as they discuss these techniques and the questions they raise, such as: How is such content being identified? Are these approaches effective? How do users respond? How can platforms be transparent and accountable for such interventions? What are the ethical and practical implications of these approaches?
Learn more about the 2021 Microsoft Research Summit: https://Aka.ms/researchsummit (opens in new tab)
- Évènement :
- Microsoft Research Summit 2021
- Piste :
- Responsible AI
- Date:
- Haut-parleurs:
- Tarleton Gillespie, Zoe Darmé, Ryan Calo, Sarita Schoenebeck, Charlotte Willner
- Affiliation:
- Microsoft Research New England, Google, University of Washington School of Law, University of Michigan, Trust & Safety Professional Association
-
-
Tarleton Gillespie
Senior Principal Researcher
-
-