banner_image

A Collaborative Approach to Assessing the Challenges of AI-Driven Harms

badge_icon Governance, Regulation and Policy

Funding Stream:

Collaboration Grant Logo Collaboration Grant

Award Details

Project Team:

Allysa Czerwinsky (Project Lead), Ashton Kingdon, Shana MacDonald, Karmvir Padda

Lead HEI:
University of Manchester
Project dates:
5 January 2026 - 31 March 2026
Geographical focus:
  • UK
  • Canada
Partner HEIs:
partner_logo
partner_logo
partner_logo

Summary

Amidst a rapid proliferation of synthetic image- and video-based content in online ecosystems, our research addresses urgent gaps in understanding AI-driven harms by examining real-world cases of extremist visual content and evaluating existing AI-powered moderation solutions. This project seeks to document how generative AI enables and accelerates visual hate speech; assess the feasibility of AI-supported content moderation tools for detecting harmful visual content; and develop evidence-based recommendations for policymakers and platform developers to mitigate these harms while preserving legitimate expression. ​

Four rapid case studies: ​​

  • visual content from ‘Unite the Kingdom’ protests​​
  • meme-based extremism in manosphere spaces​​
  • fascist aesthetics in AI-generated videos created by Sora2​​
  • opportunities for AI-supported flagging of harmful content through existing moderation tools like SightEngine.​

Collaborating with University of Southampton and Waterloo University, Canada.

Sign up to our mailing list

RAi UK will store your data in accordance with the General Data Protection Regulation 2017 (GDPR). We will not share your data with any third parties and you are given opportunities to unsubscribe at any time within the electronic communications you receive, or by emailing info@rai.ac.uk