This programme explores how AI and platform technologies can be designed to counter online harms and support PCVE efforts while maintaining autonomy, trust, and safety for diverse user groups. It integrates research on persuasive and counter-persuasive AI, value-sensitive design, and platform governance to examine how digital systems influence belief formation, vulnerability, and exposure to online extremism. The programme brings together projects supported by the Austrian Science Fund (FWF) and the Austrian Research Promotion Agency (FFG), alongside collaborations with violence-prevention NGOs and international policy initiatives working on responsible AI and the prevention of online radicalization.
This project develops a conceptual framework for understanding and governing the capabilities of emerging agentic AI systems in order to prevent malicious use, particularly in security-sensitive and PCVE contexts. Building on insights from sociotechnical risk analysis, platform governance, and early work presented at an ECSCW workshop, the project examines how autonomous agents acquire, combine, and operationalize persuasive, generative, or decision-making capabilities that may produce harmful downstream effects.
Through collaboration with researchers from Swansea University’s CYTREC centre and support from the (Terrorism and Social Media) TASM Sandpit initiative, the project explores how capability constraints, escalation safeguards, and institutional oversight can be designed to reduce misuse risks without undermining legitimate innovation. It aims to contribute a policy-oriented model of capability governance that helps organisations anticipate hazardous combinations of AI functions, assess intervention thresholds, and design preventative controls for responsible deployment.
Collaborator: Daniel E. Levenson, MA, MLA (Swansea University)
Funding support: TASM Sandpit at Swansea University
This FFG-supported feasibility study examines the technical, ethical, and operational requirements for deploying agentic AI systems in real-world PCVE and online-safety settings. The project evaluates whether conversational or counter-persuasive agents can be safely integrated into existing prevention workflows, with particular attention to usability, trust, risk mitigation, and alignment with organisational practices. Through stakeholder consultations, prototype exploration, and scenario-based assessments, the study identifies key opportunities and constraints for developing scalable, autonomy-preserving AI interventions that support violence-prevention organisations and policy stakeholders.
Funding: Austrian Research Promotion Agency (FFG)
Collaborator(s): Austrian and German Violence Prevention NGO's
Student(s): Md. Omar Khaium
This project investigates how autonomous conversational agents can counter early and mid-stage online radicalization by engaging users in structured, trust-preserving counterdialogue. Building on emerging research in counterargumentation, belief revision, (therapeutic)inoculation, and counter-persuasion, the project explores how chatbots can introduce doubt, weaken harmful narratives, and support healthier belief trajectories without compromising user autonomy.
The work combines stakeholder interviews, conceptual and value-sensitive analysis, and controlled experiments on platforms such as Discord to understand the sociotechnical dynamics of persuasive harm and safe intervention. It develops early computational models for recognising harmful influence processes, generating context-appropriate counterarguments, and managing risks such as reactance or escalation. This project contributes foundational insights for designing agentic AI systems capable of participating responsibly in Prevention and Countering Violent Extremism (PCVE) efforts and improving digital safety in online communities.
Funding: Austrian Science Fund (FWF).
Project link: FWF Research Radar
Student(s): Md. Omar Khaium
Investigating how counterargumentation, belief revision processes, and trust cues enable AI agents to challenge harmful narratives, introduce productive doubt, and support safer attitude change.
Studying how harmful beliefs evolve through long-term influence dynamics and how targeted, autonomy-preserving interventions can disrupt early radicalization pathways and reduce susceptibility to extremist content.
Applying value-sensitive design to ensure that persuasive and protective AI systems uphold user autonomy, respect diverse moral perspectives, and avoid paternalistic or deterministic assumptions in high-risk contexts.
Examining how platform design, visibility mechanisms, and moderation choices shape user trust, perceived safety, and vulnerability to manipulation, with a focus on accountability in socio-technical governance systems.
Blasiak, K. M., & Levenson, D. E. (2025). Shifting the Conversation on Malicious Use of AI: A value sensitive approach for stakeholder consensus. In Proceedings of the 23rd EUSSET Conference on Computer Supported Cooperative Work. European Society for Socially Embedded Technologies (EUSSET).
Blasiak, K. M., Risius, M., & Matook, S. (2021, December). “Social Bots for Peace”: A Dual-Process Perspective to Counter Online Extremist Messaging. In 42nd International Conference on Information Systems, ICIS 2021 TREOs:" Building Sustainability and Resilience with IS: A Call for Action". Association for Information Systems.
Blasiak, K. M., Risius, M., & Matook, S. (2021, December). "Conceptualising Social Bots for Countering Online Extremist Messages" (2021). ACIS 2021 Proceedings. 81. https://aisel.aisnet.org/acis2021/81
Capability Governance for Preventing the Malicious Use of AI
Rethinking AI Safety Through Capability Constraints: A Blue-Sky Agenda for Responsible Agent Design
Toward a Theory of Counter-Persuasion: Conceptual Foundations for AI-Mediated Counterargumentation
Embedding Human Values in Persuasive AI: A Value-Sensitive Design Approach to Countering Online Harms
I welcome collaborations on topics related to online harms, persuasive AI, platform governance, and responsible information systems.
I also supervise master’s theses in these areas.
Contact: kevin.blasiak(at)tuwien.ac.at