The Intervention Journey A Roadmap to Effective Digital Safety Measures 2025
Page 30 of 45 · WEF_The_Intervention_Journey_A_Roadmap_to_Effective_Digital_Safety_Measures_2025.pdf
Example technical interventions TABLE 1
Intervention Description Example organization(s)
AI-powered content
labelling systemsAI-powered systems that classify content into different
categories (e.g. misinformation, hate speech, explicit
material), allowing platforms to flag, rate or limit the
distribution of harmful content.YouTube (Google)
Uses AI to label content as safe or harmful, influencing
recommendations and the visibility of videos.
Real-time AI and
human-in-the-loop
moderation systemsHybrid content moderation systems where AI flags
content and human moderators make the final decision,
improving moderation accuracy for nuanced or context-
sensitive content.Twitch
Identifies potential violations of community guidelines
using a combination of automated detection and
user reporting.
Nudity detection
systemsAI-driven technology harnessing on-device machine
learning to detect and blur (when turned on) nudity
images.
Functions in secure, encrypted environments and may
default to active protection for younger users, with
prompts encouraging broader adoption among adults.Instagram (Meta platforms)
Leverages AI/ML built into the app on user devices.
When turned on, it automatically blurs images
containing nudity that are being sent and/or received in
Instagram direct messages. Teens have it turned on by
default, adults need to opt into the feature.
Directing people to safety tips when sending or
receiving these images, developed with guidance
from experts about the potential risks involved.
User identity
verificationTechnologies to verify users’ identities using
methods such as facial recognition, ID verification
or behavioural analytics.Tinder and Hinge
Uses identity verification tools to ensure users are real,
reducing catfishing and online abuse.
Blacklisted
URL systems for
illegal contentSystems that block access to known URLs containing
illegal content such as CSAM, preventing sharing or
viewing of such content.IWF
Works with internet service providers and tech
companies to block URLs hosting CSAM.
Hashing technology
for illegal content
detectionTechnologies that use hashing algorithms to detect and
prevent the spread of illegal content, such as CSAM or
terrorism-related media.PhotoDNA (Microsoft)
A hash-matching technology that can be used to detect
previously identified CSEA material.
Source: YouTube Official Blog. (2023). Our approach to responsible AI innovation; Twitch. (n.d.). Safety at Twitch; Instagram.
(2024). New Tools to Help Protect Against Sextortion and Intimate Image Abuse; Match Group. (n.d.). Safety; Internet Watch
Foundation. (n.d.). URL List; Microsoft. (n.d.). PhotoDNA.
3.2 Educational interventions
Educational interventions in digital safety consist
of structured programmes, activities or other
informational resources designed to inform and
educate individuals and groups about digital risks,
safe online practices and cybersecurity measures.
The goal of these interventions is to enhance users’
awareness, knowledge and skills for navigating the
digital world safely.
Educational interventions are crucial in raising
awareness and providing targeted knowledge.
These interventions must build on existing knowledge, progressing through different
levels – similar to a school curriculum – rather
than starting from scratch. Capacity building
for educators is equally important, requiring
investments in training and resources to equip them
to handle sensitive topics. Furthermore, harnessing
technology such as large language models (LLMs)
can enhance the learning experience by offering
safe, simulated environments where individuals can
explore scenarios without real-world risk, facilitating
practical and secure learning opportunities.
The Intervention Journey: A Roadmap to Effective Digital Safety Measures
30
Ask AI what this page says about a topic: