The Role of AI in Digital Forensics: Advancing the Fight Against Violence Against Women with Large Language Models
When a phone becomes the barrier to justice
When a woman handed her smartphone to detectives investigating her domestic abuse complaint ("Force B, Case 36”, [1]), officers promised its return within 24 hours. It took 275 days. By the time forensic analysts provided partial data, key WhatsApp messages, the victim's primary evidence of ongoing coercion, were still awaiting "deep review" [1]. That case was recorded in 2022, and since then, government and policing leaders have committed that no adult rape victim should be without a phone for more than 24 hours [2].
However, even with improvements, the scale of the challenge is stark. In the year ending March 2024, about 2.3 million adults in England and Wales experienced domestic abuse [3]; and in the year ending March 2023, 93% of domestic-abuse-related sexual-offence victims were women [4]. Digital communications saturate these cases: practitioners estimate over 90% of VAWG investigations now have a digital element [5, 6].
As a result, today’s investigations routinely involve months or years of messages across multiple apps (WhatsApp, iMessage, Instagram DMs, Snapchat, etc.). Real police and court files show how quickly volume becomes unmanageable. For example, in November 2020, Dyfed‑Powys Police reported a case in which a man was jailed after bombarding his former partner with threatening messages duplicated across SMS and WhatsApp, with messages found on the victim’s device [7]. And more recently, in August 2024, Sky News reported that Joey Essex had contacted police after receiving more than 100,000 online messages, including threats to his girlfriend, from a single alleged stalker, delivered across public and private platforms [8].
Meanwhile, HMICFRS has warned that policing is unable to keep pace with the increasing volume of digital evidence from devices which has led to considerable backlogs [9, 10]. On top of the workload problem, privacy and trust remain fragile. The government’s End-to-End Rape Review and the ICO’s 2022 opinion criticised disproportionate “digital strip-searches” of complainants [2, 11]. Finally, disclosure law adds complexity: investigators and prosecutors must pursue reasonable lines of inquiry and disclose material that might assist the defence, duties re-emphasised and updated in the Attorney General’s Guidelines (2024). Missed or late disclosure risks collapsed cases and retraumatised victims [12].
So the bottleneck is clear: mountains of messages, strict legal thresholds, and a legitimate need to minimise intrusion, all while moving quickly enough to keep victims engaged.
Can machines read first? Beyond keyword searches
What if an AI-based software could scan, sort and explain conversations at scale, surfacing what matters in hours rather than months, and doing so in a way that is auditable and privacy-conscious? With recent advances in large language models (LLMs) and Natural Language Processing (NLP), that prospect is no longer speculative. We briefly introduce one of the first pilots in UK policing where the Forensics Capability Network (FCN) and Dorset Police have partnered with the University of Warwick and the University of Leeds to actively explore AI triage to support the analysis of digital evidence and thus shrink backlogs and reduce the need to keep victims’ phones [13, 14].
Traditional approaches in policing, such as manual scrolling or keyword searches, often fall short when analysing large volumes of digital evidence. These methods can miss the subtle dynamics of abusive or controlling behaviour, and their results may vary from case to case. Meaning in VAWG-related conversations is highly contextual: a phrase that is harmless in one context may signal coercion or abuse in another. The interpretation of a message shifts depending on cultural background, family structure, personal history, and even the evolving slang or coded language used by individuals as popular culture and TV series such as ‘Adolescence’ (2025) have highlighted for the wider public [15].
To address these limitations, we designed and developed SafeSpeech, a new generation of AI tools for the analysis of textual digital evidence. Unlike traditional systems that simply flag keywords, our approach framework can quickly scan and analyse hundreds of thousands of messages in a few seconds and interpret the broader conversational context, the patterns and dynamics of abusive behaviour embedded in long, multi-platform chats.
Rethinking Context: Our Approach to NLP for VAWG
SafeSpeech is an integrated platform for analysing sexist, abusive, and coercive language at both the message and conversation level. It was built to handle what policing actually faces in VAWG cases: long, intricate chats where meaning is derived from context.
Specifically, SafeSpeech combines fine-tuned classifiers with LLMs to support three key tasks: (1) detecting harmful content at multiple granularities; (2) summarising long, multi-topic exchanges while preserving the coherence of discussions across days, weeks, or months, and focusing on harmful behaviour; and (3) explaining why the model flagged certain content, so that investigators can check, challenge, or disclose the result [17].
Fast and Fine-Grained Sexism Detection.
On the back end, SafeSpeech combines strong specialised models (such as RoBERTa and our novel ensemble “M7-FE” [18]) with recent LLMs like LLaMA, GPTs, and Gemini. On the front end, there is a lightweight browser app where analysts, and researchers in the publicly available version, can upload their own data or test existing benchmarks. Building on our recent work [20], the system also relies on data-augmentation methods to reduce sparsity in under-represented scenarios, and crucially makes operational definitions explicit. This makes it possible to identify, monitor, and potentially mitigate any bias that could otherwise remain hidden in model predictions or data labelling.
Reading conversations, not just messages.
Unlike the vast majority of existing tools in the literature, SafeSpeech does not stop at single-message labels. It first chunks long dialogues into coherent sections (so related moments are grouped even when other chatter interrupts), then runs instruction-based, toxic-aware summarisation that highlights patterns relevant to VAWG, such as escalation, monitoring, intimidation, back-handed “apologies,” etc. In addition, SafeSpeech supports speaker-conditioned summarisation, organising events by the participants involved to make roles and interactions clearer. The result is a compact, navigable storyline that points practitioners to where in the chat those behaviours occur.
Why was a flag raised?
SafeSpeech includes a soft-interpretability layer that visually highlights the phrases most influential to a model's analysis. This technique is a novel application of perplexity-gain analysis, specifically adapted to VAWG-related conversations, and gives reviewers a rapid way to assess whether the system’s focus is appropriate for the case. In the interface, these cues appear as a heatmap, accompanied by a short rationale generated by the LLM, and positioned next to the source messages for easy review [Fig. 1].
Figure 1. Heatmap for a synthetic example from SafeSpeech (public version),
Understanding People.
For research purposes only, the platform can also generate persona cues based on the Big Five traits [19], conditioned on the toxic-aware summaries. The goal is not to label individuals but to support analyst reasoning about behavioural dynamics in conversation, tracking shifts in tone, dominance, or reactivity, while clearly flagging that this is exploratory, not diagnostic.
Deployment and Future Work
The platform is currently being tested on historical cases and within safe, controlled environments. Deployment into live investigations will follow only after exhaustive technical, legal, and ethical validation. This caution is deliberate. VAWG investigations are sensitive and consequential; they demand systems that are robust, auditable, and aligned with victim-centred practice. If the standards are met, the benefits for policing, and, crucially, for the public and victims, could be significant: faster identification of relevant evidence, more focused investigative effort, and case files that present clearer, more coherent narratives to the courts.
In parallel, we are expanding collaborations with forces across the United Kingdom to specialise the system for stalking [21], and to drug-related offences, where complex, coded exchanges in messaging apps often mask supply chains and roles. The final destination is a practical, explainable AI capability that helps officers read first, intrude less, and act sooner, while strengthening the integrity of digital evidence from first report to final adjudication.
Gabriele Pergola
Assistant Professor
Department of Computer Science
References
[1] Operation Soteria. Year One Report (Accessible Version). 2022. https://www.gov.uk/government/publications/operation-soteria-year-one-reportGOV.UK
[2] HM Government. Rape Review: Progress Update (June 2022). 2022. https://assets.publishing.service.gov.uk/media/62b057fed3bf7f0afc3880a9/rape-review-progress-update-june-2022.pdf
[3] Office for National Statistics (ONS). Domestic abuse in England and Wales overview: November 2024 (latest data include the year ending March 2024). 2024. https://www.ons.gov.uk/peoplepopulationandcommunity/crimeandjustice/bulletins/domesticabuseinenglandandwalesoverview/november2024
[4] Office for National Statistics (ONS). Domestic abuse and the criminal justice system, England and Wales: November 2023 (Appendix tables). 2023. https://www.ons.gov.uk/peoplepopulationandcommunity/crimeandjustice/datasets/domesticabuseandthecriminaljusticesystemappendixtables
[5] National Police Chiefs’ Council (NPCC). Digital Forensic Science Strategy. 2020. https://www.npcc.police.uk/SysSiteAssets/media/downloads/publications/publications-log/2020/national-digital-forensic-science-strategy.pdf npcc.police.uk
[6] College of Policing. Towards a smart digital forensic advisor to support first responders in scene triage. Project overview, 2024. https://www.college.police.uk/research/projects/towards-smart-digital-forensic-advisor-support-first-responders-scene-triage-digital-evidence-across-crime-types College of Policing
[7] Dyfed-Powys Police. Jailed after threatening messages sent via SMS and WhatsApp. Press release, 27 Nov 2020. https://www.dyfed-powys.police.uk
[8] Sky News. Joey Essex contacts police after 100,000 online messages from alleged stalker. 15 Aug 2024. https://news.sky.com/story/joey-essex-police stalker-100000-messages-13165404
[9] HMICFRS. Police forces overwhelmed and ineffective in response to digital forensics, report finds. News release, 1 Dec 2022. https://www.justiceinspectorates.gov.uk/hmicfrs/news/news-feed/police-forces-overwhelmed-and-ineffective-in-response-to-digital-forensics-report-finds/
[10] HMICFRS. An inspection of how well the police and other agencies use digital forensics in their investigations. 2022. https://www.justiceinspectorates.gov.uk/hmicfrs/publications/digital-forensics/
[11] HM Government. Better protection from invasive data requests for victims of rape. News release, 17 Oct 2022. https://www.gov.uk/government/news/better-protection-from-invasive-data-requests-for-victims-of-rape
[12] Attorney General’s Office. Attorney General’s Guidelines on Disclosure (Updated 29 Feb 2024; in force 29 May 2024). 2024. https://www.gov.uk/government/publications/attorney-generals-guidelines-on-disclosure
[13]: The Independent. “AI that detects hate messages against women could also be turned on drug dealers”, July 2024. https://www.independent.co.uk/news/uk/northumbria-police-wales-england-university-of-warwick-b2580311.html
[14] Forensic Capability Network (FCN). AI can detect abusive messages 21 times faster than humans. 16 Jul 2024. https://www.fcn.police.uk/news/2024-07/ai-can-detect-abusive-messages-21-times-faster-humans fcn.police.uk
[15] Netflix. Adolescence. Limited Series. 2025.
[17] Tan, X.; Lyu, C.; Umer, H.M.; Khan, S.; Parvatham, M.; Arthurs, L.; Cullen, S.; Wilson, S.; Jhumka, A.; Pergola, G. SafeSpeech: A Comprehensive and Interactive Tool for Analysing Sexist and Abusive Language in Conversations. NAACL 2025.
[18] Khan, S.; Pergola, G.; Jhumka, A. Multilingual Sexism Identification via Fusion of Large Language Models. CLEF 2024 Working Notes (CEUR-WS Vol. 3740). 2024. https://ceur-ws.org/Vol-3740/paper-99.pdf
[19] Costa, P.T.; McCrae, R.R. The Revised NEO Personality Inventory (NEO-PI-R). In: The SAGE Handbook of Personality Theory and Assessment (Vol. 2), 2008.
[20] Khan, S.; Jhumka, A.; Pergola, G. Explaining Matters: Leveraging Definitions and Semantic Expansion for Sexism Detection. 2025.n https://aclanthology.org/2025.acl-long.809/
[21] BBC, Police force harnesses AI to help catch stalkers, May 2025. https://www.bbc.co.uk/news/articles/ckg4816gj5yo