Published July 30, 2025 | Version v11
Preprint Open

Algorithmic Proximate Harm Detection (APHD): A Forensic Framework for Quantifying Causal Exposure and Systemic Design Failure in Algorithmic Platforms

Authors/Creators

Description

Algorithmic platforms do not merely reflect risk—they actively shape it. From content curation to user matching, optimization systems determine who is seen, silenced, or re-exposed to harm. This paper introduces Algorithmic Proximate Harm Detection (APHD), a forensic scoring framework for quantifying the causal role that algorithms, users, and design systems play in digital trauma.

Unlike existing fairness audits or moderation tools, APHD produces two interpretable metrics: the Proximate Harm Score (PHS), which assesses node-level contribution to specific harm events, and the Systemic Harm Index (SHI), which captures broader structural failures in platform governance. These metrics are derived from nine normalized variables—including recurrence, exposure delta, silence suppression, and platform negligence—operating on a harm graph constructed from platform logs, moderation pathways, and social signals.

We introduce an Adaptive Multi-Stage Bat Algorithm (AMSBA) to optimize variable weighting, and use SHAP for post-hoc attribution and forensic explainability, ensuring alignment with legal evidentiary standards (e.g., Daubert, Frye). APHD also integrates a predictive behavioral module (SignatureProfiler) and a Cooperative Memory Replication (CMR) system for tracing distributed harm across linked accounts.

Beyond legal and technical accountability, APHD models the structural conditions under which digital violence becomes a public health issue—one marked by retraumatization, disengagement, and elevated risk among marginalized populations. It supports not only post-hoc analysis, but early-stage detection and traceable intervention within systems that too often amplify harm in silence.

Grounded in lived experience, graph theory, and tort logic, APHD offers a scalable, court-admissible method for detecting and explaining algorithmic harm—not only after it occurs, but as it begins to unfold.

Notes (En)

This framework is currently in prototype development with support for real-time scoring using FastAPI, Neo4j, and SHAP-based interpretability. It was inspired by my own experiences navigating queer dating platforms as a Black trans woman, where repeated exposure to aggressors—even after reporting—revealed structural failure modes that current fairness tools do not capture.

APHD is distinct from classification or moderation systems. It treats algorithmic harm as a causal, graph-encoded phenomenon—capable of being traced, quantified, and legally explained. Its architecture includes predictive behavioral profiling (SignatureProfiler), distributed harm tracing (CMR), and forensic SHAP alignment for admissibility.

The work is designed for real-world application in legal proceedings, platform reform, and public health contexts. I am currently collaborating with legal researchers and community advocates to expand empirical validation.

Files

Algorithmic_Proximate_Harm_Detection__APHD_.pdf

Files (312.7 kB)

Name Size Download all
md5:3a30fb2c7d2bb457ab977e17d919c399
312.7 kB Preview Download

Additional details

Additional titles

Alternative title (En)
Algorithmic Proximate Harm Detection (APHD): A Legal-Grade Scoring System for Identity-Based and Systemic Risk in Platform Design
Alternative title (En)
Algorithmic Proximate Harm Detection: A Graph-Theoretic Framework for Detecting, Scoring, and Explaining Digital Trauma