Demajh logo Demajh, Inc.

ROBAD: Robust Adversary-aware Local-Global Attended Bad Actor Detection Sequential Model: what it means for business leaders

ROBAD hardens social-platform safety stacks by spotting malicious users even when they rewrite or append posts to dodge filters—pairing transformer attention with adversary-aware training to cut costly manual moderation.

1. What the method is

ROBAD is a transformer-based classifier that ingests the chronological sequence of a user’s posts and predicts whether the account is benign or a bad actor. Each post is first encoded bidirectionally to capture local semantics; those embeddings feed a transformer decoder that learns long-range behaviour across the timeline. A contrastive-learning block then aligns embeddings from original and adversary-edited sequences before the prediction head, letting the model flag deceptive activity even when future posts are crafted to bypass filters.

2. Why the method was developed

Content-moderation pipelines crumble when attackers tweak wording or insert filler text, forcing expensive manual review and reputational damage. Earlier detectors captured only post-level hints or ignored adversarial threat models. ROBAD closes that gap by uniting fine-grained text understanding with global behaviour cues and training against realistic rewriting attacks, sustaining detection quality while shrinking the costly cat-and-mouse loop between platform defenders and coordinated manipulation campaigns.

3. Who should care

Trust-and-safety leaders at social networks, marketplaces and knowledge bases; security engineers fighting bots and sock-puppets; policy teams tracking influence operations; and investors evaluating compliance risk all gain from a resilient, behaviour-aware signal that pinpoints accounts threatening community integrity.

4. How the method works

The pipeline tokenises each post, passes it through a two-layer transformer encoder and stacks the resulting vectors. A causal-masked decoder attends over the entire stack to build a timeline embedding capturing style and rhythm. During training, synthetic adversarial edits—generated via PETGEN or LLaMA prompts—are appended to timelines. A contrastive loss draws embeddings of genuine and tampered sequences together within the correct class while pushing opposite classes apart, yielding a prediction head immune to realistic perturbations.

5. How it was evaluated

Experiments used Yelp fake-reviewer and Wikipedia vandal datasets. Baselines included production TIES, hierarchical RNNs, BERT-based sequence models and defence-augmented variants. Models faced PETGEN and LLaMA rewriting plus copy-append attacks. Metrics were macro-F1 before attack, F1 drop under attack and runtime. Ablations removed local-global attention or the adversary-aware contrastive term to isolate their contributions.

6. How it performed

ROBAD led clean-data benchmarks by 3–6 macro-F1 points and, under PETGEN attack, lost only 4 % F1—versus 13 % for TIES and over 20 % for vanilla BERT. Dropping the adversary-aware loss doubled the robustness hit, while omitting local-global attention cut clean accuracy by five points. Training and inference fit on a single A100 GPU with sequences up to 128 posts. (Source: arXiv 2507.15067, 2025)

← Back to dossier index