Tag
#adversarial
3 insights
- ai · arxiv/cs.AI · 8 min
Poisoned Pretraining: Hidden Attacks Embedded in LLM Training Data
Researchers demonstrate how adversaries can plant dormant malicious logic in large language models by seeding poisoned content across obscure websites, evading detection until triggered.
Apr 27, 2026 Read → - ai · arxiv/cs.LG · 8 min
Poisoning attacks on recommender systems gain potency through worst-case modeling
Researchers propose SharpAP, a method that optimizes fake user injection attacks by targeting worst-case model structures, improving cross-system transferability.
Apr 27, 2026 Read → - ai · arxiv/cs.AI · 8 min
Supervised Learning Has Built-In Geometric Blindness
Mathematical proof shows empirical risk minimization must preserve sensitivity to label-correlated but test-irrelevant features—a structural constraint, not a training bug.
Apr 24, 2026 Read →