Safety at One Shot: Patching Fine-Tuned LLMs with A Single Instance Paper • 2601.01887 • Published 4 days ago
Safety at One Shot: Patching Fine-Tuned LLMs with A Single Instance Paper • 2601.01887 • Published 4 days ago
Locket: Robust Feature-Locking Technique for Language Models Paper • 2510.12117 • Published Oct 14, 2025
StructEval: Benchmarking LLMs' Capabilities to Generate Structural Outputs Paper • 2505.20139 • Published May 26, 2025 • 19
LookAhead: Preventing DeFi Attacks via Unveiling Adversarial Contracts Paper • 2401.07261 • Published Jan 14, 2024