Cryptography Reveals Fundamental Flaws in AI Safety Filters
Cryptographers have demonstrated that external protections for AI models like ChatGPT are inherently vulnerable to bypass attacks. Using cryptographic tools such as time-lock puzzles and substitution ciphers, researchers prove that any safety filter operating with fewer computational resources than the core model can be exploited, exposing an unavoidable gap in AI security.