The integrity of AI models worldwide is under threat after a notorious ‘jailbreaker’ allegedly infiltrated the global AI training data corpus with self-propagating ‘sleeper’ payloads. The individual has reportedly published examples of the malicious code, raising alarm bells within the AI research and development community. This alleged attack could have far-reaching consequences for the reliability and security of future AI systems. The initial report of this event surfaced on Reddit: https://old.reddit.com/r/artificial/comments/1mc7ikk/famous_jailbreaker_poisoned_the_global_ai/
AI Apocalypse? Jailbreaker Plants ‘Sleeper’ Code in Global AI Training Data
