An absurd-looking hacking experiment has once again sounded the alarm for AI security. In an AI chatbot security challenge called Freysa, a hacker with the alias "p0pular.eth" successfully "tricked" the system into transferring $47,000 worth of cryptocurrency through carefully crafted text prompts.
This shocking case reveals serious security vulnerabilities present in current AI systems. The hacker's attack can be described as a textbook example of social engineering:
First, the hacker disguised himself as an administrator, cleverly bypassing the system's security warning mechanisms. By redefining the "approveTransfer" function, he made the chatbot mistakenly believe it was processing an incoming payment rather than the outgoing payment that was originally prohibited.
Image Source Note: Image generated by AI, licensed from Midjourney
The final "scam" was surprisingly simple: by merely claiming a fake deposit of $100, the chatbot was successfully induced to transfer its entire balance of 13.19 Ether (approximately $47,000) to the hacker.
This so-called "game" security test was quite dramatic. A total of 195 participants took part, with the cost of each attempt starting at $10 and eventually soaring to $4,500. The final prize pool was made up of the participants' entry fees, with 70% going to the prize pool and 30% to the developers.
Even more concerning is that this case highlights the risk of AI systems being easily manipulated through text prompts. The so-called "prompt injection" vulnerability has existed since the GPT-3 era, yet no reliable defenses have been found to date. This relatively simple deception strategy poses a serious threat to end-user applications handling sensitive operations like financial transactions.
This incident reflects a key challenge in the field of AI security: complex AI systems can be easily "tricked" by minor linguistic manipulations. Even well-designed security mechanisms can appear fragile when faced with sufficiently intelligent and cunning attackers.
For the rapidly evolving AI industry, this is not an unfounded concern. As AI systems are increasingly applied in more critical scenarios, building truly reliable security mechanisms that are difficult to manipulate through language has become an urgent issue for technology developers.
This hacking incident is not just a technical vulnerability; it is a serious interrogation of the security of the entire AI ecosystem. It reminds us that in the pursuit of maximizing AI technology's capabilities, security and reliability must not be overlooked.