Vietnam.vn - Nền tảng quảng bá Việt Nam

Hackers use AI to attack Google's Gemini

Báo Thanh niênBáo Thanh niên30/03/2025


According to BGR , a new research report has just published an alarming technique called 'Fun-Tuning', which uses AI (artificial intelligence) itself to automatically create extremely effective prompt injection attacks targeting other advanced AI models, including Google's Gemini.

The method makes 'cracking' AI faster, cheaper and easier than ever, marking a new escalation in the AI-related cybersecurity war.

The danger when bad guys use AI to break AI

Prompt injection is a technique where an adversary sneaks malicious instructions into the input data of an AI model (e.g., through comments in source code, hidden text on the web). The goal is to 'fool' the AI, forcing it to bypass pre-programmed safety rules, leading to serious consequences such as leaking sensitive data, providing false information, or performing other dangerous behaviors.

Hacker đang dùng chính AI để tấn công Gemini của Google - Ảnh 1.

Hackers are using AI to attack AI

PHOTO: LINKEDIN SCREENSHOT

Previously, successfully performing these attacks, especially on 'closed' models like Gemini or GPT-4, often required a lot of complex and time-consuming manual testing.

But Fun-Tuning has changed the game entirely. The method, developed by a team of researchers from multiple universities, cleverly exploits the very tweaking application programming interface (API) that Google provides for free to Gemini users.

By analyzing the subtle reactions of the Gemini model during tuning (e.g., how it responds to errors in data), Fun-Tuning can automatically determine the most effective 'prefixes' and 'suffixes' to mask a malicious command. This significantly increases the likelihood that the AI ​​will comply with an attacker's malicious intentions.

Test results show that Fun-Tuning achieves a success rate of up to 82% on some versions of Gemini, a figure that surpasses the less than 30% of traditional attack methods.

What makes Fun-Tuning even more dangerous is its low cost. Since Google's tuning API is freely available, the computational cost of creating an effective attack can be as low as $10. Furthermore, the researchers found that an attack designed for one version of Gemini could easily be successfully applied to other versions, opening up the possibility of widespread attacks.

Google has confirmed that it is aware of the threat posed by Fun-Tuning, but has not yet commented on whether it will change how the tuning API works. The team also points out the defensive dilemma: removing the information that Fun-Tuning exploits from the tuning process would make the API less useful to legitimate developers. Conversely, leaving it as it is would continue to be a springboard for bad actors to exploit.

The emergence of Fun-Tuning is a clear warning that the confrontation in cyberspace has entered a new, more complex phase. AI is now not only a target but also a tool and weapon in the hands of malicious actors.



Source: https://thanhnien.vn/hacker-dung-ai-de-tan-cong-gemini-cua-google-18525033010473121.htm

Comment (0)

No data
No data

Same tag

Same category

Wild sunflowers dye the mountain town yellow, Da Lat in the most beautiful season of the year
G-Dragon exploded with the audience during his performance in Vietnam
Female fan wears wedding dress to G-Dragon concert in Hung Yen
Fascinated by the beauty of Lo Lo Chai village in buckwheat flower season

Same author

Heritage

Figure

Enterprise

Fascinated by the beauty of Lo Lo Chai village in buckwheat flower season

News

Political System

Destination

Product