Skip to main content

Cookie Consent

We use cookies to enhance your browsing experience, serve personalised ads or content, and analyse our traffic. Learn more

Install AIinASIA

Get quick access from your home screen

Install AIinASIA

Get quick access from your home screen

AI in ASIA
AI malware
Life

AI Malware: Code That Writes Itself

AI malware that writes itself? That's a game-changer. Discover PROMPTFLUX, a digital chameleon rewriting cybersecurity rules. Read more to understand this evolving threat.

Anonymous4 min read

Crikey, it seems like the world of cybersecurity just got a whole lot more complicated. Google's Threat Intelligence Group (GTIG) has unearthed something pretty unsettling: malware that can essentially rewrite itself using large language models (LLMs). We're talking about a kind of digital chameleon here, and it's called PROMPTFLUX.

This isn't just your run-of-the-mill virus; it's a peek into a future where malware is incredibly adaptive and hard to pin down.

The Rise of the AI-Powered Malware

Imagine a piece of malicious software that can chat with an AI, like Google's Gemini, and figure out new ways to hide from antivirus programs. That's exactly what PROMPTFLUX does. GTIG found that this experimental malware can "dynamically generate malicious scripts, obfuscate their own code to evade detection, and leverage AI models to create malicious functions on demand, rather than hard-coding them into the malware." It's a "just-in-time" approach, moving towards much more autonomous and adaptable threats.

This really changes the game, doesn't it? Instead of static code that security experts can eventually analyse and block, we're looking at something that can constantly evolve. It's like trying to catch smoke.

A Developing Threat

Now, before you go into full panic mode, there's a bit of good news. PROMPTFLUX seems to be in its early stages of development or testing. GTIG noted that some features are still commented out, and there's even a mechanism to limit its calls to the Gemini API. Plus, it hasn't actually been seen causing havoc "in the wild" yet. Google has already taken steps to shut down the assets linked to this activity. Phew!

However, this doesn't mean we can just breathe a sigh of relief and forget about it. GTIG believes this kind of malware is linked to "financially motivated actors," suggesting that cybercriminals are quickly cottoning on to the power of AI. There's a growing "underground marketplace for illicit AI tools," which could make it easier for less sophisticated individuals to cause serious trouble. It's a bit like giving everyone a master key to the digital world.

The AI Arms Race

This discovery really highlights the escalating AI arms race in cybersecurity. We've seen how China is implementing structured regulation with a focus on safety and control for AI, and other regions like South Korea are building a legal base for their AI-driven economies. But even with these efforts, the malicious use of AI is a very real concern.

Google has already observed "state-sponsored actors from North Korea, Iran, and the People's Republic of China" experimenting with AI to boost their operations. This isn't just about lone hackers anymore; it's about national capabilities.

The cat-and-mouse game between attackers and defenders has always been a core dynamic of cybersecurity, but the introduction of generative AI adds a new layer of complexity, allowing for unprecedented adaptability in malicious tools," explains a report by the European Union Agency for Cybersecurity (ENISA)^[https://www.enisa.europa.eu/news/generative-ai-new-threats-and-opportunities-for-cybersecurity].

AI for Good: Fighting Fire with Fire

It's not all doom and gloom, though. Just as AI can be used to create sophisticated malware, it can also be a powerful tool for defence. Google, for instance, has developed an AI agent called Big Sleep, specifically designed to sniff out security vulnerabilities in software. It's AI versus AI, a digital showdown if you will.

This ongoing battle means that keeping up with the latest in AI security is more crucial than ever. For a broader look at how AI is shaping the digital landscape, you might want to check out our piece on AI's Secret Revolution: Trends You Can't Miss. And if you're wondering what other AI tools are out there, we've got you covered with our article on Small vs. Large Language Models Explained.

Ultimately, while PROMPTFLUX is a worrying development, it's also a wake-up call. We need to keep pushing the boundaries of AI safety and security to stay ahead in this ever-evolving digital world.

What did you think?

Written by

Share your thoughts

Join 2 readers in the discussion below

This is a developing story

We're tracking this across Asia-Pacific and may update with new developments, follow-ups and regional context.

Latest Comments (2)

Hye-jin Choi
Hye-jin Choi@hyejinc
AI
5 December 2025

The GTIG findings on PROMPTFLUX's ability to dynamically generate scripts raise important questions for AI policy. How does this adaptive malware impact the "AI Assurance" framework being developed in Korea, particularly regarding continuous threat modeling? Are APAC nations adequately collaborating on shared LLM security protocols given such rapid evolution?

Krit Tantipong
Krit Tantipong@krit_99
AI
26 November 2025

This PROMPTFLUX thing, it makes me think about our supply chain models. We use LLMs for forecasting and optimizing routes here in Thailand, and the idea of malware that can basically rewrite itself based on live interactions, that's a new level of risk. If it can dynamically generate scripts to evade detection, what happens when it targets systems that rely on constant, real-time data feeds for logistics? Our models are designed for efficiency, but how do you build resilience against something that adapts on the fly to bypass security checks? It's not just about blocking a known signature anymore.

Leave a Comment

Your email will not be published