Malicious AI Tools Spike 200% and Discussions on Jailbreaking Legitimate ChatGPT Grow by 52%

The cybersecurity landscape is witnessing an alarming rise in malicious artificial intelligence (AI) applications, with research reporting a 200% surge in the development and deployment of such tools.

Simultaneously, discussions surrounding the jailbreak of legitimate AI chatbots, notably OpenAI’s ChatGPT, have grown by 52%, highlighting a dual-edged threat: AI as both a vector for exploitation and an unwitting facilitator of illicit activities.

These trends shows the broader risks posed by AI-driven cyber threats, as cybercriminals now leverage advanced AI capabilities for crafting malware, phishing campaigns, and disinformation at scales never seen before.

Google News

The emergence of AI-enhanced tools as weapons in the cyber threat landscape stems from several concurrent developments.

First, the rapid democratization of AI has made its capabilities accessible to a wider audience, including malicious actors.

AI-powered tools now enable the automation of tasks traditionally requiring human effort, such as generating convincing phishing emails or bypassing CAPTCHA systems.

Second, the improved sophistication of large language models like ChatGPT has inadvertently enabled attackers to customize social engineering templates, evading traditional defense mechanisms.

Indeed, Kela researchers noted the growth of underground marketplaces where malicious developers discuss and refine these tools, with some offering “jailbreaking” techniques for legitimate AI systems to bypass programmed ethical guidelines.

From initial access on sale to a ransom attack (Source – Kela)

One particular strain of malicious AI tools discovered recently involves the creation of polymorphic malware that uses AI to evade detection by antivirus systems.

Exploiting AI’s ability

By analyzing the behavior of these tools, the malware can modify its code dynamically, altering its signature each time it executes to avoid suspicion.

For example, the following Python snippet demonstrates how attackers implement a basic code obfuscation technique to generate variable outputs dynamically:-

import random

def obfuscate_code(payload):
    key = random.randint(1, 256)
    obfuscated = ''.join(chr(ord(char) ^ key) for char in payload)
    return obfuscated, key

def deobfuscate_code(obfuscated, key):
    return ''.join(chr(ord(char) ^ key) for char in obfuscated)

# Malicious payload example
payload = "malware_payload_here"
obfuscated_payload, key = obfuscate_code(payload)
print(deobfuscate_code(obfuscated_payload, key))

These techniques are increasingly difficult to trace because they exploit AI’s ability to learn and adapt.

Kela analysts have stressed the growing use of forums and tools where attackers exchange such code samples, refining them to improve evasion further.

Top 10 groups announcing DDoS attacks were responsible for almost 50% of claims (Source – Kela)

Beyond code obfuscation, AI-powered threat actors are also employing strategies for persistence, allowing malware to remain on infected systems without detection for extended periods.

These tactics include the use of AI to monitor system health and only activate malicious operations when the device is idle, reducing activity that might alert monitoring tools.

This development not only complicates the process of detection but also extends the period during which the infected machine can be utilized for malicious purposes.

The intersection of AI and cybersecurity reveals a grim challenge for defenders.

With the exponential increase in malicious AI tools and the parallel rise in efforts to exploit legitimate AI systems for unethical purposes, robust mitigation strategies are more important than ever.

As researchers and cybersecurity professionals work to counter these evolving threats, the industry must prioritize collaboration and innovation in defense mechanisms to keep pace with attackers.

Investigate Real-World Malicious Links & Phishing Attacks With Threat Intelligence Lookup - Try for Free

Tushar Subhra Dutta
Tushar is a Cyber security content editor with a passion for creating captivating and informative content. With years of experience under his belt in Cyber Security, he is covering Cyber Security News, technology and other news.