ARACNE – LLM-based Pentesting Agent To Execute Commands on Real Linux Shell Systems

Cybersecurity researchers have unveiled a new autonomous penetration testing agent that leverages large language models (LLMs) to execute commands on real Linux shell systems.

ARACNE, as the agent is called, represents a significant advancement in automated security testing, demonstrating the potential for AI to both strengthen and potentially compromise digital infrastructure.

The agent connects to remote SSH services autonomously and executes commands to achieve specified penetration goals without human intervention.

Google News

Unlike traditional penetration testing tools that require manual operation, ARACNE plans attacks, generates shell commands, and evaluates outputs entirely on its own.

This capability showcases how LLMs can be applied to complex cybersecurity tasks with minimal oversight, raising important questions about both defensive applications and potential misuse.

Initial testing has shown ARACNE achieving a 60% success rate against autonomous defenders and nearly 58% against capture-the-flag challenges, outperforming previous state-of-the-art automated penetration testing systems.

When successful, the agent typically accomplishes its goals in fewer than five commands, demonstrating remarkable efficiency.

Analysts from the Czech Technical University in Prague, led by researchers Tomas Nieponice, Veronica Valeros, and Sebastian Garcia, identified that ARACNE’s effectiveness stems from its novel multi-LLM architecture.

The team noted this approach provides greater flexibility than existing systems while reducing reliance on extensive knowledge retrieval mechanisms.

ARACNE’s architecture

ARACNE’s architecture consists of four key components working in tandem: a planner module that creates attack strategies, an interpreter that translates plans into executable Linux commands, an optional summarizer to condense context, and a core agent that orchestrates the process and interacts with target systems.

ARACNE architecture and connection (Source – Arxiv)

This modular design enables the system to leverage different specialized LLM models for specific tasks.

The agent employs a particularly concerning jailbreak technique that effectively bypasses ethical guardrails built into commercial LLMs.

By instructing the models to “play as” an attacker in a simulated environment with “no real outcomes,” the system circumvents safety measures with approximately 95% effectiveness.

This technique, while essential for legitimate penetration testing, demonstrates how easily existing safeguards in AI systems can be bypassed.

Investigate Real-World Malicious Links & Phishing Attacks With Threat Intelligence Lookup - Try for Free

Tushar Subhra Dutta
Tushar is a Cyber security content editor with a passion for creating captivating and informative content. With years of experience under his belt in Cyber Security, he is covering Cyber Security News, technology and other news.