Lidé
Ing. Muris Sladić
Všechny publikace
VelLMes: A High-Interaction AI-Based Deception Framework
- Autoři: Ing. Muris Sladić, Ing. Veronica Valeros, Catania, C., Ing. Sebastián García, Ph.D.,
- Publikace: Proceedings of the 10th IEEE European Symposium on Security and Privacy Workshops. Cannes: IEEE Computer Society, 2025. p. 671-679. ISSN 2768-0657. ISBN 979-8-3315-9546-3.
- Rok: 2025
- DOI: 10.1109/EuroSPW67616.2025.00082
- Odkaz: https://doi.org/10.1109/EuroSPW67616.2025.00082
- Pracoviště: Katedra počítačů, Centrum umělé inteligence
-
Anotace:
There are very few SotA deception systems based on Large Language Models. The existing ones are limited only to simulating one type of service, mainly SSH shells. These systems - but also the deception technologies not based on LLMs - lack an extensive evaluation that includes human attackers. Generative AI has recently become a valuable asset for cybersecurity researchers and practitioners, and the field of cyber-deception is no exception. Researchers have demonstrated how LLMs can be leveraged to create realistic-looking honeytokens, fake users, and even simulated systems that can be used as honeypots. This paper presents an AI-based deception framework called VelLMes, which can simulate multiple protocols and services such as SSH Linux shell, MySQL, POP3, and HTTP. All of these can be deployed and used as honeypots, thus VelLMes offers a variety of choices for deception design based on the users' needs. VelLMes is designed to be attacked by humans, so interactivity and realism are key for its performance. We evaluate the generative capabilities and the deception capabilities. Generative capabilities were evaluated using unit tests for LLMs. The results of the unit tests show that, with careful prompting, LLMs can produce realistic-looking responses, with some LLMs having a 100% passing rate. In the case of the SSH Linux shell, we evaluated deception capabilities with 89 human attackers. The attackers interacted with a randomly assigned shell (either honeypot or real) and had to decide if it was a real Ubuntu system or a honeypot. The results showed that about 30% of the attackers thought that they were interacting with a real system when they were assigned an LLM-based honeypot. Lastly, we deployed 10 instances of the SSH Linux shell honeypot on the Internet to capture real-life attacks. Analysis of these attacks showed us that LLM honeypots simulating Linux shells can perform well against unstructured and unexpected attacks on the Internet, responding corr...
LLM in the Shell: Generative Honeypots
- Autoři: Ing. Muris Sladić, Ing. Veronica Valeros, Catania, C., Ing. Sebastián García, Ph.D.,
- Publikace: Proceedings - 9th IEEE European Symposium on Security and Privacy Workshops, Euro S and PW 2024. Cannes: IEEE Computer Society, 2024. p. 430-435. ISSN 2768-0657. ISBN 979-8-3503-6729-4.
- Rok: 2024
- DOI: 10.1109/EuroSPW61312.2024.00054
- Odkaz: https://doi.org/10.1109/EuroSPW61312.2024.00054
- Pracoviště: Katedra počítačů, Centrum umělé inteligence
-
Anotace:
Honeypots are essential tools in cybersecurity for early detection, threat intelligence gathering, and analysis of attacker's behavior. However, most of them lack the required realism to engage and fool human attackers long-term. Being easy to distinguish honeypots strongly hinders their effectiveness. This can happen because they are too deterministic, lack adaptability, or lack deepness. This work introduces shelLM, a dynamic and realistic software honeypot based on Large Language Models that generates Linux-like shell output. We designed and implemented shelLM using cloud-based LLMs. We evaluated if shelLM can generate output as expected from a real Linux shell. The evaluation was done by asking cybersecurity researchers to use the honeypot and give feedback if each answer from the honeypot was the expected one from a Linux shell. Results indicate that shelLM can create credible and dynamic answers capable of addressing the limitations of current honeypots. ShelLM reached a TNR of 0.90, convincing humans it was consistent with a real Linux shell. The source code and prompts for replicating the experiments have been publicly available.