Lidé

Ing. Ondřej Lukáš

Všechny publikace

Bridging the Explanation Gap in AI Security: A Task-Driven Approach to XAI Methods Evaluation

  • Autoři: Ing. Ondřej Lukáš, Ing. Sebastián García, Ph.D.,
  • Publikace: Proceedings of the 16th International Conference on Agents and Artificial Intelligence. Setúbal: Science and Technology Publications, Lda, 2024. p. 1370-1377. vol. 3. ISSN 2184-433X. ISBN 978-989-758-680-4.
  • Rok: 2024
  • DOI: 10.5220/0012475200003636
  • Odkaz: https://doi.org/10.5220/0012475200003636
  • Pracoviště: Katedra počítačů, Centrum umělé inteligence
  • Anotace:
    Deciding which XAI technique is best depends not only on the domain, but also on the given task, the dataset used, the model being explained, and the target goal of that model. We argue that the evaluation of XAI methods has not been thoroughly analyzed in the network security domain, which presents a unique type of challenge. While there are XAI methods applied in network security there is still a large gap between the needs of security stakeholders and the selection of the optimal method. We propose to approach the problem by first defining the stack-holders in security and their prototypical tasks. Each task defines inputs and specific needs for explanations. Based on these explanation needs (e.g. understanding the performance, or stealing a model), we created five XAI evaluation techniques that are used to compare and select which XAI method is best for each task (dataset, model, and goal). Our proposed approach was evaluated by running experiments for different security stakehol ders, machine learning models, and XAI methods. Results were compared with the AutoXAI technique and random selection. Results show that our proposal to evaluate and select XAI methods for network security is well-grounded and that it can help AI security practitioners find better explanations for their given tasks.

Out of the Cage: How Stochastic Parrots Win in Cyber Security Environments

  • DOI: 10.5220/0012391800003636
  • Odkaz: https://doi.org/10.5220/0012391800003636
  • Pracoviště: Katedra počítačů, Centrum umělé inteligence
  • Anotace:
    Large Language Models (LLMs) have gained widespread popularity across diverse domains involving text generation, summarization, and various natural language processing tasks. Despite their inherent limitations, LLM-based designs have shown promising capabilities in planning and navigating open-world scenarios. This paper introduces a novel application of pre-trained LLMs as agents within cybersecurity network environments, focusing on their utility for sequential decision-making processes. We present an approach wherein pre-trained LLMs are leveraged as attacking agents in two reinforcement learning environments. Our proposed agents demonstrate similar or better performance against state-of-the-art agents trained for thousands of episodes in most scenarios and configurations. In addition, the best LLM agents perform similarly to human testers of the environment without any additional training process. This design highlights the potential of LLMs to address complex decision-making tasks within cybersecurity efficiently. Furthermore, we introduce a new network security environment named NetSecGame. The environment is designed to support complex multi-agent scenarios within the network security domain eventually. The proposed environment mimics real network attacks and is designed to be highly modular and adaptable for various scenarios.

Catch Me if You Can: Improving Adversaries in Cyber-Security with Q-Learning Algorithms

  • DOI: 10.5220/0011684500003393
  • Odkaz: https://doi.org/10.5220/0011684500003393
  • Pracoviště: Katedra počítačů, Centrum umělé inteligence
  • Anotace:
    The ongoing rise in cyberattacks and the lack of skilled professionals in the cybersecurity domain to combat these attacks show the need for automated tools capable of detecting an attack with good performance. Attackers disguise their actions and launch attacks that consist of multiple actions, which are difficult to detect. Therefore, improving defensive tools requires their calibration against a well-trained attacker. In this work, we propose a model of an attacking agent and environment and evaluate its performance using basic Q-Learning, Naive Q-learning, and DoubleQ-Learning, all of which are variants of Q-Learning. The attacking agent is trained with the goal of exfiltrating data whereby all the hosts in the network have a non-zero detection probability. Results show that the DoubleQ-Learning agent has the best overall performance rate by successfully achieving the goal in 70% of the interactions.

Disrupting Active Directory Attacks with Deep Learning for Organic Honeyuser Placement

  • DOI: 10.1007/978-3-031-37320-6_6
  • Odkaz: https://doi.org/10.1007/978-3-031-37320-6_6
  • Pracoviště: Katedra počítačů, Centrum umělé inteligence
  • Anotace:
    Honeypots have been a long-established form of passive defense in a wide variety of systems. They are often used for the reliability and low false positive rate. However, the deployment of honeypots in the Active Directory (AD) systems is still limited. Intrusion detection in AD systems is a difficult task due to the complexity of the system and its design, where any authenticated account is able to query other entities in the system. Therefore, the positioning of the honeypot in such structures brings two main con trains: (i) the placement has to be organic, with similar properties to other, real entities in the structure, and (ii) the placement must not give away the nature of the honeypot to the attacker. In this work, we present a model based on a variational autoencoder capable of producing organic placements for AD structures. We show that the proposed model is capable of learning meaningful latent representations of the nodes in the AD structures and predicting new node placement with similar properties. Analysis of the latent space shows that the model can capture complex relationships between nodes with low-dimensional latent space. Our method is evaluated based on the (i) similarity with the input graphs, (ii) properties of the generated nodes, and (iii) comparison with other generative graph models. Further experiments with human attackers show that the proposed method outperforms the random honeypot placement baseline.

Deep generative models to extend active directory graphs with honeypot users

  • Autoři: Ing. Ondřej Lukáš, Ing. Sebastián García, Ph.D.,
  • Publikace: Proceedings of the 2nd International Conference on Deep Learning Theory and Applications - DeLTA. Porto: SciTePress - Science and Technology Publications, 2021. p. 140-147. vol. 1. ISSN 2184-9277. ISBN 978-989-758-526-5.
  • Rok: 2021
  • DOI: 10.5220/0010556601400147
  • Odkaz: https://doi.org/10.5220/0010556601400147
  • Pracoviště: Katedra počítačů, Centrum umělé inteligence
  • Anotace:
    Active Directory (AD) is a crucial element of large organizations, given its central role in managing access to resources. Since AD is used by all users in the organization, it is hard to detect attackers. We propose to generate and place fake users (honeyusers) in AD structures to help detect attacks. However, not any honeyuser will attract attackers. Our method generates honeyusers with a Variational Autoencoder that enriches the AD structure with well-positioned honeyusers. It first learns the embeddings of the original nodes and edges in the AD, then it uses a modified Bidirectional DAG-RNN to encode the parameters of the probability distribution of the latent space of node representations. Finally, it samples nodes from this distribution and uses an MLP to decide where the nodes are connected. The model was evaluated by the similarity of the generated AD with the original, by the positions of the new nodes, by the similarity with GraphRNN and finally by making real intruders attack the generated AD structure to see if they select the honeyusers. Results show that our machine learning model is good enough to generate well-placed honeyusers for existing AD structures so that intruders are lured into them.

Za stránku zodpovídá: Ing. Mgr. Radovan Suk