Adversarial Deep Reinforcement Learning for Cyber Security in Software Defined Networks

Saved in:
Bibliographic Details
Published in:arXiv.org (Aug 11, 2023), p. n/a
Main Author: Borchjes, Luke
Other Authors: Nyirenda, Clement, Leenen, Louise
Published:
Cornell University Library, arXiv.org
Subjects:
Online Access:Citation/Abstract
Full text outside of ProQuest
Tags: Add Tag
No Tags, Be the first to tag this record!

MARC

LEADER 00000nab a2200000uu 4500
001 2848590964
003 UK-CbPIL
022 |a 2331-8422 
035 |a 2848590964 
045 0 |b d20230811 
100 1 |a Borchjes, Luke 
245 1 |a Adversarial Deep Reinforcement Learning for Cyber Security in Software Defined Networks 
260 |b Cornell University Library, arXiv.org  |c Aug 11, 2023 
513 |a Working Paper 
520 3 |a This paper focuses on the impact of leveraging autonomous offensive approaches in Deep Reinforcement Learning (DRL) to train more robust agents by exploring the impact of applying adversarial learning to DRL for autonomous security in Software Defined Networks (SDN). Two algorithms, Double Deep Q-Networks (DDQN) and Neural Episodic Control to Deep Q-Network (NEC2DQN or N2D), are compared. NEC2DQN was proposed in 2018 and is a new member of the deep q-network (DQN) family of algorithms. The attacker has full observability of the environment and access to a causative attack that uses state manipulation in an attempt to poison the learning process. The implementation of the attack is done under a white-box setting, in which the attacker has access to the defender's model and experiences. Two games are played; in the first game, DDQN is a defender and N2D is an attacker, and in second game, the roles are reversed. The games are played twice; first, without an active causative attack and secondly, with an active causative attack. For execution, three sets of game results are recorded in which a single set consists of 10 game runs. The before and after results are then compared in order to see if there was actually an improvement or degradation. The results show that with minute parameter changes made to the algorithms, there was growth in the attacker's role, since it is able to win games. Implementation of the adversarial learning by the introduction of the causative attack showed the algorithms are still able to defend the network according to their strengths. 
653 |a Algorithms 
653 |a Software-defined networking 
653 |a Deep learning 
653 |a Networks 
653 |a Games 
653 |a Cybersecurity 
700 1 |a Nyirenda, Clement 
700 1 |a Leenen, Louise 
773 0 |t arXiv.org  |g (Aug 11, 2023), p. n/a 
786 0 |d ProQuest  |t Engineering Database 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/2848590964/abstract/embedded/ZKJTFFSVAI7CB62C?source=fedsrch 
856 4 0 |3 Full text outside of ProQuest  |u http://arxiv.org/abs/2308.04909