ResearchSpace

Applying reinforcement learning to the weapon assignment problem in air defence

Show simple item record

dc.contributor.author Mouton, H
dc.contributor.author Roodt, J
dc.contributor.author Le Roux, H
dc.date.accessioned 2011-11-28T11:37:21Z
dc.date.available 2011-11-28T11:37:21Z
dc.date.issued 2011-12
dc.identifier.citation Mouton, H, Roodt, J, Le Roux, H. 2011. Applying reinforcement learning to the weapon assignment problem in air defence. Scientia Militaria, South African Journal of Military Studies, Vol 39(2), pp 1-15 en_US
dc.identifier.issn 2224-0020
dc.identifier.uri http://scientiamilitaria.journals.ac.za/pub/article/view/115
dc.identifier.uri http://hdl.handle.net/10204/5342
dc.description Scientia Militaria, South African Journal of Military Studies, Vol 39(2), pp 123-140 en_US
dc.description.abstract The modern battlefield is a fast-paced, information-rich environment, where discovery of intent, situation awareness and the rapid evolution of concepts of operation and doctrine are critical success factors. A combination of the techniques investigated and tested in this work, together with other techniques in Artificial Intelligence (AI) and modern computational techniques, may hold the key to relieving the burden of the decision-maker and aiding in better decision-making under pressure. The techniques investigated in this article were two methods from the machine-learning subfield of reinforcement learning (RL), namely a Monte Carlo (MC) control algorithm with exploring starts (MCES), and an off-policy temporal-difference (TD) learning-control algorithm, Q-learning. These techniques were applied to a simplified version of the weapon assignment (WA) problem in air defence. The MCES control algorithm yielded promising results when searching for an optimal shooting order. A greedy approach was taken in the Q-learning algorithm, but experimentation showed that the MCES-control algorithm still performed significantly better than the Q-learning algorithm, even though it was slower. en_US
dc.language.iso en en_US
dc.publisher Scientia Militaria: Stellenbosch University en_US
dc.relation.ispartofseries Workflow request;7607
dc.subject Weapon assignment en_US
dc.subject Air defence en_US
dc.subject Reinforcement learning en_US
dc.subject Q-learning algorithm en_US
dc.subject Military en_US
dc.title Applying reinforcement learning to the weapon assignment problem in air defence en_US
dc.type Article en_US
dc.identifier.apacitation Mouton, H., Roodt, J., & Le Roux, H. (2011). Applying reinforcement learning to the weapon assignment problem in air defence. http://hdl.handle.net/10204/5342 en_ZA
dc.identifier.chicagocitation Mouton, H, J Roodt, and H Le Roux "Applying reinforcement learning to the weapon assignment problem in air defence." (2011) http://hdl.handle.net/10204/5342 en_ZA
dc.identifier.vancouvercitation Mouton H, Roodt J, Le Roux H. Applying reinforcement learning to the weapon assignment problem in air defence. 2011; http://hdl.handle.net/10204/5342. en_ZA
dc.identifier.ris TY - Article AU - Mouton, H AU - Roodt, J AU - Le Roux, H AB - The modern battlefield is a fast-paced, information-rich environment, where discovery of intent, situation awareness and the rapid evolution of concepts of operation and doctrine are critical success factors. A combination of the techniques investigated and tested in this work, together with other techniques in Artificial Intelligence (AI) and modern computational techniques, may hold the key to relieving the burden of the decision-maker and aiding in better decision-making under pressure. The techniques investigated in this article were two methods from the machine-learning subfield of reinforcement learning (RL), namely a Monte Carlo (MC) control algorithm with exploring starts (MCES), and an off-policy temporal-difference (TD) learning-control algorithm, Q-learning. These techniques were applied to a simplified version of the weapon assignment (WA) problem in air defence. The MCES control algorithm yielded promising results when searching for an optimal shooting order. A greedy approach was taken in the Q-learning algorithm, but experimentation showed that the MCES-control algorithm still performed significantly better than the Q-learning algorithm, even though it was slower. DA - 2011-12 DB - ResearchSpace DP - CSIR KW - Weapon assignment KW - Air defence KW - Reinforcement learning KW - Q-learning algorithm KW - Military LK - https://researchspace.csir.co.za PY - 2011 SM - 2224-0020 T1 - Applying reinforcement learning to the weapon assignment problem in air defence TI - Applying reinforcement learning to the weapon assignment problem in air defence UR - http://hdl.handle.net/10204/5342 ER - en_ZA


Files in this item

This item appears in the following Collection(s)

Show simple item record