liu.seSök publikationer i DiVA
Ändra sökning
Avgränsa sökresultatet
1 - 16 av 16
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Träffar per sida
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sortering
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Disputationsdatum (tidigaste först)
  • Disputationsdatum (senaste först)
Markera
Maxantalet träffar du kan exportera från sökgränssnittet är 250. Vid större uttag använd dig av utsökningar.
  • 1.
    Hayes, Conor F.
    et al.
    University of Galway.
    Rădulescu, Roxana
    Vrije Universiteit Brussel.
    Bargiacchi, Eugenio
    Vrije Universiteit Brussel.
    Källström, Johan
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Macfarlane, Matthew
    University of Amsterdam.
    Reymond, Mathieu
    Vrije Universiteit Brussel.
    Verstraeten, Timothy
    Vrije Universiteit Brussel.
    Zintgraf, Luisa M.
    University of Oxford.
    Dazeley, Richard
    Deakin University.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Howley, Enda
    University of Galway.
    Irissappane, Athirai A.
    Amazon.
    Mannion, Patrick
    University of Galway.
    Nowé, Ann
    Vrije Universiteit Brussel.
    Ramos, Gabriel
    University of Vale do Rio dos Sinos.
    Restelli, Marcello
    Politecnico di Milano.
    Vamplew, Peter
    Federation University.
    Roijers, Diederik M.
    City of Amsterdam.
    A Brief Guide to Multi-Objective Reinforcement Learning and Planning2023Ingår i: Proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS) / [ed] A. Ricci, W. Yeoh, N. Agmon, B. An, 2023, s. 1988-1990Konferensbidrag (Refereegranskat)
    Abstract [en]

    Real-world sequential decision-making tasks are usually complex, and require trade-offs between multiple–often conflicting–objectives. However, the majority of research in reinforcement learning (RL) and decision-theoretic planning assumes a single objective, or that multiple objectives can be handled via a predefined weighted sum over the objectives. Such approaches may oversimplify the underlying problem, and produce suboptimal results. This extended abstract outlines the limitations of using a semi-blind iterative process to solve multi-objective decision making problems. Our extended paper [4], serves as a guide for the application of explicitly multi-objective methods to difficult problems.

  • 2.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Model-Based Actor-Critic for Multi-Objective Reinforcement Learning with Dynamic Utility Functions2023Ingår i: Proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS), 2023, s. 2818-2820Konferensbidrag (Refereegranskat)
    Abstract [en]

    Many real-world problems require a trade-off between multiple conflicting objectives. Decision-makers’ preferences over solutions to such problems are determined by their utility functions, which convert multi-objective values to scalars. In some settings, utility functions change over time, and the goal is to find methods that can efficiently adapt an agent’s policy to changes in utility. Previous work on learning with dynamic utility functions has focused on model-free methods, which often suffer from poor sample efficiency. In this work, we instead propose a model-based actor-critic, which explores with diverse utility functions through imagined rollouts within a learned world model between interactions with the real environment. An experimental evaluation on Minecart, a well-known benchmark for multi-objective reinforcement learning, shows that by learning a model of the environment the quality of the agent’s policy is improved compared to model-free algorithms.

  • 3.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Model-Based Multi-Objective Reinforcement Learning with Dynamic Utility Functions2023Ingår i: Proceedings of the Adaptive and Learning Agents Workshop (ALA) at AAMAS 2023, 2023, s. 1-9Konferensbidrag (Refereegranskat)
    Abstract [en]

    Many real-world problems require a trade-off between multiple conflicting objectives. Decision-makers’ preferences over solutions to such problems are determined by their utility functions, which convert multi-objective values to scalars. In some settings, utility functions change over time, and the goal is to find methods that can efficiently adapt an agent’s policy to changes in utility. Previous work on learning with dynamic utility functions has focused on model-free methods, which often suffer from poor sample efficiency. In this work, we instead propose a model-based actor-critic, which explores with diverse utility functions through imagined rollouts within a learned world model between interactions with the real environment. An experimental evaluation shows that by learning a model of the environment the performance of the agent can be improved compared to model-free algorithms.

  • 4. Beställ onlineKöp publikationen >>
    Källström, Johan
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Reinforcement Learning for Improved Utility of Simulation-Based Training2023Doktorsavhandling, monografi (Övrigt vetenskapligt)
    Abstract [en]

    Team training in complex domains often requires a substantial number of resources, e.g. vehicles, machines, and role-players. For this reason, it may be difficult to realise efficient and effective training scenarios in a real-world setting. Instead, part of the training can be conducted in synthetic, computer-generated environments. In these environments trainees can operate simulators instead of real vehicles, while synthetic actors can replace human role-players to increase the complexity of the simulated scenario at low operating cost. However, constructing behaviour models for synthetic actors is challenging, especially for the end users, who typically do not have expertise in artificial intelligence. In this dissertation, we study how machine learning can be used to simplify the construction of intelligent agents for simulation-based training. A simulation-based air combat training system is used as case study. 

    The contributions of the dissertation are divided into two parts. The first part aims at improving the understanding of reinforcement learning in the domain of simulation-based training. First, a user-study is conducted to identify important capabilities and characteristics of learning agents that are intended to support training of fighter pilots. It is identified that one of the most important capabilities of learning agents in the context of simulation-based training is that their behaviour can be adapted to different phases of training, as well as to the training needs of individual human trainees. Second, methods for learning how to coordinate with other agents are studied in simplified training scenarios, to investigate how the design of the agent’s observation space, action space, and reward signal affects the performance of learning. It is identified that temporal abstractions and hierarchical reinforcement learning can improve the efficiency of learning, while also providing support for modelling of doctrinal behaviour. In more complex settings, curriculum learning and related methods are expected to help find novel tactics even when sparse, abstract reward signals are used. Third, based on the results from the user study and the practical experiments, a system concept for a user-adaptive training system is developed to support further research. 

    The second part of the contributions focuses on methods for utility-based multi-objective reinforcement learning, which incorporates knowledge of the user’s utility function in the search for policies that balance multiple conflicting objectives. Two new agents for multi-objective reinforcement learning are proposed: the Tunable Actor (T-Actor) and the Multi-Objective Dreamer (MO-Dreamer). T-Actor provides decision support to instructors by learning a set of Pareto optimal policies, represented by a single neural network conditioned on objective preferences. This enables tuning of the agent’s behaviour to fit trainees’ current training needs. Experimental evaluations in gridworlds and in the target system show that T-Actor reduces the number of training steps required for learning. MO-Dreamer adapts online to changes in users’ utility, e.g. changes in training needs. It does so by learning a model of the environment, which it can use for anticipatory rollouts with a diverse set of utility functions to explore which policy to follow to optimise the return for a given set of objective preferences. An experimental evaluation shows that MO-Dreamer outperforms prior model-free approaches in terms of experienced regret, for frequent as well as sparse changes in utility. 

    Overall, the research conducted in this dissertation contributes to improved knowledge about how to apply machine learning methods to construction of simulation-based training environments. While our focus was on air combat training, the results are general enough to be applicable in other domains. 

    Ladda ner fulltext (pdf)
    fulltext
    Ladda ner (png)
    presentationsbild
  • 5.
    Vamplew, Peter
    et al.
    Federation University Australia.
    Smith, Benjamin J.
    University of Oregon.
    Källström, Johan
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Ramos, Gabriel
    Universidade do Vale do Rio dos Sinos.
    Rădulescu, Roxana
    Vrije Universiteit Brussel.
    Roijers, Diederik M.
    Vrije Universiteit Brussel.
    Hayes, Conor F.
    University of Galway.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Mannion, Patrick
    University of Galway.
    Libin, Pieter J.K.
    Vrije Universiteit Brussel.
    Dazeley, Richard
    Deakin University.
    Foale, Cameron
    Federation University Australia.
    Scalar Reward is Not Enough2023Ingår i: Proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS), 2023, s. 839-841Konferensbidrag (Refereegranskat)
    Abstract [en]

    Silver et al.[14] posit that scalar reward maximisation is sufficient to underpin all intelligence and provides a suitable basis for artificial general intelligence (AGI). This extended abstract summarises the counter-argument from our JAAMAS paper [19].

  • 6.
    Hayes, Conor F.
    et al.
    National University of Ireland Galway, Galway, Ireland.
    Rădulescu, Roxana
    AI Lab, Vrije Universiteit Brussel, Brussels, Belgium.
    Bargiacchi, Eugenio
    AI Lab, Vrije Universiteit Brussel, Brussels, Belgium.
    Källström, Johan
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Macfarlane, Matthew
    AMLAB, University of Amsterdam, Amsterdam, The Netherlands.
    Reymond, Mathieu
    AI Lab, Vrije Universiteit Brussel, Brussels, Belgium.
    Verstraeten, Timothy
    AI Lab, Vrije Universiteit Brussel, Brussels, Belgium.
    Zintgraf, Luisa M.
    WhiRL, University of Oxford, Oxford, United Kingdom.
    Dazeley, Richard
    Deakin University, Geelong, Australia.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Howley, Enda
    National University of Ireland Galway, Galway, Ireland.
    Irissappane, Athirai A.
    University of Washington (Tacoma), Tacoma, USA.
    Mannion, Patrick
    National University of Ireland Galway, Galway, Ireland.
    Nowé, Ann
    AI Lab, Vrije Universiteit Brussel, Brussels, Belgium.
    Ramos, Gabriel
    Universidade do Vale do Rio dos Sinos, São Leopoldo, RS, Brazil.
    Restelli, Marcello
    Politecnico di Milano, Milan, Italy.
    Vamplew, Peter
    Federation University Australia, Ballarat, Australia.
    Roijers, Diederik M.
    HU University of Applied Sciences Utrecht, Utrecht, The Netherlands.
    A practical guide to multi-objective reinforcement learning and planning2022Ingår i: Autonomous Agents and Multi-Agent Systems, ISSN 1387-2532, E-ISSN 1573-7454, Vol. 36, nr 1, artikel-id 26Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Real-world sequential decision-making tasks are generally complex, requiring trade-offs between multiple, often conflicting, objectives. Despite this, the majority of research in reinforcement learning and decision-theoretic planning either assumes only a single objective, or that multiple objectives can be adequately handled via a simple linear combination. Such approaches may oversimplify the underlying problem and hence produce suboptimal results. This paper serves as a guide to the application of multi-objective methods to difficult problems, and is aimed at researchers who are already familiar with single-objective reinforcement learning and planning methods who wish to adopt a multi-objective perspective on their research, as well as practitioners who encounter multi-objective decision problems in practice. It identifies the factors that may influence the nature of the desired solution, and illustrates by example how these influence the design of multi-objective decision-making systems for complex problems.

    Ladda ner fulltext (pdf)
    fulltext
  • 7.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Granlund, R.
    RISE SICS East, Linköping, Sweden.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Design of simulation-based pilot training systems using machine learning agents2022Ingår i: Aeronautical Journal, ISSN 0001-9240, Vol. 126, nr 1300, s. 907-931, artikel-id PII S0001924022000082Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The high operational cost of aircraft, limited availability of air space, and strict safety regulations make training of fighter pilots increasingly challenging. By integrating Live, Virtual, and Constructive simulation resources, efficiency and effectiveness can be improved. In particular, if constructive simulations, which provide synthetic agents operating synthetic vehicles, were used to a higher degree, complex training scenarios could be realised at low cost, the need for support personnel could be reduced, and training availability could be improved. In this work, inspired by the recent improvements of techniques for artificial intelligence, we take a user perspective and investigate how intelligent, learning agents could help build future training systems. Through a domain analysis, a user study, and practical experiments, we identify important agent capabilities and characteristics, and then discuss design approaches and solution concepts for training systems to utilise learning agents for improved training value.

    Ladda ner fulltext (pdf)
    fulltext
  • 8.
    Vamplew, Peter
    et al.
    Federation University Australia, Ballarat, Australia.
    Smith, Benjamin J.
    Center for Translational Neuroscience, University of Oregon, Eugene, OR, USA.
    Källström, Johan
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Ramos, Gabriel
    Universidade do Vale do Rio dos Sinos, São Leopoldo, RS, Brazil.
    Rădulescu, Roxana
    AI Lab, Vrije Universiteit Brussel, Brussel, Belgium.
    Roijers, Diederik M.
    Vrije Universiteit Brussel, Brussel, Belgium and HU University of Applied Sciences Utrecht, Utrecht, The Netherlands.
    Hayes, Conor F.
    National University of Ireland Galway, Galway, Ireland.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Mannion, Patrick
    National University of Ireland Galway, Galway, Ireland.
    Libin, Pieter J. K.
    Vrije Universiteit Brussel, Brussel, Belgium and Universiteit Hasselt, Hasselt, Belgium and Katholieke Universiteit Leuven, Leuven, Belgium.
    Dazeley, Richard
    Deakin University, Geelong, Australia.
    Foale, Cameron
    Federation University Australia, Ballarat, Australia.
    Scalar reward is not enough: a response to Silver, Singh, Precup and Sutton (2021)2022Ingår i: Autonomous Agents and Multi-Agent Systems, ISSN 1387-2532, E-ISSN 1573-7454, Vol. 36, nr 2, artikel-id 41Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The recent paper “Reward is Enough” by Silver, Singh, Precup and Sutton posits that the concept of reward maximisation is sufficient to underpin all intelligence, both natural and artificial, and provides a suitable basis for the creation of artificial general intelligence. We contest the underlying assumption of Silver et al. that such reward can be scalar-valued. In this paper we explain why scalar rewards are insufficient to account for some aspects of both biological and computational intelligence, and argue in favour of explicitly multi-objective models of reward maximisation. Furthermore, we contend that even if scalar reward functions can trigger intelligent behaviour in specific cases, this type of reward is insufficient for the development of human-aligned artificial general intelligence due to unacceptable risks of unsafe or unethical behaviour.

    Ladda ner fulltext (pdf)
    fulltext
  • 9.
    Källström, Johan
    et al.
    Linköpings universitet, Tekniska fakulteten. Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem.
    Granlund, Rego
    RISE SICS East.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Design of Simulation-Based Pilot Training Systems using Machine Learning Agents2021Ingår i: Proceedings of the 32nd Congress of the International Council of Aeronautical Sciences (ICAS), Bonn: The International Council of the Aeronautical Sciences , 2021, Vol. 32, artikel-id ICAS_2020_0130Konferensbidrag (Refereegranskat)
    Abstract [en]

    The high operational cost of aircraft, limited availability of air space, and strict safety regulations make training of fighter pilots increasingly challenging. By integrating Live, Virtual, and Constructive simulation resources, efficiency and effectiveness can be improved. In particular, if constructive simulations, which provide synthetic agents operating synthetic vehicles, were used to a higher degree, complex training scenarios could be realized at low cost, the need for support personnel could be reduced, and training availability could be improved. In this work, inspired by the recent improvements of techniques for artificial intelligence, we take a user perspective and investigate how intelligent, learning agents could help build future training systems. Through a domain analysis, a user study, and practical experiments, we identify important agent capabilities and characteristics, and then discuss design approaches and solution concepts for training systems to utilize learning agents for improved training value.

  • 10.
    Källström, Johan
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Adaptive Agent-Based Simulation for Individualized Training2020Ingår i: Proceedings of the 19th International Conference on Autonomous Agents and Multiagent Systems (AAMAS) / [ed] B. An, N. Yorke-Smith, A. El Fallah Seghrouchni, G. Sukthankar, International Foundation for Autonomous Agents and Multiagent Systems (www.ifaamas.org) , 2020, s. 2193-2195Konferensbidrag (Refereegranskat)
    Abstract [en]

    Agent-based simulation can be used for efficient and effective training of human operators and decision-makers. However, constructing realistic behavior models for the agents is challenging and time-consuming, especially for subject matter experts, who may not have expertise in artificial intelligence. In this work, we investigate how machine learning can be used to adapt simulation contents to the current needs of individual trainees. Our initial results demonstrate that multi-objective multi-agent reinforcement learning is a promising approach for creating agents with diverse and adaptive characteristics, which can stimulate humans in training.

    Ladda ner fulltext (pdf)
    fulltext
  • 11.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Agent Coordination in Air Combat Simulation using Multi-Agent Deep Reinforcement Learning2020Ingår i: 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), IEEE conference proceedings, 2020, s. 2157-2164Konferensbidrag (Refereegranskat)
    Abstract [en]

    Simulation-based training has the potential to significantly improve training value in the air combat domain. However, synthetic opponents must be controlled by high-quality behavior models, in order to exhibit human-like behavior. Building such models by hand is recognized as a very challenging task. In this work, we study how multi-agent deep reinforcement learning can be used to construct behavior models for synthetic pilots in air combat simulation. We empirically evaluate a number of approaches in two air combat scenarios, and demonstrate that curriculum learning is a promising approach for handling the high-dimensional state space of the air combat domain, and that multi-objective learning can produce synthetic agents with diverse characteristics, which can stimulate human pilots in training.

    Ladda ner fulltext (pdf)
    fulltext
  • 12.
    Domova, Veronika
    et al.
    Linköpings universitet, Institutionen för teknik och naturvetenskap, Medie- och Informationsteknik. Linköpings universitet, Tekniska fakulteten.
    Gärtner, Erik
    Lund University, Lund, Sweden.
    Präntare, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Pallin, Martin
    Royal Institute of Technology, Stockholm, Sweden.
    Källström, Johan
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Korzhitskii, Nikita
    Linköpings universitet, Institutionen för datavetenskap, Databas och informationsteknik. Linköpings universitet, Tekniska fakulteten.
    Improving Usability of Search and Rescue Decision Support Systems: WARA-PS Case Study2020Ingår i: In proceedings of the 2020 25th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Vienna, Austria: IEEE conference proceedings, 2020, s. 1251-1254Konferensbidrag (Refereegranskat)
    Abstract [en]

    Novel autonomous search and rescue systems, although powerful, still require a human decision-maker involvement. In this project, we focus on the human aspect of one such novel autonomous SAR system. Relying on the knowledge gained in a field study, as well as through the literature, we introduced several extensions to the system that allowed us to achieve a more user-centered interface. In the evaluation session with a rescue service specialist, we received positive feedback and defined potential directions for future work.

    Ladda ner fulltext (pdf)
    fulltext
  • 13.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Learning Agents for Improved Efficiency and Effectiveness in Simulation-Based Training2020Ingår i: Poceedings of the 32nd annual workshop of the Swedish Artificial Intelligence Society (SAIS), 2020, s. 1-2Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    Team training in complex domains often requires a substantial amount of resources, e.g., instructors, role-players and vehicles. For this reason, it may be difficult to realize efficient and effective training scenarios in a real-world setting. Instead, intelligent agents can be used to construct synthetic, simulationbased training environments. However, building behavior models for such agents is challenging, especially for the end-users of the training systems, who typically do not have expertise in artificial intelligence. In this PhD project, we study how machine learning can be used to simplify the process of constructing agents for simulation-based training. As a case study we use a simulation-based air combat training system. By constructing smarter synthetic agents the dependency on human training providers can be reduced, and the availability as well as the quality of training can be improved.

  • 14.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Multi-Agent Multi-Objective Deep Reinforcement Learning for Efficient and Effective Pilot Training2019Ingår i: Proceedings of the 10th Aerospace Technology Congress (FT) / [ed] Ingo Staack and Petter Krus, 2019, s. 101-111Konferensbidrag (Refereegranskat)
    Abstract [en]

    The tactical systems and operational environment of modern fighter aircraft are becoming increasingly complex. Creating a realistic and relevant environment for pilot training using only live aircraft is difficult, impractical and highly expensive. The Live, Virtual and Constructive (LVC) simulation paradigm aims to address this challenge. LVC simulation means linking real aircraft, ground-based systems and soldiers (Live), manned simulators (Virtual) and computer controlled synthetic entities (Constructive). Constructive simulation enables realization of complex scenarios with a large number of autonomous friendly, hostile and neutral entities, which interact with each other as well as manned simulators and real systems. This reduces the need for personnel to act as role-players through operation of e.g. live or virtual aircraft, thus lowering the cost of training. Constructive simulation also makes it possible to improve the availability of training by embedding simulation capabilities in live aircraft, making it possible to train anywhere, anytime. In this paper we discuss how machine learning techniques can be used to automate the process of constructing advanced, adaptive behavior models for constructive simulations, to improve the autonomy of future training systems. We conduct a number of initial experiments, and show that reinforcement learning, in particular multi-agent and multi-objective deep reinforcement learning, allows synthetic pilots to learn to cooperate and prioritize among conflicting objectives in air combat scenarios. Though the results are promising, we conclude that further algorithm development is necessary to fully master the complex domain of air combat simulation.

    Ladda ner fulltext (pdf)
    fulltext
  • 15.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Reinforcement Learning for Computer Generated Forces using Open-Source Software2019Ingår i: Proceedings of the 2019 Interservice/Industry Training, Simulation, and Education Conference (IITSEC), 2019, s. 1-11, artikel-id 19197Konferensbidrag (Refereegranskat)
    Abstract [en]

    The creation of behavior models for computer generated forces (CGF) is a challenging and time-consuming task, which often requires expertise in programming of complex artificial intelligence algorithms. This makes it difficult for a subject matter expert with knowledge about the application domain and the training goals to build relevant scenarios and keep the training system in pace with training needs. In recent years, machine learning has shown promise as a method for building advanced decision-making models for synthetic agents. Such agents have been able to beat human champions in complex games such as poker, Go and StarCraft. There is reason to believe that similar achievements are possible in the domain of military simulation. However, in order to efficiently apply these techniques, it is important to have access to the right tools, as well as knowledge about the capabilities and limitations of algorithms.   

    This paper discusses efficient applications of deep reinforcement learning, a machine learning technique that allows synthetic agents to learn how to achieve their goals by interacting with their environment. We begin by giving an overview of available open-source frameworks for deep reinforcement learning, as well as libraries with reference implementations of state-of-the art algorithms. We then present an example of how these resources were used to build a reinforcement learning environment for a CGF software intended to support training of fighter pilots. Finally, based on our exploratory experiments in the presented environment, we discuss opportunities and challenges related to the application of reinforcement learning techniques in the domain of air combat training systems, with the aim to efficiently construct high quality behavior models for computer generated forces.

  • 16.
    Källström, Johan
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Heintz, Fredrik
    Linköpings universitet, Institutionen för datavetenskap, Artificiell intelligens och integrerade datorsystem. Linköpings universitet, Tekniska fakulteten.
    Tunable Dynamics in Agent-Based Simulation using Multi-Objective Reinforcement Learning2019Ingår i: Proceedings of the 2019 Adaptive and Learning Agents Workshop (ALA), 2019, 2019, s. 1-7Konferensbidrag (Refereegranskat)
    Abstract [en]

    Agent-based simulation is a powerful tool for studying complex systems of interacting agents. To achieve good results, the behavior models used for the agents must be of high quality. Traditionally these models have been handcrafted by domain experts. This is a difficult, expensive and time consuming process. In contrast, reinforcement learning allows agents to learn how to achieve their goals by interacting with the environment. However, after training the behavior of such agents is often static, i.e. it can no longer be affected by a human. This makes it difficult to adapt agent behavior to specific user needs, which may vary among different runs of the simulation. In this paper we address this problem by studying how multi-objective reinforcement learning can be used as a framework for building tunable agents, whose characteristics can be adjusted at runtime to promote adaptiveness and diversity in agent-based simulation. We propose an agent architecture that allows us to adapt popular deep reinforcement learning algorithms to multi-objective environments. We empirically show that our method allows us to train tunable agents that can approximate the policies of multiple species of agents.

    Ladda ner fulltext (pdf)
    Tunable Dynamics in Agent-Based Simulation using Multi-Objective Reinforcement Learning
1 - 16 av 16
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf