Our participation in this year's edition of the LOD conference, as previously announced in one of our blog post, proved to be an exceptionally enjoyable experience.
The systematic evaluation of auxiliary tasks in reinforcement learning published in “Improving Reinforcement Learning Efficiency with Auxiliary Tasks in Non-Visual Environments: A Comparison” by first author Moritz Lange (Dataninja-colleague from Ruhr University Bochum) generated significant interest, as did my presentation of our work “Ökolopoly: Case Study on Large Action Spaces in Reinforcement Learning”.
The quality of our collaboration in the Dataninja (RL)3-project was acknowledged: we are excited to share that the comparison of auxiliary tasks for RL won the Best Paper Award!
Set against the picturesque backdrop of the Lake District, the conference provided an ideal setting for the thought-provoking keynote speeches that spanned a wide range of topics, from neuroscience to large language models and their applications. The LOD conference is held in conjunction with the Advanced Course & Symposium on Artificial Intelligence & Neuroscience (ACAIN), a collaboration that fosters mutual respect for advancements in each respective field and promotes the exchange of valuable insights, enhancing the experience and value of both conferences.
Beyond the scientific sessions, the hikes in the hills surrounding Lake Grasmere offered a fantastic opportunity for more in-depth discussions about science and life.
View on Lake District. Taken during a hike with my colleague.
From September 6th to September 9th we held our last annual retreat of the Dataninja research training group in Krefeld.
Alongside our invited speaker's talk (Dr. Alessandro Fabris from Max Planck Institute for Security and Privacy) about algorithmic fairness, each PhD candidate from the Dataninja projects presented their progress and current investigations.
In my contribution (Raphael Engelhardt, PhD candidate from TH Köln, Campus Gummersbach, supervised by Prof. Dr. Wolfgang Konen), I spoke about our recent progress on explainability for deep reinforcement learning, published as an open access journal article.
Between talks, enough time was left for valuable discussions and informal exchanges of ideas about new approaches. As usual, meeting the other candidates to talk about the small victories and challenges of our journey towards the PhD was a great pleasure. This year's fun activity required all the combined smartness of students and professors to solve the riddles in the escape rooms.
After the success of these three days, we are especially looking forward to our closing conference next year.
Lake Windermere on a misty morning (By Mkonikkara, CC BY-SA 3.0, via Wikimedia Commons)
For the second time we (Raphael Engelhardt and Wolfgang Konen) have been given the opportunity to present our work at the Conference on machine Learning, Optimization and Data science (LOD) conference.
To this year's 9th edition, held in Grasmere, England, UK on September 22nd - 26th we have the honor to contribute even two papers stemming from the fruitful collaboration with our Dataninja-colleagues at Ruhr-University Bochum, Prof. Laurenz Wiskott and PhD student Moritz Lange:
- Our work entitled “Ökolopoly: Case Study on Large Action Spaces in Reinforcement Learning” describes how we translate the cybernetic board game Ökolopoly into the realm of reinforcement learning and evaluate various methods of handling large observation and action spaces. Large spaces pose a serious challenge to reinforcement learning and we hope our case study will provide valuable approaches to fellow researchers. Additionally we make the environment available to the scientific community with Open AI Gym compatible API.
- “Improving Reinforcement Learning Efficiency with Auxiliary Tasks in Non-Visual Environments: A Comparison”, under the first authorship of Moritz Lange, is a thorough comparison of auxiliary tasks in a variety of control and robotic tasks, and shows how agents benefit from decoupled representation learning of auxiliary tasks in complex environments.
We are very grateful for this opportunity, look forward to hear other researchers’ advances in machine learning, and interesting discussions about current research topics.
We are delighted to announce that our article “Iterative Oblique Decision Trees Deliver Explainable RL Models” was accepted and is now part of the special issue “Advancements in Reinforcement Learning Algorithms” in the MDPI journal Algorithms (impact factor 2.2, CiteScore 3.7) .
Explainability in AI and RL (known as XAI and XRL) becomes increasingly important. In our paper we investigate several possibilities to replace complex “black box” deep reinforcement learning (DRL) models by intrinsically interpretable decision trees (DTs) which require orders of magnitudes fewer parameters. A highlight of our paper is that we find on seven classic control RL problems that the DTs achieve similar reward as the DRL models, sometimes even surpassing the reward of the DRL models. The key to this success is an iterative sampling method that we have developed.
In our work, we present and compare three different methods of collecting samples to train DTs from DRL agents. We test our approaches on seven problems including all classic control environments from Open AI Gym, LunarLander, and the CartPole-SwingUp challenge. Our iterative approach combining exploration of DTs and DRL agent’s predictions, in particular, is able to generate shallow, understandable, oblique DTs that solve the challenges and even outperform the DRL agents they were trained from. Additionally we demonstrate how, given their simpler structure and fewer parameters, DTs allow for inspection and insights, and offer higher degrees of explainability.
To readers interested in explainable AI and understandable reinforcement learning in particular, we recommend to take a look at our open-access article.
The figure shows the decision surfaces of DRL models (1st column) and various DT models (2nd and 3rd column) on the environments MountainCar (upper row) and MountainCarContinuous (lower row). The little black dots visualize various episodes showing how MountainCar rolls back and forth in the valley until it finally reaches the goal on the mountain top (x=0.5). The DRL models exhibit more complicated decision surfaces, while the DT models reach the same performance (number in round brackets in the title) with simpler decision surfaces.
The second edition of the Dataninja Spring-School was held from 8th to 10th of May 2023 in Bielefeld and as a hybrid event. We had the honor and pleasure to attend talks and tutorials from renowned researchers and aspiring young scientists.
We contributed with an extended abstract and our scientific poster “Finding the Relevant Samples for Decision Trees in Reinforcement Learning” presented during Tuesday’s poster session. The opportunity for fruitful discussions and interactions with fellow PhD students from the Dataninja project was much appreciated!
The operation of nuclear power plants (NPPs) is one of the most safety-critical tasks in industry. Prior to using AI methods in this area, it should be thoroughly investigated and evaluated via simulations, whether AI can learn (e.g.´, by reinforcment learning, RL) to power up and shut down a nuclear reactor and how well such an approach meets the safety requirements. This was exactly the task of Niklas Fabig's master thesis which he conducted under the supervision of Prof. Dr. Wolfgang Konen and PhD-candidate Raphael Engelhardt as part of our (RL)^3-project as part of https://dataninja.nrw/. The works uses as a starting point a Java-based NPP simulation tool from Prof. Dr. Benjamin Weyers, University Trier (screenshot example in image). Niklas Fabig constructed first a Java-Python bridge and then conducted over 2000 RL simulation experiments under various settings. He could show that RL algorithms can learn the power-up procedure yielding high returns, but much more research is needed to reliably meet the safety requirements.
The investigation carried out by Niklas Fabig constitutes very interesting and brand-new research in this field, which has now led to winning the 3rd place in the Steinmüller Engineering Award 2023. His supervisor Wolfgang Konen was deeply impressed by the solid, comprehensive and innovative work done by Niklas Fabig and congratulates him warmly. It should be noted, that the master thesis was conducted in the Corona years 2021 - 2022 and so the supervision had to be fully online. Nevertheless, the result of the work and the motivation of Niklas Fabig was by no means less than if the supervision had taken place in presence.
View from Certosa di Pontignano
As previously announced, last week I had the pleasure to present our joint work with our partners from Ruhr-University Bochum on explainable reinforcement learning at the 8th Annual Conference on machine Learning, Optimization and Data science (LOD). The presentation sparked interesting questions and lead to inspiring discussions in the enchanting ambiance of the medieval monastery in Tuscany.
As the conference was held in conjunction with the Advanced Course & Symposium on Artificial Intelligence & Neuroscience (ACAIN), we could profit from a very stimulating interdisciplinary environment with talks, tutorials, and posters covering topics reaching from the biology of neuronal development to implementation details of different deep learning frameworks.
We are looking forward to LOD 2023!
The second edition of the annual Dataninja-Retreat took place this September in Tecklenburg. The different Dataninja projects presented their proceedings, we had the pleasure of attending a lecture by Prof. Xiaoyi Jiang, and fresh PhD graduates of the KI-Starter-project kindly shared experiences and some tips from their recently concluded PhD-journey. Last but not least it was of course a very pleasant and rare occasion for seeing each other offline, exchanging ideas, experiences, struggles, and successes.
Group picture of the participants
We are pleased to announce that we will present our research on explainable reinforcement learning at the 8th Annual Conference on machine Learning, Optimization and Data science (LOD).
Carthusian monastery in Pontignano Siena, Italy. Venue of LOD 2022
Starting with its first edition in 2015, the LOD is an established international and interdisciplinary forum for research and discussion of Deep Learning, Optimization, Big Data, and Artificial Intelligence. This year's 8th edition of the will be held online and onsite in Pontignano near Siena, Italy on September 18th - 22nd 2022.
Reading the conference's manifesto “The problem of understanding intelligence is said to be the greatest problem in science today and ‘the’ problem for this century” we find this prestigious conference to be the perfect place to present our work targeted at making deep reinforcement agents explainable.
We are very grateful for the opportunity to present our paper titled “Sample-based Rule Extraction for Explainable Reinforcement Learning”, which outlines the results of our ongoing research of inducing simple, transparent, human-readable rules from well-trained deep reinforcement learning agents. A link to the article will be added as soon as it is published. For those interested, early registration for the conference is available until Sunday July 31st, 2022.
In September 2021, shortly after the Dataninja retreat, we participated at the KI 2021 – 44th German Conference on Artificial Intelligence.
Alongside fellow members of the Dataninja research training group we, Raphael Engelhardt and Wolfgang Konen from TH Köln together with Laurenz Wiskott and Moritz Lange from RUB Bochum, presented our work on rule extraction from trained reinforcement learning agents in a poster session as part of the workshop "Trustworthy AI in the wild". The conference had to be held virtually which did not affect some very interesting discussions and exchange of new ideas.
Our poster as well as the extended abstract are available online.