website/_posts/projects/2020-05-01-FIKS.md
2024-11-10 12:17:01 +01:00

29 lines
2.2 KiB
Markdown
Raw Blame History

This file contains invisible Unicode characters

This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
layout: single
title: "AI-Fusion: Emergence Detection for Mixed MARL Systems."
categories: acoustic anomaly-detection projects
excerpt: "Bringing together agents can be an inherent safety problem. Building the basis to mix and match."
header:
teaser: assets/images/projects/robot.png
---
![logo](\assets\images\projects\robot.png){: .align-left style="padding:0.1em; width:5em"}
In cooperation with [Fraunhofer IKS](https://www.iks.fraunhofer.de/) this project aimed to investigate and detect emergent effects in multi-agent reinforcement learning scenarios, i.e., mixed-vendor autonomous systems (AI fusion).
Emergence in general refers to emerging dynamics of higher complexity (i.e., sum), which are fed by interacting entities (each other and the environment) of a specific complexity level (regarding their policies and capabilities).
![Relation emergence](/assets/images/projects/rel_emergence.png){: .align-center style="padding:0.1em; width:30em"}
In this context, we developed a full-stack, high-performance environment in Python, following the [gymnasium](https://gymnasium.farama.org/main/) specification for the training of reinforcement learning algorithms.
<div class="table-right">
![logo](\assets\images\projects\full_domain.png){: .align-right style="padding:0.5em; width:10em"}
| [GitHub Repo](https://github.com/illiumst/marl-factory-grid/) | [Read-the-docs](https://marl-factory-grid.readthedocs.io/en/latest/)  |
| [Install via PyPI](https://pypi.org/project/Marl-Factory-Grid/) | Read the Paper (TBA)  |
</div>
The main differentiation from already established MARL environments is its ability to employ various scenarios as a combination of `modules` and `configurations`. As well as the option to define per-agent observations, including placeholder and combined observation slices (grid-world). Moreover, this environment can handle multi-agent scenarios as well as sequential actions for inter-step observations.
Furthermore, we designed and implemented a [Unity demonstrator unit](https://github.com/illiumst/F-IKS_demonstrator) that can load and replay specific pre-recorded scenarios. This way, emerging unwanted and unsafe situations can be replayed and intuitively investigated.