PhD students and their research

During the fall of 2015, WASP recruited six PhD students for research within three projects: Automated Transport Systems, Localization and Scalability for Distributed Autonomous Systems, and Integrating Perception, Learning and Verification in Interactive Autonomous Systems.

Project: Automatic software diversification

Benoit Baudry is WASP Professor of Software Technology at KTH. His group investigates the automatic synthesis of diverse variants of software components. This work aims at constantly evolving distributed software applications to prevent single points of failure. Empiricism and case studies form the core scientific foundations for this work, where diversification technology is evaluated on large, real-world software systems.

Project: Self-healing software systems (PI: M. Monperrus)

Martin Monperrus is WASP Professor of Software Technology at KTH. His research group works in the field of self-healing software systems and software anti-fragility. His PhD students are setting up experiments to measure, understand and model software brittleness, and invent new techniques that enable software to be more robust, resilient, and eventually anti-fragile in ultra-open and dynamic environments.

Project: Automated Transport Systems

Linnea Persson will work on smart multi-agent systems and the corresponding data processing algorithms to provide agents with a an accurate description of their environment and its evolution, as well as optimization and control algorithms enabling agents to select and update in real-time their actions. Persson is supervised by Bo Wahlberg.

Dirk Van Dooren will develop control algorithms for autonomous vehicles connected over mobile networks. How to move executable control code through a network will be investigated. Demonstrations on control of vehicle platoons and other formations from the communication infrastructure will be targeted. Van Dooren is supervised by Karl-Henrik Johansson.

Project: Localization and Scalability for Distributed Autonomous Systems

Jiexiong Tang will create 3D models of indoor environments using small flying robots. The vision is a system that autonomously explores and maps an unknown environment and which can improve this model over time, for example, by increasing geometric accuracy, more details in textures, and by associating common structures across space and time. Tang will be supervised by Patric Jensfelt.

Håkan Carlsson will develop sensor fusion algorithms that can fuse the individual measurements from massive arrays of low-cost low-performance inertial sensors such as integrated MEMS based accelerometers and gyroscopes, in order to create high-performance virtual sensors. Special attention will be given to the application of UAV-positioning, which calls for significantly higher dynamic sensor ranges than what is achievable by any single MEMS sensor alone. Joakim Jaldén will be supervising Carlsson.

Project: Integrating Perception, Learning and Verification in Interactive Autonomous Systems

Mia Kokic will work on robotic grasping and interaction with the environment. She will integrate low level control mechanisms based on multisensory feedback with high-level, task related knowledge for generating optimised grasping and interaction strategies. Kokic’s will be supervised by Danica Kragic.

Top page top