Talks
From the quantified-self and serious gaming to AI disease predictions and oncology focused Living Labs
Date: 14 October 2024, 11:00 AM
Location: Lecture Room 7.1.1, Neherstraße 1 and Zoom
Speakers: Prof. Panagiotis Bamidis
Panagiotis Bamidis is a Professor of Medical Physics, Informatics and Medical Education and Director of the Lab of Medical Physics and Digital Innovation in the School of Medicine of the Aristotle University of Thessaloniki (AUTH), Greece and Director of the AUTH MSc course on Biomedical Engineering. He designs, implements, and evaluates Digital Health and Assistive Technologies systems that improve everyday activities of elderly or other vulnerable groups and improves their health or life quality or improves the education and training of health professionals. He is the co-ordinator of more than 12 large European projects, and the principal investigator for many national and international funded projects. He has also founded and leads Thessaloniki’s Active and Healthy Ageing action, which is a three-star reference site of the EIP-on-AHA. In 2013 he established the ThessAHALL, Thessaloniki’s
Action for Health and Well-being Living Lab. He has received Prizes for the Best Track Record in funded research projects and the Best overall high/extra-ordinary academic performance from his University.
Abstract: Quantified Self started as a collaboration of users and tool makers who shared an interest in self knowledge through self-tracking by means of simple devices. While ubiquitous computing and monitoring technologies were advancing, so where interventional initiatives in the form of serious games and computerised training. In parallel a growing interest was shown in exploiting them daily (or in individual homes) to assist different vulnerable groups and specific patient categories.
In this talk, we revisit the journey of establishing dedicated living labs to make the aforementioned measurements and recordings capable of exploring disease-related attributes. Projects dealing with cancer, cognitive decline and multiple sclerosis will be used as paradigms to demonstrate the value of disease-focused Living Labs.
Introduction into the field of risk and resilience and prevention of mental disorders
Date: 8 October 2024, 10:30 AM
Location: Lecture Room 7.1.1, Neherstraße 1 and Zoom (10:45)
Speakers: Prof. Johanna Löchner
Prof. Johanna Löchner completed her Diploma of Psychology at Philipps-Universität Marburg in 2012, including an exchange at Universidad de Monterrey, Mexico. She earned her Doctorate in 2018 at LMU Munich, focusing on prevention of depression. Her clinical training in Cognitive Behavioral Therapy took place from 2012 to 2019 at AVM, and she obtained her psychotherapy license in 2019. She has worked at LMU Munich and TU Munich in psychiatry and psychosomatics and out-patient centres for psychotherapy
treating children, adolescents and adults with different mental health disorders.
From 2018 to 2020, Prof. Löchner was a post-doctoral research fellow at LMU. Following that, she was head of the National Center for Early at Deutsches Jugendinstitut e.V. from 2020 to 2022. Currently, she is a lecturer at AVM and Assistant Professor for Mobile Assessment and Telepsychotherapy. In December 2024, she will take up a W2 tt W3 Professorship at FAU Erlangen
Abstract: Prof. Johanna Löchner will introduce the research focus of the team including the role of digital technology in the prevention and early recognition of mental illness. She will discuss how mobile apps, telepsychotherapy, and digital assessments can enhance early detection, enabling timely interventions. Additionally, she will highlight the potential of digital phenotyping - the use of data from everyday digital interactions to identify behavioral markers of mental health conditions. Prof. Löchner will also address
how these technologies can reach underserved populations and improve personalised care. Finally, she will examine key challenges, including ethical concerns and the need for further validation, while outlining future directions in the field.
Towards Low-Resource Learning Cases for
Computational Paralinguistics
Date: 24 July 2024, 11:00 AM
Location: Lecture Room 7.1.1, Neherstraße 1 and Zoom
Speakers: Associate Prof. Xinzhou Xu
Xinzhou Xu received the bachelor's degree from Nanjing University of Posts and Telecommunications, Nanjing/China, in 2009, the master's and the PhD degree from Southeast University, Nanjing/China, in 2012 and 2017, respectively. He is currently an Associate Professor with the School of Internet of Things, Nanjing University of Posts and Telecommunications, and also with the Signal Processing and Speech Communication Laboratory, Graz University of Technology, Graz/Austria. Previously, he was with the Machine Intelligence & Signal Processing group, MMK, Technical University of Munich (TUM), Munich/Germany (from 2014 to 2016), and the Chair of Complex and Intelligent Systems, the University of Passau, Passau/Germany (from 2015 to 2016), and the Chair of Embedded Intelligence for Health Care and Wellbeing, the University of Augsburg, Augsburg/Germany (in 2019 and 2023). His research interests include intelligent signal processing, computational paralinguistics, audio signal processing, and affective computing.
Abstract: Taken from the original concepts of language resource in linguistics, the resources for computational paralinguistics basically refer to the corporeal and incorporeal materials used in building paralinguistic models. From a perspective of data-driven learning, the low-resource learning cases mainly focus on how to effectively and sufficiently make use of in-hand extremely limited data for training suitable models. Hence, this talk attempts to decomposite the learning-oriented low-resource cases into multiple specific learning tasks, in line with whether or how much the limited resources are reliable enough for learning in the field of computational paralinguistics, and then, shows our recent related progress towards these tasks with possible research directions in the future. In addition, this talk also introduces other works we are currently involved in, and what we may be interested in within the scope of applied machine learning for intelligent signal processing.
Recent Work on Multimodal Emotion Recognition
Date: 2 July 2024, 10:30 AM
Location: Lecture Room 7.1.1, Neherstraße 1 and Zoom
Speakers: Prof. Jianhua Tao
Dr. Jianhua Tao is a Professor at the Tsinghua University. He is a recognized scholar in the field of speech and language processing, multimodal human-computer interaction and affective computing. He is a Fellow of the China Computer Federation (CCF), elected in 2017 “for contributions to spoken language processing and affective computing”. He has published more than 200 papers in Speech Communication, INTERSPEECH, IEEE TASLP, IEEE TIP, EURASIPJASMP, APSIPA, ISCSLP, ICASSP, Speech Prosody, ICME, ACII, etc. His recent awards include the Award of Distinguished Young Scholars of NSFC (2014), Award of National special support program for high-level person (2018), Best Paper Awards of NCMMSC (2001, 2015, 2017), Best Paper Awards of CHCI (2011, 2013, 2015, 2016), the Winner of IMAGINATION2001 (a competition organized in Eurospeech2001). He has delivered numerous invited and keynote talks, such as Speech Prosody (2012, 2018), NCMMSC (2017), etc. He has been the Leader or Chief-Scientist of several key scientific projects including Key Project of National Natural Science Foundation of China, National Key R&D Program of China, Key Project of International Cooperation, etc.
Prof. Tao was elected Chairperson of the ISCA Special Interest Group on Chinese Spoken Language Processing (ISCA SIG-CSLP) (2019-2020) and will be Technical Program Chair of INTERSPEECH2020. He was also an elected member of the Executive Committee of AAAC association (2007-2017) and served on the Steering Committee of IEEE Transactions on Affective Computing (2009-2017). He currently serves as the Subject Editor of Speech Communication, the Editorial Board Member of Journal on Multimodal User Interfaces, the Editorial Board Member of Journal of Computer Research and Development, the Director of Human Computer Interaction Committee of CSIG, the Deputy Director of Speech Dialog and Audition Committee of CCF, the Deputy Director of Speech Information Processing Committee of CSPSC, the Deputy Director of Artificial Psychology and Artificial Emotion Committee of CAAI, the Executive Board Member of CCF.
Abstract: Mulitmodal Emotion Recognition (MER) is an important technology in human-computer interaction. The talk will summarize the recent work on MER in Tsinghua University. Generally, emotion recognition involves four key components: datasets, labels features and classifiers. This report will revolve around the improvement of these four components. Regarding datasets, we organize MER2023@ACM Multimedia and MER 2024@IJCAI, and propose a large-scale 120k Chinese dataset. Regarding labels, we introduce a new task called explainable MER to improve the annotation accuracy and reliability. For features, we introduce our recent attempts in semi-supervised learning and we establish MERBench and MERTools, benchmark and toolkit covering various features, datasets, and fusion strategies. Regarding classifiers, we analyze how to automatically search for the optimal architecture, how to solve emotion recognition in conversations, and how to improve the noise robustness. Finally, we will point out some promising research directions in this field.
Novel Security User Interfaces based on Human Behavior and Physiology
Date: 20 June 2024, 11:30 AM
Location: Lecture Room 7.1.1, Neherstraße 1 and Zoom
Speakers: Prof. Dr. Florian Alt
Florian Alt is a full professor of Usable Security and Privacy at the Research Institute for Cyber Defense (CODE) in Munich. In his work, he is interested in designing secure and privacy-preserving systems that naturally blend with the way in which users interact with computing devices. Florian’s research focuses on investigating users' behavior in security contexts, on creating and enhancing security and privacy mechanisms based on users' behavior and physiology, and on understanding as well as mitigating threats emerging from novel ubiquitous technologies. Specific areas of application are Mixed Reality, Smart Homes, and Social Engineering. Florian holds a diploma in Media Informatics from LMU Munich and a Ph.D in Human-Computer Interaction from the University of Stuttgart.
Abstract: Over the past years, a global and highly professional cybercrime industry has established itself, making user-centered attacks an omnipresent threat to society. Companies, organizations, and individuals struggle to keep up and protect themselves effectively from an ever-increasing number of sophisticated attacks. The design of solid and sustainable means of protection is still a major challenge due to the need to consider the perspective of designers, software engineers, administrators, and end users alike. This talk will demonstrate how a new paradigm of human-centered security user interfaces based on knowledge about human behavior and physiology can be established. This becomes possible through sensing moving ever closer to the human body in the form of personal devices, wearables, and sensors in our environments. The talk will showcase how security mechanisms carefully integrated with interactive media systems allow a detailed understanding of human states in security-related situations to be obtained. This knowledge can be leveraged to enable personalized, contextualized, and tangible means for protection and increasing cybersecurity literacy.
Digital Phenotyping with the PhoneStudy App
Date: 18 June 2024, 2:30 PM
Location: Hörsaal A (Ismaninger Str. 22; Klinikum rechts der Isar der TUM
Speakers: Prof. Dr. Markus Bühner; M.Sc. Yannik Terhorst
Markus Bühner has been Professor of Psychological Methods and Assessment at Ludwig-Maximilian-University (LMU) of Munich since 2011. His research focuses on predicting personality, intelligence, and well-being with smartphone sensing data. For this reason, the PhoneStudy App was originally constructed. Another focus is to build a personality model with objective behavior collected via smartphone to provide an alternative model to assess personality in opposition to questionnaire data.
Yannik Terhorst is a researcher focussing on digital mental health care since his studies in psychology. His primary research areas are 1) novel assessment and diagnostic procedures (e.g., unobtrusively collected sensor data or supervised machine learning for depression screening), 2) the heterogeneity and magnitude of Internet- and mobile-based interventions effects, and 3) the acceptance and implementation of technologies for mental health.
Abstract: The use of unobtrusively collected sensor data (e.g., GPS) obtained from commonplace devices like smartphones holds promise as a valuable addition to existing tools for assessing personality, health, and human behavior (aka smart sensing). The LMU-developed PhoneStudy app provides a powerful framework to collect sensor data such as location data, language use, smartphone usage, app usage features, and ecological momentary assessments (https://phonestudy.org/en/). The present talk aims to present the framework accompanied by a presentation of the state-of-the-art evidence regarding personality and mental health research. The session will be closed by an open discussion and vision of interdisciplinary projects on how smart sensing can be combined with ongoing innovations in health informatics (e.g., smart sensing enhanced large language model for health interventions).
Presentation of the work at Fraunhofer Institute for Digital Media Technology (IDMT) Hearing, Speech and Audio Technology (HSA) in Oldenburg
Date: 24 May 2024, 10:00 AM
Location: Building 707 Room 7.1.1.
Speakers: Dr. Jens-E. Appell
Dr. Jens-E. Appell received his diploma in physics from the University of Göttingen, Germany in 1994. From 1994 to 2001 he worked as a Researcher at the Carl von Ossietzky Universität Oldenburg, Germany, on models for human auditory perception and audio signal processing and received his doctorate in 2001 with a thesis on “Loudness Models for rehabilitative Audiology”. 2001 he started at the OFFIS Institute for Information Technology as Head of the Design Center for embedded-systems and from 2003 as Director of the Embedded Hardware-/Software Systems Division. Since 2008, he founded and is since heading the Hearing, Speech and Audio Technology branch in Oldenburg of the Fraunhofer Institute for Digital Media Technology. In his work he addressed a variety of R&D topics and its applications in embedded systems, transportation, hearing aids, consumer electronics, ambient assisted living and production. Since 2020, Dr. Appell is also CEO of the FidiTec GmbH a start-up commercializing technologies in the field of Hearing, Speech and Audio Technology.
Energy-Efficient Deep Learning-Based Image Compression for LPWANs on Resource-Constrained AIoT Devices
Date: 22 January 2024, 2:30 PM
Location: Building 10 (MDSI, GALILEO Garching) + Zoom
Speakers: Nikolai Körber, M.Sc.
Nikolai Körber received his Bachelor of Science in Computer Science from the University of Applied Sciences Landshut in 2017 and his Master of Science in Computer Science from the same university in 2019. He is currently a PhD candidate in Electrical and Computer Engineering at TUM. He previously worked on several AI-related projects in industry (Fraunhofer IPA, iteratec GmbH, MHP Management- und IT Beratung GmbH) and academia (UAS Landshut, TUM). His research interests lie at the intersection of tinyML and computer vision. As part of his work, he is investigating how to enable the benefits of neural/ generative image compression on resource-constrained devices to address the high bandwidth constraints commonly encountered in sensor networks.
Abstract: Recently, there has been a rising demand for Low-Power Wide-Area Network (LPWAN)-based computer vision applications. LPWANs are specifically designed for long battery life, high transmission range and low production costs, which come however at the expense of very low bandwidths. Consequently, data compression plays a crucial role in energy-efficient image communication due to the significantly higher energy costs associated with communication compared to computation. For that, novel compute-intensive compression methods, like deep learning-based image compression techniques, are promising to further reduce the number of packets to be transmitted. Despite their superiority over hitherto established methods, such as JPEG, there is no related research that jointly addresses deep learning-based compression performance and resource efficiency on sensor platforms. Only recently, high computational power at low battery consumption has become possible by exploiting parallel ultra-low power processors like GAP8. The goal of this research is to develop robust, energy-efficient deep learning-based image compression techniques for LPWANs on resource-constrained AI-enabled IoT (AIoT) devices. Having higher compression rates while operating at low power will dramatically reduce network traffic, extend the battery life of visual IoT sensor nodes, and pave the way to a broad range of new data-intensive applications within the LPWAN/ 5G mMTC communication era.
Research Outline: AI based Quality of Service (QoS) Prediction and Repeater Placement
Date: 22 January 2024, 2:00 PM
Location: Building 10 (MDSI, GALILEO Garching) + Zoom
Speakers: Patrick Purucker, M.Sc
Patrick Purucker has received his Bachelor of Engineering in Electrical Engineering and Information Technology from the University of Applied Sciences Amberg-Weiden in 2020 followed by the Master of Science (Applied Research in Engineering Sciences) in 2021 from the same institution. He is currently working on the EU-funded projects A-IQ Ready and Archimedes focusing on resilient, AI supported wireless communications for harsh underground environments. Previously, he was working on the ADACORSA project, which was successfully completed in 2023, investigating cellular based UAS communications.
Abstract: The talk gives an overview of the research work involved in the planned doctoral study, which focuses on resilient communications for agents conducting autonomous search and rescue missions in challenging underground environments (e.g. tunnels). The research will be conducted within the framework of the projects A-IQ Ready and Archimedes. Thereby, a simulation of the wireless ad-hoc network between an operation center and search agents will be implemented to identify potential risks and tackle them with AI based algorithms. One potential approach is to develop a prediction model for the Quality of Service (QoS) in the vicinity of the agent, with the aim of adapting its trajectory to avoid communication blind spots. Furthermore, Reinforcement Learning algorithms will be researched for optimising the repeater placement within the wireless network.