Hi. I'm Jakob Karolus.

I am a researcher at the Human-Centered Ubiquitous Media group at the Institute of Informatics at the Ludwig-Maximilian University of Munich. My research focuses on establishing guidelines and techniques for Proficiency-Aware Systems based on Ubiquitous Sensing Technologies.



Thomas Kosch, Jakob Karolus, Havy Ha, and Albrecht Schmidt Your Skin Resists: Exploring Electrodermal Activity as Workload Indicator during Manual Assembly In Proceedings of the ACM SIGCHI Symposium on Engineering Interactive Computing Systems (EICS ’19). ACM, New York, NY, USA, Article 8, 1–5.

Download Cite Publisher

Matthias Hoppe, Jakob Karolus, Felix Dietz, Paweł W. Woźniak, Albrecht Schmidt, and Tonja-Katrin Machulla VRsneaky: Increasing Presence in VR Through Gait-Aware Auditory Feedback In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI ’19). ACM, New York, NY, USA, Paper 546, 1–9.

Download Cite Video Publisher


Jakob Karolus, Patrick Dabbert, Paweł W. Woźniak I Know What You Want: Using Gaze Metrics to Predict Personal Interest In The 31st Annual ACM Symposium on User Interface Software and Technology Adjunct Proceedings (UIST ’18 Adjunct). ACM, New York, NY, USA, 105-107.

Download Cite Publisher

Jakob Karolus, Hendrik Schuff, Thomas Kosch, Paweł W. Woźniak, and Albrecht Schmidt EMGuitar: Assisting Guitar Playing with Electromyography In Proceedings of the 2018 Designing Interactive Systems Conference (DIS '18). ACM, New York, NY, USA, 651-655.

Download Cite Video Publisher

Jakob Karolus and Albrecht Schmidt Proficiency-Aware Systems: Adapting to the User's Skills and Expertise In Proceedings of the 7th ACM International Symposium on Pervasive Displays (PerDis '18). ACM, New York, NY, USA, Article 33, 2 pages.

Download Cite Publisher

Calvin Liang, Jakob Karolus, Thomas Kosch, and Albrecht Schmidt On the Suitability of Real-Time Assessment of Programming Proficiency using Gaze Properties In Proceedings of the 7th ACM International Symposium on Pervasive Displays (PerDis '18). ACM, New York, NY, USA, Article 31, 2 pages.

Download Cite Publisher


Miriam Greis, Jakob Karolus, Hendrik Schuff, Paweł W. Woźniak, and Niels Henze Detecting uncertain input using physiological sensing and behavioral measurements In Proceedings of the 16th International Conference on Mobile and Ubiquitous Multimedia (MUM '17). ACM, New York, NY, USA, 299-304.

Download Cite Publisher

Jakob Karolus, Paweł W. Woźniak, Lewis L. Chuang, and Albrecht Schmidt Robust Gaze Features for Enabling Language Proficiency Awareness In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '17). ACM, New York, NY, USA, 2998-3010.

Download Cite Video Publisher


Jakob Karolus, Paweł W. Woźniak, and Lewis L. Chuang Towards Using Gaze Properties to Detect Language Proficiency In Proceedings of the 9th Nordic Conference on Human-Computer Interaction (NordiCHI '16). ACM, New York, NY, USA, Article 118, 6 pages.

Download Cite Publisher


Biying Fu, Jakob Karolus, Tobias Grosse-Puppendahl, Jonathan Hermann, and Arjan Kuijper Opportunities for Activity Recognition Using Ultrasound Doppler Sensing on Unmodified Mobile Phones In Proceedings of the 2nd International Workshop on Sensor-based Activity Recognition and Interaction (WOAR '15). ACM, New York, NY, USA, Article 8, 10 pages. Best paper award.

Download Cite Publisher

Axel Schulz, Jakob Karolus, Frederik Janssen, and Immanuel Schweizer Accurate Pollutant Modeling and Mapping: Applying Machine Learning to Participatory Sensing and Urban Topology Data In International Conference and Workshops on Networked Systems (NetSys '15). Cottbus, 2015, pp. 1-8.

Cite Publisher


Since 2018

Current technical sensor systems offer capabilities that are superior to human perception. Cameras can capture a spectrum that is wider than visible light, high-speed cameras can show movements that are invisible to the human eye, and directional microphones can pick up sounds at long distances. The vision of this project is to lay a foundation for the creation of digital technologies that provide novel sensory experiences and new perceptual capabilities for humans that are natural and intuitive to use. In a first step, the project will assess the feasibility of creating artificial human senses that provide new perceptual channels to the human mind, without increasing the experienced cognitive load. A particular focus is on creating intuitive and natural control mechanisms for amplified senses using eye gaze, muscle activity, and brain signals. The project will quantify the effectiveness of new senses and artificial perceptual aids compared to the baseline of unaugmented perception. The overall objective is to systematically research, explore, and model new means for increasing the human intake of information in order to lay the foundation for new and improved human senses enabled through digital technologies and to enable artificial reflexes.

The ground-breaking contributions of this project are (1) to demonstrate the feasibility of reliably implementing amplified senses and new perceptual capabilities, (2) to prove the possibility of creating an artificial reflex, (3) to provide an example implementation of amplified cognition that is empirically validated, and (4) to develop models, concepts, components, and platforms that will enable and ease the creation of interactive systems that measurably increase human perceptual capabilities. Further information are available at the corresponding website: ERC Amplify

2015 - 2017

The SFB-TRR 161 is a collaborative, interdisciplinary and research projects that connects 17 project teams of the University of Stuttgart, University of Konstanz and the Max-Planck Institute for Biological Cybernetics, that are working in four Project Areas, three Task Forces in the field of Visual Computing. Further information are available at the corresponding website: SFB-TRR 161.

C02: Physiologically Based Sensing and Adaptive Visualization

In this project, we research new methods and techniques for cognition-aware visualizations. The basic idea is that a cognition-aware adaptive visualization will observe the physiological response of a person while interacting with a system und use this implicit input. Electrical signals measured on the body (e.g. EEG, EMG, ECG, galvanic skin response), changes of physiological parameters (e.g. body temperature, respiration rate, and pulse) as well as the users’ gaze behavior are used to estimate cognitive load and understanding. Through experimental research, concepts and models for adaptive and dynamic visualizations will be created. Frameworks and tools will be realized and empirically validated. You can find the project here.


Jakob Karolus
Human-Centered Ubiquitous Media
Frauenlobstr. 7a
80337 Munich
jakob.karolus (at) ifi (dot) lmu (dot) de