Thomas Kosch, Jakob Karolus, Havy Ha, and Albrecht Schmidt Your Skin Resists: Exploring Electrodermal Activity as Workload Indicator during Manual Assembly In Proceedings of the ACM SIGCHI Symposium on Engineering Interactive Computing Systems (EICS ’19). ACM, New York, NY, USA, Article 8, 1–5.Download Cite Publisher
Matthias Hoppe, Jakob Karolus, Felix Dietz, Paweł W. Woźniak, Albrecht Schmidt, and Tonja-Katrin Machulla VRsneaky: Increasing Presence in VR Through Gait-Aware Auditory Feedback In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI ’19). ACM, New York, NY, USA, Paper 546, 1–9.Download Cite Video Publisher
Calvin Liang, Jakob Karolus, Thomas Kosch, and Albrecht Schmidt On the Suitability of Real-Time Assessment of Programming Proficiency using Gaze Properties In Proceedings of the 7th ACM International Symposium on Pervasive Displays (PerDis '18). ACM, New York, NY, USA, Article 31, 2 pages.Download Cite Publisher
Miriam Greis, Jakob Karolus, Hendrik Schuff, Paweł W. Woźniak, and Niels Henze Detecting uncertain input using physiological sensing and behavioral measurements In Proceedings of the 16th International Conference on Mobile and Ubiquitous Multimedia (MUM '17). ACM, New York, NY, USA, 299-304.Download Cite Publisher
Biying Fu, Jakob Karolus, Tobias Grosse-Puppendahl, Jonathan Hermann, and Arjan Kuijper Opportunities for Activity Recognition Using Ultrasound Doppler Sensing on Unmodified Mobile Phones In Proceedings of the 2nd International Workshop on Sensor-based Activity Recognition and Interaction (WOAR '15). ACM, New York, NY, USA, Article 8, 10 pages. Best paper award.Download Cite Publisher
Axel Schulz, Jakob Karolus, Frederik Janssen, and Immanuel Schweizer Accurate Pollutant Modeling and Mapping: Applying Machine Learning to Participatory Sensing and Urban Topology Data In International Conference and Workshops on Networked Systems (NetSys '15). Cottbus, 2015, pp. 1-8.Cite Publisher
Current technical sensor systems offer capabilities that are superior to human perception. Cameras can capture a spectrum that is wider than visible light, high-speed cameras can show movements that are invisible to the human eye, and directional microphones can pick up sounds at long distances. The vision of this project is to lay a foundation for the creation of digital technologies that provide novel sensory experiences and new perceptual capabilities for humans that are natural and intuitive to use. In a first step, the project will assess the feasibility of creating artificial human senses that provide new perceptual channels to the human mind, without increasing the experienced cognitive load. A particular focus is on creating intuitive and natural control mechanisms for amplified senses using eye gaze, muscle activity, and brain signals. The project will quantify the effectiveness of new senses and artificial perceptual aids compared to the baseline of unaugmented perception. The overall objective is to systematically research, explore, and model new means for increasing the human intake of information in order to lay the foundation for new and improved human senses enabled through digital technologies and to enable artificial reflexes.
The ground-breaking contributions of this project are (1) to demonstrate the feasibility of reliably implementing amplified senses and new perceptual capabilities, (2) to prove the possibility of creating an artificial reflex, (3) to provide an example implementation of amplified cognition that is empirically validated, and (4) to develop models, concepts, components, and platforms that will enable and ease the creation of interactive systems that measurably increase human perceptual capabilities. Further information are available at the corresponding website: ERC Amplify
The SFB-TRR 161 is a collaborative, interdisciplinary and research projects that connects 17 project teams of the University of Stuttgart, University of Konstanz and the Max-Planck Institute for Biological Cybernetics, that are working in four Project Areas, three Task Forces in the field of Visual Computing. Further information are available at the corresponding website: SFB-TRR 161.
In this project, we research new methods and techniques for cognition-aware visualizations. The basic idea is that a cognition-aware adaptive visualization will observe the physiological response of a person while interacting with a system und use this implicit input. Electrical signals measured on the body (e.g. EEG, EMG, ECG, galvanic skin response), changes of physiological parameters (e.g. body temperature, respiration rate, and pulse) as well as the users’ gaze behavior are used to estimate cognitive load and understanding. Through experimental research, concepts and models for adaptive and dynamic visualizations will be created. Frameworks and tools will be realized and empirically validated. You can find the project here.