“Adsum ergo sum.” I am here, therefore I am

Louis Dechamps, Marieke van Esch, Kan Ling Wo Leo, Yue Yang, Maren Hengelmolen

Blind and visually impaired people currently have inconveniences locating themselves in the indoor environment. After an inventory of the requirements of blind people, different representations do qualify for providing specific information blind people need. The main research question is: "How can blind people localise themselves (near) real-time in indoor environments with the combination of 3 representations of reality, namely (1) LiDAR point cloud matching, (2) ArcGIS Indoors and (3) Audio dynamic tactile map as the user interface?". Room detection and positioning of the user within the room are obtained by LiDAR scanning and point cloud matching. As a user interface for blind people this report proposes two deliverables: a dynamic tactile map and an added or stand-alone audible supported user interface. Preliminary results of the qualitative validation show positive outcomes. This report is a stepping stone for the possibility of integrating multiple into one device.

More information