Figures & data
Figure 1. (Left) The routes used in the two scenarios. Numbers indicate the location of objects and the order in which they had to be identified. The northern route corresponds to Scenario A (first turning left), the southern one to Scenario B (first turning right). (Right) An overview of the objects in the two scenarios.
![Figure 1. (Left) The routes used in the two scenarios. Numbers indicate the location of objects and the order in which they had to be identified. The northern route corresponds to Scenario A (first turning left), the southern one to Scenario B (first turning right). (Right) An overview of the objects in the two scenarios.](/cms/asset/3cb6c2d8-cd72-4455-9952-23a2aafdd551/hihc_a_2075627_f0001_c.jpg)
Table 1. Conditions (Belt, NoBelt) and scenarios (A or B) for each of the four groups of participants.
Figure 2. (Left) The numpad keys and their mapping to the eight vibe boards. “5” was used to initiate the vibration following a wizard-of-oz approach. (Right) The tactile belt used as part of this study, visible are two vibe boards (bottom), the power supply/battery (top left), and the microcontroller (top right).
![Figure 2. (Left) The numpad keys and their mapping to the eight vibe boards. “5” was used to initiate the vibration following a wizard-of-oz approach. (Right) The tactile belt used as part of this study, visible are two vibe boards (bottom), the power supply/battery (top left), and the microcontroller (top right).](/cms/asset/ba5058dd-8129-4ccb-bac2-532f9abade95/hihc_a_2075627_f0002_c.jpg)
Figure 3. Experimental setup used in the study. The experimenter (left) monitors the participant’s performance and triggers vibrotactile cues to guide him to the next target along the route. The participant (right) controls a general aviation aircraft and tries to identify objects in the simulated environment while wearing eye tracking glasses and a tactile belt.
![Figure 3. Experimental setup used in the study. The experimenter (left) monitors the participant’s performance and triggers vibrotactile cues to guide him to the next target along the route. The participant (right) controls a general aviation aircraft and tries to identify objects in the simulated environment while wearing eye tracking glasses and a tactile belt.](/cms/asset/a7155616-bd7d-41de-98fb-2c73d1e79894/hihc_a_2075627_f0003_c.jpg)
Figure 5. Set of sketch maps drawn by one of the participants after the flights. The identified objects are marked with crosses and numbered from 1 to 5.
![Figure 5. Set of sketch maps drawn by one of the participants after the flights. The identified objects are marked with crosses and numbered from 1 to 5.](/cms/asset/3e7596b7-088a-48c7-8193-4fb3d90fe69f/hihc_a_2075627_f0005_b.jpg)
Table 2. Number of incorrect identifications, strikes and misses per condition, as well as the average time searching for objects per condition; with their Z and p values for a Wilcoxon signed rank test.
Table 3. Numeric results for sketch maps when subdivided by testing condition.
Figure 6. Results of the “raw” NASA-TLX questionnaires for the six categories and overall value for both conditions. Whiskers indicate standard error. Statistical significance marked by: *p < 0.05.
![Figure 6. Results of the “raw” NASA-TLX questionnaires for the six categories and overall value for both conditions. Whiskers indicate standard error. Statistical significance marked by: *p < 0.05.](/cms/asset/7b46dc73-c712-4a8d-89c6-dff18615392f/hihc_a_2075627_f0006_c.jpg)
Figure 7. UEQ benchmark results for the six categories and both conditions. Whiskers indicate confidence intervals. Statistical significance marked by: *p < 0.05, **p < 0.01, ***p < 0.001.
![Figure 7. UEQ benchmark results for the six categories and both conditions. Whiskers indicate confidence intervals. Statistical significance marked by: *p < 0.05, **p < 0.01, ***p < 0.001.](/cms/asset/25c453d5-8d66-4c90-8371-bc20f476cdd2/hihc_a_2075627_f0007_c.jpg)