194
Views
0
CrossRef citations to date
0
Altmetric
Research Articles

Human Eyes Move to the Target Earlier When Performing an Aiming Task with Increasing Difficulties

, , & ORCID Icon
Pages 1341-1346 | Received 05 Jul 2022, Accepted 16 Aug 2022, Published online: 26 Aug 2022

References

  • Atkins, M. S., Tien, G., Khan, R. S. A., Meneghetti, A., & Zheng, B. (2013). What do surgeons see: Capturing and synchronizing eye gaze for surgery applications. Surgical Innovation, 20(3), 241–248. https://doi.org/10.1177/1553350612449075
  • Biswas, P., & Langdon, P. (2015). Multimodal intelligent eye-gaze tracking system. International Journal of Human-Computer Interaction, 31(4), 277–294. https://doi.org/10.1080/10447318.2014.1001301
  • de Brouwer, A. J., Flanagan, J. R., & Spering, M. (2021). Functional use of eye movements for an acting system. Trends in Cognitive Sciences, 25(3), 252–263. https://doi.org/10.1016/j.tics.2020.12.006
  • Eivazi, S., Hafez, A., Fuhl, W., Afkari, H., Kasneci, E., Lehecka, M., & Bednarik, R. (2017). Optimal eye movement strategies: A comparison of neurosurgeons gaze patterns when using a surgical microscope. Acta Neurochirurgica, 159(6), 959–966. https://doi.org/10.1007/s00701-017-3185-1
  • Fitts, P. M. (1954). The information capacity of the human motor system in controlling the amplitude of movement. Journal of Experimental Psychology, 47(6), 381–391. http://europepmc.org/abstract/MED/13174710
  • Flanagan, J. R., & Johansson, R. S. (2003). Action plans used in action observation. Nature, 424(6950), 769–771. https://doi.org/10.1038/nature01861
  • Fuchs, A. F. (1967). Saccadic and smooth pursuit eye movements in the monkey. The Journal of Physiology, 191(3), 609–631. https://doi.org/10.1113/jphysiol.1967.sp008271
  • Helsen, W. F., Elliott, D., Starkes, J. L., & Ricker, K. L. (2000). Coupling of eye, finger, elbow, and shoulder movements during manual aiming. Journal of Motor Behavior, 32(3), 241–248. https://doi.org/10.1080/00222890009601375
  • Hermens, F., Flin, R., & Ahmed, I. (2013). Eye movements in surgery: A literature review. Journal of Eye Movement Research, 6(4), 1–11. https://doi.org/10.16910/jemr.6.4.4
  • Jiang, X., Zheng, B., & Atkins, M. S. (2015). Video processing to locate the tooltip position in surgical eye-hand coordination tasks. Surgical Innovation, 22(3), 285–293. https://doi.org/10.1177/1553350614541859
  • Johansson, R. S., Westling, G., Bäckström, A., & Flanagan, J. R. (2001). Eye-hand coordination in object manipulation. The Journal of Neuroscience, 21(17), 6917–6932. https://doi.org/10.1523/JNEUROSCI.21-17-06917.2001
  • Karamchandani, U., Erridge, S., Evans-Harvey, K., Darzi, A., Hoare, J., & Sodergren, M. H. (2022). Visual gaze patterns in trainee endoscopists – A novel assessment tool. Scandinavian Journal of Gastroenterology, 57, 1–9. https://doi.org/10.1080/00365521.2022.2064723
  • Katona, J. (2014). Examination and comparison of the EEG based Attention Test with CPT and TOVA [Paper presentation]. 2014 IEEE 15th International Symposium on Computational Intelligence and Informatics (CINTI). https://doi.org/10.1109/CINTI.2014.7028659
  • Katona, J. (2021a). Analyse the readability of LINQ code using an eye-tracking-based evaluation. Acta Polytechnica Hungarica, 18(1), 193–215. https://doi.org/10.12700/APH.18.1.2021.1.12
  • Katona, J. (2021b). Clean and dirty code comprehension by eye-tracking based evaluation using GP3 eye tracker. Acta Polytechnica Hungarica, 18(1), 79–99. https://doi.org/10.12700/APH.18.1.2021.1.6
  • Katona, J. (2022a). Examination of the advantage of the clean code technique by analyzing eye movement parameters [Paper presentation]. Proceedings of ISER International Conference.
  • Katona, J. (2022b). Measuring cognition load using eye-tracking parameters based on algorithm description tools. Sensors, 22(3), 912. https://www.mdpi.com/1424-8220/22/3/912
  • Katona, J., Kovari, A., Costescu, C., Rosan, A., Hathazi, A., Heldal, I., Helgesen, C., Thill, S., & Demeter, R. (2019). The examination task of source-code debugging using GP3 eye tracker [Paper presentation]. 2019 10th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), 23–25 October 2019. https://doi.org/10.1109/CogInfoCom47531.2019.9089952
  • Katona, J., Ujbanyi, T., Sziladi, G., & Kovari, A. (2017). Examine the effect of different web-based media on human brain waves [Paper presentation]. 2017 8th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), 11–14 September 2017. https://doi.org/10.1109/CogInfoCom.2017.8268280
  • Katona, J., Ujbanyi, T., Sziladi, G., & Kovari, A. (2019). Electroencephalogram-based brain-computer interface for internet of robotic things. In R. Klempous, J. Nikodem, & P. Z. Baranyi (Eds.), Cognitive infocommunications, theory and applications (pp. 253–275). Springer International Publishing. https://doi.org/10.1007/978-3-319-95996-2_12
  • Law, B., Atkins, M. S., Kirkpatrick, A. E., & Lomax, A. J. (2004). Eye gaze patterns differentiate novice and experts in a virtual laparoscopic surgery training environment [Paper presentation]. Proceedings of the 2004 Symposium on Eye Tracking Research & Applications, San Antonio, Texas. https://doi.org/10.1145/968363.968370
  • Matsuda, A., Okuzono, T., Nakamura, H., Kuzuoka, H., & Rekimoto, J. (2021). A surgical scene replay system for learning gastroenterological endoscopic surgery skill by multiple synchronized-video and gaze representation. Proceedings of the ACM on Human-Computer Interaction, 5(EICS), 1–22. https://doi.org/10.1145/3461726
  • Sailer, U., Flanagan, J. R., & Johansson, R. S. (2005). Eye-hand coordination during learning of a novel visuomotor task. The Journal of Neuroscience, 25(39), 8833–8842. https://doi.org/10.1523/jneurosci.2658-05.2005
  • Sivananthan, A., Kogkas, A., Glover, B., Darzi, A., Mylonas, G., & Patel, N. (2021). A novel gaze-controlled flexible robotized endoscope; preliminary trial and report. Surgical Endoscopy, 35(8), 4890–4899. https://doi.org/10.1007/s00464-021-08556-1
  • Vickers, J. N. (1996). Control of visual attention during the basketball free throw. The American Journal of Sports Medicine, 24(6 Suppl), S93–S97. https://doi.org/10.1177/036354659602406S25
  • Vickers, J. N. (2007). Perception, cognition, and decision training: The quiet eye in action. Human Kinetics.
  • Vickers, J. N., Harvey, A., Snelgrove, R., Stewart, A., Arsenault, G., & Mackenzie, C. (2015). Expertise differences in quiet eye duration and performance in surgical knot tying. International Journal of Sport Psychology, 46(6), 528–541. https://doi.org/10.7352/IJSP2015.46.528
  • Wang, J., Fu, E. Y., Ngai, G., & Leong, H. V. (2022). Investigating differences in gaze and typing behavior across writing genres. International Journal of Human–Computer Interaction, 38(6), 541–561. https://doi.org/10.1080/10447318.2021.1952801
  • Wilson, M. R., Causer, J., & Vickers, J. N. (2015). Aiming for excellence: The quiet eye as a characteristic of expertise (Routledge handbook of sport expertise) (pp. 22–37). Routledge.
  • Wilson, M. R., McGrath, J. S., Vine, S. J., Brewer, J., Defriend, D., & Masters, R. S. (2011). Perceptual impairment and psychomotor control in virtual laparoscopic surgery. Surgical Endoscopy, 25(7), 2268–2274. https://doi.org/10.1007/s00464-010-1546-4
  • Zheng, B., Jiang, X., Bednarik, R., & Atkins, M. (2021). Action-related eye measures to assess surgical expertise. BJS Open, 5(5), zrab068. https://doi.org/10.1093/bjsopen/zrab068

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.