820
Views
1
CrossRef citations to date
0
Altmetric
Research Articles

Exploring the Potentials of Crowdsourcing for Gesture Data Collection

ORCID Icon, ORCID Icon, ORCID Icon & ORCID Icon
Pages 3112-3121 | Received 20 Jun 2022, Accepted 09 Feb 2023, Published online: 23 Feb 2023

References

  • Alallah, F., Neshati, A., Sheibani, N., Sakamoto, Y., Bunt, A., Irani, P., Hasan, K. (2018). Crowdsourcing vs laboratory-style social acceptability studies? Examining the social acceptability of spatial user interactions for head-worn displays. Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems.
  • Albanie, S., Varol, G., Momeni, L., Afouras, T., Brown, A., Zhang, C., Coto, E., Camgöz, N. C., Saunders, B., Dutta, A. (2021). SeeHear: Signer diarisation and a new dataset. ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
  • Amini, S., & Li, Y. (2013). CrowdLearner: Rapidly creating mobile recognizers using crowdsourcing [Paper presentation]. Proceedings of the 26th annual ACM symposium on User interface software and technology, St. Andrews, Scotland, UK. https://doi.org/10.1145/2501988.2502029
  • Bastanfard, A., Shahabipour, M., & Amirkhani, D. (2022). Crowdsourcing of labeling image objects: An online gamification application for data collection. Available at SSRN 4072474.
  • Brown, H. R., Zeidman, P., Smittenaar, P., Adams, R. A., McNab, F., Rutledge, R. B., & Dolan, R. J. (2014). Crowdsourcing for cognitive science – the utility of smartphones. PLOS One, 9(7), e100662. https://doi.org/10.1371/journal.pone.0100662
  • Cao, Z., Hidalgo, G., Simon, T., Wei, S. E., & Sheikh, Y. (2021). OpenPose: Realtime multi-person 2D pose estimation using part affinity fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(1), 172–186. https://doi.org/10.1109/TPAMI.2019.2929257
  • Caramiaux, B., Bevilacqua, F., & Tanaka, A. (2013). Beyond recognition: Using gesture variation for continuous interaction [Paper presentation]. CHI'13 Extended Abstracts on Human Factors in Computing Systems, Paris, France. https://doi.org/10.1145/2468356.2468730
  • Cheema, S., & LaViola, J. J. (2011). Wizard of Wii: Toward understanding player experience in first person games with 3D gestures [Paper presentation]. Proceedings of the 6th International Conference on Foundations of Digital Games, Bordeaux, France. https://doi.org/10.1145/2159365.2159407
  • Cooper, S., Jones, D. B., & Prys, D. (2019). Crowdsourcing the Paldaruo Speech Corpus of Welsh for speech technology. Information, 10(8), 247. https://www.mdpi.com/2078-2489/10/8/247 https://doi.org/10.3390/info10080247
  • Duarte, A., Palaskar, S., Ventura, L., Ghadiyaram, D., DeHaan, K., Metze, F., Torres, J., & Giro-I-Nieto, X. (2021). How2sign: A large-scale multimodal dataset for continuous American sign language [Paper presentation]. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition.
  • Giel, B. (2012). Sprach-und Kommunikationsförderung bei Kindern mit Down-Syndrom. Schulz-Kirchner Verlag GmbH.
  • Grove, N., & Walker, M. (1990). The Makaton Vocabulary: Using manual signs and graphic symbols to develop interpersonal communication. Augmentative and Alternative Communication, 6(1), 15–28. https://doi.org/10.1080/07434619012331275284
  • Lee, H. J., Park, E., Kim, J., Joo, H., Whang, J., Ku, J.-A., & Lee, S.-H. (2019). Development of simplified manual sign system for students with severe and multiple disabilities. Special Education Research, 18(1), 197–224. https://doi.org/10.18541/ser.2019.02.18.1.197
  • Joze, H. R. V., & Koller, O. (2018). Ms-asl: A large-scale data set and benchmark for understanding American sign language. arXiv preprint arXiv:1812.01053
  • Kajiwara, D., & Murao, K. (2019). Gesture recognition method with acceleration data weighted by sEMG [Paper presentation]. Adjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers, London, UK. https://doi.org/10.1145/3341162.3345589
  • Kang, W., Jung, I.-T., Lee, D., & Hong, J.-H. (2021). Styling words: A simple and natural way to increase variability in training data collection for gesture recognition [Paper presentation]. Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan. https://doi.org/10.1145/3411764.3445457
  • Kim, E., & Vangala, S. (2016). Vinereactor: Crowdsourced Spontaneous Facial Expression Data [Paper presentation]. Proceedings of the 2016 ACM on International Conference on Multimedia Retrieval, New York, NY, USA. https://doi.org/10.1145/2911996.2912038
  • Kleinsmith, A., & Bianchi-Berthouze, N. (2013). Affective body expression perception and recognition: A survey. IEEE Transactions on Affective Computing, 4(1), 15–33. https://doi.org/10.1109/T-AFFC.2012.16
  • Koblin, A. M. (2009). The sheep market [Paper presentation]. Proceedings of the seventh ACM conference on Creativity and cognition. https://doi.org/10.1145/1640233.1640348
  • LaViola, J. J. (2015). Context aware 3D gesture recognition for games and virtual reality [Paper presentation]. ACM SIGGRAPH 2015 Courses, Los Angeles, California. https://doi.org/10.1145/2776880.2792711
  • Li, D., Rodríguez, C., Yu, X., & Li, H. (2019). Word-level deep sign language recognition from video: A new large-scale dataset and methods comparison.
  • Masi, I., Tran, A. T., Leksut, J. T., Hassner, T., Medioni, G. G. (2016). Do We Really Need to Collect Millions of Faces for Effective Face Recognition? CoRR, abs/1603.07057. http://arxiv.org/abs/1603.07057
  • McDuff, D., Kaliouby, R. E., & Picard, R. (2012). Crowdsourcing facial responses to online videos. IEEE Transactions on Affective Computing, 3(4), 456–468. https://doi.org/10.1109/T-AFFC.2012.19
  • Mukushev, M., Ubingazhibov, A., Kydyrbekova, A., Imashev, A., Kimmelman, V., & Sandygulova, A. (2022). FluentSigners-50: A signer independent benchmark dataset for sign language processing. PLOS One, 17(9), e0273649. https://doi.org/10.1371/journal.pone.0273649
  • Noroozi, F., Corneanu, C. A., Kaminska, D., Sapinski, T., Escalera, S., & Anbarjafari, G. (2021). Survey on emotional body gesture recognition. IEEE Transactions on Affective Computing, 12(2), 505–523. https://doi.org/10.1109/TAFFC.2018.2874986
  • Puckey, J., Wouters, R. (2010). One frame off a me: A Music Video by 40575 of our friends. http://oneframeoffame.com/
  • Savargiv, M., & Bastanfard, A. (2013). Text material design for fuzzy emotional speech corpus based on Persian semantic and structure [Paper presentation]. 2013 International Conference on Fuzzy Theory and Its Applications (iFUZZY). https://doi.org/10.1109/iFuzzy.2013.6825469
  • Shi, B., Del Rio, A. M., Keane, J., Michaux, J., Brentari, D., Shakhnarovich, G., & Livescu, K. (2018). American sign language fingerspelling recognition in the wild [Paper presentation]. 2018 IEEE Spoken Language Technology Workshop (SLT). https://doi.org/10.1109/SLT.2018.8639639
  • Sigurdsson, G. A., Varol, G., Wang, X., Farhadi, A., Laptev, I., & Gupta, A. (2016). Hollywood in homes: Crowdsourcing data collection for activity understanding. In B. Leibe, J. Matas, N. Sebe, & M. Welling (Eds.), Computer Vision – ECCV 2016. Springer.
  • Spiro, I. (2012). Motion chain: A webcam game for crowdsourcing gesture collection[Paper presentation]. Proceeding of the CHI'12 Extended Abstracts on Human Factors in Computing Systems (pp. 1345–1350). Association for Computing Machinery. https://doi.org/10.1145/2212776.2212452
  • Su, H., Deng, J., & Fei-Fei, L. (2012). Crowdsourcing annotations for visual object detection. In Human Computation - Papers from the 2012 AAAI Workshop, Technical Report (pp. 40–46). (AAAI Workshop - Technical Report; Vol. WS-12-08).
  • Suzuki, N., Watanabe, Y., & Nakazawa, A. (2020). GAN-based style transformation to improve gesture-recognition accuracy. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 4(4), 1–20. https://doi.org/10.1145/3432199
  • Tan, C. T., Rosser, D., & Harrold, N. (2013). Crowdsourcing facial expressions using popular gameplay [Paper presentation]. SIGGRAPH Asia 2013 Technical Briefs, Hong Kong, China. https://doi.org/10.1145/2542355.2542388
  • Taranta, E. M., II, Simons, T. K., Sukthankar, R., & Laviola, J. J., Jr. (2015). Exploring the benefits of context in 3D gesture recognition for game-based virtual environments. ACM Transactions on Interactive Intelligent Systems, 5(1), 1–34. https://doi.org/10.1145/2656345
  • Taranta, E. M., Pittman, C. R., Oakley, J. P., Maslych, M., Maghoumi, M., & LaViola, J. J. (2020). Moving toward an ecologically valid data collection protocol for 2D gestures in video games. Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (pp. 1–11). Association for Computing Machinery. https://doi.org/10.1145/3313831.3376417
  • Vatavu, R.-D. (2017). Beyond features for recognition: Human-readable measures to understand users’ whole-body gesture performance. International Journal of Human–Computer Interaction, 33(9), 713–730. https://doi.org/10.1080/10447318.2017.1278897
  • Vatavu, R.-D., Casiez, G., & Grisoni, L. (2013). Small, medium, or large? estimating the user-perceived scale of stroke gestures. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (pp. 277–280). Association for Computing Machinery. https://doi.org/10.1145/2470654.2470692
  • Vaughan, J. W. (2017). Making better use of the crowd: How crowdsourcing can advance machine learning research. Journal of Machine Learning Research, 18(1), 7026–7071. http://jmlr.org/papers/v18/17-234.html
  • Vondrick, C., Patterson, D., & Ramanan, D. (2013). Efficiently scaling up crowdsourced video annotation. International Journal of Computer Vision, 101(1), 184–204. https://doi.org/10.1007/s11263-012-0564-1
  • Watanabe, H., & Terada, T. (2018). Improving ultrasound-based gesture recognition using a partially shielded single microphone [Paper presentation]. Proceedings of the 2018 ACM International Symposium on Wearable Computers, Singapore, Singapore. https://doi.org/10.1145/3267242.3267274
  • Wobbrock, J. O., Wilson, A. D., & Li, Y. (2007). Gestures without libraries, toolkits or training: A $1 recognizer for user interface prototypes [Paper presentation]. Proceedings of the 20th annual ACM symposium on User interface software and technology, Newport, Rhode Island, USA. https://doi.org/10.1145/1294211.1294238
  • Yang, F., Wu, Y., Sakti, S., & Nakamura, S. (2019). Make Skeleton-based Action Recognition Model Smaller, Faster and Better [Paper presentation]. ACM Multimedia Asia, Beijing, China. https://doi.org/10.1145/3338533.3366569
  • Yuhui, W., Tian, L., & Xinxiong, L. (2020). Reliability of perceived usability assessment via crowdsourcing platform: Retrospective analysis and novel feedback quality inspection method. International Journal of Human–Computer Interaction, 36(11), 1070–1084. https://doi.org/10.1080/10447318.2019.1709339
  • Zhang, Y., Cao, C., Cheng, J., & Lu, H. (2018). EgoGesture: A new dataset and benchmark for egocentric hand gesture recognition. IEEE Transactions on Multimedia, 20(5), 1038–1050. https://doi.org/10.1109/TMM.2018.2808769