Publication Cover
Bioacoustics
The International Journal of Animal Sound and its Recording
Volume 33, 2024 - Issue 1
1,445
Views
1
CrossRef citations to date
0
Altmetric
Articles

Parselmouth for bioacoustics: automated acoustic analysis in Python

ORCID Icon, ORCID Icon & ORCID Icon
Pages 1-19 | Received 03 May 2023, Accepted 22 Aug 2023, Published online: 13 Oct 2023

References

  • Abur D, Lester-Smith RA, Daliri A, Lupiani AA, Guenther FH, Stepp CE, Larson CR. 2018. Sensorimotor adaptation of voice fundamental frequency in Parkinson’s disease. PloS One. 13(1):e0191839. doi: 10.1371/journal.pone.0191839.
  • Anglada-Tort M, Harrison PM, Lee H, Jacoby N. 2023. Large-scale iterated singing experiments reveal oral transmission mechanisms underlying music evolution. Curr Biol. 33(8):1472–1486.e12. doi: 10.1016/j.cub.2023.02.070.
  • Anikin A, Pisanski K, Reby D. 2022. Static and dynamic formant scaling conveys body size and aggression. R Soc Open Sci. 9(1):211496. doi: 10.1098/rsos.211496.
  • Araya‐Salas M, Smith‐Vidaurre G, Golding N. 2017. warbleR: an R package to streamline analysis of animal acoustic signals. Methods Ecol Evol. 8(2):184–191. doi: 10.1111/2041-210X.12624.
  • Beguš G, Leban A, Gero S. (2023). Approaching an unknown communication system by latent space exploration and causal inference. arXiv preprint arXiv:2303.10931.
  • Boë L-J, Berthommier F, Legou T, Captier G, Kemp C, Sawallis TR, Becker Y, Rey A, Fagot J, Reby D. 2017. Evidence of a vocalic proto-System in the baboon (Papio papio) suggests pre-hominin speech precursors. PloS One. 12(1):e0169321. doi: 10.1371/journal.pone.0169321.
  • Boersma P. 1993. Accurate short-term analysis of the fundamental frequency and the harmonics-to-noise ratio of a sampled sound. Proceedings of the institute of phonetic sciences (Vol. 17, No. 1193, p. 97–110).
  • Boersma P. 2001. Praat, a system for doing phonetics by computer. Glot International. 5(9):341–345.
  • Boersma P, Weenink D. 2023. Praat: doing phonetics by computer [Computer program]. Version 10.3.17. [accessed 2023 Sept 10]. http://www.praat.org/.
  • Bogdanchikov A, Zhaparov M, Suliyev R. 2013. Python to learn programming. J Phys. 423( 1):12027. 1.
  • Boussalis C, Coan TG, Holman MR, Müller S. 2021. Gender, candidate emotional expression, and voter reactions during televised debates. Am Polit Sci Rev. 115(4):1242–1257. doi: 10.1017/S0003055421000666.
  • Brückl M. 2012. Vocal tremor measurement based on autocorrelation of contours. INTERSPEECH 2012 ISCA's 13th Annual Conference Portland; Sep 9–13; Portland, OR, USA; p. 715–718. doi: 10.21437/Interspeech.2012-223.
  • Childers DG. 1978. Modern spectrum analysis. IEEE Press.
  • Choi HS, Lee J, Kim W, Lee J, Heo H, Lee K. 2021. Neural analysis and synthesis: reconstructing speech from self-supervised representations. Adv Neural Inf Process Syst. 34:16251–16265.
  • Claidiere N, Smith K, Kirby S, Fagot J. 2014. Cultural evolution of systematically structured behaviour in a non-human primate. Proc R Soc B. 281( 1797): 20141541.
  • Costantini G, Cesarini V, Di Leo P, Amato F, Suppa A, Asci F, Saggio G. 2023. Artificial intelligence-based voice assessment of patients with Parkinson’s disease off and on treatment: machine vs deep-learning comparison. Sensors. 23(4):2293. doi: 10.3390/s23042293.
  • Elemans CP, Rasmussen JH, Herbst CT, Düring DN, Zollinger SA, Brumm H, Švec JG. 2015. Universal mechanisms of sound production and control in birds and mammals. Nat Commun. 6(1):8978. doi: 10.1038/ncomms9978.
  • Fagot J, Bonté E. 2010. Automated testing of cognitive performance in monkeys: use of a battery of computerized test systems by a troop of semi-free-ranging baboons (Papio papio). Behav Res Methods. 42(2):507–516. doi: 10.3758/BRM.42.2.507.
  • Fagot J, Paleressompoulle D. 2009. Automatic testing of cognitive performance in baboons maintained in social groups. Behav Res Methods. 41(2):396–404. doi: 10.3758/BRM.41.2.396.
  • Fitch WT, de Boer B, Mathur N, Ghazanfar AA. 2016a. Monkey vocal tracts are speech-ready. Sci Adv. 2(12):e1600723. doi: 10.1126/sciadv.1600723.
  • Grainger J, Dufau S, Montant M, Ziegler JC, Fagot J. 2012. Orthographic processing in baboons (Papio papio). Sci. 336(6078):245–248. doi: 10.1126/science.1218152.
  • Harrison P, Marjieh R, Adolfi F, van Rijn P, Anglada-Tort M, Tchernichovski O, Larrouy-Maestri P, Jacoby N. 2020. Gibbs sampling with people. Adv Neural Inf Process Syst. 33:10659–10671.
  • Ikuma T, McWhorter AJ, Oral E, and Kunduk M. 2023. Formant-aware spectral analysis of Sustained vowels of pathological breathy voice. J Voice. doi:10.1016/j.jvoice.2023.05.002.
  • Jadoul Y, Thompson B, De Boer B. 2018. Introducing Parselmouth: a python interface to Praat. J Phon. 71:1–15.
  • Keller PE, König R, Novembre G. 2017. Simultaneous cooperation and competition in the evolution of musical behavior: sex-related modulations of the singer’s formant in human chorusing. Front Psychol. 8:1559. doi:10.3389/fpsyg.2017.01559.
  • Laban G, George JN, Morrison V, Cross ES. 2020. Tell me more! Assessing interactions with social robots from speech. Paladyn. 12(1):136–159. doi: 10.1515/pjbr-2021-0011.
  • Laban G, Morrison V, Kappas A, Cross ES. 2022. Informal caregivers disclose increasingly more to a social robot over time. CHI '22: CHI Conference on Human Factors in Computing Systems; 29 April 2022–5 May 2022; New Orleans, LA, USA; p. 1–7. doi: 10.1145/3491101.
  • Lapp S, Rhinehart T, Freeland‐Haynes L, Khilnani J, Syunkova A, Kitzes J. 2023. OpenSoundscape: an open‐source bioacoustics analysis package for Python. Methods Ecol Evol. 14(9):2321–2328. doi: 10.1111/2041-210X.14196
  • Leonetti S, Jadoul Y, Torres Borda L, de Reus K, Rasilo H, Salazar Casals A, Ravignani A. 2022. Noise-dependent vocal plasticity in harbour seal and grey seal pups. The European Conference on Behavioural Biology (ECBB 2022); July 20– 23; Groningen, the Netherlands.
  • Madsen PT, Siebert U, Elemans CP. 2023. Toothed whales use distinct vocal registers for echolocation and communication. Science. 379(6635):928–933. doi: 10.1126/science.adc9570.
  • McComb K, Shannon G, Sayialel KN, Moss C. 2014. Elephants can determine ethnicity, gender, and age from acoustic cues in human voices. Proc Natl Acad Sci. 111(14):5433–5438.
  • Mészárosová E. 2015. Is python an appropriate programming language for teaching programming in secondary schools. Int J Inf Commun Technol Educ. 4(2):5–14. doi: 10.1515/ijicte-2015-0005.
  • Nicholson D. 2023. Crowsetta: a Python tool to work with any format for annotating animal vocalizations and bioacoustics data. J Open Source Softw. 8(84):5338. doi: 10.21105/joss.05338.
  • Nicholson D, Cohen Y. 2022. vak (0.6.0). Zenodo. doi: 10.5281/zenodo.6808839.
  • Nunes CEP, Nevard L, Montealegre-Z F, Vallejo-Marín M. 2021. Variation in the natural frequency of stamens in six morphologically diverse, buzz-pollinated, heterantherous Solanum taxa and its relationship to bee vibrations. Bot J Linn Soc. 197(4):541–553. doi: 10.1093/botlinnean/boab044.
  • Peirce J, Gray JR, Simpson S, MacAskill M, Höchenberger R, Sogo H, Lindeløv JK, Lindeløv JK. 2019. PsychoPy2: experiments in behavior made easy. Behav Res Methods. 51(1):195–203. doi: 10.3758/s13428-018-01193-y.
  • Pisanski K, Jones BC, Fink B, O’Connor JJ, DeBruine LM, Röder S, Feinberg DR. 2016a. Voice parameters predict sex-specific body morphology in men and women. Anim Behav. 112:13–22. doi:10.1016/j.anbehav.2015.11.008.
  • Pisanski K, Mora EC, Pisanski A, Reby D, Sorokowski P, Frackowiak T, Feinberg DR. 2016b. Volitional exaggeration of body size through fundamental and formant frequency modulation in humans. Sci Rep. 6(1):34389. doi: 10.1038/srep34389.
  • Półrolniczak E, Kramarczyk M. 2023. Acoustic analysis of the influence of warm-up on singing voice quality. J Voice. doi: 10.1016/j.jvoice.2023.02.017.
  • Poupard M, Best P, Schlüter J, Symonds H, Spong P, Lengagne T, Soriano T, Glotin H. 2019. Large-scale unsupervised clustering of Orca vocalizations: a model for describing Orca communication systems. 2nd International Workshop on Vocal Interactivity in-and-between Humans, Animals and Robots; Aug 29–30; London, UK.
  • Rasilo H, Jadoul Y. 2020. Discovering articulatory speech targets from synthesized random babble. In Proc. Interspeech OCT 25–29, 2020, 3715–3719. Shanghai, China; p. 3715–3719. doi: 10.21437/Interspeech.2020-3186.
  • Rasilo H, Räsänen O, Laine UK. 2013. Feedback and imitation by a caregiver guides a virtual infant to learn native phonemes and the skill of speech inversion. Speech Commun. 55(9):909–931.
  • Ravignani A, Herbst CT. 2023. Voices in the ocean. Sci. 379(6635):881–882. doi: 10.1126/science.adg5256.
  • Ravignani A, Kello CT, de Reus K, Kotz SA, Dalla Bella S, Méndez-Aróstegui M, Rapado-Tamarit B, Rubio-Garcia A, de Boer B. 2019. Ontogeny of vocal rhythms in harbor seal pups: an exploratory study. Curr Zool. 65(1):107–120. doi: 10.1093/cz/zoy055.
  • Rhinehart T, Lapp S, Kitzes J. 2022. Identifying and building on the current state of bioacoustics software. J Acoust Soc Am. 151(4):A27–A27. doi: 10.1121/10.0010544.
  • Riad R, Titeux H, Lemoine L, Montillot J, Bagnou JH, Cao XN, Bachoud-Lévi AC. 2020. Vocal markers from sustained phonation in Huntington’s disease. Proc. Interspeech 2020; p. 1893–1897. https://dblp.org/rec/conf/interspeech/RiadTLMBCDB20.html.
  • Riede T, Bronson E, Hatzikirou H, Zuberbühler K. 2005. Vocal production mechanisms in a non-human primate: morphological data and a model. J Hum Evol. 48(1):85–96. doi: 10.1016/j.jhevol.2004.10.002.
  • Rocha LH, Ferreira LS, Paula BC, Rodrigues FH, Sousa-Lima RS. 2015. An evaluation of manual and automated methods for detecting sounds of maned wolves (Chrysocyon brachyurus illiger 1815). Bioacoustics. 24(2):185–198. doi: 10.1080/09524622.2015.1019361.
  • Rose SJ, Allen D, Noble D, Clarke JA. 2018. Quantitative analysis of vocalizations of captive Sumatran tigers (Panthera tigris sumatrae). Bioacoustics. 27(1):13–26. doi: 10.1080/09524622.2016.1272003.
  • Schultebraucks K, Yadav V, Shalev AY, Bonanno GA, Galatzer-Levy IR. 2022. Deep learning-based classification of posttraumatic stress disorder and depression following trauma utilizing visual and auditory markers of arousal and mood. Psychol Med. 52(5):957–967. doi: 10.1017/S0033291720002718.
  • Seyfarth RM, Cheney DL. 2003. Meaning and emotion in animal vocalizations. Ann N Y Acad Sci. 1000(1):32–55. doi: 10.1196/annals.1280.004.
  • Shyam M. (2022). shyamblast/Koogu: v0.7.1 (v0.7.1). Zenodo. doi: 10.5281/zenodo.7275319.
  • Singla YK, Shah J, Chen C, Shah RR 2022. What do audio transformers hear? Probing their representations for language delivery & structure. 2022 IEEE International Conference on Data Mining Workshops (ICDMW); November; 28 Nov 2022–01 Dec 2022; Orlando, FL, USA; p. 910–925. doi: 10.1109/ICDMW58026.2022.00120.
  • Sueur J, Aubin T, Simonis C. 2008. Seewave, a free modular tool for sound analysis and synthesis. Bioacoustics. 18(2):213–226. doi: 10.1080/09524622.2008.9753600.
  • Taylor AM, Charlton BD, Reby D. 2016. Vocal production by terrestrial mammals: source, filter, and function. In: Suthers R, Fitch W, Fay R, and Popper A, editors. Vertebrate sound production and acoustic communication. Springer handbook of auditory research. Vol. 53. Cham: Springer. doi: 10.1007/978-3-319-27721-9_8.
  • Taylor AM, Reby D. 2010. The contribution of source–filter theory to mammal vocal communication research. J Zool. 280(3):221–236. doi: 10.1111/j.1469-7998.2009.00661.x.
  • Torres Borda L, Jadoul Y, Rasilo H, Salazar Casals A, Ravignani A. 2021. Vocal plasticity in harbour seal pups. Philos Trans R Soc B. 376(1840):20200456.
  • Townsend SW, Manser MB, Hauber M. 2013. Functionally referential communication in mammals: the past, present and the future. Ethology. 119(1):1–11. doi: 10.1111/eth.12015.
  • Tsfasman M, Saravanan A, Viner D, Goslinga D, De Wolf S, Raman C, Oertel C. 2021. Towards a real-time measure of the perception of anthropomorphism in human-robot interaction. Proceedings of the 2nd ACM Multimedia Workshop on Multimodal Conversational AI; p. 13–18.
  • Ulloa JS, Haupert S, Latorre JF, Aubin T, Sueur J. 2021. Scikit-maad: an open-source and modular toolbox for quantitative soundscape analysis in Python. Methods Ecology Evol. 12(12):2041–210X.13711. doi: 10.1111/2041-210X.13711.
  • Urrutia A, Bánszegi O, Szenczi P, Hudson R. 2022. Scaredy-cat: assessment of individual differences in response to an acute everyday stressor across development in the domestic cat. Appl Anim Behav Sci. 256:105771. doi:10.1016/j.applanim.2022.105771.
  • Vahedian-Azimi A, Keramatfar A, Asiaee M, Atashi SS, Nourbakhsh M. 2021. Do you have COVID-19? An artificial intelligence-based screening tool for COVID-19 using acoustic parameters. J Acoust Soc Am. 150(3):1945–1953. doi: 10.1121/10.0006104.
  • van Rijn P, Larrouy-Maestri P. 2023. Modelling individual and cross-cultural variation in the mapping of emotions to speech prosody. Nat Hum Behav. 7(3):1–11. doi: 10.1038/s41562-022-01505-5.
  • Wirth C, Warren JD. 2020. Spatial and temporal variation in toadfish (Opsanus tau) and cusk eel (Ophidion marginatum) mating choruses in eelgrass (Zostera marina) beds in a shallow, temperate estuary. Bioacoustics. 29(1):61–78. doi: 10.1080/09524622.2018.1542631.
  • Zhao Y, Ando A, Takaki S, Yamagishi J, Kobashikawa S. (2019). Does the Lombard Effect Improve Emotional Communication in Noise? — Analysis of Emotional Speech Acted in Noise. Proc. Interspeech September 15–19, 2019, 3292–3296, doi: 10.21437/Interspeech.2019-1605.