1,254
Views
0
CrossRef citations to date
0
Altmetric
Articles

‘All possible sounds’: speech, music, and the emergence of machine listening

Pages 253-281 | Received 14 Nov 2022, Accepted 21 Mar 2023, Published online: 10 Apr 2023

References

  • ABC News. 2020. “Apple Wants to Listen to You Wash Your Hands, and Help You Lose Your Car Keys.” ABC News, June 23. https://www.abc.net.au/news/2020-06-23/apple-wwdc-2020-ios-14-digital-key-and-hand-washing-arm-chips/12383124
  • AI Now Institute. 2018. “AI Now Report 2018.”
  • Amazon dir. 2019. Acoustic Event Detection with Alexa Guard. https://www.youtube.com/watch?v=-nKelNVVblM
  • Anderson, Chris. 2008. “The End of Theory: The Data Deluge Makes the Scientific Method Obsolete.” Wired, June 23. https://www.wired.com/2008/06/pb-theory/
  • Andrejevic, Mark. 2004. Reality TV: The Work of Being Watched. Critical Media Studies. Lanham, MD: Rowman & Littlefield Publishers.
  • Andrejevic, Mark. 2020a. Automated Media. New York, NY: Routledge.
  • Andrejevic, Mark. 2020b. “Data Civics: A Response to the ‘Ethical Turn.’” Television & New Media 21 (6): 562–567. doi:10.1177/1527476420919693.
  • Audio Analytic. 2018. “Audio Analytic.” Audio Analytic. https://www.audioanalytic.com/
  • Auditory. 1995. “[2nd CFP] IJCAI-95 Workshop on CASA.” http://www.auditory.org/mhonarc/1995/msg00006.html
  • Barlow, W.H. 1878. “The Logograph.” Journal of the Society of Telegraph Engineers 7 (21): 65–68. doi:10.1049/jste-1.1878.0006.
  • Barthes, Roland. 1991. The Responsibility of Forms: Critical Essays on Music, Art, and Representation. Translated by Richard Howard. Reprint ed. Berkeley: University of California Press.
  • Beck, Ulrich. 2009. World at Risk. Cambridge: Polity Press.
  • Beck, John, and Ryan Bishop. 2020. “Technocrats of the Imagination: Art, Technology, and the Military-Industrial Avant-Garde.” In Technocrats of the Imagination. Duke University Press. doi:10.1515/9781478007326.
  • Beck, Jacob, Barbara Hope, and Azriel Rosenfeld. 1986. Human and Machine Vision. Vol. 8. New York: Academic Press.
  • Bell, Eamonn. 2019. The Computational Attitude in Music Theory. Columbia University.
  • Bijsterveld, Karin. 2008. Mechanical Sound: Technology, Culture, and Public Problems of Noise in the Twentieth Century. Cambridge, Mass: Inside Technology.
  • Bijvoet, Marga. 1990. “How Intimate Can Art and Technology Really Be? A Survey of the Art and Technology Movement of the Sixties.“ In Culture, Technology and Creativity in the Late Twentieth Century, edited by Hayward Philip, 15–38. London: John Libbey.
  • Born, Georgina. 1995. Rationalizing Culture: IRCAM, Boulez, and the Institutionalization of the Musical Avant-Garde. Berkeley: Univ of California Press.
  • Brand, Stewart. 1987. The Media Lab: Inventing the Future at MIT. New York: Penguin.
  • Bregman, Albert S. 1984. “Auditory Scene Analysis.” In IEEE Conference on Pattern Recognition, Montreal, 168–175.
  • Bregman, Albert S. 1990. Auditory Scene Analysis: The Perceptual Organization of Sound. Cambridge, Mass: MIT Press.
  • Brewster, Ben. 1972. “Introduction to Marx’s ‘Notes on Machines.’” Economy and Society 1 (3): 235–243. doi:10.1080/03085147200000013.
  • Brown, Guy J. 1992. “Computational Auditory Scene Analysis: A Representational Approach”. The Journal of the Acoustical Society of America 94 (4): 2454–2454. doi:10.1121/1.407441.
  • Brown, Guy, and Martin Cooke. 1994. “Computational Auditory Scene Analysis.” Compter Speech and Language 8 (4): 297–336. doi:10.1006/csla.1994.1016.
  • Brown, Judith C., Andrea Hodgins-Davis, and Patrick J. O. Miller. 2006. “Classification of Vocalizations of Killer Whales Using Dynamic Time Warping.” The Journal of the Acoustical Society of America 119 (3): EL34–40. doi:10.1121/1.2166949.
  • Brown, Judith C., and Patrick J. O. Miller. 2007. “Automatic Classification of Killer Whale Vocalizations Using Dynamic Time Warping.” The Journal of the Acoustical Society of America 122 (2): 1201–1207. doi:10.1121/1.2747198.
  • “CASA Workshop Summary.” 1996. November 30. https://web.archive.org/web/19961130120944/http:/sound.media.mit.edu/~dfr/casa/summary.html
  • CCRMA. 1992. “Center for Computer Research in Music and Acoustics: Research Overview.”
  • Cherry, E.C. 1953. “Some Experiments on the Recognition of Speech with One and with Two Ears.” The Journal of the Acoustical Society of America 25 (5): 975–979. doi:10.1121/1.1907229.
  • Couldry, Nick, and Ulises Mejias. 2018. “Data Colonialism: Rethinking Big Data’s Relation to the Contemporary Subject.” Television & New Media 20 (4): 1–14. doi:10.1177/1527476418796632.
  • Crawford, Kate, and Vladan Joler. 2018. “Anatomy of an AI System.” Anatomy of an AI System. http://www.anatomyof.ai
  • Crawford, Kate, and Trevor Paglen. 2019. “Excavating AI: The Politics of Training Sets for Machine Learning.” https://excavating.ai
  • David, E., and O. Selfridge. 1962. “Eyes and Ears for Computers.” Proceedings of the IRE 50 (5): 1093–1101. doi:10.1109/JRPROC.1962.288011.
  • DCASE. 2021. “DCASE 2021 Workshop Statistics.” https://dcase.community/documents/workshop2021/dcase2021_statistics.pdf
  • Dean, Roger T. 2009. The Oxford Handbook of Computer Music. Oxford: Oxford University Press.
  • Denes, Peter. 1960. “Automatic Speech Recognition: Experiments with a Recogniser Using Lingustic Statistics.” Contract No. AF 61(514)-1176. Air Force Cambridge Research Center: United States Air Force Air Research and Development Command.
  • DESED. 2020. “Domestic Environment Sound Event Detection Dataset.” https://project.inria.fr/desed/
  • Diduck, Ryan. 2018. Mad Skills: MIDI and Music Technology in the Twentieth Century. London: Repeater.
  • Dockray, Sean. 2018. Learning from YouTube. Video essay.
  • Dockray, Sean, James E. K. Parker, and Joel Stern. 2020. “Machine Listening, a Curriculum.” Machine Listening Curriculum. https://machinelistening.exposed/curriculum/
  • Doherty, Aiden R, Alan F Smeaton, Keansub Lee, and Daniel P W Ellis. 2007. “Multimodal Segmentation of Lifelog Data.” In Proceedings of the 8th International Conference on Computer-Assisted Information Retrieval RIAO 2007, May 18. Pittsburgh.
  • Downie, J Stephen. 2003. “Music Information Retrieval.” In Annual Review of Information Science and Technology, edited by Blaise Cronin, 295–340. Medford: Information Today.
  • Downie, Stephen, and Michael Nelson. 2000. “Evaluation of a Simple and Effective Music Information Retrieval Method.” In Proceedings of the 23rd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval - SIGIR ’00, 73–80. Athens, Greece: ACM Press. doi:10.1145/345508.345551.
  • Dyson, Frances. 2006. Frances Dyson, and Then It Was Now : Enduring Rhetorics. Edited by Clarisse Bardiot. Fondation Langlois. https://www.fondation-langlois.org/html/e/page.php?NumPage=2144
  • Edwards, Paul N. 1997. The Closed World: Computers and the Politics of Discourse in Cold War America. Cambridge, Mass: MIT Press.
  • Ellis, Daniel P W. 1996. Prediction-Driven Computational Auditory Scene Analysis. MIT.
  • Ellis, Daniel P W. 2001. “Detecting Alarm Sounds.” 4. https://www.ee.columbia.edu/~dpwe/pubs/crac01-alarms.pdf
  • Ellis, Daniel P W. 2002. “The Listening Machine: Sound Source Organization for Multimedia Understanding.” Electrical Engineering 19.
  • Ellis, Dan. 2018. “Recognizing Sound Events.” John Hopkins: Center for Language and Speech Processing, October 4. https://jh.hosted.panopto.com/Panopto/Pages/Viewer.aspx?id=4a7e392c-5163-41a6-8229-aadc01099e63
  • Ernst, Wolfgang. 2021. “The Media Epistemic Value of Sonic Analytics Tools. A Commentary.” Internet Histories 5 (1): 48–56. doi:10.1080/24701475.2020.1862528.
  • Estabrooks, Maurice. 1995. Electronic Technology, Corporate Strategy, and World Transformation. Westport: Quorum Books.
  • Feldman, Jessica. 2016. “‘The Problem of the Adjective’: Affective Computing of the Speaking Voice.” Transposition, no. 6, December. doi:10.4000/transposition.1640.
  • Font, Frederic, Annamaria Mesaros, Daniel P. W. Ellis, Eduardo Fonseca, Magdalena Fuentes, and Benjamin Elizalde. 2021. “Proceedings of the 6th Workshop on Detection and Classication of Acoustic Scenes and Events (DCASE 2021),” November. doi:10.5281/ZENODO.5770113.
  • Foucault, Michel. 2000. “The order of things: An archaeology of the human sciences.” In Posthumanism, 27–29. Palgrave, London.
  • Friedland, Gerard, Paris Smaragdis, Josh McDermott, and Raj Bhisha. 2018. “Audition for Multimedia Computing.” In Frontiers of Multimedia Research, edited by Shih-Fu Chang, 416. San Rafael, CA: Association for Computing Machinery.
  • Fry, D. B., and P. Denes. 1958. “The Solution of Some Fundamental Problems in Mechanical Speech Recogntion.” Language and Speech 1 (1): 35–58. doi:10.1177/002383095800100104.
  • Fuchs, Christian, and Vincent Mosco, eds. 2016. Marx in the Age of Digital Capitalism. Vol. 80. Leiden, Boston: Brill.
  • Gemmeke, Jort F., Daniel PW Ellis, Dylan Freedman, Aren Jansen, Wade Lawrence, R. Channing Moore, Manoj Plakal, and Marvin Ritter. 2017. “Audio Set: An Ontology and Human-Labeled Dataset for Audio Events.” In Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference On, New Orleans, 776–780. IEEE.
  • Goldenfein, Jake. 2019. “The Profiling Potential of Computer Vision and the Challenge of Computational Empiricism.” Association for Computing Machinery 27:110–119.
  • Goto, Masataka, and Satoru Hayamizu. 1999. “A Real-Time Music Scene Description System: Detecting Melody and Bass Lines in Audio Signals.” In IJCAI-99 Workshop on Computational Auditory Scene Analysis, Stockholm, 10.
  • Goto, Masataka, and Yoichi Muraoka. 1998. “An Audio-Based Real-Time Beat Tracking System and Its Applications.” In Proceedings of International Computer Music Conference, Ann Arbor.
  • Guzmán, Adolfo. 1968. “Decomposition of a Visual Scene into Three-Dimensional Bodies.” In Proceedings of the December 9-11, 1968, Fall Joint Computer Conference, Part I San Francisco, 291–304. doi:10.1145/1476589.1476631.
  • Halkias, Xanadu C., and Daniel P.W. Ellis. 2006. “Call Detection and Extraction Using Bayesian Inference.” Applied Acoustics 67 (11–12): 1164–1174. doi:10.1016/j.apacoust.2006.05.006.
  • Harcourt, Bernard E. 2015. Exposed: Desire and Disobedience in the Digital Age. Cambridge, Mass: Harvard University Press.
  • Härmä, Aki. 2003. “Automatic Identification of Bird Species Based on Sinusoidal Modeling of Syllables.” ICASSP. doi:10.1109/ICASSP.2003.1200027.
  • Hawley, Michael. 1993. Structure of Sound. Cambridge, Mass: MIT.
  • House, Brian. 2017. “Machine Listening: Wavenet, Media Materialism, and Rhythmanalysis.“ Machine Research 6 (1): 9.
  • Hurlbut, J Benjamin. 2018. “Control Without Limits in the New Biology.” In Gene Editing, Law, and the Environment: Life Beyond the Human, edited by Irus Braverman, 77–94. New York: Routledge.
  • Hvistendahl, Mara. 2020. “How a Chinese AI Giant Made Chatting—and Surveillance—Easy.” Wired. https://www.wired.com/story/iflytek-china-ai-giant-voice-chatting-surveillance/
  • Jasanoff, Sheila, and Sang-Hyun Kim. 2015. Dreamscapes of Modernity: Sociotechnical Imaginaries and the Fabrication of Power. Chicago: University of Chicago Press.
  • Jones, Nicholaus A. 2018. “Listening to the Frontend: United States Patent US 10,020,004 B2,” 14.
  • Kahn, Jonathan. 2013. Race in a Bottle: The Story of BiDil and Racialized Medicine in a Post-Genomic Age. New York: Columbia University Press.
  • Kahn, Douglas, and Marvin Minsky. 1988. “Minsky and Artificial Intelligence.” EAR.
  • Kang, Edward B. 2023. “Ground Truth Tracings (GTT): On the Epistemic Limits of Machine Learning.” Big Data & Society 10 (1): 205395172211461. doi:10.1177/20539517221146122.
  • Kang, Edward B., and Simogne Hudson. 2022. “Audible Crime Scenes: ShotSpotter as Diagnostic, Policing, and Space-Making Infrastructure.” Science, Technology, & Human Values, 016224392211432, December. doi:10.1177/01622439221143217.
  • Kennedy, Lyndon S, and Daniel P W Ellis. 2004. “ Laughter Detection in Meetings.“ NIST ICASSP 2004 Meeting Recognition Workshop, Montreal Canada.
  • Lakhotia, Kushal, Evgeny Kharitonov, Wei-Ning Hsu, Yossi Adi, Adam Polyak, Benjamin Bolte, Tu-Anh Nguyen, et al. 2021. “Generative Spoken Language Modeling from Raw Audio.” ArXiv: 2102.01192 [Cs], September. http://arxiv.org/abs/2102.01192
  • Lawrence, H.M. 2019. “Siri Disciplines“. In Your Computer is on Fire edited by Mullaney, Thomas S, Peters, Benjamin, Hicks, Mar, Philip, Kavita. Cambridge, Mass: MIT Press.
  • Lee, Keansub, and Daniel P. W. Ellis. 2010. “Audio-Based Semantic Concept Classification for Consumer Video.” IEEE Transactions on Audio, Speech, and Language Processing 18 (6): 1406–1416. doi:10.1109/TASL.2009.2034776.
  • Lewis, George. 2000. “Too Many Notes.” Leonardo Music Journal 10: 33–39. doi:10.1162/096112100570585.
  • Lewis, George. 2018. “Technosphere Magazine: 5. Rainbow Family.” Technosphere Magazine. /p/5-Rainbow-Family-5Aj9nAxzG6zFRAAd9icEvH
  • Lewis, George E. 2018. Why Do We Want Our Computers to Improvise? Edited by Roger T. Dean and Alex McLean. Vol. 1. Oxford University Press. doi:10.1093/oxfordhb/9780190226992.013.29
  • Li, Xiaochang. 2017. Divination Engines: A Media History of Text Prediction. New York: New York Univesrity.
  • Li, Xiaochang, and Mara Mills. 2019. “Vocal Features: From Voice Identification to Speech Recognition by Machine.” Technology and Culture 60 (2S): S129–60. doi:10.1353/tech.2019.0066.
  • Licklider, J.C.R. 1960. “Man-Computer Symbiosis.” IRE Transactions on Human Factors in Electronics HFE-1 (1): 4–10. doi:10.1109/THFE2.1960.4503259.
  • Lynskey, Orla. 2019. “Grappling with ‘Data Power’: Normative Nudges from Data Protection and Privacy.” Theoretical Inquiries in Law 20 (1): 189–220. doi:10.1515/til-2019-0007.
  • Lyon, Richard F. 1978. “Sig_Proc_Model_of_Hearing-Lyon1978.Pdf.”
  • Lyon, Richard F. 2010. “Machine Hearing: An Emerging Field.” IEEE Signal Processing Magazine, September, 6. doi:10.1109/MSP.2010.937498.
  • Lyon, Richard F. 2017. Human and Machine Hearing: Extracting Meaning from Sound. Cambridge University Press. doi:10.1017/9781139051699.
  • “Machine Listening Lab.” 2018. Machine Listening Lab. http://machine-listening.eecs.qmul.ac.uk/
  • MacKenzie, Donald. 1984. “Marx and the Machine.” Technology and culture 25 (3): 473–502. doi:10.2307/3104202.
  • Maier, Stefan. 2018. “Technosphere Magazine: 1. WaveNet: On Machine and Machinic Listening.” Technosphere Magazine. /p/1-WaveNet-On-Machine-and-Machinic-Listening-a2mD8xYCxtsLqoaAnTGUbn
  • Malkin, Robert G. 2006. “Machine Listening for Context-Aware Computing.”
  • Marx, Karl. 1976. Capital: A Critique of Political Economy. Vol. 1. Translated by Ben Fowkes. London: Penguin.
  • Mathews, Max, and John R Pierce, eds. 1989. Current Directions in Computer Music Research. Cambridge, Mass: MIT Press.
  • Mattern, Shannon. 2020. “Urban Auscultation; or, Perceiving the Action of the Heart.” Places Journal, April. doi:10.22269/200428.
  • McQuillan, Dan. 2018a. “Data Science as Machinic Neoplatonism.” Philosophy & Technology 31 (2): 253–272. doi:10.1007/s13347-017-0273-3.
  • McQuillan, Dan. 2018b. “Mental Health and Artificial Intelligence: Losing Your Voice | OpenDemocracy.” https://www.opendemocracy.net/en/digitaliberties/mental-health-and-artificial-intelligence-losing-your-voice-poem/
  • Mellinger, David K., and Christopher W. Clark. 2000. “Recognizing Transient Low-Frequency Whale Sounds by Spectrogram Correlation.” The Journal of the Acoustical Society of America 107 (6): 3518–3529. doi:10.1121/1.429434.
  • Mesaros, Annamaria, Toni Heittola, and Tuomas Virtanen. 2016. “TUT Database for Acoustic Scene Classification and Sound Event Detection.” In 2016 24th European Signal Processing Conference (EUSIPCO), 1128–1132. Budapest, Hungary: IEEE. doi:10.1109/EUSIPCO.2016.7760424.
  • Mills, Mara. 2010. “Deaf Jam.” Social Text 28 (1): 35–58. doi:10.1215/01642472-2009-059.
  • Mills, M. 2011. “On Disability and Cybernetics: Helen Keller, Norbert Wiener, and the Hearing Glove.” Differences 22 (2–3): 74–111. doi:10.1215/10407391-1428852.
  • Mills, Mara. 2011a. “Do Signals Have Politics? Inscribing Abilities in Cochlear Implants.” The Oxford Handbook of Sound Studies, December. doi:10.1093/oxfordhb/9780195388947.013.0077.
  • Mills, Mara. 2011b. “Hearing Aids and the History of Electronics Miniaturization.” IEEE Annals of the History of Computing 33 (2): 24–45. doi:10.1109/MAHC.2011.43.
  • Minsky, Marvin. 1975. “A Framework for Representing Knowledge.” In The Psychology of Computer Vision, edited by P H Winston. New York: McGraw Hill.
  • Minsky, Marvin. 1982. “Music, Mind, and Meaning.” In Music, Mind, and Brain: The Neuropsychology of Music edited by M. Clynes, 19. New York: .Springer.
  • Minsky, Marvin. 1986. The Society of Mind. New York: Simon and Schuster.
  • Mody, Cyrus C. M., and Andrew J. Nelson. 2013. “‘A Towering Virtue of Necessity’: Interdisciplinarity and the Rise of Computer Music at Vietnam-Era Stanford.” Osiris 28 (1): 254–277. doi:10.1086/671380.
  • Moorer, J A. 1975. On the Segmentation and Analysis of Continuous Musical Sound by Digital Computer. .
  • Mumford, Lewis. 1966. The Myth of the Machine: Technics and Human Development. Vol. 1. New York: Harcourt.
  • Negroponte, Nicolas. 1970. The Architecture Machine: Toward a More Human Environment. Cambridge, Mass: MIT Press.
  • Nest. 2021. “Nest Cam Indoor.” Nest Cams. https://www.nestcamera.net/nest-cam-indoor/
  • Parker, James E. K., and Lawrence Abu Hamdan. 2022. “Forensic Listening as Machine Listening.” Disclaimer. https://disclaimer.org.au/contents/forensic-listening-as-machine-listening
  • Pfeifer, Michelle 2021. Listening for the Border - Affective Objectivity and Border Sonics (Unpublished Draft). .
  • Phan, Thao. 2019. “Amazon Echo and the Aesthetics of Whiteness.” Catalyst: Feminism, Theory, Technoscience 5 (1): 1–38. doi:10.28968/cftt.v5i1.29586.
  • Pieraccini, Roberto. 2012. The Voice in the Machine: Building Computers That Understand Speech. Cambridge, Mass: MIT Press.
  • Pieraccini, Roberto. 2021. AI Assistants. Cambridge, Mass: MIT Press.
  • Pierce, John R. 1969. “Whither Speech Recognition?” The Journal of the Acoustical Society of America 46 (4B): 1049–1051. doi:10.1121/1.1911801.
  • Pierce, J. R., and Mary E. Shannon. 1949. “Composing Music by a Stochastic Process.” Technical Memorandum (Bell Telephone Laboratories, November 15).
  • Reuters. 2021. “Voice Assistant Maker SoundHound to Go Public via $2 Bln SPAC Deal.” Reuters, November 16, sec. Technology. https://www.reuters.com/technology/voice-ai-platform-soundhound-go-public-via-21-bln-spac-merger-2021-11-16/
  • Rice, Tom. 2015. “Hearing.” In Keywords in Sound, edited by David Novak and Matt Sakakeeny, 99–111. Durham, N.C: Duke University Press.
  • Rosenthal, David F., and Hiroshi G. Okuno, eds. 1998. Computational Auditory Scene Analysis. Computational Auditory Scene Analysis. Mahwah, NJ: Lawrence Erlbaum Associates Publishers.
  • Rovner, Lisa, dir. 2020. Sisters with Transistors. https://sisterswithtransistors.com/
  • Rowe, Robert. 1991. Machine Listening and Composing: Making Sense of Music with Cooperating Real-Time Agents. Cambridge, Mass: MIT.
  • Rowe, Robert. 1992. “Machine Listening and Composing with Cypher.” Computer Music Journal 16 (1): 43. doi:10.2307/3680494.
  • Rowe, Robert. 1993. Interactive Music Systems: Machine Listening and Composing. Cambridge, Mass: MIT Press.
  • Rowe, Robert. 1999. “The Aesthetics of Interactive Music Systems.” Contemporary Music Review 18 (3): 83–87. doi:10.1080/07494469900640361.
  • Rowe, Robert. 2001. Machine Musicianship. Cambridge: MIT Press.
  • “SAPA Workshops - Index.” n.d. Accessed 6 May 2022. https://www.sapaworkshops.org/
  • Schroeder, Manfred Robert. 1985. Speech and Speaker Recognition. Vol. 12. Basel: Karger Medical and Scientific Publishers.
  • Schuller, Björn W. 2014. Intelligent Audio Analysis. Berlin: Springer Science & Business Media.
  • Schuller, Björn W., Alican Akman, Yi Chang, Harry Coppock, Alexander Gebhard, Alexander Kathan, Esther Rituerto-González, Andreas Triantafyllopoulos, and Florian B. Pokorny. 2022. “Climate Change & Computer Audition: A Call to Action and Overview on Audio Intelligence to Help Save the Planet.” ArXiv:2203.06064 [Cs], March. http://arxiv.org/abs/2203.06064
  • Schuller, Björn W., Dagmar M. Schuller, Kun Qian, Juan Liu, Huaiyuan Zheng, and Xiao. Li. 2020. “COVID-19 and Computer Audition: An Overview on What Speech & Sound Analysis Could Contribute in the SARS-CoV-2 Corona Crisis.” ArXiv:2003.11117 [Cs, Eess] 3, March. http://arxiv.org/abs/2003.11117
  • Scott, Alan. 1997. “Modernity’s Machine Metaphor.” The British journal of sociology 48 (4): 561. doi:10.2307/591596.
  • Semel, Beth Michelle. 2019. Speech, Signal, Symptom: Machine Listening and the Remaking of Psychiatric Assessment. MIT.
  • ShotSpotter. 2022. “Shotspotter: Cities.” ShotSpotter. https://www.shotspotter.com/cities/
  • Snell, John. 1977. “Editorial Introduction.” Computer Music Journal 1 (1): 2.
  • Snell, John. 2006. “How Did” Computer Music Journal“Come to Exist?” Computer Music Journal 30 (1): 10–20. doi:10.1162/comj.2006.30.1.10.
  • Steinbeck, Paul. 2018. “George Lewis’s Voyager.” In The Routledge Companion to Jazz Studies, edited by Nicholas Gebhardt, Nichole Rustin-Paschal, and Tony Whyton, 1st ed., 261–270. Routledge. doi:10.4324/9781315315805-25
  • Sterne, Jonathan. 2003. The Audible Past: Cultural Origins of Sound Reproduction. Durham and London: Duke University Press.
  • Sterne, Jonathan. 2012. The Sound Studies Reader. New York: Routledge.
  • Sterne, Jonathan. 2022. “Is Machine Listening Listening?” Communication +1 9: 5.
  • Sterne, Jonathan, and Elena Razlogova. 2021. “Tuning Sound for Infrastructures: Artificial Intelligence, Automation, and the Cultural Politics of Audio Mastering.” Cultural Studies 35 (4–5): 750–770, March 1–21. doi:10.1080/09502386.2021.1895247.
  • Sterne, Jonathan, and Mehak Sawhney. 2022. “The Acousmatic Question and the Will to Datafy.” The Lancet HIV 9 (6): e404–413. doi:10.1016/S2352-3018(22)00046-7.
  • Stowell, Dan, Dimitrios Giannoulis, Emmanouil Benetos, Mathieu Lagrange, and Mark D. Plumbley. 2015. “Detection and Classification of Acoustic Scenes and Events.” IEEE Transactions on Multimedia 17 (10): 1733–1746. doi:10.1109/TMM.2015.2428998.
  • Szendy, Peter. 2017. All Ears : The Aesthetics of Espionage. Books at JSTOR Demand Driven Acquisitions. New York, NY: Fordham University Press.
  • Turner, Fred. 2006. From Counterculture to Cyberculture: Stewart Brand, the Whole Earth Network, and the Rise of Digital Utopianism. Chicago: University of Chicago Press.
  • Turner, Fred. 2008. “Romantic Automatism: Art, Technology, and Collaborative Labor in Cold War America.” Journal of Visual Culture 7 (1): 5–26. doi:10.1177/1470412907087201.
  • Vercoe, Barry. 1984. “The Synthetic Performer in The Context of Live Performance.” In International Computer Music Conference Proceedings, Paris. http://hdl.handle.net/2027/spo.bbp2372.1984.026
  • Vercoe, Barry. 1990. “Synthetic Listeners and Synthetic Performers.” In Proceedings, International Symposium on Multimedia Technology and Artificial Intelligence (Computerworld 90), Kobe Japan, 136–141.
  • Vetter, Grant. 2012. The Architecture of Control: A Contribution to the Critique of the Science of Apparatuses. Winchester: John Hunt Publishing.
  • Wang, Avery Li-Chun. 2003. “An Industrial-Strength Audio Search Algorithm,“ International Society for Music Information Retrieval, Washington and Baltimore.
  • Wang, Wenwu. 2010. Machine Audition: Principles, Algorithms and Systems. 1st ed. Hershey, PA: IGI Global.
  • Wang, Deliang, and Guy J. Brown. 2006. “Fundamentals of Computational Auditory Scene Analysis.” In Computational Auditory Scene Analysis, 1–44. New Jersey: Wiley.
  • Wark, McKenzie. 2019. Capital Is Dead. London, New York: Verso.
  • Weber, Max. 2013. From Max Weber: Essays in Sociology. London: Routledge.
  • Whittaker, Meredith. 2021. “The Steep Cost of Capture.” Interactions 28 (6): 50–55. doi:10.1145/3488666.
  • Wold, Erling, Thom Blum, Doug Keislar, and JAmes Wheaton. 1996. “Content-Based Classification, Search, and Retrieval of Audio | IEEE MultiMedia.” IEEE MultiMedia 3 (3): 3. doi:10.1109/93.556537.
  • Wolfinger, Kirk, dir. 1990. AT&T Dawn Of Speech Recognition Technology: Employee Video. https://www.youtube.com/watch?v=GDtEkxUH7qE
  • Xiang, Ning, and Gerhard M. Sessler, eds. 2015. Acoustics, Information, and Communication: Memorial Volume in Honor of Manfred R. Schroeder. Cham: Springer International Publishing. doi:10.1007/978-3-319-05660-9.
  • Yu, Haizi, and Lav R. Varshney. 2017. “On ‘Composing Music by a Stochastic Process’: From Computers That Are Human to Composers That Are Not Human.” EEE Information Theory Society Newsletter, December.