Sonoscopy of a Panacoustic

 2021  ·  Book  ·  Podcast

This research focuses on an emerging field of computer science called Machine Listening. After teaching our machines to see, the next step is to enable them to hear and make sense of the sounds in our environments, thereby expanding the scope of contemporary surveillance.

This book does not provide direct access to its content (available in the form of an audio file), but it is a visualization of how an automatic listening program analyzes sounds and deduces information about their nature. I recorded myself reading its content in various locations (at home, in the city, in a bar, in a cathedral…) using a microphone that captured both my voice and the surrounding environment. I then submitted this recording to a pre-trained neural network for audio pattern recognition, accessible as open source, to experiment with automatic identification of sound environments. This customized program transcribes all the sound events, such as speech, a drawer opening and closing, or a meow, while attributing a recognition rate to them. The textual interpretation it provides of the recording is the sole composition of the pages of the book.

Upon analyzing the recognized sounds, I noticed that biases had crept in. For example, my voice is always labeled as a male voice: “Male speech,” because the program was trained on a sound database where male voices are predominant.

The text resulting from this operation can be read as follows: the recording is divided into blocks. A block represents 10 seconds of the recording. The first block, therefore, is the analysis of the first 10 seconds of the recording.

‘Metadata’ visible on the first line groups the metadata of the file, such as its name, block length, block size, sampling rate, number of blocks, and the parameters I pre-specified for the analysis.

‘Raw_datas’ is the list by block of all the sounds successively recognized in the recording. A block is formatted as follows: [['Speech', 0.7751136], ['Music', 0.21910243], ['Inside, small room', 0.13027504], ['Male speech, man speaking', 0.11066271]], enclosed in brackets.

‘Tags’ indicates the recurrence of each identified sound in the different blocks and its recognition rate. For example: ‘Walk, footsteps’: [0.15982443, 0, 0, 0, …], means that footsteps are recognized at 15.98% in block 1, but are not recognized in blocks 2 and 3, and so on.


This project was exhibited during the Regional Festival of Hybrid Arts and Digital Cultures as part of the Human Tech Days for the exhibition un Cabinet de curiosités numériques  organized by La Labomedia at 108 in Orléans from June 19 to June 23, 2023, 

At the Centre de Création Contemporaine (CCCOD) in Tours during the exhibition U.S.B #6 nous vivons à la lisière,  from March 24 to May 21, 2023,

At the ESAD Orléans for the exhibition U.S.B #3 ∙ Void draw () {Carte blanche aux diplômé∙es 2022},  from November 17 to December 1, 2022.

Podcast : Sonoscopy of a Panacoustic (2021)



Exhibition un Cabinet de curiosités numériques organized by La Labomedia for the Regional Festival of Hybrid Arts and Digital Cultures as part of the Human Tech Days.
June 19-23, 2023, at Le 108 in Orléans.

Photo : La Labomedia
Photo : Paul de Lanzac