Next Article in Journal
Remote Thermoelastic Characterization of Candidate Structural and Protective Coatings for Lead-Bismuth Eutectic Cooled Nuclear Reactors
Next Article in Special Issue
A Study of the Dynamic Response of Carbon Fiber Reinforced Epoxy (CFRE) Prepregs for Musical Instrument Manufacturing
Previous Article in Journal
Experimental Investigation of the Natural Bonding Strength between Stay-In-Place Form and Concrete in FRP-Concrete Decks/Beams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Design and Application of the BiVib Audio-Tactile Piano Sample Library

by
Stefano Papetti
1,†,
Federico Avanzini
2,† and
Federico Fontana
3,*,†
1
Institute for Computer Music and Sound Technology (ICST), Zurich University of the Arts, CH-8005 Zurich, Switzerland
2
LIM, Department of Computer Science, University of Milan, I-20133 Milano, Italy
3
HCI Lab, Department of Mathematics, Computer Science and Physics, University of Udine, I-33100 Udine, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2019, 9(5), 914; https://doi.org/10.3390/app9050914
Submission received: 11 January 2019 / Revised: 20 February 2019 / Accepted: 26 February 2019 / Published: 4 March 2019
(This article belongs to the Special Issue Musical Instruments: Acoustics and Vibration)

Abstract

:
A library of piano samples composed of binaural recordings and keyboard vibrations has been built, with the aim of sharing accurate data that in recent years have successfully advanced the knowledge on several aspects about the musical keyboard and its multimodal feedback to the performer. All samples were recorded using calibrated measurement equipment on two Yamaha Disklavier pianos, one grand and one upright model. This paper documents the sample acquisition procedure, with related calibration data. Then, for sound and vibration analysis, it is shown how physical quantities such as sound intensity and vibration acceleration can be inferred from the recorded samples. Finally, the paper describes how the samples can be used to correctly reproduce binaural sound and keyboard vibrations. The library has potential to support experimental research about the psycho-physical, cognitive and experiential effects caused by the keyboard’s multimodal feedback in musicians and other users, or, outside the laboratory, to enable an immersive personal piano performance.

1. Introduction

During instrumental performance musicians are exposed to auditory, visual and also somatosensory cues. This multisensory experience has been studied since long [1,2,3,4,5], however the specific interaction between sound and vibrations has been object of systematic research since the 1980’s [6,7,8,9,10,11,12], when tactile and force feedback cues started to be recognized to have a prominent role in the complex perception-action mechanisms occurring during musical instrument playing [13]. More recently, research on the somatosensory perception of musical instruments has been consolidated, as testified by the emerging “musical haptics” topic [14].
This increased interest is partly due to the increased availability of accurate yet affordable sensors and actuators, capable of recording touch gestures and rendering vibrations and force, respectively. Using these devices, complex experimental settings can be realized to measure and deliver multisensory information in a musical instrument, often as a result of a real-time analysis and synthesis process taking place during the performance [15,16,17]. Once integrated with traditional audio microphone and loudspeaker systems, touch sensors and actuators can be employed first to investigate the joint role of the auditory and somatosensory modality in the perception of a musical instrument, and then to realize novel musical interfaces and instruments building on the lessons of the previous investigation. Through this process, richer or even unconventional feedback cues can be conveyed to the performer, with the aim of increasing engagement, and hence the initial acceptability and subsequent playability of the new instrument [18,19,20,21].
In this scenario, the availability of multimodal databases combining and synchronizing different streams of information (audio, video, kinematic data of the instrument and performer in action, physiological signals, interactions among musicians etc.) is increasingly recognized as an essential asset for studying music performance. Recent examples include the “multimodal string quartet performance dataset” (QUARTET) [22], the “University of Rochester multimodal music performance dataset” (URMP) [23], the “database for emotion analysis using physiological signals” (DEAP) [24], the “TU-Note violin sample library” [25]. Furthermore, initiatives aiming at systematizing the creation of these databases have recently appeared, such as RepoVizz [26], a framework for storing, browsing, and visualizing synchronous multimodal data. Overall, this scenario suggests an increasing attention to the design of musical instrument sound databases, motivated by the concrete possibility for their content to be reproduced in instrumental settings including not only loudspeakers and headphones, but also haptic and robotic devices.
The piano represents an especially relevant case study not only for its importance in the history of Western musical tradition, but also for its potential in the musical instruments market due to the universality of the keyboard interface, a feature that has traditionally induced novel musical instrument makers to propose conservative instances of the standard piano keyboard whenever this interface made possible to control even revolutionary sound synthesis methods [27]. It is no accident that, also in recent years, sales of digital pianos and keyboard synthesizers have shown a growing trend as opposed to other instrument sales (https://www.namm.org/membership/global-report). For the same reason, researchers on new musical instruments have steadily elected the piano keyboard as the platform of choice for designing expansions of the traditional paradigm, affording a performer to accurately play two local selections of 88 available tones with the desired amplitude and temporal development [28,29].
When playing an acoustic piano, the performer is exposed to a variety of auditory, visual, somatosensory, and vibrotactile cues that combine and integrate to shape the pianist’s perception–action loop. The present authors are involved in a long-term research collaboration around this topic, with a focus on two main aspects. The first one is the tactile feedback produced by keyboard vibrations that reach the pianist’s fingers after keystrokes and remains active until key release. The second one is the auditory spatial information in the sound field radiated by the instrument at the performer’s head position. Binaural piano tones are offered by a few audio plugin developers (e.g., Modartt Pianoteq (https://www.pianoteq.com/)) and digital piano manufacturers (e.g., Yamaha Clavinova (https://europe.yamaha.com/en/products/musical_instruments/pianos/clavinova/)), but they have limited flexibility of use. Free binaural piano samples can be found, too, such as the “binaural upright piano” library, (https://www.michaelpichermusic.com/binaural-upright-piano) which however offers only three dynamic levels (as opposed to the ten levels provided by the present dataset). More generally concerning the presentation of audio-tactile piano tones, the existing literature is scarce and provides mixed if not contradictory results about the actual perceptibility and possible relevance of this multisensory information [8]. Specific discussions of these aspects have been provided in previously published studies, regarding both sound localization [30] and vibration perception [12] on the acoustic piano. As a notable result of these studies, a digital piano prototype was developed that reproduces various types of vibrations [20] including those generated by acoustic pianos as an unavoidable by-product of their sound production.
Across the years, this research has resulted in the production of an extensive amount of experimental data, most of which resulting from highly accurate measurements with calibrated devices. In an effort to provide public access to such data, the authors of this paper present a dataset of audio-tactile piano samples organized as libraries of synchronized binaural sound and vibration signals.
The dataset contains samples relative to all 88 tones, played at different dynamics on two instruments: a grand and an upright piano. A preliminary version was presented at a recent conference [31], and has now been updated by a new release containing additional upright piano sounds along with the binaural impulse responses of the room in which the same piano was recorded. In order to use this dataset, it is necessary to take into account the recording conditions and, on the user side, to take control of the rendering system in an effort to match the reproduction to the characteristics of the original recorded signals. For this reason, Section 2 describes the hardware/software recording setup and the organization of samples into libraries for the free version of a popular music software sampler. An explanation about how to reproduce the database is provided in Section 3. Section 4 suggests some applications of this library, based on the authors’ past design experiences with multimodal piano samples, and on more experiences that are foreseen for future research.
Among past experiences, the most prominent have consisted of studies on the role of haptic feedback during the performance, both vibratory and somatosensory: the former concerns the perception of keyboard vibrations, their accurate reproduction and their effects on the performance [12,20]; the latter concerns the influence of actively playing on the keyboard in the auditory localization process of piano tones [30,32]. Such experiences let to the decision of sharing the dataset with the scientific community, with the goal of fostering the research on the role of vibrations and tone localization in the pianist’s perceived instrument quality (both not completely understood yet), as well as adding knowledge about the importance at cognitive level of multisensory feedback for its use in the design of novel keyboard interfaces.

2. Creation of the BiVib library

The BiVib (binaural and vibratory) sample library is a collection of high-resolution audio files (.wav format, 24-bit @ 96 kHz) containing binaural piano sounds and keyboard vibrations, coming along with documentation and project files for its reproduction through a free music software sampler. The dataset, whose core structure is illustrated in Table 1, is made available through an open-access data repository (https://zenodo.org/record/2573232) and released under a Creative Commons (CC BY-NC-SA 4.0) license.

2.1. Recording Hardware

The samples were recorded from two Yamaha Disklavier pianos—a grand model DC3 M4 located in Padova (PD), Italy, and an upright model DU1A with control unit DKC-850 located in Zurich (ZH), Switzerland. Disklaviers are Musical Instrument Digital Interface (MIDI from here on)-compliant acoustic pianos equipped with sensors for recording keystrokes and pedaling, and operating electromechanical motors for playback. The grand piano is located in a large laboratory space (approximately 6 × 4  m), while the upright piano is in an acoustically treated small room (approximately 4 × 2  m).
Binaural audio recordings made use of dummy heads for acoustic measurements, with slightly different setups in PD and ZH: the grand piano was recorded with a KEMAR 45BM (GRAS Sound & Vibration A/S, Holte, Denmark), whereas the upright piano with a Neumann KU 100 (Georg Neumann GmbH, Berlin, Germany). Both mannequins were placed in front of the pianos, approximately where the pianist’s head is located on average (see Figure 1).
The dummy heads were connected to the microphone inputs of two professional audio interfaces, a RME Fireface 800 (PD, gain set to + 40  dB) (Audio AG, Haimhausen, Germany) and a RME UCX (ZH, gain set to + 20  dB) (Audio AG, Haimhausen, Germany). The pair of condenser microphones inside the dummy heads were respectively driven by two 26CB preamplifiers supplied by a 12AL power module (PD), and by 48V phantom power provided by the audio interface (ZH).
Three configurations of the keyboard lid were selected for each piano. The grand piano (PD) was measured with the lid closed, fully open, and removed (i.e., detached from the instrument). The upright piano was recorded with the lid closed, in semi-open position (see Figure 1), and fully open. Different lid configurations in fact add insight on the role of the mechanical noise coming from the moving keys to the creation of transient cues of lateralization in the sound field reaching the performer’s ears [30]. As a result, three sets of binaural samples were recorded for both pianos—one set for each lid position.
Vibration recordings were acquired with a Wilcoxon Research 736 (Wilcoxon Sensing Technologies Inc., Amphenol, MD, USA) piezoelectric accelerometer connected to a Wilcoxon Research iT100M (Wilcoxon Sensing Technologies Inc., Amphenol, MD, USA) intelligent transmitter, whose AC-coupled output fed one line input of the audio interface. The accelerometer was manually attached with double-sided adhesive tape to each key in sequence, as shown in Figure 2. Its center was positioned 2 cm far from the key edge, where most vibration modes are radiated efficiently, and where pianists typically put their fingertip.

2.2. Recording Software

Ten values of dynamics were chosen between MIDI key velocity 12 and 111 by evenly splitting this range into eleven intervals. This choice was motivated by a previous study by the present authors, which reported that both Disklaviers produced inconsistent dynamics outside this velocity range [12]. In general, the servo-mechanics of computer-controlled pianos fall short (to an extent that depends on the model) of providing a reliable response at extreme dynamic values [33]. Holding this assumption, two different software setups were used respectively for sampling sounds and vibrations.
Binaural samples were recorded via an automatic procedure programmed in SuperCollider. (A programming environment for sound processing and algorithmic composition: http://supercollider.github.io/.) The recording sessions took place at night time, thus minimizing unwanted noise coming from human activity in the building. On the grand piano, note lengths were determined algorithmically depending on their dynamics and pitch, ranging from 30 s when A0 was played at key velocity 111, to 10 s when C8 was played at key velocity 12. In fact, notes of increasing pitch and/or decreasing dynamics have shorter decay times. These durations allow each note to fade out completely, while minimizing silent recordings and the overall duration of the recording sessions—still, each session lasted approximately six hours.
On the contrary, an undocumented protection mechanism on the upright piano prevents its electromechanical system from holding down the keys for more than about 17 s, thus making a complete decay impossible for some notes, especially at low pitches and high dynamics. In this case, for the sake of simplicity all tones were recorded for just as long as possible.
Vibration samples were recorded through a less sophisticated procedure. A digital audio workstation (DAW) software was used to play back all notes in sequence at the same ten MIDI velocity values as those used for the binaural audio recordings, using a constant duration of 16 s. This duration in fact is certainly greater than the time taken by any key vibration to decay below sensitivity thresholds [12,34].

2.3. Sample Pre-Processing

Because of the mechanics of piano keyboards and the intrinsic limitations of computer-controlled electro-mechanical actuation, a systematic delay is introduced while reproducing MIDI note ON messages, which mainly varies with key dynamics. For this reason, all recorded samples started with silence of varying durations, which had to be removed in view of their use in a sampler (see Section 2.4). Given the number of files that had to be pre-processed (880 for each set), an automated procedure was implemented in SuperCollider to cut the initial silence of each audio sample. The procedure analyzed the amplitude envelope, detected the position of the largest peak, and finally applied a short fade-in starting a few milliseconds before the peak.
Additionally, vibration signals presented abrupt onsets in the first 200–250 ms right after the starting silence, as a consequence of the initial fly of the key and of the following impact against the piano keybed (see Figure 3). These onsets are not related to keyboard vibrations, and therefore they had to be removed.
As such onset profiles showed large variability, in spite of several tests made in MATLAB, no reliable automated procedure could be realized for editing the vibration samples. A manual approach was employed instead: files were imported in a sound editor, their waveform was zoomed in and played back, and the onset was cut off.

2.4. Sample Library Organization

The sample library was organized for playback with the ‘Kontakt Player’ software—a free version of Native Instruments’ Kontakt sampler, (https://www.native-instruments.com/en/products/komplete/samplers/kontakt-5-player/) available for Windows and Mac OS systems. The full version of Kontakt 5 was instead used to develop Kontakt project files as described below. The resulting library is organized into several folders, named ‘Instruments’, ‘Multis’, ‘Resources’, and ‘Samples’.
The ‘Samples’ folder—whose total size amounts to about 65 GB—held separate subfolders for the binaural and vibration samples, respectively, which contain further subfolders for each sample set (see Table 1), for example ‘grand-open’ under the ‘binaural’ folder. Independent of their type, sample files were named according to the following mask:
[note][octave #]_[lower MIDI velocity]_[upper MIDI velocity].wav,
where [note] follows the English note-naming convention, [octave #] ranges from 0 to 8, [lower MIDI velocity] equals the MIDI velocity (range 12–111) used during recording and is the smallest velocity value mapped to that sample in Kontakt (see below), [upper MIDI velocity] is the largest velocity value mapped to that sample in Kontakt. As an example, the file A4_100_110.wav corresponds to the note A from the 4th octave (fundamental frequency 440 Hz) recorded at MIDI velocity 100, and mapped to the velocity range 100–110 in Kontakt. Since the lowest recorded velocity value was 12, for consistency no samples were mapped to the velocity range 1–11 in Kontakt.
Following the terminology used in Kontakt, each instrument reproduces a sample set (e.g., binaural recording of the grand piano with lid open), while each multi combines two instruments respectively reproducing one binaural and one vibration sample set belonging to the same piano. The two instruments in each multi are configured so as to receive MIDI input data on channel 1, thus playing back at once, while their respective outputs are routed to different virtual channels in Kontakt: binaural samples are routed to a pair of stereo channels (numbered 1–2), while vibration samples are played through a mono channel (numbered 3). In this way, when using audio interfaces offering more than two physical outputs, it is possible to render synchronized binaural and vibrotactile cues at the same time by routing the audio signal to headphones and, in parallel, the vibration signal to an amplifier driving one or more actuators.
In each instrument, sample mapping was implemented relying on the ‘auto-map’ feature found in the full version of Kontakt: this parses file names and uses the recognized tokens for assigning samples to e.g., a pitch and velocity range. The chosen file naming template made it straightforward to batch-import the samples.
The amplitude of the recorded signals was not altered, that is, no dynamic processing or amplitude normalization was applied. The volume of all Kontakt instruments was set to 0 dB. Because of this setting and of the adopted velocity mapping strategy, sample playback was kept as transparent as possible for simplifying the setup of acoustic and vibratory analysis procedures, experiments and interactive applications (see Section 3 and Section 4).

3. Application of the BiVib Library

Binaural piano tones such as those offered by Yamaha digital pianos or the Modartt Pianoteq software synthesizer are not fully suitable for research purposes due to being undocumented, hence non reproducible, acquisition procedures and/or post-processing of the sound signals. Moreover, the present authors have no evidence of public samples including piano keyboard vibrations. Thus, BiVib fills two gaps found in the datasets currently available for the reproduction of piano feedback.

3.1. Sample Reproduction

Experiments and applications requiring the use of calibrated data need exact reconstruction of the measured signals at the reproduction side: acoustic pressure for binaural sounds, and acceleration for keyboard vibrations. Obviously the reproduction must take place on a set-up in which neither autonomous sounds nor vibrations are present. For instance, the reproduction of vibrations could take place on a weighted MIDI keyboard (such as those found in digital pianos), while binaural sounds may be rendered through headphones. Figure 4 (left) shows one such setting, in which a commercial digital piano is augmented through the setup schematized as in Figure 4 (right). Note that the bottom of the digital piano has been reinforced by substituting the keybed with a thicker wooden panel, to form what we will call a haptic digital piano from here.
Knowledge of the recording equipment’s nominal specifications enable this reconstruction. Such specifications are summarized in a companion document in the ‘Documentation’ folder. Two examples are listed below:
  • Vibration accelerations in m/s2 are calculated from the vibration samples by making use of the nominal sensitivity values of the audio interface and accelerometer: voltage values are reconstructed from the recorded digital signals (represented by values between −1 and 1) by means of the line input sensitivity of the audio interface, whose full scale level (0 dBFS) was set to + 19 dBu (reference voltage 0.775 V) for recording. Finally, voltage values are transformed into proportional acceleration values through the sensitivity constant of the accelerometer, equal to 10.2 mV/m/s2 at 25 °C, with a flat ( ± 5 %) frequency response in the range 10–15,000 Hz.
    As an example, the digital value 0.001 is equivalent to:
    20 × log 10 ( | 0.001 | ) = - 60 dBFS
    and the corresponding voltage value is calculated based on the 0 dBFS level ( + 19 dBu, with 0.775 V reference) as follows:
    0.775 × 10 ( ( 19 - 60 ) / 20 ) = 6.9 mV .
    Finally, the matching acceleration value is obtained as:
    6.9 mV 10.2 mV / m / s 2 = 0.667 m / s 2 .
  • Analogously, acoustic pressure values in Pa are obtained from the binaural samples by making use of the relevant nominal specification values of the microphone input chain (microphone, pre-amplifier if present, audio interface input). For the upright piano: microphone sensitivity of 20 mV/Pa, audio interface input gain set to + 26 dB for an equivalent 0 dBFS level of - 16 dBu (reference 0.775 V). For the grand piano: microphone sensitivity 50 mV/Pa, preamplifier gain - 0.35 dB, and audio interface input gain set to + 40 dB for an equivalent 0 dBFS level of - 19 dBu (reference 0.775 V).
    As an example, the digital value 0.001 , equivalent to - 60 dBFS as seen in Equation (1), translates to the following voltages:
    0.775 × 10 ( ( - 16 - 60 ) / 20 ) = 0.12 mV
    for the upright piano, or
    0.775 × 10 ( ( - 19 - 60 + 0.35 ) / 20 ) = 0.09 mV
    for the grand piano. The equivalent acoustic pressures are then computed as:
    0.12 mV 20 mV / Pa = 0.006 Pa ( about 49.5 dB SPL )
    for the upright piano, or
    0.09 mV 50 mV / Pa = 0.0018 Pa ( about 39 dB SPL )
    for the grand piano.
Once the original measurements have been reconstructed, physical quantities are, in principle, ready for presentation of sounds through headphones and vibrations through tactile actuators, holding the issues and limitations that are listed in the next section.

3.2. Sample Equalization and Limitations

BiVib users should hear the same sounds as the binaural pressure signals measured by the dummy head microphones, and should receive vibrations at the fingers which are identical to the acceleration signals measured by the accelerometer.
The audio reproduction condition can be satisfied by using headphones or earphones that bypass the outer ear and provide a frequency response that is as flat as possible. In this way, the performer would listen to stereo sounds that contain the binaural information created by the dummy head. Note that these sounds also contain the contribution of room reverberation, as the pianos could not be located inside anechoic rooms during the recording sessions: especially the laboratory hosting the grand piano was moderately reverberant, and no reverberation data could be collected for this room at the time of the acquisition of samples. Conversely, the upright piano (ZH) benefited of a silent studio room whose impulse responses have been measured and recently included in BiVib (folder ‘IR’). More precisely, responses from two source points generated by a pair of Genelec 8040A loudspeakers were taken in correspondence of the two ear canal entrances of the Neumann KU 100 dummy head with both outer ears removed: As shown in Figure 5, the loudspeakers were placed symmetrically at both sides of the upright piano, 80 cm above the floor, pointing toward the dummy head (angle with the vertical plane parallel to the piano equal to 47°) which was positioned 55 cm far from the upright panel of the piano. Logarithmic sweeps were synthesized in the Audacity audio editor using the Aurora modules [35], then reproduced using the loudspeakers, and finally deconvolved again with Aurora to find four room transfer functions forming the source-ear transfer matrix. Such transfer functions were included in the dataset and can be used to remove the echoes of the recording room through inversion of the transfer matrix, using standard deconvolution techniques [36]. Also, because of the low energy of these echoes, binaural recordings were intentionally left unprocessed, thus allowing musicians and performers to use them as they are, while leaving any decision on their possible manipulation to advanced users.
The vibration reproduction condition may be ideally satisfied by implementing a weighted MIDI keyboard (such as those found on digital pianos) where each key is provided with an actuator, similar to the prototype technology described in [37]. Even in the unlikely case that such actuators would offer a flat frequency response over the range of interest for tactile sensation [38], vibration reproduction would be altered by the shape, material and construction of the keyboard. That requires the implementation of a compensation procedure—similar to that described in the previous paragraph to remove room acoustic components—that would enable the accurate reproduction of the vibrations recorded by the accelerometer on each key. Unfortunately, on such a keyboard there would be 88 touch points requiring deconvolution. In alternative, two arrays of 88 transfer characteristics can be formed each by measuring vibrations on the same touch points, after exciting the digital piano body with a corresponding pair of tactile transducers.
Figure 6 (left) [12] shows a detail of the haptic digital piano that the authors realized according to the latter solution, where two transducers were mounted below the keybed, one mid-way the keyboard length and another approximately at one quarter way the same length—see also Figure 4 (right).
By playing an impulse simultaneously on both of them, advanced BiVib users should accurately measure on each key the corresponding impulse response, and then (with Aurora or similar tools) design an equalizer which on average flattens the “coloured” response affecting the vibrations due to their path from the audio interface to the keyboard. Since such measures depend on the user’s musical keyboard, they could not be included in BiVib as the binaural room transfer functions were instead. However, in the aforementioned digital piano customization it was experimentally observed by the authors that the tactile transducers shown in Figure 6 (left) were mostly responsible for the inclusion of distinct spectral peaks affecting all the key responses. Hence, the average equalization curve shown in Figure 6 (right) was estimated, and proved to be effective for flattening the peaks measured on an evenly-spaced subset of the keys. A similar curve can be approximated by BiVib users by designing an equalizer that flattens the frequency response of the tactile transducers equipping the setup. This response is normally reported in a good quality transducer’s data sheet.

4. Experiments, Applications and Future Work

BiVib has been originally created to support multisensory experiments in which precise control had to be maintained over the simultaneous auditory and vibrotactile stimuli reaching a performing pianist, particularly when judging the perceived quality of an instrument. In a recent paper, the authors were able to confirm subjective vibrotactile frequency thresholds of active touch [34] by conducting tests with pianists who were asked to detect vibrations at the piano keyboard [12]. A comparison between such thresholds and the spectrum of a lower fortissimo A0 tone is shown in Figure 7. This demonstrates a previously unreported results: during active playing, pianists are able to integrate tactile sensation that would be imperceptible in passive touch conditions [38].
A precise manipulation of the intensity relations between piano sound and vibrations may be used to investigate the existence of cross-modal effects occurring during piano playing. Such effects have been discovered as part of a more general multisensory integration mechanism [39] that under certain conditions can increase the perceived intensity of auditory signals [40], or conversely enhance touch perception [41]. Aiming to understand whether piano keyboard vibrations impact the perceived quality of the instrument and, as a secondary effect, the quality of a performance, the authors have first observed significant differences in the perceived quality of vibrating vs. silently playing Disklavier pianos. This observation marks a point in favor of the former, especially since that pianists involved in the experiment reported to be unaware of the existence of vibratory feedback [12]. In other words, while most pianists preferred the vibrating instrument, they did not consciously realize that their decision was caused by the vibrations produced (or not produced) by the instrument during the performance. Based on this result, a further experiment was designed for the haptic digital piano using diverse types of tactile feedback, synthesized by manipulating the BiVib samples. The test aimed at investigating possible consequences of the vibrotactile feedback on the pianist’s playing experience (qualitative effect) and on the performance in terms of timing and dynamics accuracy (quantitative effect). Cross-modal effects resulting from varying the tactile feedback of the keyboard were observed, still these preliminary results are far from giving a systematic view about the impact of the different sensory channels on the pianist’s playing experience, and especially on the accuracy of execution [20].
Another potential use of BiVib is in the investigation of binaural spatial cues for the acoustic piano. Using the recommendations given in the previous section, single tones of the upright piano can in fact be accurately cleared of the room echoes, and then be reproduced in the position of the ear entrance. The existence of localization cues in piano sounds has not been completely understood yet. Even in pianos where these cues are reported to be audible by listeners, their exact acoustic origin is still an open question [42]. Moreover, visual cues of self-moving keys (a condition possible on Disklavier pianos) producing the corresponding tones, as well as somatosensory cues occurring during active piano playing, may have an influence on localization judgments [30,32].
One further direction which may take advantage of BiVib deals with research in cognitive neuroscience: recently, pianists and their instrument served as key subjects for understanding diverse aspects of brain and motor development [43,44]. In this context, the contribution of the auditory, visual and tactile sensory modalities to this development have not been ascertained yet. Such knowledge could be of help not only to capture more general aspects of the development of human senses, but also to guide a perceptually and cognitively informed design of novel keyboard interfaces.
In summary, future research that can be conducted in the laboratory using BiVib includes tests aiming at conclusively understanding whether (i) vibrations affect the performance on the keyboard, and whether (ii) auditory lateralization is able to guide piano tone localization. If so, such multisensory cues may substantially contribute to the sense of engagement and, hence, improve the quality of the performance and make the learning curve of a keyboard interface more acceptable.
Outside the laboratory, the library can reward musicians who simply wish to use its sounds. On the one hand, the grand piano recordings contain echoes that bring distinct cues of the room where they have been recorded. In this sense they are ready for use, although labeled by a precise acoustic footprint. On the other hand, the upright piano recordings are much more anechoic and, hence, far less difficult to spatialize than piano recordings taken outside an acoustically controlled room [45]. Consequently, they can be easily imported and conveniently personalized by users through artificial reverberation.

5. Conclusions

BiVib provides a unique set of multimodal piano data recorded using high-quality equipment in controlled conditions through reproducible computer-controlled procedures. Since its original release, the library has been enriched with binaural responses of the room where the upright piano was recorded. We hope that a successful use of the BiVib dataset, in conjunction with this documentation and through publicly available projects for the Kontakt software sampler, will facilitate further research in piano acoustics, performance, and new musical interface design also for educational purposes.

Author Contributions

Conceptualization, F.A., F.F. and S.P.; methodology, F.A., F.F. and S.P.; software, S.P.; resources, F.A., F.F. and S.P.; data curation, S.P.; writing—original draft preparation, S.P.; writing—review and editing, F.A., F.F. and S.P.

Funding

This research was partially funded by Swiss National Science Foundation grants number 150107 and 178972.

Acknowledgments

This research received support by project AHMI (Audio-haptic modalities in musical interfaces, 2014–2016), and HAPTEEV (Haptic technology and evaluation for digital musical interfaces 2018–2022), both funded by the Swiss National Science Foundation. Support came also from the PRID project SMARTLAND, funded by the University of Udine. The Disklavier grand model DC3 M4 located in Padova was made available by courtesy of the Sound and Music Processing Lab (SaMPL), a project of the Conservatory of Padova funded by Cariparo Foundation (thanks in particular to Nicola Bernardini and Giorgio Klauer). Gianni Chiaradia donated the wooden panel used to build the haptic digital piano. The authors would like to thank several students and collaborators who contributed to the development of this work along the years, in chronological order: Francesco Zanini, Valerio Zanini, Andrea Ghirotto, Devid Bianco, Lorenzo Malavolta, Debora Scappin, Mattia Bernardi, Francesca Minchio, Martin Fröhlich.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Turner, E.O. Touch and Tone-Quality: The Pianist’s Illusion. Music. Times 1939, 80, 173–176. [Google Scholar] [CrossRef]
  2. Palmer, C.F.; Jones, R.K.; Hennessy, B.L.; Unze, M.G.; Pick, A.D. How is a trumpet known? The “basic object level” concept and perception of musical instruments. Am. J. Psychol. 1989, 102, 17–37. [Google Scholar] [CrossRef] [PubMed]
  3. Galembo, A.; Askenfelt, A. Quality assessment of musical instruments—Effects of multimodality. In Proceedings of the 5th Triennial Conference of the European Society for the Cognitive Sciences of Music (ESCOM5), Hannover, Germany, 8–13 September 2003. [Google Scholar]
  4. Hodges, D.A.; Hairston, W.D.; Burdette, J.H. Aspects of Multisensory Perception: The Integration of Visual and Auditory Information in Musical Experiences. Ann. N. Y. Acad. Sci. 2005, 1060, 175–185. [Google Scholar] [CrossRef] [PubMed]
  5. Campbell, D.M. Evaluating musical instruments. Phys. Today 2014, 67, 35–40. [Google Scholar] [CrossRef] [Green Version]
  6. Marshall, K.; Genter, B. The musician and the vibrational behavior of a violin. J. Catgut Acoust. Soc. 1986, 45, 28–33. [Google Scholar]
  7. Suzuki, H. Vibration and sound radiation of a piano soundboard. J. Acoust. Soc. Am. 1986, 80, 1573–1582. [Google Scholar] [CrossRef]
  8. Askenfelt, A.; Jansson, E.V. On Vibration Sensation and Finger Touch in Stringed Instrument Playing. Music Percept. 1992, 9, 311–349. [Google Scholar] [CrossRef] [Green Version]
  9. Keane, M.; Dodd, G. Subjective Assessment of Upright Piano Key Vibrations. Acta Acust. United Acust. 2011, 97, 708–713. [Google Scholar] [CrossRef]
  10. Saitis, C. Evaluating Violin Quality: Player Reliability and Verbalization. Ph.D. Thesis, Department of Music Research, McGill University, Montreal, QC, Canada, 2013. [Google Scholar]
  11. Wollman, I.; Fritz, C.; Poitevineau, J. Influence of vibrotactile feedback on some perceptual features of violins. J. Acoust. Soc. Am. 2014, 136, 910–921. [Google Scholar] [CrossRef] [PubMed]
  12. Fontana, F.; Papetti, S.; Järveläinen, H.; Avanzini, F. Detection of keyboard vibrations and effects on perceived piano quality. J. Acoust. Soc. Am. 2017, 142, 2953–2967. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. O’Modhrain, S.; Gillespie, B.R. Once More, with Feeling: The Dynamics of Performer-Instrument Interaction. In Musical Haptics; Papetti, S., Saitis, C., Eds.; Springer: Berlin, Germany, 2018. [Google Scholar]
  14. Papetti, S.; Saitis, C. (Eds.) Musical Haptics; Springer: Berlin, Germany, 2018. [Google Scholar]
  15. O’Modhrain, S.; Chafe, C. Incorporating Haptic Feedback into Interfaces for Music Applications. In Proceedings of the World Automation Conference ISORA, Maui, HI, USA, 11–16 June 2000. [Google Scholar]
  16. Marshall, M.T.; Wanderley, M.M. Vibrotactile feedback in digital musical instruments. In Proceedings of the 2006 Conference on New Interfaces for Musical Expression, Paris, France, 4–8 June 2006; pp. 226–229. [Google Scholar]
  17. Overholt, D.; Berdahl, E.; Hamilton, R. Advancements in Actuated Musical Instruments. Organ. Sound 2011, 16, 154–165. [Google Scholar] [CrossRef]
  18. Birnbaum, D.M.; Wanderley, M.M. A systematic approach to musical vibrotactile feedback. In Proceedings of the International Computer Music Conference, Copenhagen, Denmark, 27–31 August 2007. [Google Scholar]
  19. Michailidis, T.; Bullock, J. Improving performers’ musicality through live interaction with haptic feedback: A case study. In Proceedings of the Sound and Music Computing (SMC), Padova, Italy, 6–9 July 2011. [Google Scholar]
  20. Fontana, F.; Avanzini, F.; Järveläinen, H.; Papetti, S.; Klauer, G.; Malavolta, L. Rendering and Subjective Evaluation of Real vs. Synthetic Vibrotactile Cues on a Digital Piano Keyboard. In Proceedings of the Sound and Music Computing Conference, Maynooth, Ireland, 26 July–1 August 2015; pp. 161–167. [Google Scholar]
  21. Young, G.W.; Murphy, D.; Weeter, J. A Qualitative Analysis of Haptic Feedback in Music Focused Exercises. In Proceedings of the New Interfaces for Musical Expression NIME’17, Copenhagen, Denmark, 15–18 May 2017; pp. 204–209. [Google Scholar]
  22. Maestre, E.; Papiotis, P.; Marchini, M.; Llimona, Q.; Mayor, O.; Pérez, A.; Wanderley, M.M. Enriched Multimodal Representations of Music Performances: Online Access and Visualization. IEEE MultiMed. 2017, 24, 24–34. [Google Scholar] [CrossRef]
  23. Li, B.; Liu, X.; Dinesh, K.; Duan, Z.; Sharma, G. Creating a Multitrack Classical Music Performance Dataset for Multimodal Music Analysis: Challenges, Insights, and Applications. IEEE Trans. Multimed. 2019, 21, 522–535. [Google Scholar] [CrossRef]
  24. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. Deap: A database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef]
  25. Von Coler, H. TU-Note Violin Sample Library—A Database of Violin Sounds with Segmentation Ground Truth. In Proceedings of the 21st International Conference on Digital Audio Effects (DAFx-18), Aveiro, Portugal, 4–8 September 2018; pp. 312–317. [Google Scholar]
  26. Mayor, O.; Llop, J.; Maestre Gómez, E. RepoVizz: A multi-modal on-line database and browsing tool for music performance research. In Proceedings of the 12th International Society for Music Information Retrieval Conference (ISMIR 2011), Miami, FL, USA, 24–28 October 2011. [Google Scholar]
  27. Moog, R.A.; Rhea, T.L. Evolution of the Keyboard Interface: The Bösendorfer 290 SE Recording Piano and the Moog Multiply-Touch-Sensitive Keyboards. Comput. Music J. 1990, 14, 52–60. [Google Scholar] [CrossRef]
  28. McPherson, A.P.; Gierakowski, A.; Stark, A.M. The space between the notes: Adding expressive pitch control to the piano keyboard. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Paris, France, 27 April–2 May 2013; ACM: New York, NY, USA, 2013; pp. 2195–2204. [Google Scholar]
  29. d’Alessandro, N.; Tilmanne, J.; Moreau, A.; Puleo, A. AirPiano: A Multi-Touch Keyboard with Hovering Control. In Proceedings of the International Conference on New Interfaces for Musical Expression; Berdahl, E., Allison, J., Eds.; Louisiana State University: Baton Rouge, LA, USA, 2015; pp. 255–258. [Google Scholar]
  30. Fontana, F.; Scappin, D.; Avanzini, F.; Bernardi, M.; Bianco, D.; Klauer, G. Auditory, Visual And Somatosensory Localization Of Piano Tones: A Preliminary Study. In Proceedings of the Sound Music Computing Conference (SMC), Espoo, Finland, 5–8 July 2017; pp. 254–260. [Google Scholar]
  31. Papetti, S.; Avanzini, F.; Fontana, F. BIVIB: A Multimodal Piano Sample Library of Binaural Sounds and Keyboard Vibrations. In Proceedings of the International Conference on Digital Audio Effects (DAFx-18), Aveiro, Portugal, 4–8 September 2018; pp. 237–243. [Google Scholar]
  32. Fontana, F.; Avanzini, F.; Papetti, S. Evidence of lateralization cues in grand and upright piano sounds. In Proceedings of the 15th International Conference on Sound and Music Computing (SMC2018), Limassol, Cyprus, 4–7 July 2018; pp. 80–84. [Google Scholar]
  33. Goebl, W.; Bresin, R. Measurement and reproduction accuracy of computer-controlled grand pianos. J. Acoust. Soc. Am. 2003, 114, 2273–2283. [Google Scholar] [CrossRef] [PubMed]
  34. Papetti, S.; Järveläinen, H.; Giordano, B.L.; Schiesser, S.; Fröhlich, M. Vibrotactile sensitivity in active touch: Effect of pressing force. IEEE Trans. Haptics 2017, 10, 113–122. [Google Scholar] [CrossRef] [PubMed]
  35. Farina, A. Simultaneous Measurement of Impulse Response and Distortion with a Swept-Sine Technique. In Proceedings of the 108th Convention of the Audio Engineering Society, Paris, France, 19–22 February 2000. [Google Scholar]
  36. Kirkeby, O.; Nelson, P.A.; Hamada, H.; Orduña-Bustamante, F. Fast deconvolution of multichannel systems using regularization. IEEE Trans. Speech Audio Process. 1998, 6, 189–194. [Google Scholar] [CrossRef] [Green Version]
  37. Papetti, S.; Schiesser, S.; Fröhlich, M. Multi-point vibrotactile feedback for an expressive musical interface. In Proceedings of the International Conference on New Interfaces for Musical Expression; Berdahl, E., Allison, J., Eds.; Louisiana State University: Baton Rouge, LA, USA, 2015; pp. 235–240. [Google Scholar]
  38. Verrillo, R.T. Vibration sensation in humans. Music Percept. 1992, 9, 281–302. [Google Scholar] [CrossRef]
  39. Kayser, C.; Petkov, C.I.; Augath, M.; Logothetis, N.K. Integration of Touch and Sound in Auditory Cortex. Neuron 2005, 48, 373–384. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Gillmeister, H.; Eimer, M. Tactile enhancement of auditory detection and perceived loudness. Brain Res. 2007, 1160, 58–68. [Google Scholar] [CrossRef] [PubMed]
  41. Ro, T.; Hsu, J.; Yasar, N.E.; Elmore, L.C.; Beauchamp, M.S. Sound enhances touch perception. Exp. Brain Res. 2009, 195, 135–143. [Google Scholar] [CrossRef] [PubMed]
  42. Askenfelt, A. Observations on the transient components of the piano tone. STL-QPSR 1993, 34, 15–22. Available online: http://www.speech.kth.se/qpsr/show_by_author.php?author=Askenfelt%2C+A (accessed on 1 March 2019).
  43. Kalra, S.; Jain, S.; Agarwal, A. Infrared Powered Vibro-Haptic Piano Training System for the Visually Impaired. Int. J. Adv. Soft Comput. Appl. 2017, 9, 49–68. [Google Scholar]
  44. Alves-Pinto, A.; Ehrlich, S.; Cheng, G.; Turova, V.; Blumenstein, T.; Lampe, R. Effects of short-term piano training on measures of finger tapping, somatosensory perception and motor-related brain activity in patients with cerebral palsy. Neuropsychiatr. Dis. Treat. 2017, 13, 2705. [Google Scholar] [CrossRef] [PubMed]
  45. Rocchesso, D. Spatial effects. In Digital Audio Effects; Zölzer, U., Ed.; John Wiley & Sons: Chirchester Sussex, UK, 2002; pp. 137–200. [Google Scholar]
Figure 1. Binaural audio recording setup in Zurich (ZH). Piano lid in ‘semi-open’ position.
Figure 1. Binaural audio recording setup in Zurich (ZH). Piano lid in ‘semi-open’ position.
Applsci 09 00914 g001
Figure 2. Vibration recording setup. A Wilcoxon Research 736 accelerometer is attached with adhesive tape to a key that is being played remotely via MIDI control.
Figure 2. Vibration recording setup. A Wilcoxon Research 736 accelerometer is attached with adhesive tape to a key that is being played remotely via MIDI control.
Applsci 09 00914 g002
Figure 3. Vibration signal recorded on the grand Disklavier by playing the note A2 at MIDI velocity 12. Picture from [12].
Figure 3. Vibration signal recorded on the grand Disklavier by playing the note A2 at MIDI velocity 12. Picture from [12].
Applsci 09 00914 g003
Figure 4. Left: haptic digital piano customization for use with binaural and vibratory (BiVib). Right: schematic of a possible reproduction setting. Pictures from [20].
Figure 4. Left: haptic digital piano customization for use with binaural and vibratory (BiVib). Right: schematic of a possible reproduction setting. Pictures from [20].
Applsci 09 00914 g004
Figure 5. Upright piano recording setup.
Figure 5. Upright piano recording setup.
Applsci 09 00914 g005
Figure 6. Left: tactile transducer mounted below the haptic digital piano keybed. Right: average spectral flattening curve. Pictures from [20].
Figure 6. Left: tactile transducer mounted below the haptic digital piano keybed. Right: average spectral flattening curve. Pictures from [20].
Applsci 09 00914 g006
Figure 7. Magnitude spectrum of the vibration signal at the A0 key, recorded on the upright Disklavier playing at MIDI velocity 111. The dash-dotted curve depicts the reference vibrotactile threshold for passive touch [38], while the two horizontal dashed lines represent the minimum and maximum thresholds recently measured under active touch conditions [34]. Picture adapted from [12].
Figure 7. Magnitude spectrum of the vibration signal at the A0 key, recorded on the upright Disklavier playing at MIDI velocity 111. The dash-dotted curve depicts the reference vibrotactile threshold for passive touch [38], while the two horizontal dashed lines represent the minimum and maximum thresholds recently measured under active touch conditions [34]. Picture adapted from [12].
Applsci 09 00914 g007
Table 1. Binaural and vibratory (BiVib) core structure. Piano lid configurations are given in square brackets.
Table 1. Binaural and vibratory (BiVib) core structure. Piano lid configurations are given in square brackets.
Disklavier GrandDisklavier Upright
Sample sets
(.wav files)
Binaural (closed)
Binaural (open)
Binaural (removed)
Keyboard vibration
Binaural (closed)
Binaural (semi-open)
Binaural (open)
Keyboard vibration
Sampler projects
(Kontakt multis)
Binaural (closed) + vibration
Binaural (open) + vibration
Binaural (removed) + vibration
Binaural (closed) + vibration
Binaural (semi-open) + vibration
Binaural (open) + vibration

Share and Cite

MDPI and ACS Style

Papetti, S.; Avanzini, F.; Fontana, F. Design and Application of the BiVib Audio-Tactile Piano Sample Library. Appl. Sci. 2019, 9, 914. https://doi.org/10.3390/app9050914

AMA Style

Papetti S, Avanzini F, Fontana F. Design and Application of the BiVib Audio-Tactile Piano Sample Library. Applied Sciences. 2019; 9(5):914. https://doi.org/10.3390/app9050914

Chicago/Turabian Style

Papetti, Stefano, Federico Avanzini, and Federico Fontana. 2019. "Design and Application of the BiVib Audio-Tactile Piano Sample Library" Applied Sciences 9, no. 5: 914. https://doi.org/10.3390/app9050914

APA Style

Papetti, S., Avanzini, F., & Fontana, F. (2019). Design and Application of the BiVib Audio-Tactile Piano Sample Library. Applied Sciences, 9(5), 914. https://doi.org/10.3390/app9050914

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop