Next Article in Journal
Sensors for Wheelchair Tennis: Measuring Trunk and Shoulder Biomechanics and Upper Extremity Vibration during Backhand Stroke
Next Article in Special Issue
A Model for Estimating Tactile Sensation by Machine Learning Based on Vibration Information Obtained while Touching an Object
Previous Article in Journal
Characterization of COVID-19’s Impact on Mobility and Short-Term Prediction of Public Transport Demand in a Mid-Size City in Spain
Previous Article in Special Issue
Wearable Urban Mobility Assistive Device for Visually Impaired Pedestrians Using a Smartphone and a Tactile-Foot Interface
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Audio-Tactile Rendering: A Review on Technology and Methods to Convey Musical Information through the Sense of Touch

by
Byron Remache-Vinueza
1,2,*,
Andrés Trujillo-León
1,3,
Mireya Zapata
4,
Fabián Sarmiento-Ortiz
2 and
Fernando Vidal-Verdú
1,3
1
Departamento de Electrónica, Universidad de Málaga, 29071 Málaga, Spain
2
SISAu Research Group, Facultad de Ingeniería y Tecnologías de la Información y la Comunicación, Universidad Tecnológica Indoamérica, Quito 170103, Ecuador
3
Instituto de Investigación Biomédica de Málaga (IBIMA), 29010 Málaga, Spain
4
Research Center of Mechatronics and Interactive Systems—MIST, Facultad de Ingeniería y Tecnologías de la Información y la Comunicación, Universidad Tecnológica Indoamérica, Quito 170103, Ecuador
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(19), 6575; https://doi.org/10.3390/s21196575
Submission received: 9 August 2021 / Revised: 24 September 2021 / Accepted: 27 September 2021 / Published: 30 September 2021
(This article belongs to the Special Issue Tactile Sensing and Rendering for Healthcare Applications)

Abstract

:
Tactile rendering has been implemented in digital musical instruments (DMIs) to offer the musician haptic feedback that enhances his/her music playing experience. Recently, this implementation has expanded to the development of sensory substitution systems known as haptic music players (HMPs) to give the opportunity of experiencing music through touch to the hearing impaired. These devices may also be conceived as vibrotactile music players to enrich music listening activities. In this review, technology and methods to render musical information by means of vibrotactile stimuli are systematically studied. The methodology used to find out relevant literature is first outlined, and a preliminary classification of musical haptics is proposed. A comparison between different technologies and methods for vibrotactile rendering is performed to later organize the information according to the type of HMP. Limitations and advantages are highlighted to find out opportunities for future research. Likewise, methods for music audio-tactile rendering (ATR) are analyzed and, finally, strategies to compose for the sense of touch are summarized. This review is intended for researchers in the fields of haptics, assistive technologies, music, psychology, and human–computer interaction as well as artists that may make use of it as a reference to develop upcoming research on HMPs and ATR.

1. Introduction

Human beings perceive the world and interact with it through the senses. The sense of touch, for instance, enables humans to perceive temperature and roughness of a surface. According to the authors of [1], tactile stimuli are perceived through a series of complex mechanisms which are part of the somatosensory system. More specifically, mechanoreceptors allow us to feel tactile stimulation in the skin, whereas proprioceptors in joints, muscles, and ligaments enable kinesthesia (the ability to feel weight, location, and position of limbs) [2]. There are four channels associated with mechanoreception: Pacinian (P), Non-Pacinian I (NP-I), Non-Pacinian II (NP-II), and Non-Pacinian III (NP-III) which activate depending on the frequency of the stimuli and according to what is shown in Table 1. Moreover, it is well known that fast afferent I (FAF-I) and fast afferent II (FAFII) physiological types, into the P and NP-I channels, respectively, are responsible for processing vibrotactile stimuli [3]. The limits for vibrotactile stimuli perception have been investigated and established from values as low as 0.3 Hz to a maximum of 1000 Hz [4,5,6]. Best perceived frequencies varies depending on the part of the body and the intensity of the vibrations (i.e., amplitude). In [7], the researchers compared thresholds for the fingertip, forearm, and abdomen, and found out that the fingertip is more sensitive as it allows best perception of stimuli around the same frequency but at lower intensity; less energy is required for vibrotactile stimuli to be perceived in the fingertip. Furthermore, an overall peak frequency of 250 Hz has been suggested by the authors of [8]; however, as sensitivity varies from one part of the body to another, it would probably be better to establish an overall best sensitivity range such as the one proposed by [5], ranging from 50 Hz to 500 Hz, or depending on the part of the body such as the range for the head suggested by [9] which lays between 32 Hz and 64 Hz.
Another parameter that defines vibrotactile perception is spatial resolution, which refers to the ability to recognize stimuli presented in different parts of the body. According to Goldstein [10], spatial resolution also varies depending on the location of the stimuli and ranges, on average, from 10 mm in the hand and lips, to over 40 mm in the back and calf. Moreover, location recognition of two vibratory stimuli depends on two temporal factors: duration of stimuli and inter-stimulus onset asynchrony (ISOA) (i.e., when each actuator is activated and deactivated), and acuity decreases as number of stimuli presented increases [3]. Thus, it is clear that frequency, amplitude, location in the body, number of stimuli, and ISOA are important parameters to take into account when determining vibrotactile thresholds, and should be explored in accordance to the application or experimental setup. To find out more about tactile perception thresholds check the works in [7,10,11,12]. Other relevant parameters that affect vibrotactile perception are frequency change discrimination (pitch change) and signal rendering, which will be later discussed in the context of audio-tactile rendering (ATR).

1.1. Vibrotactile Feedback

Touch has been successfully exploited as an alternative channel to communicate information by means of haptic interfaces. Human–computer haptic interfaces (HCHIs) are devices that offer the user force feedback to enable interaction with technology [13]. This force may be generated by vibrating actuators in contact with the skin. Probably, the most common implementation is vibratory stimuli in cellphones to notify users regarding incoming messages, incoming calls, and alarms. HCHIs are also used as sensory substitution systems (SSSs) to convey information to the hearing impaired or blind individuals [3]. For modality translation to be successful, it is necessary to follow a systematic process that begins with the identification of key information features from the parent modality, then the information processing or mapping is performed, and finally information is presented adequately in the alternative modality [14,15]. Feedback may be available to configure parameters of vibrotactile stimuli based on user perception. When the haptic interface conveys tactile information through vibrations it may also be called vibrotactile interface [16], vibrotactile display [17], or vibrotactile sensory substitution system [18], and the information flow between it and the user is usually known as vibrotactile feedback. Some references to review the use of the concept of vibrotactile feedback are found in [2,3,8,9,19].

1.2. Musical Haptics

The term musical haptics was first introduced by Papetti in 2018 [20] and refers to the use of force feedback to enhance the digital musical instrument (DMI) playing or the music listening experience. As proposed in [21], haptic technology is suitable to communicate musical information and enrich the experience of listening to music, live or recorded. For the case of music listening, vibrotactile feedback is the usual selected mechanism to convey musical information through the skin. Although in [5] the author observed that not every aspect of music may be mapped from audio to vibrations, various researchers (see, e.g., in [2,22,23]) have demonstrated that strategies may be outlined to extend the possibilities to render music characteristics to vibratory stimuli. In addition, the Haptic and Audio Interaction Design Workshop (HAID), discontinued since 2013, was reactivated in 2019 due to the increased necessity of a venue to connect the research fields of haptics, audio, and human–computer interaction (HCI) [24]. According to the authors of [20], there are research opportunities in musical haptics to enhance vibrotactile feedback, specially on sensors, actuators, portability, energy efficiency and signal processing. On the other hand, by the time of this review there is no established methodology to evaluate the user experience, so future consensus would be necessary to make musical haptic technology scalable.

1.3. Structure of the Review

In Section 2, a previous publication with review content about musical haptics is shortly described and then the tools and strategies for searching are outlined. In Section 3, types of haptic music players (HMPs) are described and systematized. In Section 4, methods for tactile rendering of music are studied. In Section 5, strategies to compose for the sense of touch are contrasted. Finally, in Section 6, a discussion regarding current challenges and future work is added, and the most important conclusions are drawn.

2. Methodology

2.1. Previous Reviews

The book “Musical Haptics” [20], published in 2018, compiles works on haptic technology applied to DMIs and one focused on a HMP named “Auditory-Tactile Experience of Music” by Merchel and Altinsoy [25]. This work focuses on the evaluation of music listening experience with vibrotactile feedback using a haptic seat. Other related works are referenced in [14,15,18,26,27,28,29,30,31,32,33,34]; but no systematic exploration was observed. In addition, it was verified that about 51% of the HMP literature was published in the last 5 years (Figure 1), and the literature covered in [20] represents just ~23% of the total literature published at this time.

2.2. Review Method

To narrow the search process, musical haptics was divided into two groups: haptic displays to enhance music playing of DMIs and haptic displays to play haptic music (i.e., HMPs), as shown in Figure 2. This review focuses on the right branch of the proposed classification corresponding to HMPs. Due to the trans-disciplinary nature of the topic, databases Scopus, Web of Science and PubMed as well as search engine Google Scholar, were used to look for literature. The keywords used to search are: haptics, music, vibrotactile, player, skin, composition, haptic chair, haptic wearable, sensory substitution, and the following combinations:
  • Musical haptics;
  • Haptic music;
  • Haptic music player;
  • Musical haptic wearables;
  • Vibrotactile music;
  • Vibrotactile composition;
  • Vibrotactile music composition;
  • Vibrotactile music player;
  • Skin music;
  • Skin music player;
  • Music sensory substitution system.
To extract information of literature Mendeley web importer for Google Chrome was used. Metadata were managed in the library of Mendeley (i.e., web and desktop) and synced with Overleaf for referencing in the manuscript. This workflow allowed reliable and efficient reference management.

3. Haptic Music Player (HMP)

3.1. Architecture

To provide vibrotactile rendering of musical elements four main components are necessary: audio signal processing software, digital to analog converter (DAC), amplifier(s), and actuator(s). The information flow starts with an audio signal which may be extracted from the local storage or from an audio transducer (i.e., a microphone). The audio signal is then processed to extract musical information and then is translated to vibratory signals. The intensity of signals is controlled using an amplifier and the resulting analog signals are finally sent to the vibrating actuators that will be in contact with the user’s skin. The interface may provide interaction options such as knobs or buttons so that the user can have certain level of control over the vibrating stimuli [26,35]. Depending on the goals of the research the installation may have sensory modality variations: haptic feedback, haptic-auditory feedback, haptic-visual feedback, or haptic-auditory-visual feedback. After reviewing the literature, it was found that actuators and signal processing are the main focus of research. Actuators will be studied according to the technologies and the way they are attached to the skin. Signal processing will be analyzed in Section 4 and Section 5, according to the techniques used to translate music features to vibrotactile stimuli and as a strategy to compose for the sense of touch.

3.2. Actuators

Actuators are vibrating devices that are in contact with the user to convey information by means of vibrotactile stimuli. The types of actuators usually encountered in musical vibrotactile displays are listed in the following.

3.2.1. Voice Coil Actuator (VCA)

VCAs are the most commonly used actuators for conveying musical information through vibrotactile stimuli, as the audio signal may be used directly to activate them with little or no additional signal processing; pitch translates to frequency and loudness translates to intensity of vibrations [8]. VCAs are efficient low-cost devices that have good response to amplitude changes of signals [2].

3.2.2. Linear Resonant Actuator (LRA)

LRAs are suitable for vibrotactile applications as they work in similar way as loudspeakers and VCAs [36]. They have a smaller frequency response bandwidth [37], thus suitability for a given application must be evaluated taking into account tactile perception thresholds.

3.2.3. Eccentric Rotating Mass (ERM)

ERMs are rotating motors with an eccentric mass attached to the shaft. These type of actuators are small and lightweight, which make them suitable to enhance portability of the haptic interface [18]. However, independent control of amplitude and frequency of vibrations is not possible. Although the investigators in [18] consider response of ERMs enough to convey information even through clothes, Hwang, Lee, and Choi [37] state that response to changes in dynamics of ERMs is slow. Thus, mapping of signals carrying temporal components of music such as rhythm to ERMs has to be performed carefully, as may not be accurate [8].

3.2.4. Piezoelectric Actuator

Piezoelectric actuators, or piezo-buzzers, vibrate as a result of alternating displacements generated by changes in electric field. Even though they require specific electronics to be driven [20], piezoelectric actuators are flexible, low-cost, energy efficient, and have a wide operating frequency band.

3.2.5. Dual Mode Actuator (DMA)

DMAs are small actuators that provide vibrotactile stimuli composed of two fundamental frequencies. The authors of [37] demonstrated that it is possible to render complex signals that users are able to recognize and evaluate as better compared with LRAs. In addition, computation for activation of DMA is easier and more efficient.

3.2.6. Loudspeakers

Although loudspeakers are not thought as actuators, they generate vibrations that can convey musical information through the skin as demonstrated in [33,38,39,40,41]. Frequency response of loudspeakers by far overcome upper tactile perception threshold (e.g., up to 20 kHz), so subwoofers are usually selected for haptic applications. Working principle is similar to that of VCAs but the vibrations are obtained in a diaphragm. Pitch and loudness are also translated to frequency and intensity, respectively [8]. Loudspeakers also allow independent control of frequency and intensity [3], which is desirable to convey signals that carry more complex musical information such as timbre or melody.
Check Section 3.3 for references on types of actuators implemented by type of actuator attachment mechanism.

3.2.7. Actuator Selection

According to Giordano and Wanderley [3], there are four main criteria to select actuators for musical applications: the role of the vibrotactile stimuli in the interface, size, energy consumption, and the kind of information that will be conveyed. For the case of HMPs, vibrations are intended to stimulate the skin according to a musical composition to convey music to the hearing impaired, to enhance the music listening experience, or to convey vibrotactile music compositions (VMCs). Regarding size and energy consumption, bigger actuators such as speakers and some voice coil actuators may need amplification which results in higher energy consumption. In addition, according to Petry, Huber, and Nanayakkara [8], resolution of stimuli perception may increase as the area covered by the actuator increases. However, as explained in [15], the increment in area in contact with the skin at some point will result on diffusion of stimuli, thus recognition of location of vibrotactile source may be more difficult [9]. Therefore, a balance between size and stimuli location will allow high resolution tactile rendering. Furthermore, enhancing performance of actuators at low frequencies is desirable as the frequency range for best perceptible vibrations is lower than that of audio. Innovations such as the string-motor actuator proposed in [42] allows to cover a wider area efficiently but effectiveness to render specific music features, such as melody or timbre, remains unclear. On the other hand, the DMA proposed in [37] allows complex signal mapping covering a smaller area, which suggests that balance between size and stimuli resolution is achievable. Due to the difference on sensitivity of human skin along the body, choice must also be based on a careful analysis of frequency perception thresholds and responsiveness of the actuator. Although no specific strategy to select actuators for HMPs has been encountered by the time of this review, design considerations found in [43,44] are useful for preliminary design. Selection of actuators represents an opportunity for future research, specially for comparison between types of actuators, advantages, and limitations regarding vibrotactile rendering of musical features. In summary, best actuator selection combines low-cost and lightweight design, enhanced performance at low frequencies and a good balance between size and vibrotactile stimulation.

3.3. Actuator Attachment Mechanism

To communicate vibrating stimuli, actuators are attached to the user in different ways and in different parts of the body. Contact mechanisms may be organized into three groups: haptic installations, haptic wearables, and hybrid.

3.3.1. Haptic Music Player-Installation (HMP-I)

HMP-Is are fixed setups for analyzing psychophysical responses of hearing-impaired users to musical vibrotactile stimuli, although there are experimental setups designed to study responses of hearing individuals also. Haptic chairs are the most common installations found in the literature. The concept of using a chair as an HMP was patented by Komatzu in 2002 [45] and has been explored since then by various authors, as shown in Table 2. The primary advantage of using a chair as a medium of vibrotactile communication is the extended area available to spatialize stimuli. Space can be used as a vibrotactile music composition resource to compensate the limitations of tactile perception [15]. While sitting, the user is provided with vibrotactile feedback underside of the seat, the back, the seat arms, or the feet, while other sensory channels may complement feedback such as auditory, visual or both, depending on the purpose of the investigation. Some limitations of haptic chairs are portability and customization, as the user may not be able to change the location of the chair or the location of the actuators with ease. Important factors to be considered in the design of a haptic chair are bone conduction and airborne conduction, as they may be augmented through the structure of the chair. Perception of vibrotactile stimuli can be affected by bone conduction [46,47], and should be considered mainly if psychophysical evaluations are involved. High-frequency and high-intensity vibrations may produce not only audible resonance, but also vibration of internal organs such as the viscera whose excitation has been associated with emotion [15]. Bone conduction may be treated as interference in some cases, for instance when the vibrotactile stimuli needs to be assessed isolated from auditory perception (i.e., audible stimuli resulting from bone conduction may reach the ear and affect the results) [23]; however, it may be also desirable, for instance, to enhance the music listening experience as proposed by Sakuragi in [46]. These parameters have been taken into account in design stages in [25,30,48], as the body related transfer function (BRTF) similar to the head related transfer function (HRTF) in auditory studies. Other types of installations are multimodal platforms for user experience evaluation [49] and fixed desktop devices such as knobs [50] and buttons [51] for vibrotactile music perception evaluation, as shown in Figure 3.

3.3.2. Haptic Music Player-Wearable Device (HMP-WD)

The notion of wearability of HMPs was considered by Gunther in 2001 in his proposed vibrotactile compositional tool “Skinscape” [27]. Although the idea was not implemented, it evolved to a wearable whole body musical haptic interface presented in 2003 by Gunther and O’Modhrain [15]. Nevertheless, in 2018, Turchet [58] introduced the concept of musical haptic wearables for audiences or MHWAs, where vibrotactile feedback was transmitted via haptic garments that members of an audience could wear. In this review, the concept of HMP-WDs is proposed to generalize its implementation not only for listening experiences but also as SSSs for hearing impaired individuals. Even though Turchet describes basic requirements of MHWAs, such as embedded intelligence and wireless connectivity, some prototypes needed desktop installations because of the research stage at which the projects were, such as Gunther’s “Cutaneous grooves” whole body haptic interface, while others are musical haptic wearables intended for experimentation purposes only. Consequently, most works described in this section are not necessarily portable at the moment of publication, but have clear opportunities to lately become HMP-WDs. Starting from prototypes that cover a small skin surface, there are bracelets (Figure 4a), designed to be worn on the wrist (see, e.g., in [19,21,59,60]); gloves and mobile device mockups (Figure 4b), designed to be worn or held on the hands (see, e.g., in [37,61,62,63,64]); belts (Figure 4c), designed to be worn surrounding the body from the chest to the abdomen (see, e.g., in [8,35,41,42]); and jackets (Figure 4d), designed to be worn on the upper body with actuators usually located on the back, the front and the superior limbs (e.g., [38,65]). Other variations are whole body suits [15], and headphone type displays [39], but instances are scarce. As well as for HMP-Is, design of HMP-WDs requires consideration of frequency perception thresholds according to the part of the body where the device will be attached. Auditory feedback may be provided with almost no effect on portability but if adding visual feedback the user would be constrained to a reduced space, such as virtual reality (VR) or augmented reality (AR) installations. For effective portability, special attention must be focused on energy consumption to extend at maximum battery lifetime. Strategies such as pulse-width modulation (PWM) for efficient signal rendering may be combined with lightweight efficient actuators such as DMAs; although resulting increased noise should be assessed [36]. Additionally, if it is expected that the user wears the HMP, the appearance and user interface capabilities of the device are relevant and should be considered, as they are in [35,42,65]. Finally, bone and airborne conduction should be considered for high frequency-energy vibrations or when HMP-WDs are attached near the head or internal organs, as may influence vibrotactile perception. Belts and jackets are of particular consideration as they cover larger areas. Alternatively, Yamazaki et al. [42] showed that larger areas of the body may be covered using strings attached to vibrating motors instead of using the motors themselves to convey the vibrotactile stimuli. This is an innovative implementation that results in better transmission of low frequency vibrations than that obtained with linear actuators. Nevertheless, additional psychophysical exploration would be necessary considering tactile rendering of more specific musical features and the BRTF. Table 3 shows a systematic overview of research on HMP-WDs.

3.3.3. Haptic Music Player-Hybrid (HMP-H)

HMP-H are setups that combine actuator contact mechanisms from HMP-Is and HMP-WDs. The use of bracelets, sleeves, and belts extends the area covered by fixed installations. In [75], for instance, electronic music artist Martin Garrix performs in a hybrid installation that features platforms, vibrating objects, touchable speakers, and jackets in order to convey musical information to hearing impaired individuals. Participants reported the experience as positive and energetic. Although expanding the area of a haptic installation may be desirable, for research purposes it may evolve to a HMP-I or HMP-WD. For instance, the model human cochlea proposed in [28] and ref. [14] started as a wearable device but ended up as a haptic chair [54], while the hybrid installation of Gunther [27] later became a wearable haptic whole body suit. Hybrid installations may include vibrotactile music input devices (VMIDs) such as the Vibrochord, designed and tested in [76,77], or the mobile device mock-up investigated in [62]. A VMID allows a performer to play vibrotactile music while the user perceives the stimuli in real-time. This vibrotactile music device is different from a DMI as it is intended for vibrotactile music execution or composition, and neither auditory stimuli is generated nor acoustic musical instrument sound simulated. Table 4 shows a systematic overview of research on HMP-H.
According to the literature reviewed in Table 2, Table 3 and Table 4, HMPs have evolved consistently during the last three decades. Although first ideas focused on fixed installations such as haptic chairs, it is clear that the tendency has been towards wearable technologies; about 54% of the total of publications corresponds to HMP-WDs with about 64% of them published in the last 5 years.

4. Audio-Tactile Rendering

High-quality rendering of audio features to vibrotactile stimuli would generate robust HMPs to aid the hearing impaired to feel music through touch and will effectively enhance the music listening experience of hearing individuals. Although translation of music features to vibrotactile stimuli is neither a straightforward nor a constrained task, it is possible to outline relevant mapping considerations. These considerations may be grouped according to the musical feature to be translated, being the most explored: rhythm, pitch, melody, timbre, and loudness.

4.1. Tactile Rendering of Rhythm

Rhythm may be defined as an auditory or visual pattern that repeats on time [77]. This ubiquitous musical feature may be perceived by multiple sensory channels such as visual, auditory, and touch. Moreover, rhythm recognition enhances when visual feedback is offered to the user [18,60]. Translation from auditory rhythmic patterns to vibrotactile stimuli in HMPs has been studied by various researchers (check Table 2, Table 3 and Table 4). It has been suggested that the sense of touch is able to recognize rhythm with ease [3,61], and that rhythmic patterns in vibrotactile music contributes to the experience in a greater proportion compared to other musical features [32,78]. Depending on the kind of music, rhythmic patters may have more presence in a specific frequency band, so one way to enhance vibrotactile rhythmic information is using filters [32,37]. For instance, in jazz music the bass or drums usually mark the rhythmic baseline, therefore a low-pass (high-shelf) filter may be used to cut high frequencies and enhance bass tones that carry rhythm. However, use of filters may affect the quality of the final vibrotactile composition [63]. Perception of rhythm with vibrotactile stimuli may depend on the type and size of actuator used. In [34], for instance, the authors found that covering a greater area of the body with the actuator allows users to better feel rhythm and energy of music. Rhythm may also be conceived from scratch. In [35], the authors create vibrotactile rhythmic patterns without processing audio signals from music. In this case, signals are synthesized to create short high intensity pulses which are later sequenced to create vibrotactile compositions. Other ways to create rhythmic patterns is using software instruments such as those found in digital audio workstations (DAW), where highly rhythmic instruments can be selected to create independent tracks [47], enabling a kind of vibrotactile orchestration tool. However, control over the signals sent to the actuators may be limited. On the other hand, when considering a HMP for hearing impaired individuals, it will be important to determine how they perceive rhythm as may be different to what a hearing person perceives [19]. The authors of [40] suggest that hearing-impaired users are able to identify rhythmic patterns in vibrotactile music, but no sufficient psychophysical evidence is presented to demonstrate that perception of vibrotactile rhythm is similar in deaf and hearing users.

4.2. Tactile Rendering of Pitch

Rendering pitch to vibrotactile stimuli is a complex task as touch has frequency perception limitations that have already described. A simple way to translate pitch and loudness to vibrotactile stimuli is using speakers or VCAs which directly convert pitch to frequency and loudness to intensity of vibrations [8]. However, frequency response of these actuators overpass skin perception thresholds, so information embedded in high-frequency bands (i.e., over 1000 Hz) might be lost. Moreover, pitch discrimination is not constant and depends on frequency. The just noticeable difference (JND) between pitches varies as frequency varies; if frequency increases, the JND between pitches also increases. Thus, during design stages it would be important to consider that lower frequency bands will require a reduced pitch band, and vice versa [2]. If some frequencies are not perceivable because of a lack of intensity, the user may be provided with intensity control in an interface, such as in the Emoti-Chair [54], but independent control of intensity at a given frequency may require complex control protocols so that frequency remains between perceivable ranges. In addition, to overcome frequency limitations of touch, alternatives such as specialization of pitch have been proposed [32]. In [28] the authors named the spatialization of frequencies FM (Frequency Model). In the FM, audio signals have to be filtered to obtain frequency bands that are sent to different groups of actuators (see Figure 5). Results suggest that this method allows better perception of elements and emotional content of music than raw signals sent directly to the actuators. Moreover, in [39] the authors suggest that performance of the FM for hearing impaired users would increase if visual feedback is added. Consonance between pitches is another important feature of music, although dissonance can also be used as a composing resource. It has been studied in [57], and results show that users may process vibrotactile consonance in similar way as the auditory channel, supporting the results obtained in [64]. It was also found that evaluation of consonance enhances when vibrotactile feedback is presented in a wider area, which agrees with what was stated in [34]. Another concept introduced in [22] to relate pitch with touch is the use of tactile metaphors. The authors found that association between tactile metaphors such as sharpness, roughness, softness, weight, heat, and wetness, and musical characteristics such as pitch, loudness, timbre, and its combinations, is in fact possible. For instance “higher pitches may be described as sharper, rougher, harder, colder, drier and lighter than lower pitches”. Tactile metaphors might be of special value for the hearing impaired as they might have no experience with music but have experience with, for instance, textures that may be rendered or even synthesized from scratch, as suggested in [16]. Although, tactile metaphors remain unexplored as a method to map musical information to vibrotactile stimuli, they represent opportunities for future research.

4.3. Tactile Rendering of Melody

Melody builds up as a suitable combination of pitch changes over time. Therefore, most of the limitations for pitch conversion also apply to melody. Strategies already summarized for vibrotactile pitch may be combined to render vibrotactile melody. For instance, spatialization of frequencies in different parts of the body varying on time, will result in a practical representation of melody [20], and may include the implementation of spatio-temporal tactile illusions such as phantom motion [65]. However, the lack of frequency content will remain. Such as for rhythm, pitch and melody may be synthesized using software instruments from DAWs [47], but the lack of signal control will also remain, as well as frequency constrains due to perceptual limitations of touch. It might be more valuable to explore specific characteristics of melody to render it effectively to vibrotactile stimuli. For instance, one important characteristic of melody is interval, or the distance between notes. In [69], it was found that participants are able to discriminate intervals with changes in frequency of about 8 Hz, but this value depends on the location of the vibrotactile stimuli. The authors conclude that it is easier for participants to recognize larger intervals than smaller ones, which suggests that touch resolution for pitch discrimination is lower than that of the auditory. Another more complex strategy is to extract melodic content of audio and perform signal processing. One clear methodology implemented to extract melodic content of music and translate it to vibrotactile stimuli is presented in [40]. It consists on using algorithms to extract melodic features of audio (i.e., music information retrieval) and convert them to a MIDI (Musical Instrument Digital Interface) representation that then is executed, audio file is filtered and finally sent to the actuators contained in a bracelet named the Auris Bracelet. One key aspect of this HMP-H is the use of the Auris Bracelet to present melodic content along with the Auris Chair which presents vibrotactile feedback from filtered audio, as shown in Figure 6; this combination allows to convey musical information even encoded in frequencies over 1 kHz, as suggested by the researchers. Translating pitch and melody represents a challenge due to the tonal content of these musical features, and even more for hearing impaired users who may have a completely different conception of what music is [8].

4.4. Tactile Rendering of Timbre

Timbre allows the listener to differentiate between tones played from one or another musical instrument. Timbre relies on the frequency content (i.e., spectral content) of audio signals, and therefore tactile rendering represents a challenge. Although the reduced tactile perception band will affect the recognition of small spectral content variations, individuals are able to recognize timbre of rendered audio signals from different musical instruments (e.g., piano, cello, or trombone) with vibrotactile stimuli only [79]. Moreover, the sense of touch is able to recognize waveform of signals [15], where the mechanoreceptors work as tactile filters that aid in the process [6]. This recognition ability may be used as a tool to render texture of sound (i.e., timbre) as vibrotactile texture. In [79], the researchers rendered different signals varying waveform, envelope, fundamental frequency, harmonics, duration, and ISOA, and found that individuals are able to differentiate these representations of timbre; participants with hearing impairment present the same ability. Another innovative method to render timbre was proposed in [32], where the researchers measure the amount of noise present in the audio signal, what they call a bark-based spectral flatness measure, and use it to perform an interpolation between a 500 Hz sine tone and a white noise signal. The result is a vibrotactile signal that represents timbre. However, evaluation of participants did not focus on timbre discrimination but on overall quality of the vibrotactile stimuli, so further exploration would be required. As earlier stated, timbre is closely associated with type of musical instrument. Such as for frequency, the authors of [28] proposed the TM (Track Model), a model that consists on splitting music into independent audio tracks, each containing different musical instruments, and then send the signals to groups of actuators (see Figure 7). Although this model ignores the frequency content that defines timbre of instruments, the researchers demonstrated that evaluation of perception was better than that obtained with the FM. The TM offers opportunities for future work as demonstrated by Hashizume et al. [38] who tested the method in a multi-modal experimental setup. Tactile metaphors is another resource that represent an opportunity for timbre rendering but, as mentioned in Section 4.2, this method remains unexplored. Thus, it is clear that the parent audio signal as well as control over signal waveform aids to reach tactile rendering. Moreover, variations on signal envelope: attack, decay, sustain, and release (ADSR), may offer a wider set of opportunities to obtain high fidelity vibrotactile rendering of timbre. Innovative methods may be proposed but it is necessary to present evidence that supports successful tactile rendering.

4.5. Tactile Rendering of Loudness

In traditional music notation loudness is a key feature to consider [65]. Subjectively, vibrotactile loudness is a variable corresponding to the distance that the skin is displaced by the stimuli [2]. Tactile rendering of loudness may be straightforward as can be mapped directly to the intensity of the actuators [8], and the use of software instruments and MIDI representations allows easy control [55]. Representation of loudness may even be enhanced by adding visual feedback in the form of brightness variation of light [18]. On the other hand, there are psychophysical implications that may affect the interaction between loudness and music tactile rendering. According to the work in [2], loudness is independent of frequency in the range of 20 to 40 Hz, while in [20] it was found that perception of loudness may be affected at low frequencies, which adds complexity to the definition of a suitable bandwidth to render musical information. In addition, according to Verillo [4], there are effects such as summation (i.e., perceptual increase of loudness) that occurs when two stimuli are presented in the same psychophysical channel (i.e., Pacinian or Non-Pacinian), or suppression (i.e., perceptual decrease of loudness) in a second vibrotactile stimuli when two tones are presented in independent psychophysical channels, which represents additional complexity for tactile rendering of loudness, specially when the HMP has arrays of actuators that present multiple vibrotactile stimuli.

5. Vibrotactile Music Composition (VMC)

Although music for the ears follows not only well established but also evolving methodologies, VMC is still being explored. VMCs are usually dedicated to hearing impaired users and transmitted by means of HMPs, but may also be enjoyed by hearing individuals. In [15], the authors propose a list of features that may constitute the foundations of tactile compositions: frequency, duration, intensity, waveform, spectral content, and space. Most of these elements have already been studied in Section 4, thus this section focuses on describing the strategies that have been implemented to compile these features and create VMCs.

5.1. Tactile Illusions

A resource used to compose for the sense of touch are spatio-temporal tactile illusions. Although there is no concise relation of tactile illusions with any musical element, the apparent sensations that may be created in the skin using vibrotactile stimuli offers a conceptual tool to convey meaningful information to the user. Apparent movement or phantom motion, for instance, may be correlated to music dynamics, emotion, or musician movements such as movement in dance [15]. Tactile illusions have been explored in the context of VMC by various authors [3,15,21,27,33,65,67], who agree that tactile illusions have good potential as a compositional resource.

5.2. Real-Time VMC

Real-time VMC has also been studied. In this case, a vibrotactile musical input device (VMID) is required. The VMID designed in [62], for instance, is a Nintendo Switch device with a tactile keyboard that renders vibrotactile signals presented to the performer by means of a VCA. One octave of notes are assigned to the keys from C-131 Hz to C-262 Hz that allows the performer to create complex vibrotactile compositions. Although there was no evidence to validate the performance of this VMID, it represents a suitable proposal for future exploration. Another device that has been designed is the Vibrochord [66] a piano-like VMID conceived exclusively to play vibrotactile compositions (see Figure 8). In this work, the authors propose an octave vibrotactile scale based on traditional western music, divided in numbers: 1—40 Hz, 2—56 Hz, 3—78.4 Hz, 4—109.76 Hz, 5—153.66 Hz, 6—215.12 Hz, 7—301.18 Hz, 8—421.65 Hz. To represent vibrotactile loudness, the keys of the Vibrochord respond to pressure, the harder they are played the higher is the intensity of vibrations, allowing the user greater expressiveness. The researchers have shown that vibrotactile music composition may require specific technology to allow more expressiveness and precision in real time playing. However, effects such as adaptation and learnability must be considered as the evaluation of performer or audience perception might be affected.

5.3. VMC from Scratch

As mentioned before, some musical features may be rendered from scratch. This method was used by Vallgarda et al. in [35] to create VMCs that are played through the Hedonic Haptic Player [68]. The compositions were created in the basis of two musical features: rhythm and dynamics, rendered by modifying waveform, event length, wave amplitude, length in % of the original, silent event, and modulation frequency (see Figure 9). The researchers show how these compositions based on the interaction of vibrotactile stimuli, generates meaningful positive responses from users, more than just feeling one discrete vibration. Although the authors do not mention it, the results obtained suggest the perception of a combination of tactile illusions, which have already been considered as a powerful resource for VMC. Results obtained by Vallgarda et. al. are supported in [67], where VMCs were rendered to be played on a vibrotactile garment named Ilinx [80] and created from the point of view of artists. In [78], composers were provided with compositional resources and used them to build up tactile compositions. Resources comprise segments of signals with different waveforms (e.g., sine, saw, and square) and frequencies so that the artist creatively combines them to end up with an original tactile composition, that is later experienced by an audience in the Emoti-chair [54]. Although vibrotactile compositions were tested for audio visual environments, the method can be expanded to create VMCs for the hearing impaired; similar to what was implemented in [35,67].

6. Discussion and Conclusions

In this review, tactile rendering technology and methods to convey musical features through the skin have been systematically explored. The strategy to search relevant literature has been presented and the method used to manage documentation metadata was described. During the first stage of the review it was found that linking web-based tools such as Scopus, Mendeley, and Overleaf allows the researcher reliable and efficient organization of information.
Transmission of musical information through the skin has been accomplished by implementing vibrating technologies in devices whose design has evolved during the last twenty-eight years. The fundamental architecture of a HMP was described, and it was suggested that selection of actuators is important due to the availability of different types of them which have advantages and limitations regarding vibrotactile communication of musical elements. It was proposed that the ideal actuator should combine low-cost and lightweight design, good performance at low frequencies, and a balance between size and vibrotactile stimulation. VCAs and ERMs are the most used actuators (see Table 2, Table 3 and Table 4) and new concepts such as the DMA [37] and the string-motor device [42] are being studied. However, there are other technologies like mid-air ultrasound haptic feedback that demonstrated to be effective for frequency discrimination tasks [81] and may be investigated for touchless applications to explore music in museums or music stores and avoid disease transmission, for instance.
HMPs were classified depending on the way vibrotactile stimuli are presented to the user. Literature was organized into three groups: HMP installations (HMP-Is), wearable HMPs (HMP-WDs), and hybrid HMPs (HMP-H). Moreover, information was systematically allocated in tables according to year of publication, type of contact mechanism, type of actuator, type of stimuli presented, and musical features explored. Although preliminary research on HMPs was focused on fixed installations, it was found that ~54% of total publications by 2020 corresponds to HMP-WDs (see Figure 1), suggesting that HMP technology evolves towards wearability and portability. It is congruent not only with the evolution of electronics towards smaller and more efficient devices but also with the idea that listening to music is a day by day experience; something that everyone wants to enjoy anywhere at any moment. On the other hand, most of the installations such as chairs and platforms (see Table 2), bigger than a common portable music player, have been necessary to explore perception of various parts of the body simultaneously, and effects such as bone conduction and stimuli spatialization. Although these prototypes are neither wearable nor portable, they may evolve towards further application in medicine for instance to complement music therapy for the Alzheimer [82], dementia [83] or cancer [84,85] treatment; or entertainment such as amusement parks, movie theaters, video games, and virtual and augmented reality installations.
Regarding ATR, rhythm, loudness, and pitch are the predominant features explored in the literature as rendering seems to be straightforward (see Table 2, Table 3 and Table 4). Although direct rendering of pitch and loudness to frequency and intensity of vibrations is feasible, respectively, information may be lost due to frequency perception limitations of touch. Further, other musical elements such as melody and timbre require intensive signal processing such as music information retrieval techniques and in some cases the introduction of concepts such as spatialization of pitch or tactile metaphors, to avoid frequency discrimination limitations. Resources such as software instruments and MIDI representations may aid in the process. Thus, tactile rendering of musical features represents a challenge and a vast field for future research. Indeed, it seems to be clear that research on HMPs, and specifically ATR, has to consider that playing music is not only a whole complex performance perceived by the ear and the entire body through vibrations, but also the result of the summation of individual events, performances, instruments, colors, notes, rhythms and every musical feature. Therefore, music can be composed and decomposed, its vibratory nature and its relation with the human skin unveiled during this process, and might be different for the case of hearing impaired individuals who may perceive vibrotactile music in a different way.
Furthermore, a tactile composition was defined as a way to compose for the sense of touch. Strategies implemented to create VMCs were summarized and contrasted (see Table 2, Table 3 and Table 4). It was found that creating a compilation of sequences or patterns that change in time results more meaningful for users, instead of discrete stimuli presented solely. Some researchers used spatiotemporal tactile illusions with special focus on creating apparent movement that allowed composers and artists to expand creativity (see Section 5.1). VMCs may be created and later played through an HMP such as the Hedonic Haptic Player [35], the Emoti-chair [54] or the Ilinx [80], or may be played in real time using VMIDs such as the Vibrochord [66] or a mobile device (e.g., Nintendo Switch) [62]. VMC represents a new tool for composers. Nevertheless, more scientific evidence is required to overcome tactile perception limitations and psychophysical effects on users. VMC has been used as a novel resource for artists to compose for the sense of touch, and therefore an opportunity for hearing impaired users to perceive musical compositions by means of vibrotactile stimuli, but the lack of understanding of the human response to vibrotactile stimuli arrangements and the unavailability of suitable technology still restricts the development of novel applications. Accordingly, exploration of human skin limitations in the musical context should progressively be expanded to the use of more complex VMC arrangements and technologies.
Besides, vibrotactile music interaction between an audience and the performer has been explored, enabling a new communication channel to interact in a live concert installation [21]. This idea may be expanded towards interaction not only between the audience and the performer but also between groups of audiences, between individuals or even between concert venues, which opens up a new paradigm for event designers, musicians and artists in general.
During literature exploration, investigations with sound scientific foundation have been encountered with a clear hypothesis and clear objectives, which usually ends up unveiling the limitations of the human tactile sense or the pros and cons of some technology application. However, there are projects that seem to bypass this requirement and proceed to a conceptual and more aesthetic implementation, dismissing the limitations of the human perception capabilities and technology implications. It would be relevant to soon provide a clear general methodology for HMP research that enables effective and robust research to later develop a well-established starting point for technology development and implementation.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript (in order of appearance):
PPacinian
NP-INon-Pacinian I
NP-IINon-Pacinian II
NP-IIINon-Pacinian III
FAF-IFast Afferent I
FAF-IIFast Afferent II
ISOAInter-Stimulus Onset Asynchrony
ATRAudio-Tactile Rendering
HCHIHuman–Computer Haptic Interface
SSSSensory Substitution System
DMIDigital Musical Instrument
HAIDHaptic and Audio Interaction Design
HCIHuman Computer Interaction
HMPHaptic Music Player
DACDigital to Analog Converter
VCAVoice Coil Actuator
LRALinear Resonant Actuator
ERMEccentric Rotating Mass
DMADual Mode Actuator
VMCVibrotactile Music Composition
HMP-IHaptic Music Player - Installation
BRTFBody Related Transfer Function
HRTFHead Related Transfer Function
ATFSAudio-Tactile Frequency Synchronism
FMFrequency Model
TMTrack Model
VMLEVibrotactile Music with/without Listening Experience
ATFMAudio-Tactile Frequency Matching
F-EQFrequency Equalization
MCMelodic Consonace
HMP-WDHaptic Music Player - Wearable Device
MHWAMusical Haptic Wearables for Audiences
VRVirtual Reality
ARAugmented Reality
PWMPulse-Width Modulation
FDFrequency Discrimination
APIAudience-Performer Interaction
EEGElectroencephalogram
VMIDVibrotactile Music Input Devices
JNDJust Noticeable Difference
MIDIMusical Instrument Digital Interface

References

  1. Schneider, A.; Feussner, H. Mechatronic Support Systems and Robots. In Biomedical Engineering in Gastrointestinal Surgery; Academic Press: Cambridge, MA, USA, 2017; pp. 387–441. [Google Scholar] [CrossRef]
  2. Birnbaum, D.M.; Wanderley, M.M. A Systematic Approach to Musical Vibrotactile Feedback. In Proceedings of the International Computer Music Conference, Copenhagen, Denmark, 27–31 August 2007; pp. 397–404. [Google Scholar]
  3. Giordano, M.; Wanderley, M.M. Perceptual and Technological Issues in the Design of Vibrotactile-Augmented Interfaces for Music Technology and Media. In Proceedings of the International Workshop on Haptic and Audio Interaction Design, Daejeon, Korea, 18–19 April 2013; pp. 89–98. [Google Scholar]
  4. Verrillo, R.T. Vibration Sensation in Humans. Music Percept. 1992, 9, 281–302. [Google Scholar] [CrossRef]
  5. Chafe, C. Tactile Audio Feedback. In Proceedings of the 19th International Computer Music Conference (ICMC 1993), Tokyo, Japan, 10–15 September 1993; pp. 76–79. [Google Scholar]
  6. Young, G.W.; Murphy, D.; Weeter, J. Vibrotactile Discrimination of Pure and Complex Waveforms. In Proceedings of the 12th International Conference in Sound and Music Computing, Maynooth, Ireland, 30 July–1 August 2015; pp. 359–362. [Google Scholar]
  7. Jones, L.A.; Sarter, N.B. Tactile Displays: Guidance for Their Design and Application. Hum. Factors 2008, 50, 90–111. [Google Scholar] [CrossRef]
  8. Petry, B.; Huber, J.; Nanayakkara, S. Scaffolding the Music Listening and Music Making Experience for the Deaf. In Cognitive Science and Technology; Springer: Singapore, 2018; pp. 23–48. [Google Scholar] [CrossRef]
  9. Mirzaei, M.; Kan, P.; Kaufmann, H. EarVR: Using Ear Haptics in Virtual Reality for Deaf and Hard-of-Hearing People. IEEE Trans. Vis. Comput. Graph. 2020, 26, 2084–2093. [Google Scholar] [CrossRef] [PubMed]
  10. Goldstein, E.B. Sensation and Perception, 8th ed.; Cengage: Boston, MA, USA, 2010; p. 492. [Google Scholar]
  11. Abraira, V.E.; Ginty, D.D. The Sensory Neurons of Touch. Neuron 2013, 79, 618–639. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. McGlone, F.; Wessberg, J.; Olausson, H. Discriminative and Affective Touch: Sensing and Feeling. Neuron 2014, 82, 737–755. [Google Scholar] [CrossRef] [Green Version]
  13. Hayes, L. Vibrotactile Feedback-Assisted Performance. In Proceedings of the 2011 Conference on New Interfaces for Musical Expression, NIME, Oslo, Norway, 30 May–1 June 2011. [Google Scholar]
  14. Karam, M.; Russo, F.; Fels, D. Designing the Model Human Cochlea: An Ambient Crossmodal Audio-Tactile Display. IEEE Trans. Haptics 2009, 2, 160–169. [Google Scholar] [CrossRef]
  15. Gunther, E.; O’Modhrain, S. Cutaneous Grooves: Composing for the Sense of Touch. J. New Music Res. 2003, 32, 369–381. [Google Scholar] [CrossRef] [Green Version]
  16. Chang, A.; O’Sullivan, C. An Audio-Haptic Aesthetic Framework Influenced by Visual Theory. In Proceedings of the International Workshop on Haptic and Audio Interaction Design HAID 2008, Jyväskylä, Finland, 15–16 September 2008; pp. 70–80. [Google Scholar] [CrossRef] [Green Version]
  17. Giordano, M.; Wanderley, M.M. A Learning Interface for Novice Guitar Players Using Vibrotactile Stimulation. In Proceedings of the 8th Sound and Music Computing Conference (SMC2011), Padova, Italy, 6–9 July 2011. [Google Scholar] [CrossRef]
  18. Petry, B.; Illandara, T.; Nanayakkara, S. MuSS-bits: Sensor-Display Blocks for Deaf People to Explore Musical Sounds. In Proceedings of the 28th Australian Conference on Computer-Human Interaction-OzCHI ’16, Tasmania, Australia, 29 November–2 December 2016; ACM Press: New York, New York, USA, 2016; pp. 72–80. [Google Scholar] [CrossRef]
  19. Petry, B.; Illandara, T.; Elvitigala, D.S.; Nanayakkara, S. Supporting Rhythm Activities of Deaf Children using Music-Sensory-Substitution Systems. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–26 April 2018; pp. 1–10. [Google Scholar] [CrossRef]
  20. Musical Haptics; Papetti, S.; Saitis, C. (Eds.) Springer Series on Touch and Haptic Systems; Springer International Publishing AG: Cham, Switzerland, 2018. [Google Scholar] [CrossRef] [Green Version]
  21. Turchet, L.; West, T.; Wanderley, M.M. Touching the Audience: Musical Haptic Wearables for Augmented and Participatory Live Music Performances. Pers. Ubiquitous Comput. 2020, 25, 749–769. [Google Scholar] [CrossRef]
  22. Eitan, Z.; Rothschild, I. How Music Touches: Musical Parameters and Listeners’ Audio-Tactile Metaphorical Mappings. Psychol. Music. 2011, 39, 449–467. [Google Scholar] [CrossRef]
  23. Romagnoli, M.; Fontana, F.; Sarkar, R. Vibrotactile Recognition by Western and Indian Population Groups of Traditional Musical Scales Played with the Harmonium. In Proceedings of the International Workshop on Haptic and Audio Interaction Design, Kusatsu, Japan, 25–26 August 2011; pp. 91–100. [Google Scholar] [CrossRef]
  24. Pietrzak, T.; Wanderley, M.M. Haptic and Audio Interaction Design. J. Multimodal User Interfaces 2020, 14, 231–233. [Google Scholar] [CrossRef]
  25. Merchel, S.; Altinsoy, M.E. Auditory-Tactile Experience of Music. In Musical Haptics; Springer: Cham, Switzerland, 2018; pp. 123–148. [Google Scholar] [CrossRef] [Green Version]
  26. Daub, M.; Altinsoy, M.E. Audiotactile Simultaneity Perception of Musical-Produced Whole-Body Vibrations. In Proceedings of the Joint Congress CFA/DAGA’04, Strasbourg, France, 22–25 March 2004; pp. 111–112. [Google Scholar]
  27. Gunther, E. Skinscape: A Tool for Composition in the Tactile Modality; Massachusetts Institute of Technology: Cambridge, MA, USA, 2001. [Google Scholar]
  28. Karam, M.; Russo, F.; Branje, C.; Fels, D.I. Towards A Model Human Cochlea: Sensory Substitution for Crossmodal Audio-Tactile Displays. In Proceedings of the Graphics Interface Conference, Windsor, ON, Canada, 28–30 May 2008; pp. 267–274. [Google Scholar] [CrossRef]
  29. Merchel, S.; Altinsoy, M.E. Auditory-Tactile Music Perception. Proc. Mtgs. Acoust. 2013, 19, 015030. [Google Scholar] [CrossRef] [Green Version]
  30. Merchel, S.; Altinsoy, M.E. Vibratory and Acoustical Factors in Multimodal Reproduction of Concert DVDs; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2009; Volume 5763, pp. 119–127. [Google Scholar] [CrossRef]
  31. Nanayakkara, S.; Taylor, E.; Wyse, L.; Ong, S.H. An Enhanced Musical Experience for the Deaf: Design and Evaluation of a Music Display and a Haptic Chair. In Proceedings of the Conference on Human Factors in Computing Systems, Boston, MA, USA, 4–9 April 2009; pp. 337–346. [Google Scholar] [CrossRef]
  32. Jack, R.; Mcpherson, A.; Stockman, T. Designing Tactile Musical Devices with and for Deaf Users: A Case Study. In Proceedings of the International Conference on the Multimedia Experience of Music, Sheffield, UK, 23–25 March 2015; pp. 1–7. [Google Scholar]
  33. Hayes, L. Skin Music (2012). In Proceedings of the 2015 ACM SIGCHI Conference on Creativity and Cognition, Glasgow, UK, 22–25 June 2015; pp. 359–360. [Google Scholar] [CrossRef]
  34. Yamazaki, Y.; Mitake, H.; Hasegawa, S. Tension-Based Wearable Vibroacoustic Device for Music Appreciation. In Proceedings of the EuroHaptics: International Conference on Human Haptic Sensing and Touch Enabled Computer Applications, London, UK, 4–7 July 2016; pp. 273–283. [Google Scholar] [CrossRef]
  35. Vallgårda, A.; Boer, L.; Cahill, B. The Hedonic Haptic Player. Int. J. Des. 2017, 11, 17–33. [Google Scholar] [CrossRef] [Green Version]
  36. Pezent, E.; Cambio, B.; O’Malley, M.K. Syntacts: Open-Source Software and Hardware for Audio-Controlled Haptics. IEEE Trans. Haptics 2020, 14, 225–233. [Google Scholar] [CrossRef] [PubMed]
  37. Hwang, I.; Lee, H.; Choi, S. Real-Time Dual-Band Haptic Music Player for Mobile Devices. IEEE Trans. Haptics 2013, 6, 340–351. [Google Scholar] [CrossRef] [PubMed]
  38. Hashizume, S.; Sakamoto, S.; Suzuki, K.; Ochiai, Y. LIVEJACKET: Wearable Music Experience Device with Multiple Speakers; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2018; Volume 10921, pp. 359–371. [Google Scholar] [CrossRef]
  39. Trivedi, U.; Alqasemi, R.; Dubey, R. Wearable Musical Haptic Sleeves for People with Hearing Impairment; ACM International Conference Proceeding Series; Association for Computing Machinery: New York, NY, USA, 2019; pp. 146–151. [Google Scholar] [CrossRef]
  40. Alves Araujo, F.; Lima Brasil, F.; Candido Lima Santos, A.; de Sousa Batista Junior, L.; Pereira Fonseca Dutra, S.; Eduardo Coelho Freire Batista, C. Auris System: Providing Vibrotactile Feedback for Hearing Impaired Population. BioMed Res. Int. 2017, 2017, 1–9. [Google Scholar] [CrossRef] [PubMed]
  41. Branje, C.; Maksimouski, M.; Karam, M.; Fels, D.I.; Russo, F. Vibrotactile Display of Music on the Human Back. In Proceedings of the 2010 Third International Conference on Advances in Computer-Human Interactions, Washington, DC, USA, 10–16 February 2010; pp. 154–159. [Google Scholar] [CrossRef]
  42. Yamazaki, Y.; Mitake, H.; Oda, R.; Wu, H.H.; Hasegawa, S.; Takekoshi, M.; Tsukamoto, Y.; Baba, T. Hapbeat: Single DOF Wide Range Wearable Haptic Display; ACM SIGGRAPH 2017 Emerging Technologies; ACM: New York, NY, USA, 2017; pp. 1–2. [Google Scholar] [CrossRef]
  43. Choi, S.; Kuchenbecker, K.J. Vibrotactile Display: Perception, Technology, and Applications. Proc. IEEE 2013, 101, 2093–2104. [Google Scholar] [CrossRef]
  44. Marshall, M.T. Physical Interface Design for Digital Musical Instruments. Ph.D. Thesis, McGill University, Montreal, QC, Canada, 2008. [Google Scholar]
  45. Komatsu, A. Method for Expressing Vibratory Music and Apparatus Therefor; Acouve Laboratory, Inc.: Tokyo, Japan, 2002. [Google Scholar]
  46. Sakuragi, R.; Ikeno, S.; Okazaki, R.; Kajimoto, H. CollarBeat: Whole Body Vibrotactile Presentation via the Collarbone to Enrich Music Listening Experience. In Proceedings of the International Conference on Artificial Reality and Telexistence and Eurographics Symposium on Virtual Environments, ICAT-EGVE 2015, Kyoto, Japan, 28–30 October 2015; pp. 141–146. [Google Scholar] [CrossRef]
  47. Tranchant, P.; Shiell, M.M.; Giordano, M.; Nadeau, A.; Peretz, I.; Zatorre, R.J. Feeling the Beat: Bouncing Synchronization to Vibrotactile Music in Hearing and Early Deaf People. Front. Neurosci. 2017, 11, 507. [Google Scholar] [CrossRef]
  48. Altinsoy, M.E.; Merchel, S. Cross-Modal Frequency Matching: Sound and Whole-Body Vibration; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2010; Volume 6306, pp. 37–45. [Google Scholar] [CrossRef]
  49. Frid, E.; Lindetorp, H. Haptic Music: Exploring Whole-Body Vibrations and Tactile Sound for a Multisensory Music Installation. In Proceedings of the Sound and Music Computing Conference, Torino, Italy, 24–26 June 2020; pp. 68–75. [Google Scholar]
  50. Allen, M.; Gluck, J.; MacLean, K.; Tang, E. An Initial Usability Assessment for Symbolic Haptic Rendering of Music Parameters. In Proceedings of the 7th International Conference on Multimodal Interfaces—ICMI ’05, Torento, Italy, 4–6 October 2005; p. 244. [Google Scholar] [CrossRef]
  51. Hopkins, C.; Maté-Cid, S.; Fulford, R.; Seiffert, G.; Ginsborg, J. Vibrotactile Presentation of Musical Notes to the Glabrous Skin for Adults with Normal Hearing or a Hearing Impairment: Thresholds, Dynamic Range and High-Frequency Perception. PLoS ONE 2016, 11, e0155807. [Google Scholar] [CrossRef] [Green Version]
  52. Darrow, A.A. The Effect of Vibrotactile Stimuli via the SOMATRON™ on the Identification of Pitch Change by Hearing Impaired Children. J. Music Ther. 1992, 29, 103–112. [Google Scholar] [CrossRef]
  53. Karam, M.; Nespoli, G.; Russo, F.; Fels, D.I. Modelling perceptual elements of music in a vibrotactile display for deaf users: A field study. In Proceedings of the 2nd International Conferences on Advances in Computer-Human Interactions, ACHI 2009, Cancun, Mexico, 1–7 February 2009; pp. 249–254. [Google Scholar] [CrossRef]
  54. Karam, M.; Branje, C.; Nespoli, G.; Thompson, N.; Russo, F.A.; Fels, D.I. The Emoti-Chair: An Interactive Tactile Music Exhibit. In Proceedings of the 28th of the International Conference Extended Abstracts on Human Factors in Computing Systems—CHI EA ’10, New York, NY, USA, 10–15 April 2010; p. 3069. [Google Scholar] [CrossRef]
  55. Nanayakkara, S.C.; Wyse, L.; Ong, S.H.; Taylor, E.A. Enhancing Musical Experience for the Hearing-Impaired Using Visual and Haptic Displays. Hum. Comput. Interact. 2013, 28, 115–160. [Google Scholar] [CrossRef]
  56. Yamazaki, R.; Ohkura, M. Affective Evaluation While Listening to Music with Vibrations to the Body. In Advances in Intelligent Systems and Computing; Springer: Berlin/Heidelberg, Germany, 2019; Volume 774, pp. 379–385. [Google Scholar] [CrossRef]
  57. Fontana, F.; Camponogara, I.; Cesari, P.; Vallicella, M.; Ruzzenente, M. An Exploration on Whole-Body and Foot-Based Vibrotactile Sensitivity to Melodic Consonance. In Proceedings of the International Conference on Systems, Man, and Cybernetics, Hamburg, Germany, 31 August–3 September 2016; pp. 143–150. [Google Scholar]
  58. Turchet, L.; Barthet, M. Demo of Interactions Between a Performer Playing a Smart Mandolin and Audience Members Using Musical Haptic Wearables. In Proceedings of the International Conference on New Interfaces for Musical Expression, Blacksburg, VA, USA, 3–6 June 2018; pp. 82–83. [Google Scholar]
  59. La Versa, B.; Diamanti, L.; Peruzzi, I.; Zemolin, M. MUVIB: Music and Vibration. In Proceedings of the 2014 ACM International Symposium on Wearable Computers Adjunct Program—ISWC ’14 Adjunct, New York, NY, USA, 13–17 September 2014; pp. 65–70. [Google Scholar] [CrossRef]
  60. Florian, H.; Mocanu, A.; Vlasin, C.; Machado, J.; Carvalho, V.; Soares, F.; Astilean, A.; Avram, C. Deaf people feeling music rhythm by using a sensing and actuating device. Sens. Actuators Phys. 2017, 267, 431–442. [Google Scholar] [CrossRef]
  61. Mazzoni, A.; Bryan-Kinns, N. Mood Glove: A Haptic Wearable Prototype System to Enhance Mood Music in Film. Entertain. Comput. 2016, 17, 9–17. [Google Scholar] [CrossRef] [Green Version]
  62. Nakada, K.; Onishi, J.; Sakajiri, M. An Interactive Musical Scale Presentation System via Tactile Sense Through Haptic Actuator. In Proceedings of the HAVE 2018—IEEE International Symposium on Haptic, Audio-Visual Environments and Games, Proceedings, Dalian, China, 20–21 September 2018; pp. 13–16. [Google Scholar] [CrossRef]
  63. Okazaki, R.; Kuribayashi, H.; Kajimoto, H. The Effect of Frequency Shifting on Audio-Tactile Conversion for Enriching Musical Experience. Lect. Notes Electr. Eng. 2015, 277, 45–51. [Google Scholar] [CrossRef]
  64. Yoo, Y.; Hwang, I.; Choi, S. Consonance of Vibrotactile Chords. IEEE Trans. Haptics 2014, 7, 3–13. [Google Scholar] [CrossRef]
  65. West, T.J.; Bachmayer, A.; Bhagwati, S.; Berzowska, J.; Wanderley, M.M. The Design of the Body:Suit:Score, a Full-Body Vibrotactile Musical Score; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Luxembourg, 2019; Volume 11570, pp. 70–89. [Google Scholar] [CrossRef]
  66. Branje, C.; Fels, D.I. Playing Vibrotactile Music: A Comparison between the Vibrochord and a Piano Keyboard. Int. J. Hum. Comput. Stud. 2014, 72, 431–439. [Google Scholar] [CrossRef]
  67. Hattwick, I.; Franco, I.; Giordano, M.; Egloff, D.; Wanderley, M.M.; Lamontagne, V.; Arawjo, I.; Salter, C.; Martinucci, M. Composition Techniques for the Ilinx Vibrotactile Garment. In Proceedings of the International Computer Music Conference, Denton, TX, USA, 25 September–1 October 2015; pp. 420–423. [Google Scholar]
  68. Boer, L.; Cahill, B.; Vallgårda, A. The Hedonic Haptics Player: A Wearable Device to Experience Vibrotactile Compositions. In Proceedings of the 2016 ACM Conference Companion Publication on Designing Interactive Systems—DIS ’17 Companion, Edinburgh, UK, 10–14 June 2017; pp. 297–300. [Google Scholar]
  69. Egloff, D.C.; Wanderley, M.M.; Frissen, I. Haptic Display of Melodic Intervals for Musical Applications. In Proceedings of the 2018 IEEE Haptics Symposium (HAPTICS), San Francisco, CA, USA, 25–28 March 2018; Volume 2018, pp. 284–289. [Google Scholar] [CrossRef]
  70. Yamazaki, Y.; Mitake, H.; Takekoshi, M.; Tsukamoto, Y.; Baba, T.; Hasegawa, S. Hapbeat: Tension-Based Wearable Vibroacoustic Device. In Lecture Notes in Electrical Engineering; Springer: Berlin/Heidelberg, Germany, 2018; Volume 432, pp. 387–392. [Google Scholar] [CrossRef]
  71. Sharp, A.; Houde, M.S.; Bacon, B.A.; Champoux, F. Musicians Show Better Auditory and Tactile Identification of Emotions in Music. Front. Psychol. 2019, 10. [Google Scholar] [CrossRef]
  72. Sharp, A.; Bacon, B.A.; Champoux, F. Enhanced Tactile Identification of Musical Emotion in the Deaf. Exp. Brain Res. 2020, 238, 1229–1236. [Google Scholar] [CrossRef]
  73. Lucia, M.J.; Revuelta, P.; Garcia, A.; Ruiz, B.; Vergaz, R.; Cerdan, V.; Ortiz, T. Vibrotactile Captioning of Musical Effects in Audio-Visual Media as an Alternative for Deaf and Hard of Hearing People: An EEG Study. IEEE Access 2020, 8, 190873–190881. [Google Scholar] [CrossRef]
  74. Enriquez, K.; Palacios, M.; Pallo, D.; Guerrero, G. SENSE: Sensory Component VR Application for Hearing Impaired People to Enhance the Music Experience. In Proceedings of the Iberian Conference on Information Systems and Technologies, CISTI, Seville, Spain, 24–27 June 2020; Volume 2020. [Google Scholar] [CrossRef]
  75. Garrix, M. Music Lifts You Up: Concert for the Deaf. 2016. Available online: https://www.youtube.com/watch?v=vGF1KlaGa1E (accessed on 1 June 2021).
  76. Branje, C. The Vibrochord: Investigating a Vibrotactile Musical Instrument. Ph.D. Thesis, University of Toronto, Toronto, ON, Canada, 2014. [Google Scholar]
  77. Jiam, N.T.; Limb, C.J. Rhythm Processing in Cochlear Implant-Mediated Music Perception. Ann. N. Y. Acad. Sci. 2019, 1453, 22–28. [Google Scholar] [CrossRef]
  78. Baijal, A.; Kim, J.; Branje, C.; Russo, F.; Fels, D.I. Composing Vibrotactile Music: A Multi-Sensory Experience with the Emoti-Chair. In Proceedings of the Haptics Symposium, Vancouver, BC, Canada, 4–7 March 2012; pp. 509–515. [Google Scholar] [CrossRef] [Green Version]
  79. Russo, F.A.; Ammirante, P.; Fels, D.I. Vibrotactile Discrimination of Musical Timbre. J. Exp. Psychol. Hum. Percept. Perform. 2012, 38, 822–826. [Google Scholar] [CrossRef] [Green Version]
  80. Lamontagne, V.; Hattwick, I.; Franco, I.; Giordano, M.; Egloff, D.; Martinucci, M.; Salter, C.; Wanderley, M.M. The Ilinx Garment: Whole-body tactile experience in a multisensorial art installation. In Proceedings of the 21st International Symposium on Electronic Arts, Vancouver, BC, Canada, 14–18 August 2015; ACM: New York, NY, USA, 2015. [Google Scholar]
  81. Carter, T.; Seah, S.A.; Long, B.; Drinkwater, B.; Subramanian, S. UltraHaptics: Multi-Point Mid-Air Haptic Feedback for Touch Surfaces. In Proceedings of the 26th Annual ACM Symposium on User Interface Software and Technology, St. Andrews, UK, 8 October 2013; pp. 505–514. [Google Scholar] [CrossRef] [Green Version]
  82. Clark, C.N.; Warren, J.D. Music, memory and mechanisms in Alzheimer’s disease: Figure 1. Brain 2015, 138, 2122–2125. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  83. Gómez-Romero, M.; Jiménez-Palomares, M.; Rodríguez-Mansilla, J.; Flores-Nieto, A.; Garrido-Ardila, E.; González-López-Arza, M. Benefits of music therapy on behaviour disorders in subjects diagnosed with dementia: A systematic review. Neurología 2017, 32, 253–263. [Google Scholar] [CrossRef] [PubMed]
  84. Barrera, M.E.; Rykov, M.H.; Doyle, S.L. The effects of interactive music therapy on hospitalized children with cancer: A pilot study. Psycho-Oncology 2002, 11, 379–388. [Google Scholar] [CrossRef] [PubMed]
  85. Bradt, J.; Dileo, C.; Magill, L.; Teague, A. Music interventions for improving psychological and physical outcomes in cancer patients. Cochrane Database Syst. Rev. 2016, CD006911. [Google Scholar] [CrossRef]
Figure 1. Number of HMP publications since 1992 until 2020.
Figure 1. Number of HMP publications since 1992 until 2020.
Sensors 21 06575 g001
Figure 2. Preliminary classification of musical haptics.
Figure 2. Preliminary classification of musical haptics.
Sensors 21 06575 g002
Figure 3. Examples of haptic player installations (HMPIs): (a) Deaf user experiencing music in a haptic chair, retrieved from in [8]. (b) Multimodal platform, retrieved from in [49]. (c) Desktop haptic knob installation, retrieved from in [50]. (d) Desktop haptic button installation, retrieved from in [51].
Figure 3. Examples of haptic player installations (HMPIs): (a) Deaf user experiencing music in a haptic chair, retrieved from in [8]. (b) Multimodal platform, retrieved from in [49]. (c) Desktop haptic knob installation, retrieved from in [50]. (d) Desktop haptic button installation, retrieved from in [51].
Sensors 21 06575 g003
Figure 4. Examples of haptic music player-wearable devices (HMP-WDs): (a) Deaf user experiencing music with a haptic bracelet MuSS-Bits, retrieved from in [19]. (b) Haptic glove, retrieved from in [61]. (c) Haptic belt Hedonic Haptic Player, retrieved from in [35]. (d) Haptic jacket Body:Suit:Score, retrieved from in [65].
Figure 4. Examples of haptic music player-wearable devices (HMP-WDs): (a) Deaf user experiencing music with a haptic bracelet MuSS-Bits, retrieved from in [19]. (b) Haptic glove, retrieved from in [61]. (c) Haptic belt Hedonic Haptic Player, retrieved from in [35]. (d) Haptic jacket Body:Suit:Score, retrieved from in [65].
Sensors 21 06575 g004
Figure 5. Spatialization of frequencies in a haptic chair, retrieved from in [32].
Figure 5. Spatialization of frequencies in a haptic chair, retrieved from in [32].
Sensors 21 06575 g005
Figure 6. The Auris system for tactile rendering of melodic content, retrieved from in [40].
Figure 6. The Auris system for tactile rendering of melodic content, retrieved from in [40].
Sensors 21 06575 g006
Figure 7. Different instrument tracks for TM implementation in the Model Human Cochlea, retrieved from in [14].
Figure 7. Different instrument tracks for TM implementation in the Model Human Cochlea, retrieved from in [14].
Sensors 21 06575 g007
Figure 8. The Vibrochord, a vibrotactile musical input device VMID, retrieved from [66].
Figure 8. The Vibrochord, a vibrotactile musical input device VMID, retrieved from [66].
Sensors 21 06575 g008
Figure 9. Example of rhythmic VMC created for the Hedonic Haptic Player, retrieved from in [35].
Figure 9. Example of rhythmic VMC created for the Hedonic Haptic Player, retrieved from in [35].
Sensors 21 06575 g009
Table 1. Vibrotactile channel characteristics, adapted from [2].
Table 1. Vibrotactile channel characteristics, adapted from [2].
Psychophysical ChannelPNP-INP-IINP-III
Full namePacinianNon-Pacinian INon-Pacinian IINon-Pacinian III
Psychological typeFAF-IIFAF-ISA-IISA-I
Fiber innervation density (fingertip, per cm2)211404970
Subjective sensation“vibration”“flutter”unknown“pressure”
Frequency range40–500 Hz2–40 Hz100–500 Hz0.4–3 Hz
Prime sensitivity range 1 250–300 Hz25–40 Hz150–400 Hz0.4–1 Hz
Shape of frequency response functionU-shapeFlat 2 U-shapeFlat
1 Defined as best frequencies to lower threshold of perception; 2 Notch at 30 Hz.
Table 2. Overview of research on haptic music player installations (HMP-Is).
Table 2. Overview of research on haptic music player installations (HMP-Is).
HMP-IYearContact MechanismType of ActuatorStimuliFeatures Explored
SOMATRON [52]1992MattressSpeaker, SubwooferVibrotactile, AuditoryPitch
Vibratory Music (Patent) [45]2002ChairN/A 1 N/AN/A
Audiotactile Simultaneity [26]2004ChairN/AVibrotactile, AuditoryATFS 2
Symbolic Haptic Rendering [50]2005KnobN/AVibrotactileTempo, energy
Model Human Cochlea (Design) [14]2009ChairVoice coilVibrotactile, AuditoryFM 3 , TM 4 , VMLE 5
Multimodal reproduction [30]2009SeatVoice coilVibrotactile, AuditoryVMLE, BRTF
Music Display and Haptic Chair [31]2009ChairSpeakerVibrotactile, AuditoryVMLE
Model Human Cochlea [53]2009ChairVoice coilVibrotactile, AuditoryFM, TM, VMLE
Emoti Chair [54]2010ChairVoice coilVibrotactileFM, TM, VMLE
Whole Body Vibration [48]2010ChairVoice coilVibrotactile, AuditoryATFM 6 , BRTF
Auditory-Tactile Music [29]2013ChairVoice coilVibrotactile, Auditory, VisualBRTF, VMLE
Haptic Display [55]2013ChairSpeakerVibrotactile, Auditory, VisualVMLE,
Tactile Musical Device [32]2015ChairVoice coil, SubwooferVibrotactileLoudness, Pitch, Rhythm, Timbre, VMLE
Skin Music [33]2015ChairVoice coilVibrotactile, AuditoryVMLE
Musical Notes to the Skin [51]2016Button, platformVoice CoilVibrotactile, Auditory, VisualPitch
Feeling the Beat [47]2017PlatformVoice coilVibrotactile, AuditoryRhythm, Tempo, Beat Synchronization
Auditory-Tactile Experience of Music [25]2018ChairVoice coilVibrotactile, AuditoryATFS, BRTF, VMLE
Music with Vibrations [56]2019ChairVoice coilVibrotactile, AuditoryF-EQ 7 , VMLE,
Vibrotactile Consonance [57]2019ChairVoice coilVibrotactileMC 8
Haptic Music [49]2020PlatformVoice coilVibrotactile, Auditory, VisualBRTF, Frequency, VMC 9 , VMLE
1 Not applicable. 2 Audio-tactile frequency synchronism. 3 Frequency model: spatialization of frequencies. 4 Track model: spatialization of tracks. 5 Vibrotactile music with/without listening experience: explores overall perception of vibrotactile music, with or without music listening. 6 Audio-tactile frequency matching. 7 Frequency equalization: signal processing to control intensity of particular frequencies. 8 Melodic consonance. 9 Vibrotactile music composition.
Table 3. Overview of research on haptic music player-wearable devices (HMP-WDs).
Table 3. Overview of research on haptic music player-wearable devices (HMP-WDs).
HMP-WDYearContact MechanismType of ActuatorStimuliFeatures Explored
Cutaneous Grooves [15]2003Whole body suitVoice coil, SubwooferVibrotactile, AuditoryVMC
Model Human Cochlea [28]2008BeltSpeakerVibrotactile, AuditoryFM, TM, VMLE
Vibrotactile display [41]2010BeltSpeakerVibrotactileFD 1
Vibrotactile Music System (Design) [66]2011N/AN/AN/AVMLE
Vibrotactile Music System [66]2012Not availableNot availableNot availableFD, Interval Size, Pitch Direction
Dual Band HMP [37]2013Mobile device mock-upDMAVibrotactile, AuditoryF-EQ, Rhythm
MUVIB [59]2014BraceletVoice coilVibrotactile, AuditoryIntensity
Vibrotactile Chords [64]2014Mobile device mockupVoice coilVibrotactileMC
Vibrotactile Composition [67]2015Jacket, legginsERMVibrotactileVMC
CollarBeat [46]2015Collar, beltVoice coilVibrotactile, AuditoryVMLE
Audio-Tactile Conversion [63]2015Mobile device mock-upVoice coilVibrotactile, AuditoryFD
Vibroacoustic Device for Music [34]2016BeltMotor-String, voice coilVibrotactile, AuditoryAmplitude, Frequency, Rhythm, VMLE,
MuSS-Bits [18]2016Bracelet, magnetic, beltERMVibrotactile, VisualRhythm, VMLE
Mood Glove [61]2016GloveVoice CoilVibrotactile, Auditory, VisualRhythm, VMLE
Feeling Music [60]2017BraceletVoice coilVibrotactile, VisualRhythm
Hedonic Haptic Player: Design [68]2017BeltVoice coilVibrotactileVMC
Hedonic Haptic Player [35]2017BeltVoice coilVibrotactileRhythm, VMC, VMLE
Hapbeat Test [42]2017BeltMotor-String, voice coilVibrotactile, AuditoryAmplitude, Frequency, VMLE
Haptic Melodic Interval [69]2018Belt, mobile device mockupVibrotactileMelodic interval
Hapbeat Re-Design [70]2018BeltMotor-StringVibrotactileN/A
Music Sensory Substitution System [19]2018BraceletERMVibrotactile, VisualRhythm
Musical Scale Through Haptic Actuator [62]2018Mobile device mock-up, VMIDVoice coilVibrotactile, VisualMelody, Pitch, Timming, VME
Musical Haptic Wearables [58]2018Belt (armband)N/AVibrotactile, Auditory, VisualAPI 2 , VMLE
LIVEJACKET [38]2018JacketPiezoelectric, SubwooferVibrotactile, AuditoryTM, VMLE
Musical Haptic Sleeve [39]2019SleeveSpeakerVibrotactile, AuditoryVMLE
Body:Suit:Score [65]2019Whole body suitERMVibrotactile, AuditoryPitch, FM, Tempo, VME
Tactile Identification in Music [71]2019GloveVoice coilVibrotactile, AuditoryVMLE
Tactile Musical Emotion [72]2020GloveVoice coilVibrotactileTimbre, VMLE
Touching the audience [21]2020Bracelet, jacketERMVibrotactile, Auditory, VisualAPI, VMLE
Vibrotactile Captioning [73]2020GloveVoice coilVibrotactile, Auditory, VisualEEG 3 , Tempo, VMLE,
SENSE [74]2020GloveVoice coilVibrotactile, VisualVMLE
1 Frequency Discrimination; 2 Audience-Performer Interaction; 3 Vibrotactile perception supported with electroencephalogram.
Table 4. Overview of research on haptic music players-hybrid (HMP-H).
Table 4. Overview of research on haptic music players-hybrid (HMP-H).
HMP-H)YearContact MechanismType of ActuatorStimuliFeatures Explored
Skinscape [27]2001Chair, braceletVoice coil, SubwooferVibrotactile, AuditoryVMC
Vibrochord Vs. Piano [66]2014VMID + ChairN/AVibrotactileVMC, VME, VMLE
Vibrochord Design [76]2014Vibrotactile music input device + ChairN/AVibrotactileVMC, VME, VMLE
Concert for the Deaf [75]2016Platform, JacketSpeakerVibrotactile, VisualVMLE
Auris System [40]2017Chair, braceletVoice coil, SpeakerVibrotactileFrequency mapping, Rhythm, VMLE, EEG
Scaffolding the Music [8]2018Chair, beltChair with voice coil, Belt with ERMVibrotactile, VisualPitch, Frequency, Rhythm, VMLE
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Remache-Vinueza, B.; Trujillo-León, A.; Zapata, M.; Sarmiento-Ortiz, F.; Vidal-Verdú, F. Audio-Tactile Rendering: A Review on Technology and Methods to Convey Musical Information through the Sense of Touch. Sensors 2021, 21, 6575. https://doi.org/10.3390/s21196575

AMA Style

Remache-Vinueza B, Trujillo-León A, Zapata M, Sarmiento-Ortiz F, Vidal-Verdú F. Audio-Tactile Rendering: A Review on Technology and Methods to Convey Musical Information through the Sense of Touch. Sensors. 2021; 21(19):6575. https://doi.org/10.3390/s21196575

Chicago/Turabian Style

Remache-Vinueza, Byron, Andrés Trujillo-León, Mireya Zapata, Fabián Sarmiento-Ortiz, and Fernando Vidal-Verdú. 2021. "Audio-Tactile Rendering: A Review on Technology and Methods to Convey Musical Information through the Sense of Touch" Sensors 21, no. 19: 6575. https://doi.org/10.3390/s21196575

APA Style

Remache-Vinueza, B., Trujillo-León, A., Zapata, M., Sarmiento-Ortiz, F., & Vidal-Verdú, F. (2021). Audio-Tactile Rendering: A Review on Technology and Methods to Convey Musical Information through the Sense of Touch. Sensors, 21(19), 6575. https://doi.org/10.3390/s21196575

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop