1. Introduction
TV broadcasting, which is typically accessible to everyone, is an important medium for delivering crucial information, such as important news and government policy announcements, to the public. During the ongoing battle against COVID-19, we have often needed to listen to government briefings about the virus. When delivering such important information, a sign language interpreter is present next to the speaker so that DHH people can also understand the information. In 2020, the U.S. government was sued by the DHH association for not providing sign language interpreters during a COVID-19 briefing in the U.S. [
1].
It is recognized that sign language should be provided as a basic service when delivering socially and politically important information. As shown in
Figure 1, Korean sign language services are currently provided during the government’s COVID-19 briefings or major news sessions by major broadcasters in Korea. Although Korean sign language interpreters appear along with the speaker in the government’s COVID-19 briefings [
2], they are usually placed in a circle or square in the lower right corner of the screen with the size of about a quarter of the vertical length of the screen in usual TV news programs [
3]. These screens are viewed by both people with normal hearing and those who are DHH together. As shown in
Figure 1b, a Korean sign language interpreter is usually presented in the lower right corner in a very small size, which leads to a question of whether their hands, a crucial means of communication in sign language, are sufficiently recognizable to DHH people. In situations where sign language services are becoming the norm, the research team of this study attempted to identify and solve the potential problems lying in the current sign language services from the perspective of DHH people.
DHH individuals do not have a visibly distinguishable disability compared to others; what they suffer is problems due to their difficulties in hearing. The cognitive ability of DHH people is not different from that of non-disabled people [
4,
5]. However, Dyer et al. stated through several research examples [
6,
7,
8], “In most deaf people, reading and writing skills fail to achieve levels appropriate to the age and intelligence of the student, typically lagging their peers by several years in the final years of obligatory schooling”. In addition, he mentioned that the reason for their low functional literacy was because most DHH people have restricted access to spoken language forms [
9]. This is because there is a big difference between acquiring a language through hearing and acquiring a language through sight. Because of the loss of hearing, it is difficult for DHH people to learn language in a similar way to non-disabled people [
10]. Auer and Bernsteins suggested through experiments that “the acquisition of a speech vocabulary available by eye alone not only is smaller but also shows different distributional characteristics than that for heard speech” [
11]. Thus, because DHH people rely on sight to acquire information, their literacy ability is significantly lower than non-disabled people [
9,
12].
With regards to the language and information acquisition of DHH people, Korean Sign Language [
13], a study on information accessibility of the disabled [
14,
15,
16], a study on subtitles and visualization for DHH people [
17,
18,
19,
20,
21], etc., and research on sign language interpretation screens using eye trackers have also been conducted [
22,
23]. These studies have suggested diverse aspects related to the language and information acquisition of DHH people. First, the language used by non-disabled people is different from sign language in terms of grammar and word order. Therefore, the subtitles displayed following the grammar and word order of the language used by non-disabled people would be difficult to understand for those DHH people whose mother language is a sign language. Second, there are many DHH people whose literacy or ability to communicate in sign language is poor due to lack of proper sign language education. It is therefore important to give DHH people equal opportunities in terms of easy access to information acquisition via TV broadcasts. Through sign language services that provide information at a level comprehensible by non-disabled people as well, DHH people will be given the opportunity to understand and be exposed to grammar, word order, and various words of the language used by DHH people. Sign language services in TV broadcasts not only provide opportunities for DHH individuals, as citizens, not to be alienated from important information but also empowers them to improve their language skills. This will have the effect of giving a practice opportunity to match the language of non-disabled people with the language of DHH people.
DHH people who have lost hearing rely mostly on sight to receive information. In order to allow DHH people to better acquire information from TV broadcasts, the arrangement of visual elements is important for delivering information to them.
In Korea, the Korea Communications Commission and Telecommunications Technology Association have provided Korea Smart Sign Language Broadcasting Services as personalized services for DHH people since 2020. This service can be viewed on major channels via IPTV. Through these services, it is possible for DHH individuals to watch TV programs with adjustable screen size, location, subtitles, etc., for the convenience of information acquisition, as shown in
Figure 2. These Korean sign language broadcasts allow DHH individuals to freely turn on and off the sign language screen and adjust the size and position of the screen as well. In order to use the smart sign language broadcasting service, however, a TV that supports such functions is required. In normal TV sets without the required functions, it is impossible to adjust the screen elements for the sign language.
In this study, we tried to examine the arrangement of information for DHH people on the screen of ordinary TVs not supporting screen element adjusting controls. First, user experiments were conducted with various element placement conditions for DHH individuals to define the characteristics of information delivery types and major visual elements for information delivery for DHH individuals, and to find out how to arrange these elements so that both DHH and non-disabled people can acquire information efficiently. Based on the results, screen arrangement options for visual information delivery elements for DHH people were proposed.
3. Preferences Evaluation of Information Delivery Elements of TV Broadcasts for DHH Individuals
3.1. Elements of Information Delivery and Design of Arrangement Conditions
Experiments for all broadcast types were divided into Test Step 1 and Test Step 2 to conduct the experiment in two stages. In Test Step 1, the arrangement options of visual components were tested by each broadcasts’ characteristics, and in Test Step 2, the preference for the presence of subtitles was tested by adding subtitles to each option. For videos fully meeting the conditions for the experiments, a total of 23 experimental videos were created by collecting existing TV clips, filming the sign language interpreter’s movements corresponding to the content, and editing the footage as per each of the experimental conditions.
3.1.1. Experiment Design for News
For news, the screen was designed focusing on the visibility of the sign language interpreter delivering the content as shown in
Table 2. In step 1, the visibility of the interpreter was related to the size and placement of the interpreter. In this type of broadcast, four cases were designed by differentiating the visual components, such as size of the interpreter (small or large size) and background (with or without background), with the most widely used right side layout being set as default. The small size type of the interpreter is the size currently used by TV broadcasts. In step 2, subtitles were added to each of the four cases to examine whether the presence of subtitles was preferred or not.
3.1.2. Experiment Design for Panel Discussion
The biggest problem for DHH individuals watching panel discussions on TV is that it is difficult to know who is speaking and whose comments are being shown in the subtitles. Therefore, in the panel discussion, the focus was on the speaker-specific sign language and the arrangement of subtitles to allow the audience to easily recognize who’s speaking. As seen in
Table 3, in step 1, a total of three basic options were provided: (1) one sign language interpreter is provided for each speaker, (2) multiple sign language interpreters are placed horizontally, and (3) multiple sign language interpreters are placed vertically. After selecting an option in step 1, in step 2, an option of subtitles or individual subtitles for each speaker were also added. In the experiment, video clips of the panel discussion that dealt with politics, society, and entertainment were used without distinction. In the case of a sign language interpreter provided for each speaker, a circular background was used for the horizontal arrangement and a square background was used for the vertical arrangement for efficient use of the screen. In the case of individual subtitles in step 2, the background color and position of the subtitles for each speaker were set different to make it easier to visually distinguish them.
3.1.3. Experiment Design for Weather Forecasts
In the weather forecast type, the entire screen consists of weather data and a weather caster who explains the data that appears on one side of the screen, so the placement (location) of the sign language interpreter is important. With this in mind, in step 1, three basic options were determined: (1) a sign language interpreter is placed on the right side of the weather forecast screen, (2) a sign language interpreter is placed on the left side, and (3) a sign language interpreter and the data screen are presented separately. In step 2, subtitles were added to each option. In the case of a split screen in step 2, both sign language interpreters and subtitles were separated from the weather forecast screen in order to exclude the elements that could interfere with the weather forecast screen.
Table 4 presents the examples of the experiment design for the weather forecast type.
3.2. Experiment Environment
This experiment was conducted for 30 DHH participants in two laboratories as shown in
Figure 9 at the shelter for the DHH people run by the Gwangju city branch of the Korean Association of the Deaf. The subjects of the experiment were people with hearing impairment over the age of 20; the age range of subjects was from 30 to 60 and there were 24 females and 6 males. Regarding educational background, there were those without education (2), middle school graduates or lower (8), high school graduates (11), and college graduates (2). The experiment consisted of a total of three people: a subject, a sign language interpreter to help guide the experiment progress, and an investigator. Tobii Nano, an eye tracker, was attached to a 27-inch monitor to collect gaze information from the subject during the experiment.
Prior to the experiment, a survey was conducted on the level of the use of sign language and Hangeul (Korean language) among the subjects. The results showed that 93.5% of the 30 subjects had used sign language for more than 10 years, but only 74.2% of them had received regular sign language education before. A total of 45.2% of the subjects answered that there was an expression they did not know when using sign language. In addition, when asked about the use of character language, 71% of them answered that they had received Hangeul education and used Hangeul for more than 10 years. However, 83.9% answered that they felt difficulties due to the different systems of Hangeul and Korean sign language.
In this experiment, the subject was asked to select the most preferred subordinate option after watching the step 1 video clips for each broadcast type as described above. In step 2, they watched a video with subtitles added to the layout type selected in step 1. After completing step 2, they were asked to evaluate their preferences for all layout types that they watched in the 2 steps of the experiments. In addition, individual interviews were conducted to investigate the reasons for their preferred layout type. Each subject was asked to evaluate two types at random. An average of 20 subjects responded per type. For the eye tracking of the subjects, the eye movement path of the subjects and the frequency of gaze for each component of the screen were identified by using the screen-attached eye tracker “Tobii nano”, which is a convenient device for tracking the user’s gaze on the monitor screen [
24].
Initially, the experiment was designed with each subject answering a total of three phases of questions: one for each type (news, panel discussion, weather forecast). However, this took more than one and a half hours since the majority of the subjects were found to experience difficulties understanding and answering the questions after watching the video clips. A sign language interpreter helped them understand the experiments, but the experiment did not progress as smoothly as expected. Considering the fatigue of the subjects, the experiment design was changed so that one subject watched two clips for each type, and an average of 20 subjects participated in watching and answering sessions for each video type. Compared to the subjects with normal hearing, the experiments on the subjects with hearing loss took two to three times longer.
4. Experimental Results
The questions presented to the subjects for the evaluation of preference for various layout components are shown in the
Table 5 below. Among these, question 5 is to select the screen that was most uncomfortable to see, and, unlike questions 1 to 4, it is a negative question.
Prior to the comparative analysis for each type, we evaluated the inter-scorer reliability of the response’s scores for each question. Acceptable Cronbach’s alpha for three types (news, panel discussion, and weather forecasts) were obtained as 0.76, 0.68, and 0.77, respectively.
4.1. Results of Step 1 by Type
4.1.1. News
The results of the survey conducted after the subjects watched the video clips containing accidents (news) are as follows.
For the first question (“On which screen were the information delivery elements arranged most efficiently?”), as shown in
Figure 10, 14 subjects selected clip no. 4, 5 subjects selected clip no. 3, and 1 subject each selected clip no. 1 and no. 2. For the second question, (“On which screen were the information delivery elements arranged most properly and harmoniously?”), 15 subjects selected clip no. 4 and 5 selected clip no. 3. For question no. three, (“On what screen is the sign language interpreter most clearly visible?”), 13 subjects selected clip no. 4, 7 subjects selected clip no. 3, and no one selected clip no. 1. This suggests that the majority preferred to have a large sign language interpreter on screen without a background.
For the fourth question (“Which screen do you want to continue to use?”), 14 subjects selected clip no. 3, 5 subjects selected clip no. 4, and 1 subject each selected clip no. 1 and no. 2. For the fifth question, (“Which screen was the most difficult to understand the content?”), the overwhelming majority of the subjects chose clip no. 1, which presented the layout with a background and a sign language interpreter in small size. Considering that the majority of subjects selected clip no. 1 and no. 2, it was found that the size of the sign language interpreter was a more important consideration than the presence of a background.
To summarize, as shown in
Figure 11, the average value of news type preferences showed that the best screen layout for the news type was clip no. 4 with a large-sized sign language interpreter and no background of the interpreter, and the worst screen was clip no. 1, which is similar to the one currently shown on TV programs today.
In an individual interview after the preferences survey, as presented in
Table 6, the subjects presented the following opinions on the screen components of news type and the proposed screen layout.
Figure 12 shows the result of the gaze-tracking hit map in news cases. In the subtitled and most preferred layout, gaze was mostly focused on the subtitles and sign language interpreter. The subjects often identified professional words from the subtitle since sign language is limited in the sufficient delivery of the information. In the interview, a subject mentioned that it is essential to read subtitles because sign language does not contain many professional words used in news broadcasts. On the other hand, subjects least preferred watching the main content and the subtitled layout rather than watching the subtitles. If the presentation of the interpreter is too small, the subjects gave up trying to figure out the full content and, thus, watched fewer subtitles.
4.1.2. Panel Discussion
As shown in
Figure 13, in the panel discussion type, 12 subjects selected clip no. 1 and 8 subjects selected clip no. 2 for the first question (“On which screen were the information delivery elements arranged most efficiently?”). For the second question, (“On which screen were the information delivery elements arranged most properly and harmoniously?”), 11 subjects selected clip no. 2 and 8 selected clip no. 1. For the third question, (“On which screen is the sign language interpreter most clearly visible?”), 10 subjects each selected clip no. 1 and no. 2, respectively.
For the fourth question (“Which screen do you want to continue to use?”), 11 subjects selected clip no. 1, and 8 subjects selected clip no. 2. For the fifth question, (“Which screen was the most difficult to understand the content?”), clip no. 3 was selected the most, where the sign language screens were provided for each speaker and arranged vertically. A total of 14 subjects chose clip no. 3, and 5 selected clip no. 5.
When looking at the average value of preference in the panel discussion type, as shown in
Figure 14, basic type 1 and type 2 showed a similar level of preferences with 11.2 and 11.0, respectively. In an interview after the experiment, some subjects commented that there is a shortage of sign language interpreters in reality, and it is wasteful and not economical to have multiple interpreters on the same screen all at once. Clip no. 3, in which multiple sign language interpreters were vertically arranged separately, got a low score because it was difficult and confusing to match the speaker and the interpreter.
In an individual interview after the preferences survey, as presented in
Table 7, the subjects presented the following opinions on the screen components of panel discussion cases and the proposed screen layout.
Figure 15 shows examples of the gaze-tracking hit map for the panel discussion case. In the most preferred layouts, subjects fairly focused on the interpreter, however, in the worst layouts, the gaze was scattered and there was much transition between speakers and the corresponding interpreters to figure out who is speaking.
4.1.3. Weather Forecasts
For the first question (“On which screen were the information delivery elements arranged most efficiently?”), as shown in
Figure 16, 14 subjects selected clip no. 3, 5 subjects selected clip no. 1, and 3 subject selected clip no. 2. For the second question, (“On which screen were the information delivery elements arranged most properly and harmoniously?”) and the third question (“On which screen is the sign language interpreter most clearly visible”), 12 subjects selected clip no. 3 and 5 each selected clip no. 1 and no. 2. For the fourth question (“Which screen do you want to continue to use?”), 13 subjects selected clip no. 3, 5 chose clip no. 1, and 4 selected clip no. 2. For the fifth question, (“Which screen was the most difficult to understand the content?”), 8 subjects each chose clip no. 1 and no. 2 in which the interpreter was arranged differently. Clip no. 3, which is a split arrangement of the data screen and sign language interpreter, was selected by 6 subjects, showing that all three types of clips were preferred by a similar number of the subjects.
As for the average value of the preference of the weather forecast type, as shown in
Figure 17, clip no. 3, with the weather description part split from the large-sized sign language interpreter, showed the highest preference, and clip no. 2, with the sign language screen being placed on the opposite side of the weather caster, was at the lowest in preference. In the case of clip no. 1, which has a slightly higher score than clip no. 2, the subjects’ opinions were that the lower right arrangement of the sign language interpreter (like the current broadcast sign language services) is familiar, and it is easy to see as the interpreter was in the same direction as the weather caster.
They answered that it would be difficult to see if the weather data were too small because the entire screen is a visual reference containing information. They also answered that the visibility of the weather data was more important than the sign language interpreter or subtitles for weather report broadcasts. In addition, they gave the opinions that the sign language screen should not cover the hands of the weather caster because a weather caster’s hands indicating the weather data serves an important role in weather forecasts.
Figure 18 shows the result of the gaze-tracking hit map in the weather forecast case. The subjects mostly focused on the interpreter and sometimes watched subtitles or the weather information. Compared to the other two cases, the dependency on the subtitle is small, since the presentation and content of the weather information is very formal and easy.
In individual interviews, the subjects presented the following opinions on the screen components of weather report clips and the proposed screen layout as presented in
Table 8.
4.2. Results of Experiments with Subtitles
In response to the question, “Choose the one with or without subtitles that you think is useful, and why?”, the subjects answered that they preferred the presence of subtitles for all three types of clips, as shown in
Figure 19.
Regarding subtitles, as presented in
Table 9, the subjects responded that subtitles and sign language should be provided together to help people with hearing impairment acquire information, and that subtitles also help them to learn unknown words or vocabulary. Some mentioned the problem of the arrangement and speed of subtitles, saying that the speed of subtitles was too fast for DHH individuals and hard to see at the same time with sign language. Therefore, many commented that it would be better to provide 2–3 lines of subtitles instead of one line, allowing the subtitles to stay on the screen longer, so that people do not miss the content while watching it alternately with the sign language.
Regarding the experiments with subtitles added to the three types of video clips, all subjects answered that the subtitles are necessary for the news clips. They responded that there may be unfamiliar expressions if only a sign language interpreter is provided, but when subtitles and sign language interpretation are provided together, this gives them an opportunity to understand or learn difficult words. Looking at the eye tracker data analysis results for news type footage, some subjects looked at the subtitles more than the sign language interpreters.
For the panel discussion type, three conditions were tested: (1) No subtitles, (2) One representative subtitle, and (3) Individual subtitles for each of the multiple speakers. As shown in
Figure 20, 2 out of 21 subjects answered that it was easier to see the clips without subtitles, and 19 answered that they liked it better when subtitles are provided. Among the 19 subjects who answered they preferred the presence of the subtitles, 6 said they preferred one integrated subtitle and 13 preferred individual subtitles for each speaker.
Many answered that arranging multiple sign language interpreters for each speaker, as proposed in the basic type in step 1 for the panel discussion clip, was confusing and financially not feasible. The subjects’ opinion was that it would be very good to have one sign language interpreter and individual subtitles for each speaker. A case where there is only one sign language interpreter performing while placing multiple subtitles separately for each speaker in the panel discussion, as shown in
Figure 21, showed the highest preference at 62%. Most of the subjects responded that it was less confusing and easy to understand which speaker was talking when subtitles appeared for each speaker. There was one opinion that it would be nice to provide subtitles in the form of speech bubbles.
In the weather forecast type, most of the subjects preferred the presence of subtitles, while 9% of the respondents preferred the clip with no subtitles. They said the reason is that subtitles are not necessary because the sign language interpretation and the reference weather report data provided can be enough, to some extent, when it comes to weather reports.
In conclusion, based on the results of experiments and subjects’ opinions, the role or necessity of subtitles can be identified. Therefore, subtitles are necessary and should be provided in 2–3 lines and stay longer on the screen so that DHH individuals can read them along with a sign language interpretation. In panel discussions with multiple speakers appearing, the need for individual subtitles was also found.
5. Design Proposal of TV Broadcasting for DHH Individuals
With the eye tracker data collected on the screen navigation from the three types of TV broadcasting, it was found that DHH participants viewed the sign language interpreter more than the reference video or subtitles in all cases. In addition, DHH participants tended to look at the face of the sign language interpreter the most, because not only the interpreter’s hand shapes but also the facial expressions of the interpreter are very important when comprehending the message. This is because, even for the same hand shapes, the viewer can acquire additional information (i.e., whether the content is positive or negative) depending on the interpreter’s facial expression. As shown in
Figure 22, the sign language interpreter is currently broadcast in about a quarter of the size of the screen. This is not the appropriate size for viewers to see facial expressions. Therefore, the sign language interpreter should be placed as large as possible on the screen. Moreover, since DHH participants do not prefer the background behind the interpreter, the interpreter would be better placed without a background.
Based on the results from the experiments conducted above, the basic conditions for the information delivery component layout design for DHH individuals are presented in
Table 10.
Based on the above results, the layout design for each information type for DHH individuals was organized.
In the case of the news type, the opinion of most of the subjects that the size of the sign language interpreter was too small was respected. For news type clips, the sign language interpreter is the most important component in information delivery because DHH individuals acquire information through sign language rather than the reference footages provided. As shown in
Figure 23, the interpreter size is enlarged to 1/3 the height of the full screen so that sign language can be seen clearly, improving visibility compared to existing ones. The background of the interpreter has been removed so that the reference footage is not unnecessarily blocked, and the viewers can feel less stuffy watching the screen. For the subtitles, two lines are provided at a time allowing the subtitles not to pass too quickly so that DHH individuals can look at both sign language and subtitles together.
For panel discussions with multiple speakers, the subjects’ opinion that multiple interpreters appearing can be confusing and impractical was reflected in the proposed layout. As seen in
Figure 24, only one interpreter is placed on the screen. Subtitles are provided for each speaker so the viewers can clearly understand each speaker’s comment. In the layout, it is important that the subtitles are not blocking the mouth or hands of the speakers. It is easy to identify the speaker when the subtitles are color-coded by speaker, so translucent subtitle boxes of various colors have been used.
For weather forecast types where the weather data and caster are presented in the same screen, it is recommended to reduce the size of the weather data section and make it separate from the sign language interpreter, as shown in
Figure 25, so that the interpreter does not block the weather data section. A blank space is created on the screen because of this. In this blank space, a sign language interpreter and subtitles are placed. Both the weather caster and sign language interpreter are placed on the right side on the screen so that the movement of the viewers’ eyes can be minimized when watching the weather forecast. The sign language interpreter should be careful not to block the weather caster and his/her hands or indications. Although some answered that subtitles are not highly necessary for weather forecast types, the overall survey results showed that most of the respondents think that subtitles are required. Therefore, two to three lines of subtitles are placed on the screen that do not block other information on the screen.
One general cautionary note for all broadcast types is that improving the visibility of sign language interpreters should be the priority, and subtitles are considered next. Although 29% of the subjects answered in the pre-interview that it is important and necessary that sign language and subtitles appear together, the results of eye tracker analysis showed that the time the subjects stared at the subtitles was shorter than the time they spent looking at the sign language interpreter. From this, it can be assumed that sign language acquired as the native language is the main tool for DHH individuals to acquire key information. Therefore, it is necessary to increase the visibility of sign language interpreters and provide subtitles to expand information acquisition opportunities for DHH people who have acquired sign language as their native language.
6. Conclusions
This study was conducted to analyze and find solutions to the difficulties experienced by DHH people in viewing public TV broadcasts alongside non-disabled people. The process of discovering and resolving these problems was the starting point for understanding how DHH individuals communicate information. Through research and experimentation, three important considerations have been identified from this study. The first is the importance of subtitles. Contrary to the assumptions of non-disabled people that DHH people do not see subtitles but resort only to sign language, they are watching subtitles and sign language together, and the subtitles can also serve as an opportunity for DHH people to learn. This is a problem that DHH people experience because the language system of sign language is different from that of Hangeul. This also indicates that providing subtitles can serve as an opportunity for DHH people to understand the language system of non-disabled people.
Second, sign language can be understood accurately when a viewer looks at the hand shapes and facial expressions of the performer together due to the linguistic limitations of sign language, which conveys meaning only by hands. Third, DHH participants find the background of the sign language interpreter uncomfortable. The investigator (and as a designer) assumed that distinguishing the sign language interpreter by using a background would increase visibility from the screen, but DHH participants felt that the interpreter is confined to a small space when there is a background. Some also responded that the background color made their eyes tired, thereby interfering with watching sign language. Contrary to the general conception that there would be no problem for DHH people in viewing TV broadcasts simply by placing a sign language interpreter on the screen, the results of this study found that a more delicate screen layout design is necessary for DHH people.
It is expected that this study will serve as a helpful guide in providing better sign language services for TV broadcasts that can be conveniently viewed by both DHH people and non-disabled people. However, the experiments were conducted only on DHH people in this study. Therefore, in the future study, it is necessary to analyze the preference and satisfaction of non-disabled people who watch sign language services for TV broadcasts as well. To that end, the scope of this study should be further expanded to find ideal TV news layout designs that can be enjoyed both by DHH people and non-disabled people. Additionally, many other factors, e.g., background with illumination variations or variations of the sign interpreter in age and sex, could be investigated as a part of the future work to find a better design for TV broadcasts.