Next Article in Journal
On-Board Data Management Layer: Connected Vehicle as Data Platform
Next Article in Special Issue
Integration of Extended Reality and a High-Fidelity Simulator in Team-Based Simulations for Emergency Scenarios
Previous Article in Journal
Forecasting PM2.5 Concentration Using a Single-Dense Layer BiLSTM Method
Previous Article in Special Issue
Planar-Equirectangular Image Stitching
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Production of Mobile English Language Teaching Application Based on Text Interface Using Deep Learning

1
Department of Computer Engineering, Graduate School, Hansung University, Seoul 02876, Korea
2
Division of Computer Engineering, Hansung University, Seoul 02876, Korea
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(15), 1809; https://doi.org/10.3390/electronics10151809
Submission received: 29 June 2021 / Revised: 26 July 2021 / Accepted: 26 July 2021 / Published: 28 July 2021
(This article belongs to the Special Issue LifeXR: Concepts, Technology and Design for Everyday XR)

Abstract

:
This paper proposes a novel text interface using deep learning in a mobile platform environment and presents the English language teaching applications created based on our interface. First, an interface for handwriting texts is designed with a simple structure based on a touch-based input method of mobile platform applications. This input method is easier and more convenient than the existing graphical user interface (GUI), in which menu items such as buttons are selected repeatedly or step by step. Next, an interaction that intuitively facilitates a behavior and decision making from the input text is proposed. We propose an interaction technique that recognizes a text handwritten on the text interface through the Extended Modified National Institute of Standards and Technology (EMNIST) dataset and a convolutional neural network (CNN) model and connects the text to a behavior. Finally, using the proposed interface, we create English language teaching applications that can effectively facilitate learning alphabet writing and words using handwriting. Then, the satisfaction regarding the interface during the educational process is analyzed and verified through a survey experiment with users.

1. Introduction

User interfaces for human–computer interactions (HCIs) have evolved in user-oriented, easy, and convenient ways, and are considered as an important technology in various platform and application areas. In recent years, along with the advancement of virtual reality (VR) and augmented reality (AR) technologies, interest has been increasing in immersive interactions and interfaces for immersive content.
Shneiderman [1] proposed a direct manipulation to provide a user interface similar to the way a user manipulates objects by hand to act and perform tasks in a real-world environment. In HCI, interfaces are provided in a form similar to the way humans behave in and manipulate the real environment and have been developed to facilitate natural interactions. Interactions with computers should consider the input device type and method and the use of the five senses, such as vision, sound, and touch, which are properly reflected in human emotion, decision making, and behavioral responses. Accordingly, the user will experience a range of emotional experiences, such as immersion, presence, and satisfaction, along with the one-dimensional function of receiving information. Some studies have analyzed user interfaces in this aspect [2,3].
Furthermore, in the field of immersive contents using VR and AR, studies have been conducted to design and analyze user interfaces (gaze, hand, walking, etc.) that provide satisfying presence and experience in interactions between the user and the virtual environment [4,5,6,7,8,9]. Studies have been also conducted to design an optimized interface for a given user platform and experiential environment in terms of interactions in an asymmetric virtual environment, considering the collaboration in a multi-user virtual experience environment or the user participation on various platforms (VR, AR, personal computer (PC), mobile, etc.) [10,11,12]. Recently, through convergence with machine learning, new interfaces have been developed to facilitate the user’s behavior and decision making more conveniently in an intuitive structure [6,13]. Studies on novel interfaces based on user experience and satisfaction are regarded as important in the process of developing content and applications for various platforms and experiential environments.
User interfaces are used when creating applications for various fields, including in education, games, the military, and manufacturing. In particular, studies have been conducted for immersive education applications that facilitate training and learning in a highly immersive user environment using VR and AR technologies. Recently, with the advancement of smart devices, there has been a growing interest in educational applications based on various platforms, such as mobile devices, tablets, and VR equipment as well as PCs [14,15,16]. Therefore, further studies are required for user interfaces that consider user immersion and convenience in educational applications.
With the aim of providing a new environment for educational experiences and a satisfying interface experience, this study designs a new interface suitable for mobile platform applications and uses it to create an English language teaching application. The key objectives of this study are as follows.
1.
A deep-learning-based text interface is proposed through the Extended Modified National Institute of Standards and Technology (EMNIST) dataset and a convolutional neural network (CNN) to control behaviors in an intuitive structure based on touch-based input processing in the mobile platform environment;
2.
English language teaching applications that reflect immersive interactions are created to facilitate behaviors directly from the text interface by connecting the user’s handwritten texts to alphabet writing and word learning.

2. Related Work

In traditional classroom education, learners participate relatively passively in the educator-centered one-way class. This case Z may lead to limited educational effects because it is difficult to arouse interest in learning in the educational environment. Therefore, triggering the interest of learners and interacting effectively are important factors in improving the quality of education. Furthermore, the educational environment should be provided in real or simulated realistic situations, particularly for young learners who lack life experience. For instance, the educational effect can be improved if pictures and audio are used to simulate real-world situations [17]. Hence, various applications have been developed for education using digital devices, such as computers, and related studies have been conducted.
Vernadakis et al. [18] conducted studies to analyze the potential benefits of education using digital devices, such as computers. They confirmed that graphical user interface (GUI)-based education with animations and interactive multimedia content can improve the educational effects on learning phonemes, words, writing, etc., by increasing the interest of learners and facilitating self-directed learning that best suits their learning speed. In another study [19], the effect on language learning was analyzed by focusing on a mobile-based educational application program. Studies that systematically analyzed the relationship between the use of computer-aided tools and learning were conducted from various perspectives. For example, methods of integrating language learning using various digital tools, such as blogs, iPad apps, and online games, the benefits (motivation, interaction, collaboration, etc.) provided by digital tools, and the digital technologies used by teachers in daily teaching were analyzed [20]. Furthermore, a study applied game elements to education [21], and another study investigated the impact of e-books containing animation elements on the language development of learners [22]. It was confirmed that the elements of computing tools, such as multimedia content, interaction, game, and collaboration, can increase the educational effect through benefits, such as increased attention and motivation in education. Particularly, they have a positive effect on supporting various learning activities, such as text comprehension, vocabulary, speaking, listening, and writing.
In recent years, immersive media technologies, such as VR and AR, have been used in the development of teaching applications in not only language teaching but also various industrial sectors, including the military and manufacturing. Lele [14] suggested a direction for applying VR to the military field. Alhalabi [15] applied VR to engineering education and based on the use/non-use of VR and independent devices without a tracking system, they conducted a comparative study on the effect of VR systems on engineering education. Jensen and Konradsen [23] attempted to update the knowledge for the use of a head-mounted display (HMD) in education and training. They examined various situations in which the HMD is useful for acquiring skills and analyzed the motion sickness, technical issues, immersive experiences, etc., in comparison with traditional education. In addition, a new method was proposed to use AR in teaching English to preschool children [17], and major AR-based learning activities, such as word spelling games, word knowledge activities, and location-based word activities, were analyzed. They analyzed how to apply the benefits of AR to support major activities of language learning and proposed a direction in this regard [24]. Doolani et al. [16] examined methods of applying extended reality (XR) to education in the manufacturing sector. As such, production technologies of educational applications using digital devices are evolving into the XR technology beyond VR and AR. However, along with the evolution of the platforms, there is a need to conduct studies in a new direction suitable for platforms and technologies to develop user interfaces that can increase the convenience and immersion as an approach toward increasing the learner-centric educational effect.
Various studies have designed interfaces considering the satisfaction, experience, and immersion from the user’s perspective, and accordingly proposed immersive interactions [2,3]. Some studies related to user interfaces have been conducted in conjunction with various input processing methods, including the gaze, hand gesture, and controller, with the advancement of immersive media technologies, such as VR and AR [5,9,25]. In addition to the studies on VR user-oriented immersive interactions, some studies have proposed user interfaces having intuitive structures to improve the presence of experiencers in the immersive experience environment of HMD users, in which the immersion is relatively low. For example, a hand tracking device was used to design an interface, in which the PC users participate directly in an asymmetric virtual environment where PC-based non-HMD users and HMD-based VR users experience together [12]. As another example, a study designed a touch interface suitable for mobile devices in a virtual environment in which the mobile-platform-based AR users and HMD-based VR users go through the experience together [26]. Recently, studies have been conducted to propose easy, fast, and convenient hand interfaces by applying deep learning technology to directly connect gestures and actions in conventional GUI-based decision-making and action-performing processes [6]. Moreover, there is a growing interest in studies that apply deep learning to interfaces [13]. A study on deep learning applications has proposed a new direction that can provide more beneficial and scientific education than conventional education models by applying deep learning to not only an interface but also an education model [27]. In this study, we specifically set a topic for English language teaching based on a mobile platform and apply deep learning to design an interface that can be extended to digital teaching applications and AR.

3. Text Interface Using Deep Learning

The proposed interface has a structure that can be applied to mobile platform applications and uses handwritten text based on touch inputs. Under the premise of convenient and highly accessible interactions, the proposed interface uses deep learning to perform intuitive actions with objects or a virtual environment. An integrated development environment using the Unity3D engine [28] is implemented with the aim of designing an interface that can be extended to not only mobile applications but also AR applications. Figure 1 shows the structure of the proposed text interface using deep learning.

3.1. Touch-Based Text Interface

Mobile platform applications are generally based on an interface method in which a hand touches the screen directly. Input types include touch, double touch, drag, swipe, and long press, and interactions and actions are performed through the GUI. The input method using touch is advantageous in that direct interactions with virtual objects are performed, but it has a limitation in representing various actions other than object selections. Furthermore, GUI-based interactions represented in graphics, such as buttons, images, and menus, have a disadvantage in that, as the functions increase, the screen becomes more complex and the users need to learn more manuals.
In this study, the conventional touch inputs and GUI are used minimally to design a highly accessible interface. Furthermore, we propose a text interface in which handwritten texts are drawn to perform various actions intuitively in the process of the user interacting with the environment or objects.
Figure 2 shows a text interface using a GUI in the Unity3D development environment, and the proposed text interface comprises a virtual joystick that controls the movement of characters, the output image of the handwritten text drawing result, and an action execution button. The key is the process of processing the handwritten text drawing, and a text interface is implemented using the Unity3D engine’s Line Renderer properties and InputManager. From the moment a finger touches the screen to the moment the finger disengages, the finger coordinates on the screen are added as the positions connecting a line to draw the handwriting. The text interface procedure of Algorithm 1 shows the process of drawing handwritten text by touching the screen and sets the font color and line width through the Line Renderer property. In this case, the attribute value was similarly set in consideration of the font attribute of EMNIST dataset. The RenderTexture image is expressed in the GUI to check the handwriting being drawn by the user in real time. The user can write easily and conveniently.

3.2. Interaction: From Text to Behavior

The touch-based text interface in the mobile platform environment aims to deliver the user’s thought and decisions directly to the virtual environment. Hence, we propose interactions that are connected to behaviors accurately with a fast and intuitive structure while minimizing the process of performing behaviors through a conventional GUI of conversational systems.
From the text interface, the user’s handwritten text is recognized, and based on the recognized result, the behavior is performed in real time without requiring a separate GUI, thereby providing convenient and highly immersive interactions.
Algorithm 1: Process of text to behavior interaction.
1: procedure TEXT INTERFACE
2:    if touch state then
3:       pos[] ← draw screen points.
4:       The coordinates where the fingertip is drawing in the touch state are stored in
   the pos array.
5:       The pos array coordinates are connected as a line.
6:     else
7:        Draw screen points are cleared.
8:     end if
9:     The handwritten text composed by connected lines is stored as an image.
10: end procedure
11: i m g d r a w ← image of handwritten text drawn using text interface.
12: pr ← maximum probability value for inference
13: procedure BEHAVIOR PROCESS( i m g d r a w , pr)
14:      l← learning data using CNN based on the EMNIST dataset.
15:      t← the text recognition result of CNN with an inference value of pr or higher via
   setting i m g d r a w as an input node and using l.
16:      bv[] ← a behavior array corresponding to the text.
17:      Active(bv[t]).
18: end procedure

3.2.1. Text Recognition Using EMNIST and CNN

The Modified National Institute of Standards and Technology (MNIST) dataset contains images of handwritten numbers and it is used as a standard benchmark for learning, classification, and computer-vision systems. The MNIST dataset has been built based on the National Institute of Standards and Technology (NIST) dataset and is characterized by easy-to-understand and intuitive characteristics, relatively small size and storage requirements, easy accessibility to the database itself, and ease of use.
In this study, we design the recognition of handwritten text input received from the text interface by using the EMNIST, which also considers the handwritten alphabet by expanding the MNIST dataset composed of numbers [29,30]. The EMNIST dataset image composed of a collection of handwritten alphanumeric is converted to a 28 × 28 format. The training set has 697,932 images and the test set has 116,323 uppercase and lowercase letters and numbers 0–9 that map to their classes. Each image represents an intensity value of 784 pixels (28 × 28 grayscale pixels) in the range 0–255. The test and train images have a white foreground and a black background. This dataset consists of a total of 62 combinations of 10 numbers from 0 to 9 and 26 letters, each with uppercase and lowercase letters [31]. Figure 3 shows a portion of the EMNIST dataset consisting of numbers and the alphabet.
A CNN, which is efficient for image recognition (inference), is used to provide a natural and accurate interaction that connects the text to behavior. The user’s handwriting is accurately recognized through the processing of learning and inferring the EMNIST dataset in the CNN. The CNN structure is based on the CNN model proposed by Mor et al. [31] for EMNIST classification for the purpose of developing a handwritten text classification app for Android. The diagram of the CNN model used in this study stores the handwritten image input through the text interface as 256 × 256. The CNN structure consists of a total of eight layers, from the reshape layer that converts the input image to a 28 × 28 image, two convolutional layers, a Maxpooling layer, a flatten layer that converts a two-dimensional image into a one-dimensional array, a dense and drop-out layer, and the last output layer. To design an interface used in creating conversational content on the mobile platform, this study uses TensorFlowSharp [32] in the Unity3D engine and performs the inference from the EMNIST dataset learned based on the user’s handwriting input received from the interface [6]. Figure 4 shows the process of recognizing and processing a handwritten text using the text interface based on CNN and EMNIST dataset in the Unity3D engine development environment.

3.2.2. Behavior Process

Based on the text interface using a deep learning model, the user interacts with the virtual environment through a direct behavior in an intuitive structure. The texts mapped to behaviors are defined based on 62 text characters. When the user enters one of the defined texts in handwriting, the handwritten text is inferred through deep learning, and the behavior corresponding to the input text is performed in real time without requiring a separate GUI.
Algorithm 1 summarizes the process of receiving a handwritten text as an input from the text interface and inferring the handwriting to connect to behavior.

4. English Language Teaching Application

4.1. Design

The proposed interface aims to facilitate decision making and behaviors easily and conveniently in conversational applications on mobile platforms with a more intuitive structure compared with that of GUI-based interactions. This process is assumed to provide interesting experiences and immersions in the applications along with a satisfying interface experience. This novel interface design is expected to further enhance the benefits of GUI-based digital content, thus having positive effects on education. Therefore, along with the user evaluation of the proposed interface, this study aims to create English language teaching applications with a new conversational structure optimized for the interface.
The application is planned and produced as a casual game by using texts defined in advance and defining English word object selection behaviors corresponding to the predefined texts. Figure 5 shows animal objects against a nature background, based on a casual concept using the scenes in the applications created in this study. The learner plays the game by writing the animal names while exploring the scenes from the first-person perspective (Supplementary Video S1).

4.2. Production

4.2.1. Process

The English language teaching applications use graphic resources (animal objects: Cartoon Animals [33], background: Cartoon Town and Farm [34], button GUI: Game GUI Buttons [35]) provided by the asset store in the Unity3D engine development environment to create the scenes and use the functions and components provided by the engine to implement the conversational structure and interactions. The application GUI comprises a virtual joypad, text drawing, and behavior execution button, as shown in Figure 2. First, the user’s movement is implemented through the virtual joypad input, which is processed such that the camera follows the character from the first-person perspective. In the moving process, if an animal object approaches, the gaze of the animal is directed toward the user (first-person camera), and the text drawing and behavior execution button are activated. Here, if the behavior execution button is pressed after writing by hand the first letter of the animal name in the text interface, the whole word and spoken word are output along with a disappearing motion of the animal object, which is then finally removed. If the user inputs an answer different from the animal’s letter, an incorrect answer sound is output. Figure 6 summarizes the flow of performing the English word teaching process through the proposed interface and English language teaching applications (Supplementary Video S1).
In this study, we examine the production of conversational-structure-based mobile applications and the production of AR applications on the mobile platform to validate their applicability to immersive contents.

4.2.2. Mobile

The mobile English language teaching application produced in this study has screens of the landscape orientation in the Android operating system, and the scenes were designed to increase the interest and immersion during the education process by placing casual game elements with background sounds. Figure 7 shows some results of the produced application.

4.2.3. Augmented Reality

In a situation where there is a growing interest in producing educational applications using immersive media technologies, such as VR and AR, we have used a mobile AR that is highly accessible and easy to use in the proposed educational application. This study uses Vuforia to combine the mobile AR technology with the proposed educational application. Vuforia is a mobile-platform-specific AR software development tool used for the development of AR applications [36]. Vuforia uses computer vision technology to provide marker-based as well as marker-less tracking methods. Vuforia uses various types of markers, such as single images, cubic objects (six images), cylinder objects, and 3D objects. It can generate various virtual scenes combined with the real world by analyzing the real-world images captured using a camera, GPS, and accelerometer, and calculating spatial information.
In this study, we produce an AR application using image targets to provide easy access and usability. As Vuforia is provided as a single function in the Unity3D engine, it can be easily used by developers. Therefore, Vuforia can be easily expanded by adding an AR camera and image targets in the production process of a mobile application. Figure 8 shows some results of applying it to the production of an AR teaching application. In the case of AR application, the flow of content is the same as for mobile platforms. However, there is a difference in that education and experience can be conducted from various angles and viewpoints desired by the user.

5. Experimental Results and Analysis

For the proposed text interface and interaction, a deep learning model was implemented using Anaconda 3, Conda 4.6.12, and TensorFlow 1.13.0. The recognition and inference in the learning model were implemented using the TensorFlowSharp 1.15.1 plug-in with the Unity3D engine. Furthermore, the interface-technology-applied mobile English language teaching application was produced using Unity3D 2019.2.3f1 (64 bit). Here, Vuforia engine 8.5.9 was used for the production of the AR application. The PC environment used to create and test the applications included Intel Core i7-6700, 16 GB RAM, and Geforce GTX 1080 GPU. Furthermore, a Galaxy Note 9 mobile device was used to run and test the produced applications.
The satisfaction with the proposed interface technology was comparatively analyzed with the users, and survey experiments were conducted to determine the experience provided by the English language teaching applications. There were 10 survey participants (seven males and three females) between the ages of 23 and 39. The survey participants consisted of students or researchers majoring in computer science who had experience in creating or using digital content such as games and education. The key objectives of the surveys were to analyze the satisfaction provided by the proposed interface and check the educational experience of the English language teaching applications.
The first survey experiment was the analysis of satisfaction regarding the text interface. The proposed interface aims to provide a highly immersive experience environment with easy and convenient use through direction interactions based on a minimized GUI and touch inputs. Accordingly, we conducted a survey experiment using Arnold Lund’s USE (Usefulness, Satisfaction, and Ease of use) questionnaire [37]. It comprises 30 questions on a seven-point scale (1∼7) based on four items, namely, usefulness, ease of use, ease of learning, and satisfaction. Table 1 presents the statistical data based on the survey results. A high satisfaction score of approximately 6.0 was recorded for the four items. In particular, the survey results show that there are benefits such as easy and convenient learning and using because texts are input by directly drawing handwritten texts on the screen, without having to learn how to use the GUI of the mobile applications. However, a low score was recorded for the usefulness because the users who were familiar with the conventional GUI method experienced awkwardness and unfamiliarity with the process of interacting through the new interface method. Nevertheless, as the learning and using processes were easy, the users provided an opinion that they could experience immersion with high satisfaction eventually after adaptation. However, since it is not a comparison experiment with the existing GUI, it is only confirmed that a satisfactory interface experience can be provided to the user despite the new interface from the point of view of the proposed interface.
The second survey experiment was the analysis of the user experience for the English language teaching applications. To this end, we used a five-point scale (0∼4) to record the responses based on items such as the core module, positive experience, and social interaction module, using a game experience questionnaire (GEQ) [38]. Table 2 summarizes the survey results, in which the mobile and AR applications were classified. First, the survey result on the overall experience confirms that the English language teaching applications provided interesting experiences similarly on all the platforms. The comparison results of specific items show that the visual differences of the applications provided through the combination of reality and virtual world have positive effects on increasing immersion or concentration. However, the immersive experience environment (AR) sometimes disturbed the learning or had a negative effect because of inconvenient manipulations or errors in the controlling process. In other words, there is a difference in the experience environment in which mobile can experience content from a specified composition and viewpoint, whereas AR can explore and control scenes or objects from a desired angle and various distances (near or far). Due to these characteristics, positive effects on immersion and tension, as well as negative effects such as dizziness and confusion, were shown.
A comprehensive analysis of the survey results shows that the proposed interface has the effect of enhancing the experience of the English language teaching applications. If the advantages of immersive technology, such as VR and AR, are used appropriately, good educational effects will be produced based on high immersion.

6. Limitation and Discussion

This study confirmed interface satisfaction and application experience through a survey experiment. However, there are limitations in that the sample size is small as the number of participants is around 10 and the age range of the participants is narrow. In particular, in the case of an English language teaching application, even though it is an educational environment for preschool users, preschool children are not considered in the survey.
As for the problem that the survey subjects of the educational applications for preschool users were not preschool children, we intend to recruit multiple participants between the ages of 5 and 8 and analyze the educational effects or usability of the proposed applications. In addition, the AR application focuses on a new experience environment and various viewpoint changes, rather than the combination of real and virtual ones. Therefore, an AR application will be possible to design to increase the advantages of AR Characteristic by reducing virtual background objects and focusing on virtual animal objects centered on English words.
Finally, based on application production and survey results, it is judged that the proposed interface can be applied to the existing commercial English language teaching applications [39,40] or to create new commercial education applications.

7. Conclusions

In this paper, we proposed a novel interface using deep learning that is easy and convenient to use with an intuitive structure for interactions that can improve the educational effects by triggering immersion and interest in the education process in mobile platform environments. First, we designed an interface that can interact with the virtual environment using a minimal GUI by drawing handwritten texts based on the touch-based input method of the mobile platform. Then, we implemented an interaction technology using the EMNIST dataset and a CNN to represent decision making and behaviors accurately and effectively by recognizing and inferring the user’s handwriting. Based on the proposed interface and interaction technology, we developed English language teaching applications that can facilitate effective word education. Furthermore, we suggested a development direction by applying the AR technology with the mobile platform application. A survey through the USE questionnaire was conducted to analyze the proposed interface, and a survey experiment was conducted to analyze the user experience for the English language teaching applications. The results confirmed that the users were satisfied with both the proposed interface and the educational experience in the process of using the applications.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/electronics10151809/s1, Video S1: Production of Mobile English Language Teaching Application Based on Text Interface Using Deep Learning.

Author Contributions

Conceptualization, Y.C. and J.K.; methodology, J.K.; software, Y.C. and J.K.; validation, J.K.; formal analysis, Y.C. and J.K.; investigation, Y.C.; writing—original draft preparation, J.K.; writing—review and editing, Y.C. and J.K.; visualization, Y.C.; supervision, J.K.; project administration, J.K.; funding acquisition, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by Hansung University for Jinmo Kim.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shneiderman Direct manipulation: A step beyond programming languages. Computer 1983, 16, 57–69. [CrossRef] [Green Version]
  2. Lindley, S.E.; Le Couteur, J.; Berthouze, N.L. Stirring up experience through movement in game play: Effects on engagement and social behaviour. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI ’08, Florence, Italy, 5–10 April 2008; ACM: New York, NY, USA, 2008; pp. 511–514. [Google Scholar]
  3. Nesbitt, K.V.; Hoskens, I. Multi-Sensory Game Interface Improves Player Satisfaction but Not Performance. In Proceedings of the Ninth Conference on Australasian User Interface-Volume 76, AUIC ’08, Wollongong, New South Wales, Australia, 22–25 January 2008; Australian Computer Society, Inc.: Darlinghurst, Australia, 2008; pp. 13–18. [Google Scholar]
  4. Vinayagamoorthy, V.; Garau, M.; Steed, A.; Slater, M. An Eye Gaze Model for Dyadic Interaction in an Immersive Virtual Environment: Practice and Experience. Comput. Graph. Forum 2004, 23, 1–11. [Google Scholar] [CrossRef]
  5. Kim, M.; Kim, J.; Jeong, K.; Kim, C. Grasping VR: Presence of Pseudo-Haptic Interface Based Portable Hand Grip System in Immersive Virtual Reality. Int. J. Hum. Comput. Interact. 2020, 36, 685–698. [Google Scholar] [CrossRef]
  6. Kang, T.; Chae, M.; Seo, E.; Kim, M.; Kim, J. DeepHandsVR: Hand Interface Using Deep Learning in Immersive Virtual Reality. Electronics 2020, 9, 1863. [Google Scholar] [CrossRef]
  7. Sebastian, M.; Maximilian, B.; Lukas, W.; Lung-Pan, C.; Floyd, M.F.; Patrick, B. VirtualSpace-Overloading Physical Space with Multiple Virtual Reality Users. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (CHI ’18), Montreal QC, Canada, 21–26 April 2018; ACM: New York, NY, USA, 2018; pp. 241:1–241:10. [Google Scholar]
  8. Lee, J.; Jeong, K.; Kim, J. MAVE: Maze-based immersive virtual environment for new presence and experience. Comput. Anim. Virtual Worlds 2017, 28, e1756. [Google Scholar] [CrossRef]
  9. Joo, H.; Simon, T.; Sheikh, Y. Total Capture: A 3D Deformation Model for Tracking Faces, Hands, and Bodies. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8320–8329. [Google Scholar]
  10. Lacoche, J.; Pallamin, N.; Boggini, T.; Royan, J. Collaborators Awareness for User Cohabitation in Co-located Collaborative Virtual Environments. In Proceedings of the 23rd ACM Symposium on Virtual Reality Software and Technology, VRST ’17, Gothenburg, Sweden, 8–10 November 2017; ACM: New York, NY, USA, 2017; pp. 15:1–15:9. [Google Scholar]
  11. Grandi, J.G.; Debarba, H.G.; Maciel, A. Characterizing Asymmetric Collaborative Interactions in Virtual and Augmented Realities. 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR), Osaka, Japan, 23–27 March 2019; pp. 127–135. [Google Scholar]
  12. Lee, J.; Jeong, K.; Kim, J. RoleVR: Multi-experience in immersive virtual reality between co-located HMD and non-HMD users. Multim. Tools Appl. 2020, 79, 979–1005. [Google Scholar] [CrossRef]
  13. Kim, J. VIVR: Presence of Immersive Interaction for Visual Impairment Virtual Reality. IEEE Access 2020, 8, 196151–196159. [Google Scholar] [CrossRef]
  14. Lele, A. Virtual reality and its military utility. J. Ambient. Intell. Humaniz. Comput. 2013, 4, 17–26. [Google Scholar] [CrossRef]
  15. Alhalabi, W. Virtual reality systems enhance students’ achievements in engineering education. Behav. Inf. Technol. 2016, 35, 919–925. [Google Scholar] [CrossRef]
  16. Doolani, S.; Wessels, C.; Kanal, V.; Sevastopoulos, C.; Jaiswal, A.; Nambiappan, H.; Makedon, F. A Review of Extended Reality (XR) Technologies for Manufacturing Training. Technologies 2020, 8, 77. [Google Scholar] [CrossRef]
  17. Chen, Y.; Zhou, D.; Wang, Y.; Yu, J. Application of Augmented Reality for Early Childhood English Teaching. In Proceedings of the 2017 International Symposium on Educational Technology (ISET), Hong Kong, China, 27–29 June 2017; pp. 111–115. [Google Scholar]
  18. Vernadakis, N.; Andreas, A.; Tsitskari, E.; Zachopoulou, E. The Use of Computer Assisted Instruction in Preschool Education: Making Teaching Meaningful. Early Child. Educ. J. 2005, 33, 99–104. [Google Scholar] [CrossRef]
  19. Yang, J. Mobile Assisted Language Learning: Review of the Recent Applications of Emerging Mobile Technologies. Engl. Lang. Teach. 2013, 6, 19–25. [Google Scholar] [CrossRef]
  20. Colwell, J.; Hutchison, A.C. Supporting Teachers in Integrating Digital Technology Into Language Arts Instruction to Promote Literacy. J. Digit. Learn. Teach. Educ. 2015, 31, 56–63. [Google Scholar] [CrossRef]
  21. Hung, H.; Yang, J.C.; Hwang, G.; Chu, H.; Wang, C. A scoping review of research on digital game-based language learning. Comput. Educ. 2018, 126, 89–104. [Google Scholar] [CrossRef]
  22. Zucker, T.A.; Moody, A.K.; McKenna, M.C. The Effects of Electronic Books on Pre-Kindergarten-to-Grade 5 Students’ Literacy and Language Outcomes: A Research Synthesis. J. Educ. Comput. Res. 2009, 40, 47–87. [Google Scholar] [CrossRef]
  23. Lasse, J.; Flemming, K. A Review of the Use of Virtual Reality Head-Mounted Displays in Education and Training. Educ. Inf. Technol. 2018, 23, 1515–1529. [Google Scholar]
  24. Fan, M.; Antle, A.N.; Warren, J.L. Augmented Reality for Early Language Learning: A Systematic Review of Augmented Reality Application Design, Instructional Strategies, and Evaluation Outcomes. J. Educ. Comput. Res. 2020, 58, 1059–1100. [Google Scholar] [CrossRef]
  25. Li, Y.; Huang, J.; Tian, F.; Wang, H.; Dai, G. Gesture interaction in virtual reality. Virtual Real. Intell. Hardw. 2019, 1, 84–112. [Google Scholar] [CrossRef]
  26. Yunsik, C.; Jiewon, K.; Jaekyung, J.; Jongchan, P.; Mingyu, K.; Jinmo, K. X-person asymmetric interaction in virtual and augmented realities. Comput. Animat. Virtual Worlds 2020. [Google Scholar] [CrossRef]
  27. Yang, S.; Zhang, M. Application of Brain Neural Network in Personalized English Education System. Int. J. Emerg. Technol. Learn. (IJET) 2018, 13, 15–22. [Google Scholar] [CrossRef]
  28. Unity Technologies. Unity Engine. Available online: https://unity3d.com (accessed on 14 July 2021).
  29. Gregory, C.; Saeed, A.; Jonathan, T.; André, V. EMNIST: Extending MNIST to handwritten letters. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 2921–2926. [Google Scholar]
  30. Alejandro, B.; Yago, S.; Pedro, I. A Survey of Handwritten Character Recognition with MNIST and EMNIST. Appl. Sci. 2019, 9, 3169. [Google Scholar]
  31. Mor, S.S.; Solanki, S.; Gupta, S.; Dhingra, S.; Jain, M.; Saxena, R. HANDWRITTEN TEXT RECOGNITION: With Deep Learning and Android. Int. J. Eng. Adv. Technol. (IJEAT) 2019, 8, 819–825. [Google Scholar]
  32. Google Brain. TensorFlow. Available online: https://www.tensorflow.org/ (accessed on 28 July 2021).
  33. Zealous Interactive. Cartoon Animals. Available online: https://assetstore.unity.com/packages/3d/characters/animals/cartoon-animals-56146 (accessed on 1 September 2020).
  34. Manufactura K4. Cartoon Town and Farm. Available online: https://assetstore.unity.com/packages/3d/environments/cartoon-town-and-farm-17254 (accessed on 1 February 2018).
  35. Rizwan Ashraf. Game GUI Buttons. Available online: https://assetstore.unity.com/packages/2d/gui/icons/game-gui-buttons-96277 (accessed on 8 August 2017).
  36. PTC, Inc. Vuforia Engine. Available online: https://developer.vuforia.com (accessed on 27 May 2021).
  37. Lund, A. Measuring Usability with the USE Questionnaire. Usability User Exp. Newsl. Stc Usability Sig 2001, 8, 3–6. [Google Scholar]
  38. Ijsselsteijn, W.A.; de Kort, Y.A.W.; Poels, K. The Game Experience Questionnaire: Development of a self-report measure to assess the psychological impact of digital games. Tech. Univ. Eindh. 2013, 1–9. [Google Scholar]
  39. Duolingo. Duolingo. Available online: https://en.duolingo.com/ (accessed on 26 July 2021).
  40. Hodoo Labs. Hodoo English. Available online: https://hodooenglish.com/ (accessed on 7 July 2021).
Figure 1. Structure of the proposed text interface using deep learning.
Figure 1. Structure of the proposed text interface using deep learning.
Electronics 10 01809 g001
Figure 2. Text interface composition and processing process.
Figure 2. Text interface composition and processing process.
Electronics 10 01809 g002
Figure 3. Examples of EMNIST dataset [30].
Figure 3. Examples of EMNIST dataset [30].
Electronics 10 01809 g003
Figure 4. Recognition and processing process of handwritten text with the Unity3D engine.
Figure 4. Recognition and processing process of handwritten text with the Unity3D engine.
Electronics 10 01809 g004
Figure 5. Composition of scenes in the English language teaching application [33,34,35].
Figure 5. Composition of scenes in the English language teaching application [33,34,35].
Electronics 10 01809 g005
Figure 6. English language teaching process of text-interface-based application [33,34,35].
Figure 6. English language teaching process of text-interface-based application [33,34,35].
Electronics 10 01809 g006
Figure 7. Production results of the mobile platform application [28,33,34,35].
Figure 7. Production results of the mobile platform application [28,33,34,35].
Electronics 10 01809 g007
Figure 8. Production results of the AR application [28,33,34,35,36].
Figure 8. Production results of the AR application [28,33,34,35,36].
Electronics 10 01809 g008
Table 1. Analysis results of satisfaction with the proposed text interface.
Table 1. Analysis results of satisfaction with the proposed text interface.
Proposed Interface
Mean (SD)
usefulness5.863 (0.641)
ease of use5.964 (0.715)
ease of learning6.175 (0.537)
satisfaction6.300 (0.627)
Table 2. User experience analysis results for English language teaching applications using the text interface.
Table 2. User experience analysis results for English language teaching applications using the text interface.
MobileAR
Mean(SD)
comprehensive experience2.876 (0.341)2.989(0.319)
competence3.200 (0.566)3.140 (0.614)
sensory and immersion3.317 (0.491)3.433 (0.436)
flow3.100 (0.553)3.320 (0.360)
tension/annoyance1.200 (0.452)1.400 (0.593)
challenge2.260 (0.614)2.460 (0.514)
negative affect0.875 (0.646)1.075 (0.372)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cho, Y.; Kim, J. Production of Mobile English Language Teaching Application Based on Text Interface Using Deep Learning. Electronics 2021, 10, 1809. https://doi.org/10.3390/electronics10151809

AMA Style

Cho Y, Kim J. Production of Mobile English Language Teaching Application Based on Text Interface Using Deep Learning. Electronics. 2021; 10(15):1809. https://doi.org/10.3390/electronics10151809

Chicago/Turabian Style

Cho, Yunsik, and Jinmo Kim. 2021. "Production of Mobile English Language Teaching Application Based on Text Interface Using Deep Learning" Electronics 10, no. 15: 1809. https://doi.org/10.3390/electronics10151809

APA Style

Cho, Y., & Kim, J. (2021). Production of Mobile English Language Teaching Application Based on Text Interface Using Deep Learning. Electronics, 10(15), 1809. https://doi.org/10.3390/electronics10151809

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop