Next Article in Journal
The Coach–Athlete Relationship and Self-Talk in Turkish Athletes
Next Article in Special Issue
Effect of Flipped Teaching on Cognitive Load Level with Mobile Devices: The Case of a Graphic Design Course
Previous Article in Journal
The Business Model in Sustainability Transitions: A Conceptualization
Previous Article in Special Issue
A Topic-Based Bibliometric Review of Computers in Human Behavior: Contributors, Collaborations, and Research Topics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vocabulary Learning Based on Learner-Generated Pictorial Annotations: Using Big Data as Learning Resources

1
Department of English Language Education, The Education University of Hong Kong, Hong Kong, China
2
Department of Computing and Decision Sciences, Lingnan University, Hong Kong, China
*
Author to whom correspondence should be addressed.
Sustainability 2021, 13(11), 5767; https://doi.org/10.3390/su13115767
Submission received: 4 April 2021 / Revised: 11 May 2021 / Accepted: 19 May 2021 / Published: 21 May 2021

Abstract

:
This research discusses the potential of using big data for vocabulary learning from the perspective of learner-generated pictorial annotations. Pictorial annotations lead to effective vocabulary learning, the creation of which is however challenging and time-consuming. As user-generated annotations promote active learning, and in the big data era, data sources in social media platforms are not only huge but also user-generated, the proposal of using social media data to establish a natural and semantic connection between pictorial annotations and words seems feasible. This research investigated learners’ perceptions of creating pictorial annotations using Google images and social media images, learners’ evaluation of the learner-generated pictorial annotations, and the effectiveness of Google pictorial annotations and social media pictorial annotations in promoting vocabulary learning. A total of 153 undergraduates participated in the research, some of whom created pictorial annotations using Google and social media data, some evaluated the annotations, and some learned the target words with the annotations. The results indicated positive attitudes towards using Google and social media data sets as resources for language enhancement, as well as significant effectiveness of learner-generated Google pictorial annotations and social media pictorial annotations in promoting both initial learning and retention of target words. Specifically, we found that (i) Google images were more appropriate and reliable for pictorial annotations creation, and therefore they achieved better outcomes when learning with the annotations created with Google images than images from social media, and (ii) the participants who created word lists that integrate pictorial annotations were likely to engage in active learning when they selected and organized the verbal and visual information of target words by themselves and actively integrated such information with their prior knowledge.

1. Introduction

Big data is defined as large-volume, high-velocity, and great-variety information assets that require advanced techniques and technologies for data collection, storage, management, and analysis [1,2]. Based on previous initiatives of defining big data, Gandomi and Haider [3] summarized five characteristics of big data, namely volume, velocity, variety, veracity, and value. Specifically, (1) volume refers to the magnitude of data, (2) velocity refers to the rate of data generation and the speed of data processing, (3) variety refers to the structural heterogeneity and complexity of data, (4) veracity refers to the unreliability inherent in some data sources, and (5) value refers to the attribute that large volumes of data accumulate high value despite of the low value density [3]. The attitude of the academic community towards the use of big data in education is positive in general [4]. It is widely agreed that the implementation of big data in education is associated with promising opportunities such as improved quality of academic programs, developed understanding of students’ learning processes, customized teaching and learning for student needs, enhanced learning experiences, reduced dropout rates, and improved learning performance [5]. A number of researchers also believe that big data and learning analytics promise teachers and students a new age of education by assisting educators to collect information in real time, track learning processes, assess learning progresses, predict future performance, spot potential issues, and provide personalized instruction [6,7]. Furthermore, the use of big data in education has been extended to other sub-research areas. For example, the virtual reality (VR) and augmented reality (AR) techniques have been widely employed for providing immersive and interactive learning experiences for science education and language learning [8,9,10]. As suggested by Pikhart and Klímová [11], the recent advancement of artificial intelligence (AI) and big data techniques will facilitate the rapid transformation of e-learning systems and applications to a new era called eLearning 4.0.
In the field of language education and research, the emerging interest in big data and learning analytics has also led to great opportunities, one of which is personalization in language learning [4]. The other is the use of big data as learning resources for language education, for example, the use of videos, songs, movies, and social media. Social media is regarded as a platform to transform teaching and learning practices [12]. It allows students to learn regardless of time and space and integrates formal and informal learning [13]. It is the most attractive platform to acquire knowledge and information for Generation Z (i.e., learners born after the year 2000) [11]. It also serves as a learning tool for information sharing and collaborative learning [14]. Studies have examined the diverse use of social media tools in education aspects. It was found that social media platforms were effective learning tools for improving English for EFL teachers and students at Hong Kong secondary schools [15]. Similarly, Lai and Tai [16] found common use of self-initiated social media activities in language education, which can improve the learning motivation and language proficiency.
Many applications of big data in language education are associated with multimedia learning, and a considerable proportion of the practices of integrating multimedia into language learning resources is associated with annotations for vocabulary learning. The literature also indicates that multimedia annotations are very effective in promoting vocabulary learning [17]. In this study, we focus on pictorial annotations, which is a kind of multimedia annotations, and refer to the use of images from different sources (e.g., Google image search or social media platforms) for interpreting and depicting semantic meanings, connotations, and denotations of the target words. However, few practices and investigations of pictorial annotations for vocabulary learning involve students’ proactive participation in the generation of pictorial annotations, although the practices of keeping word lists with textual annotations are common among language learners [18]. This indicates an inconsistency among the common practices of student-generated word lists with textual annotations and the widely acknowledged advantages of pictorial annotations. That is, on the one hand, the academic community emphasizes the effectiveness of pictorial annotations in promoting vocabulary learning and the importance of active learning; on the other hand, most language learners create their own wordlists with textual annotations, rather than pictorial annotations. We therefore asked learners to create pictorial annotations and investigated the effectiveness of learning with the pictorial annotations in the present project.
Specifically, we asked learners to search for images from Google images and social media for the annotation creation, as these two datasets are the two main sources of images. We asked two groups of learners to create pictorial annotations using these two image sources and another group of learners to evaluate these annotations. In this way, we analyzed students’ perceptions of creating pictorial annotations with the two image sources and their evaluations of these two types of pictorial annotations. To investigate the effectiveness of the two types of pictorial annotations in promoting vocabulary learning, we also asked two groups of learners to learn with the two types of pictorial annotations, with a baseline of the vocabulary learning performance of a group of learners who learned with textual annotations. Thus, we investigated the application of big data in language education from the qualitative dimension in terms of learners’ perceptions of using big data in creation of pictorial annotations and from the quantitative dimension in terms of learners’ performance of vocabulary learning using pictorial annotations created with big data in this research. Our research questions are listed as follows.
  • Which source of big data do students prefer for pictorial annotation creation, Google images or images from social media? What may be the possible explanations for such results?
  • Which type of pictorial annotations do students highly rate, the annotations created with Google images or those with images from social media? What may be the possible explanations for such results?
  • Which type of pictorial annotations promote better vocabulary learning, the annotations created with Google images or those with images from social media? What may be the possible explanations for such results?

2. Literature Review

2.1. Big Data in Language Learning

Big data has been commonly used to refer to the extremely rapid growth of data volume on the Web, although there is no universal and consistent definition of the term in the data science communities [19,20]. From a technological perspective, the main characteristics of big data are V3 (i.e., high Velocity, high Volume, and high Variety) as big data is “driven in part by social media and the Internet of Things (IoT) phenomenon” [21] (p. 293). Specifically, social media platforms such as YouTube, Twitter, and Facebook create a paradigm of user-generated data, while IoT technologies such as wireless communications, 5G technologies, and GPS have established the infrastructure for empowering the data access, generation, and organization. How do big data grow? “2.5 quintillion bytes of data are created and 90 percent of the data in the world today were produced within the past two years” [22] (p. 97). More recently, the value of big data has been acknowledged and further applied to various domains including healthcare [23], manufacturing [24], digital media [25], the e-learning environment, [26] and so on.
The research on big data in language education can be mainly categorized into two classes. One category of research studies focuses on learning analytics for language learners, the data of which include learning behavioral data, learning logs, learner profiles, learning preferences, performance, and so on. For example, Hsiao, Lan, Kao, and Li [27] visualized the learning paths by collecting the data of learning processes from college students who were studying the Mandarin vocabularies in the virtual world. Zou and Xie [28] have developed a personalized vocabulary learning system for facilitating the recommendations of learning tasks through a learner profile established according to the technique feature analysis theory. Another category of research studies aims to integrate big data platforms and/or big data sources into the language learning processes. For example, Liu and Lan [29] examined the interaction patterns of negotiation on Facebook between EFL learners and native speakers. Xie, Zou, Lau, Wang, and Wong [30] identified the preferred topics of language learners in their social media platforms for generating personalized vocabulary tasks. Pikhart and Klímová [11] studied eLearning 4.0, which utilized the state-of-the-art big data and AI techniques, and used this new paradigm in second language acquisition for Generation Z language learners.

2.2. Google and Social Media

Google and social media platforms are quite different in terms of information creation and retrieval. Google employs a crawler to collect information sources from millions of websites, organizes information sources in the database, and extracts key features for establishing indices for keyword searches. Essentially, the core idea of a Google search algorithm like PageRank is to identify the most authoritative information sources for responding to an issued query [31]. In contrast, social media platforms allow users to create, share, and disseminate their own content, including individual opinions, life stories, and so on. The information sources at the platforms are personalized and subjective [32]. Therefore, the information retrieval in the social media platforms aims to access the highly personalized and interested information [33,34]. Due to their differences in information retrieval and creation, we investigate and compare the impact of the use of images from Google and social media platforms as pictorial annotations for vocabulary learning in this study.

2.3. Multimedia Annotations for Vocabulary Learning

The importance of annotations (or glosses) for effective vocabulary uptake from reading has been widely acknowledged in the research community of second language vocabulary acquisition [35,36]. With the advent of technology-enhanced language learning, annotations have also developed from the traditional format, which involves only short textual definitions that aim to facilitate reading comprehension, to multimedia formats that present various aspects of word knowledge through “different modalities (textual, visual, and auditory) and modes (video, picture, and text)” [37] (p. 136). Researchers in general argue that pictorial annotations, compared to textual annotations, are more effective in enhancing reading comprehension, promoting students’ uptake of word meanings, and meeting learner needs and preferences [38,39,40,41]. Specifically, Chun and Plass [42] found that it was easier to learn words associated with images or actual objects than words without. Jones and Plass [43] also reported that dual annotations, which involve both text and images, were significantly more effective than single annotations that involve only text or images. In addition to the two basic types of annotations, textual only and textual plus pictorial annotations, Yeh and Wang [44] further investigated textual plus pictorial plus audio annotations and compared these three annotation types, the results of which showed that textual plus multimedia annotations were the most effective. Similarly, Turk and Ercetin [45] emphasized that presenting visual and verbal information simultaneously led to better learning than allowing students to select the types of annotations they want. Interestingly, studies that compared the annotations that integrate static pictures to the annotations that involve dynamic animations revealed controversial findings. Ikeda [46] and Lin and Chen [47] found annotations with dynamic pictures more effective, yet Ariew and Ercetin [48] argued the opposite. Additionally, Boers et al. [17] found that pictorial annotations induced more attention from learners to target words and hence promoted better learning.
The large number of studies on the benefits of multimedia (text plus static pictures) annotations also induces numerous attempts that aim to explain the effectiveness of multimedia annotations, such as Mayer’s generative theory of multimedia learning [49,50,51,52], Schmidt’s noticing hypothesis [53,54,55], and Craik and Lockhart’s levels of processing theory [56], three of the most frequently referred to theoretical frameworks. Although many studies have been conducted to investigate the pictorial annotations provided by teachers or material designers, little is known about learner-generated pictorial annotations. Additionally, language learners are often encouraged to keep notes of unfamiliar words that they encounter while reading and go over them periodically to obtain a strong and durable memory trace [18]. As one of the most widely adopted word learning strategies, such a practice of creating personalized word lists, however, does not normally involve multimodality. Word lists are usually only comprised of the target words and simplified definitions, with sample phrases or sentences being included sometimes. Few learners try to, or are advised to, integrate images into their word lists, even though the effects of pictorial annotations on vocabulary acquisition have been widely acknowledged. There is therefore a call for research that examines learner-generated word lists that integrate pictorial annotations.

3. Method

3.1. Research Design

Motivated by previous research studies on big data in language learning, Google and social media, and multimedia annotations for vocabulary learning as introduced in Section 2, we can find that theories about big data, language acquisition, and multimedia learning form the methodological background of this study. Moreover, as many studies in Section 2 used the mixed research method, which integrated the quantitative and qualitative data, this method was employed in this study.
A total of 153 undergraduates, who were speakers of Chinese and learned English as a foreign language, participated in the research. They were upper intermediate learners according to the common European framework of reference for languages. These students were randomly assigned into five groups and completed different learning tasks. Their ages ranged from 18 to 22 and they had learned English for around 8 years on average. They were all non-English majors in a local university in Hong Kong. Among these participants, 90 of them were female students and were equally and randomly assigned to the five groups. The experiment was conducted at a local university in Hong Kong in 2019.
The 30 students in Group 1 (the annotation-creation group) created two sets of multimedia-annotations for 10 target words, one set using images from social media data, and the other Google data. The target words included burglarize, grin, inflammation, ostensible, procrastination, rake, shatter, shiver, tumble, and wrath. These words are all tangible to the senses and can be easily imagined. Moreover, according to Davis’s Corpus of Contemporary American English (COCA) [57], these words are out of the 6000 most frequently used words; thus, they are unlikely to be unknown to the participants who are intermediate English learners and know approximately 3000 words. The textual annotations for the words were provided in the worksheets, and the participants searched for appropriate and accurate images that depicted semantic meanings of the words from Google images and three social media platforms including Twitter, Instagram, and Facebook. In sum, 600 learner-created pictorial annotations were created, with each participant generating 10 annotations for the 10 target words using Google data and 10 annotations using social media data. These participants were interviewed about their experiences of annotation creation and image search afterwards. Guided questions included whether it was easy to search for images that appropriately depicted target word meanings, and whether the images presented by the data sources were relevant, reliable, appropriate, and interesting. The participants were also encouraged to share their overall perceptions, feelings, and experiences freely. The interviews were audiotaped to record all details.
To examine to what extent the images searched from Google Images and social media depict the meanings of the target words vividly, a group of 30 participants evaluated the 600 learner-created pictorial annotations and rated them from 1 (very inappropriate and inaccurate) to 5 (very appropriate and accurate). These participants in Group 2 (the annotation-evaluation group) were trained to evaluate the appropriateness and accuracy of pictorial annotations before conducting the evaluation by two experts in English language education. They were also interviewed for the investigation of their evaluation of the pictorial annotations.
Based on the scores of the leaner-generated pictorial annotations, as given by the participants in Group 2, the two experts selected two sets of pictorial annotations for the 10 target words. One set of pictorial annotations were created using Google images, and the other using social media images. The selection criterion is that the set of pictorial annotations had the highest score from the two perspectives, with appropriateness and accuracy in depicting the meanings of the target words.
To evaluate the effectiveness of these two sets of pictorial annotations in promoting vocabulary learning, two groups of participants were asked to learn with the annotations, and their learning outcomes were measured. There were 31 participants in Group 3 (the experiment group who learned with Google pictorial annotations) and 32 in Group 4 (the experiment group who learned with social media pictorial annotations).
As a control group, another group of 30 students learned with simple textual annotations, which were the same as those provided to the participants in Group 1 and used by those in Group 3 and 4. All three groups of participants learned the target words by reading a text in which the 10 target words were integrated and learned the words with associated annotations. The annotations for Group 5 (the control group) were simple textual annotations, those for Group 3 were Google pictorial annotations, and those for Group 4 were social media pictorial annotations.
The participants decided how they would like to search and select the images, and the search and selection procedures were fully dependent on their preferences. They were encouraged to freely search images by issuing the queries to Google (or the search portals in the social media platforms, depending on their assigned groups), and then select the images that were best to match and express the meaning of a target word.
These three groups of participants’ prior knowledge of the target words was measured through a pre-test before the intervention, and their initial learning and retention of the words were measured through an immediate post-test right after the intervention, and a delayed post-test 1 week later. Additionally, these participants were interviewed about their experiences of learning the target words with the annotations.

3.2. Assessment and Scoring

We conducted a pre-test to evaluate the participants’ prior knowledge of the target words and immediate and delayed post-tests to examine the development of their word knowledge. These participants were from Group 3 (the experiment group who learned with Google pictorial annotations), Group 4 (the experiment group who learned with social media pictorial annotations), and Group 5 (the control group who learned with textual annotations). To evaluate their development of receptive and productive knowledge of the meanings of the target words after learning with the pictorial annotations, we applied Folse’s modified version [58] of Paribakht and Wesche’s vocabulary knowledge scale [59] as the assessment tool and adopted Zou’s scoring system [60,61]. A meaning was graded zero if it is incorrect and a full score if it is correct. A sentence was graded zero if it is used semantically incorrect, a half score if it is semantically correct but grammatically incorrect (e.g., ‘Don’t procrastination. It’s harmful’.), and a full score if it is semantically and grammatically correct (e.g., ‘Procrastination can lead to feelings of guilt’). Two trained raters scored the answers independently, and the Pearson’s r for the inter-rater reliability was 0.99 for the pre-test, 0.93 for the immediate post-test, and 0.95 for the delayed post-test.

3.3. Collection and Analysis of Interview Data

We collected three types of interview data in this research. Firstly, we interviewed the participants in Group 1 (the annotation-creation group) to collect data of their experiences of annotation creation and image search. The guided questions for the interview focused on what the participants thought about the idea of using Google data and social media data as resources for language enhancement, what they felt about searching for images from Google data and social media data, and how they perceived the images as presented by Google images and social media.
We also collected data of the participants’ evaluation of the learner-generated pictorial annotations created with social media data and Google data through interviewing the participants in Group 2 (the annotation-evaluation group) and the two experts in English language education. The guided questions for the interview focused on the general appropriateness and accuracy of the images from Google data and social media data for depicting the meanings of the target words.
Moreover, we collected data of the participants’ experiences of learning the target words with different types of annotations through interviews after the delayed post-test. Nine participants from Group 3 (the experiment group who learned with Google pictorial annotations), Group 4 (the experiment group who learned with social media pictorial annotations), and Group 5 (the control group who learned with textual annotations) were interviewed, with three from each group. The guided questions for the participants in the two experimental groups focused on whether the pictorial annotations facilitated their learning, and how they felt about learning with the images from Google data and social media data.
The sample guided questions are listed in Appendix A. We transcribed the interview data verbatim first and then analyzed them to understand features of Google data and social media data. We then conducted text analysis for the transcribed texts by finding evidences the common perceptions, feelings, and opinions from the participants about learning procedures. Following Zou’s previous study [58], all data were skimmed first to note distinct features, focusing on which relevant literature were reviewed. This process is cyclical, throughout which we repeatedly re-examine the data and literature.

4. Results

4.1. Perceptions of Google Data and Social Media Data for Multimedia-Annotations Creation

The interview results of the 30 students in Group 1 (the annotation-creation group) showed five main perceptions of the participants concerning the use of Google data and social media data for multimedia-annotations creation (see Table 1). Firstly, over 90% of participants considered the idea of using Google and social media data as resources for language enhancement interesting and creative, and secondly, over 80% of them considered it feasible and reliable. Other positive comments concerning the use of Google and social media data for language education, which appeared very frequently in the interview transcripts, included “I like this way of learning” and “it’s awesome”. One participant also suggested, “Learning should be integrated into life, and if social media can be well used for learning, it will be very helpful, especially for learners with low motivation”. Such results indicated that the participants were aware of the potential benefits of using big data for language education.
Moreover, the participants generally agreed that it was easy to create pictorial annotations using Google and social media data. Specifically, almost all participants found it easy to create pictorial annotations through searching for images that appropriately depict target word meanings from Google data, while a lower percentage of them felt it easy to do so using social media data (slightly below 70%). Furthermore, over 85% of participants considered the images presented by Google images relevant to the keywords that they entered for search, but lower than 65% of participants felt so while searching for images from social media data. Additionally, over 75% of participants considered Google data appropriate, yet only 40% of the participants described social media data as such. One participant commented, “I can see clear differences between Google images and social media images; Google is more reliable, and social media is very personal”. These results indicated that overall, the participants showed positive attitudes towards developing language learning resources using Google and social media data, but they considered Google data more appropriate, relevant, and easy to use than social media data.

4.2. Scores of Learner-Generated Google Pictorial Annotations and Social Media Pictorial Annotations

The descriptive statistics for the scores of learner-generated annotations, as displayed in Table 2, demonstrated that the mean score of Google pictorial annotations (M = 4.21, SD = 0.67) was higher than that of social media pictorial annotations (M = 3.42, SD = 0.70). To further test whether Google pictorial annotations was significantly better than social media pictorial annotations, an independent samples test was applied, the results of which indicated statistically significant differences (t = 14.04, df = 598, p < 0.001, r = 0.49), as presented in Table 3. Such results indicated that the participants considered the learner-generated Google pictorial annotations much more appropriate and accurate than social media pictorial annotations.
It is also noteworthy that the interview results of the 30 students in Group 1 (the annotation-creation group), the 30 participants in Group 2 (the annotation-evaluation group), and the two experts in English language education all indicated differences between the annotations for the six verbs and those for the three nouns and one adjective. This is perhaps because the meanings of the verbs (i.e., burglarize, grin, rake, shatter, shiver, and tumble) are more concrete than those of the nouns and adjectives (i.e., inflammation, procrastination, wrath, and ostensible) in this research. To test this hypothesis, we examined the scores of the learner-generated annotations for the verbs and the nouns and adjective. The results, as shown in Table 4, showed great differences, and the scores of the annotations for the verbs were much higher (M = 4.19, SD = 0.67) than those of the annotations for the nouns and adjective (M = 3.25, SD = 0.62). As a follow-up test of whether significant differences existed between the scores of the annotations for verbs and those for the nouns and adjective, we conducted another independent samples t-test, the results of which showed statistical significance (t = 17.22, df = 598, p < 0.001, r = 0.58) (see Table 5). That is, the participants considered the learner-generated annotations for words with concrete meanings much more appropriate and accurate than those for words with abstract meanings.

4.3. Learning Performance of the Participants Who Learned with Textual Annotations, Google Pictorial Annotations, and Social Media Pictorial Annotations

The descriptive statistics, as shown in Table 6, demonstrate three main findings. Firstly, all three types of annotations were conducive to effective learning of the target words from both perspectives of immediate learning (the means of the immediate post-tests were 8.34, 7.23, and 6.09, respectively) and delayed retention (the means of the delayed post-tests were 6.84, 5.50, and 3.93, respectively), considering that the participants’ pre-knowledge of the words was close to zero. Secondly, the participants who learned with pictorial annotations achieved better learning outcomes than those who learned with textual annotations, as indicated by their scores on the immediate and delayed post-tests. Thirdly, the participants who learned with Google pictorial annotations had the highest scores in both immediate and delayed post-tests.
Further, we conducted two one-way ANOVA tests, which were widely used in other similar studies [51,52] to examine whether any significant differences existed among the effectiveness of the three annotation types in promoting vocabulary learning. The data met all assumptions for performing ANOVA.
Concerning the participants’ scores in the immediate post-test, the significance values of the tests of normality of the data were 0.07, 0.053, and 0.07, respectively, which were greater than 0.05 and indicated that the data were normally distributed. The data also met the requirements of homoscedasticity, homogeneity of variances and regression slopes, etc. As presented in Table 7, the results of the one-way ANOVA test of the participants’ scores in the immediate post-test showed statistically significant differences among the three groups (F = 13.07, p < 0.001). Table 8 also shows that Google pictorial annotations promoted significantly more effective vocabulary learning than both social media pictorial annotations (p = 0.02) and textual annotations (p < 0.001). The differences between the effectiveness of social media pictorial annotations and that of textual annotations were also statistically significant (p = 0.04). In sum, the mean differences among the three groups were significant at the 0.05 level.
Concerning the participants’ scores in the delayed post-test, the significance values of the tests of normality of the data were 0.058, 0.06, and 0.12, respectively, indicating that the data were normally distributed. The data also met the requirements of homoscedasticity, homogeneity of variances and regression slopes, etc. As presented in Table 9, significant differences were identified among the three groups (F = 15.32, p < 0.001). The results shown in Table 10 also suggested that after 1 week, the participants who learned with Google pictorial annotations still performed significantly better than those who learned with social media pictorial annotations (p = 0.03) and textual annotations (p < 0.001). In addition, the participants who learned with social media pictorial annotations performed significantly better than those who learned with textual annotations (p = 0.01). In sum, the mean differences among the three groups were significant at the 0.05 level.
The participants in Group 3 (the experiment group who learned with Google pictorial annotations) and Group 4 (the experiment group who learned with social media pictorial annotations) also reported in the interview that the Google and social media images were conducive to their learning of the target words.
In summary, the results of this research indicated that the participants considered the idea of using Google and social media data sets as resources for language enhancement interesting and creative. They also felt that Google data were more appropriate, relevant, and easy to use than the social media data. The scores of the learner-generated Google pictorial annotations and those of the social media pictorial annotations, as given by the participants in the annotation-evaluation group, also showed that the participants considered the learner-generated Google pictorial annotations much more appropriate and accurate than social media pictorial annotations. Moreover, the participants’ learning performance indicated that Google pictorial annotations were significantly more effective than social media pictorial annotations in promoting learners’ initial learning and retention of the target words.

5. Discussion

5.1. Perceptions of Using Google Data and Social Media Data for Learning Resource Creation

We found that the participants in the annotation-creation group considered Google data more appropriate, relevant, and easy to use than social media data in this research. The participants in the annotation-evaluation group also rated the annotations created with Google images higher, compared to those created with social media images. A possible reason for such results is that social media data are more subjective. Typical social media platforms (e.g., Facebook, Instagram, Twitter, blogs, and YouTube) are all designed to allow people to share their personal feelings, ideas, and life experiences, so the social media data are closely associated with the context in which the users create them, the correct interpretation of which requires comprehensive consideration of the many factors that are related to it. De-contextualization of social media images tends to result in loss of meanings or misinterpretation, as the situations, events, feelings, or information related to the images all contribute to the expression of particular meanings in certain contexts. Thus, when social media images are used in other contexts that are different from the original contexts where the images are created, the information may not be conveyed accurately. However, this is common when social media data are used for learning resource creation. Moreover, different users may interpret an image from varied perspectives. The same image may seem appropriate to some learners for the expression of a word’s meaning, but inappropriate to others. For example, some participants considered the following pictorial annotations, which were created by a participant using an image from Twitter in our research, to be very accurate and appropriate, yet some felt they were very inaccurate and inappropriate (see Figure 1).
Compared to the social media images, the participants showed more positive attitudes towards Google images, considering them less subjective and more reliable. This is likely because the Google ranking systems consist of a whole series of algorithms, through which they can sort billions of data to identify the most relevant results, with factors such as webpage relevance and usability, source expertise, content quality, context, and setting being taken into account [59]. Thus, Google images tend to be more independent from the contexts of the web pages from where the images originate compared to social media images. The participants therefore felt that Google images were more appropriate and reliable for pictorial annotations creation, and they achieved better outcomes when learning with the annotations created with Google images.
Another noteworthy finding is that the learner-generated annotations for words with concrete meanings were much more highly rated than the annotations for words with abstract meanings. This indicates that it is easier to create annotations for words with concrete meanings than those with abstract concepts, and the multimedia-annotation-enhanced vocabulary learning approach is more appropriate for the learning of words with concrete meanings. We also noticed in this research that some participants tried to concretize words of abstract concepts through additional self-generated explanations in the textual glosses, an example of which is presented in Figure 2. This facilitates learners’ contextualization of the abstract concepts and comprehension of the word meanings to some extent.

5.2. Effectiveness of Google Pictorial Annotations and Social Media Pictorial Annotations in Promoting Language Learning

The results of this research showed that pictorial annotations were significantly more effective than textual annotations in promoting learners’ initial learning and retention of the target words. This aligns with the multimedia principle, which argues that people can learn more deeply from text and images than from text alone [60]. It is also found that Google pictorial annotations were significantly more effective than social media pictorial annotations. This is likely because the Google pictorial annotations can better assist learners to understand the meanings of the target words, as Google images depend less on the contexts of the web pages from where the images originate. Such findings indicate that the effectiveness of pictorial annotations in promoting learning are associated with the quality of the annotations, one indicator of which is the extent to which the images can depict meanings of the target words accurately and appropriately. Given the difficulty of depicting words of abstract concepts, the strategy of concretizing these words through additional self-generated contextualized explanations in the textual glosses seems feasible. Many participants in this research did apply this strategy in annotation creation and found it helpful. The participants who evaluated the learner-generated annotations also commented that the annotations, which are associated with contextualized explanations of the images, could convey meanings of the target words more accurately and seemed more conducive to effective learning.

5.3. Potential of Using Big Data as Language Learning Resources

The learners were aware of the potential of using big data for language education in general. Most participants in this research considered the idea of using Google and social media data sets as resources for language enhancement interesting, creative, and feasible. Many of them also mentioned in the interview that images could facilitate them to visualize the target words and learn better. Moreover, many learners have the habit of creating their own word lists, the words in which are normally the unfamiliar words that they have encountered in reading [18]. Therefore, it seems feasible to ask learners to create multimedia word lists rather than textual word lists, taking into account the results of this research, which indicate the participants’ interests in using Google and social media images to create pictorial annotations.
Additionally, active word learning happens when learners proactively connect verbal and visual information of target vocabulary with their prior knowledge and create their own lexical associations [49,50,62]. The participants who created word lists that integrate pictorial annotations were likely to engage in active learning when they selected and organized the verbal and visual information of target words by themselves and actively integrated such information with their prior knowledge. These features are unique for the learning activities that involve learner-generated pictorial annotations and play an important role in promoting effective learning. The research results also showed that the participants repeatedly evaluated what images better matched the target words, and no one mentioned anything like heavy cognitive load or unaffordable mental efforts. Thus, the potential of using big data as language learning resources is significant.

5.4. Challenges of Using Big Data as Language Learning Resources

The quality of big data is unstable, leading to challenges of effective use of big data for high quality education. The results of this research showed that the scores of Google pictorial annotations were higher than social media pictorial annotations, and the learners who learned with Google pictorial annotations achieved better learning outcomes than those who learned with social media pictorial annotations, indicating that the quality of learning resources has influence on the learning outcomes. Moreover, the interview results of this research showed that the participants were aware of the possible influences of the data quality on learning outcomes. Many participants reported concerns that their annotations may not be professional enough to be qualified learning resources. Thus, because of the unstable quality of big data and importance of the quality of learning resources, quality assurance is essential for the use of big data as language learning resources. However, it is difficult, if not impossible, to control the quality of user-created big data, and it is consequently challenging to assure the quality of learner-created pictorial annotations. The underlying reason is that, as discussed in Section 2.2, Google is used to identify the most authoritative information sources for responding to an issued query, whereas social media platforms mainly access and share the highly personalized and interested information.

6. Conclusions

We investigated the use of big data as learning resources for language education in this research from three perspectives—learners’ perceptions of creating pictorial annotations using Google images and social media images, learners’ evaluation of the learner-generated pictorial annotations, and the effectiveness of Google pictorial annotations and social media pictorial annotations in promoting vocabulary learning. The results indicated positive attitudes towards using Google and social media data sets as resources for language enhancement, as well as significant effectiveness of learner-generated Google pictorial annotations and social media pictorial annotations in promoting both initial learning and retention of target words. Specifically, we have found that (i) Google images were more appropriate and reliable for pictorial annotations creation, and therefore they achieved better learning outcomes than annotations created with images from social media, and (ii) the participants who created word lists that integrate pictorial annotations were likely to engage in active learning when they selected and organized the verbal and visual information of target words by themselves and actively integrated such information with their prior knowledge.
One important implication of this study is that a computer-assisted tool is necessary to help teachers and learners identify the relevant resources according to their preferences for creating annotations for vocabulary as there has been many irrelevant and noisy resources provided to the learners by Google and social media platforms. Another implication is that the learner engagement is an essential factor for developing learning tasks with big data. The learner-generated pictorial annotation is a good example for engaging learners, which involves a deep level of processing target words by searching, filtering, comparing, and ranking candidate pictures.
This research is limited in that it only investigated the use of big data for vocabulary learning, rather than the general development of language skills. However, vocabulary knowledge is the fundamental element of language knowledge, so the effectiveness of big data for vocabulary knowledge development can, to some extent, indicate the potential of the effectiveness of big data for language enhancement. Moreover, this research focused on the use of big data for the creation of pictorial annotations, while big data can be applied in many other language learning areas through a wide range of approaches; thus, more in-depth understanding of the potential of big data for language education ought to involve a larger number of studies from more varied viewpoints. Moreover, the current study does not provide a feasible solution for improving data quality of social media pictorial annotations. It is feasible to define a conceptual framework for both learners and teachers for selecting high-quality learning resources. In addition, the adoption of hardware devices including mobile phones [63], wearable devices [64] and so on can be an important factor for further investigation. Future research is therefore advised to be conducted in such directions.

Author Contributions

Conceptualization, D.Z. and H.X.; methodology, D.Z. and H.X.; software, H.X.; validation, D.Z. and H.X.; formal analysis, D.Z.; investigation, D.Z.; resources, H.X.; data curation, D.Z.; writing—original draft preparation, D.Z.; writing—review and editing, D.Z. and H.X.; visualization, D.Z. and H.X.; supervision, D.Z. and H.X.; project administration, H.X.; funding acquisition, H.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Lam Woo Research Fund (LWI20011) at Lingnan University, Hong Kong, and the APC was funded by Lam Woo Research Fund (LWI20011) at Lingnan University, Hong Kong.

Institutional Review Board Statement

The work had passed relevant ethical review processes, and the code of ethics and the privacy rights of human subjects were observed throughout the project. The data contains no personal, or personally identifiable information.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

Di Zou’s work is supported by the Internal Research Grant (RG15/20-21R), The Education University of Hong Kong. Haoran Xie’s work is supported by the Lam Woo Research Fund (LWI20011) of Lingnan University, Hong Kong.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Sample guided questions in the open-end interview:
  • To what extent do you agree that it is easy to search for images that can appropriately depict the target word meanings?
  • To what extent do you agree that the images presented by the data sources are relevant, reliable, appropriate, and interesting?
  • To what extent do you agree that the images from Google data and social media data can depict the meanings of the target words accurately and properly?
  • To what extent do you agree that the pictorial annotations can facilitate your learning of the target vocabulary?
  • What factors or features related to the pictorial annotations do you find useful for your learning of the target vocabulary?
  • What are your feelings, perceptions, or experiences about the learning process?
  • To what extent are you satisfied with the learning approach?
  • Do you have any other comments on the learning process and approach?

References

  1. Chen, H.; Chiang, R.H.; Storey, V.C. Business Intelligence and Analytics: From Big Data to Big Impact. MIS Q. 2012, 36. [Google Scholar] [CrossRef]
  2. Kwon, O.; Lee, N.; Shin, B. Data Quality Management, Data Usage Experience and Acquisition Intention of Big Data Analytics. Int. J. Inf. Manag. 2014, 34, 387–394. [Google Scholar] [CrossRef]
  3. Gandomi, A.; Haider, M. Beyond the Hype: Big Data Concepts, Methods, and Analytics. Int. J. Inf. Manag. 2015, 35, 137–144. [Google Scholar] [CrossRef] [Green Version]
  4. Godwin-Jones, R. Scaling Up and Zooming in: Big Data and Personalization in Language Learning. Lang. Learn. Technol. 2017, 21, 4–15. [Google Scholar]
  5. Daniel, B.K. Big Data in Higher Education: The Big Picture. In Big Data and Learning Analytics in Higher Education; Springer: Cham, Switzerland, 2017; pp. 19–28. [Google Scholar] [CrossRef]
  6. Cope, B.; Kalantzis, M. Big Data Comes to School: Implications for Learning, Assessment, and Research. AERA Open 2016, 2, 1–19. [Google Scholar] [CrossRef] [Green Version]
  7. Tulasi, M.S. Advanced Analysis of Large Scale Data Using Big Data Technologies. 2018. Available online: http://hdl.handle.net/10603/239753 (accessed on 9 November 2019).
  8. Jong, M.S.Y.; Tsai, C.C.; Xie, H.; Wong, F.K.-K. Integrating Interactive Learner-Immersed Video-Based Virtual Reality into Learning and Teaching of Physical Geography. Br. J. Educ. Technol. 2020, 51, 2064–2079. [Google Scholar] [CrossRef]
  9. Chen, M.P.; Wang, L.C.; Zou, D.; Lin, S.Y.; Xie, H.; Tsai, C.C. Effects of Captions and English Proficiency on Learning Effectiveness, Motivation and Attitude in Augmented-Reality-Enhanced Theme-Based Contextualized EFL Learning. Comput. Assist. Lang. Learn. 2020, 1–31. [Google Scholar] [CrossRef]
  10. Huang, X.; Zou, D.; Cheng, G.; Xie, H. A Systematic Review of AR and VR Enhanced Language Learning. Sustainability 2021, 13, 4639. [Google Scholar] [CrossRef]
  11. Pikhart, M.; Klímová, B. eLearning 4.0 as a Sustainability Strategy for Generation Z Language Learners: Applied Linguistics of Second Language Acquisition in Younger Adults. Societies 2020, 10, 38. [Google Scholar] [CrossRef]
  12. Zhang, R.; Zou, D. Types, Purposes, And Effectiveness of State-Of-The-Art Technologies for Second and Foreign Language Learning. Comput. Assist. Lang. Learn. 2020. [Google Scholar] [CrossRef]
  13. Greenhow, C.; Robelia, E. Informal Learning and Identity Formation in Online Social Networks. Learn. Media Technol. 2009, 34, 119–140. [Google Scholar] [CrossRef]
  14. Su, F.; Zou, D. Technology-Enhanced Collaborative Language Learning: Theoretical Foundations, Technologies, and Implications. Comput. Assist. Lang. Learn. 2020. [Google Scholar] [CrossRef]
  15. Tso, A.W.B. Learning English as a Foreign Language through Social Media: Perspectives from Hong Kong Adolescents. In Shaping the Future of Education, Communication and Technology; Springer: Singapore, 2019; pp. 107–115. [Google Scholar]
  16. Lai, C.; Tai, C.P. Types of Social Media Activities and Hong Kong South and Southeast Asians Youth’s Chinese Language Learning Motivation. System 2021, 97, 102432. [Google Scholar] [CrossRef]
  17. Boers, F.; Warren, P.; Grimshaw, G.; Siyanova-Chanturia, A. On the Benefits of Multimodal Annotations for Vocabulary Uptake from Reading. Comput. Assist. Lang. Learn. 2017, 30, 709–725. [Google Scholar] [CrossRef] [Green Version]
  18. Hulstijn, J.H. Intentional and Incidental Second-Language Vocabulary Learning: A Reappraisal of Elaboration, Rehearsal and Automaticity. In Cognition and Second Language Instruction; Robinson, P., Ed.; Cambridge University Press: Cambridge, United Kingdom, 2001; pp. 258–286. [Google Scholar]
  19. Lynch, C. Big Data: How Do Your Data Grow? Nature 2008, 455, 28. [Google Scholar] [CrossRef]
  20. Lohr, S. The Age of Big Data. New York Times. 2012. Available online: http://euler.mat.ufrgs.br/~viali/estatistica/estatistica/relacionados/08_Big%20Data%20Impact%20in%20the%20World.pdf (accessed on 9 November 2019).
  21. Lee, I. Big Data: Dimensions, Evolution, Impacts, and Challenges. Bus. Horiz. 2017, 60, 293–303. [Google Scholar] [CrossRef]
  22. Wu, X.; Zhu, X.; Wu, G.Q.; Ding, W. Data Mining with Big Data. IEEE Trans. Knowl. Data Eng. 2013, 26, 97–107. [Google Scholar] [CrossRef]
  23. Groves, P.; Kayyali, B.; Knott, D.; Kuiken, S.V. The “Big Data” Revolution in Healthcare: Accelerating Value and Innovation. 2016. Available online: hdl.handle.net/11146/465 (accessed on 9 November 2019).
  24. Opresnik, D.; Taisch, M. The Value of Big Data in Servitization. Int. J. Prod. Econ. 2015, 165, 174–184. [Google Scholar] [CrossRef]
  25. Mahrt, M.; Scharkow, M. The Value of Big Data in Digital Media Research. J. Broadcasting Electron. Media 2013, 57, 20–33. [Google Scholar] [CrossRef]
  26. Anshari, M.; Alas, Y.; Sabtu, N.P.H.; Hamid, M.S.A. Online Learning: Trends, Issues and Challenges in the Big Data Era. J. E-Learn. Knowl. Soc. 2016, 12, 121–134. [Google Scholar] [CrossRef]
  27. Hsiao, I.Y.T.; Lan, Y.J.; Kao, C.-L.; Li, P. Visualization Analytics for Second Language Vocabulary Learning in Virtual Worlds. Educ. Technol. Soc. 2017, 20, 161–175. [Google Scholar]
  28. Zou, D.; Xie, H. Personalized Word-Learning Based on Technique Feature Analysis and Learning Analytics. J. Educ. Technol. Soc. 2018, 21, 233–244. [Google Scholar]
  29. Liu, S.H.J.; Lan, Y.J. Learning to Learn Collaboratively on Facebook—A Pilot Study. In Workshop Proceedings of the 22nd International Conference on Computers in Education, ICCE 2014, Nara, Japan, 30 November 2014–4 December 2014; Asia-Pacific Society for Computers in Education: Taiwan, China, 2014; pp. 664–669. [Google Scholar]
  30. Xie, H.; Zou, D.; Lau, R.Y.; Wang, F.L.; Wong, T.L. Generating Incidental Word-Learning Tasks Via Topic-Based and Load-Based Profiles. IEEE Multimed. 2015, 23, 60–70. [Google Scholar] [CrossRef]
  31. Page, L.; Brin, S.; Motwani, R.; Winograd, T. The Pagerank Citation Ranking: Bringing Order to the Web. Stanf. InfoLab. 1999. Available online: http://ilpubs.stanford.edu:8090/422/1/1999-66.pdf (accessed on 9 November 2019).
  32. Rieh, S.Y. Judgment of Information Quality and Cognitive Authority in the Web. J. Am. Soc. Inf. Sci. Technol. 2002, 53, 145–161. [Google Scholar] [CrossRef]
  33. Luo, L.; Xie, H.; Rao, Y.; Wang, F.L. Personalized Recommendation by Matrix Co-Factorization with Tags and Time Information. Expert Syst. Appl. 2019, 119, 311–321. [Google Scholar] [CrossRef]
  34. Xie, H.; Li, X.; Wang, T.; Chen, L.; Li, K.; Wang, F.L.; Cai, Y.; Li, Q.; Min, H. Personalized Search for Social Media via Dominating Verbal Context. Neurocomputing 2016, 172, 27–37. [Google Scholar] [CrossRef]
  35. Li, T.; Zou, D.; Weng, X.; Wang, M.; Xie, H.; Wang, F. Acceptance and Implication of Smartphones for Informal Language Learning in Foreign Study Tours. Int. J. of Mobile Learn. and Org. 2020, 14, 456–477. [Google Scholar] [CrossRef]
  36. Tu, Y.; Zou, D.; Zhang, R. A Comprehensive Framework for Designing and Evaluating Vocabulary Learning App. Int. J. of Mobile Learn. and Org. 2020, 14, 370–397. [Google Scholar] [CrossRef]
  37. Mohsen, M.A.; Balakumar, M. A Review of Multimedia Glosses and Their Effects on L2 Vocabulary Acquisition in CALL Literature. ReCALL 2011, 23, 135–159. [Google Scholar] [CrossRef]
  38. Jacobs, G.M.; Dufon, P.; Fong, C.H. L1 And L2 Vocabulary Glosses in L2 Reading Passages: Their Effectiveness for Increasing Comprehension and Vocabulary Knowledge. J. Res. Read. 1994, 17, 19–28. [Google Scholar] [CrossRef]
  39. Abraham, L. Computer-Mediated Glosses in Second Language Reading Comprehension and Vocabulary Learning: A Meta-Analysis. Comput. Assist. Lang. Learn. 2008, 12, 199–226. [Google Scholar] [CrossRef]
  40. Mohsen, M.A. Effects of Help Options in a Multimedia Listening Environment on L2 Vocabulary Acquisition. Comput. Assist. Lang. Learn. 2016, 29, 1220–1237. [Google Scholar] [CrossRef]
  41. Khezrlou, S.; Ellis, R.; Sadeghi, K. Effects of Computer-Assisted Glosses on EFL Learners’ Vocabulary Acquisition and Reading Comprehension in Three Learning Conditions. System 2017, 65, 104–116. [Google Scholar] [CrossRef]
  42. Chun, D.M.; Plass, J.L. Effects of Pictorial Annotations on Vocabulary Acquisition. Mod. Lang. J. 1996, 80, 183–198. [Google Scholar] [CrossRef]
  43. Jones, L.C.; Plass, J.L. Supporting Listening Comprehension and Vocabulary Acquisition in French with Pictorial Annotations. Mod. Lang. J. 2002, 86, 546–561. [Google Scholar] [CrossRef]
  44. Yeh, Y.; Wang, C.W. Effects of Multimedia Vocabulary Annotations and Learning Styles on Vocabulary Learning. CALICO J. 2003, 21, 131–144. [Google Scholar] [CrossRef] [Green Version]
  45. Türk, E.; Erçetin, G. Effects of Interactive Versus Simultaneous Display of Multimedia Glosses on L2 Reading Comprehension and Incidental Vocabulary Learning. Comput. Assist. Lang. Learn. 2014, 27, 1–25. [Google Scholar] [CrossRef]
  46. Ikeda, N. Effects of Different Types of Images on the Understanding of Stories: Basic Research to Develop Japanese Teaching Materials for Use on the Internet. System 1999, 27, 105–118. [Google Scholar] [CrossRef]
  47. Lin, H.; Chen, T. Reading Authentic EFL Text Using Visualization and Advance Organizers in a Multimedia Learning Environment. Lang. Learn. Technol. 2007, 11, 83–106. [Google Scholar]
  48. Ariew, R.; Ercetin, G. Exploring the Potential of Hypermedia Annotations for Second Language Reading. Comput. Assist. Lang. Learn. 2004, 17, 237–259. [Google Scholar] [CrossRef]
  49. Mayer, R.E. Multimedia Learning: Are We Asking the Right Questions? Educ. Psychol. 1997, 32, 1–19. [Google Scholar] [CrossRef] [Green Version]
  50. Mayer, R.E. Multimedia Learning; Cambridge University Press: New York, NY, USA, 2001. [Google Scholar] [CrossRef]
  51. Mayer, R.E.; Lee, H.; Peebles, A. Multimedia Learning in a Second Language: A Cognitive Load Perspective. Appl. Cogn. Psychol. 2014, 28, 653–660. [Google Scholar] [CrossRef]
  52. Plass, J.L.; Chun, D.M.; Mayer, R.E.; Leutner, D. Cognitive Load in Reading a Foreign Language Text with Multimedia Aids and the Influence of Verbal and Spatial Abilities. Comput. Hum. Behav. 2003, 19, 221–243. [Google Scholar] [CrossRef]
  53. Schmidt, R. The Role of Consciousness in Second Language Learning. Appl. Linguist. 1990, 11, 129–158. [Google Scholar] [CrossRef]
  54. Schmidt, R. Consciousness, Learning and Interlanguage Pragmatics. In Interlanguage Pragmatics; Kasper, G., Blum-Kulka, S., Eds.; Oxford University Press: Oxford, UK, 1993; pp. 21–42. [Google Scholar]
  55. Schmidt, R. (Ed.) Consciousness and Foreign Language Learning: A Tutorial on the Role of Attention and Awareness in Learning. In Attention and Awareness in Foreign Language Learning; University of Hawaii: Honolulu, HI, USA, 1995; pp. 1–63. [Google Scholar]
  56. Craik, F.I.; Lockhart, R.S. Levels of Processing: A Framework for Memory Research. J. Verbal Learn. Verbal Behav. 1972, 11, 671–684. [Google Scholar] [CrossRef]
  57. Davies, M. Corpus of Contemporary American English (1990–2012). 2012. Available online: corpus.byu.edu/coca/ (accessed on 28 August 2019).
  58. Folse, K. The Effect of Type of Written Exercise on L2 Vocabulary Retention. TESOL Q. 2006, 40, 273–293. [Google Scholar] [CrossRef]
  59. Paribakht, T.S.; Wesche, M. Vocabulary enhancement activities and reading for meaning in second language vocabulary acquisition. Second Lang. Vocab. Acquis. Ration. Pedagog 1997, 55, 174–200. [Google Scholar]
  60. Zou, D. Vocabulary Acquisition Through Cloze Exercises, Sentence-Writing and Composition-Writing: Extending the Evaluation Component of the Involvement Load Hypothesis. Lang. Teach. Res. 2017, 21, 54–75. [Google Scholar] [CrossRef]
  61. Zou, D. Comparing Dictionary-Induced Vocabulary Learning and Inferencing in The Context of Reading. Lexikos. 2016, 26, 372–390. [Google Scholar] [CrossRef] [Green Version]
  62. Channell, J. Psycholinguistic Considerations in the Study of L2 Vocabulary Acquisition. Vocab. Lang. Teach. 1988, 83–96. [Google Scholar]
  63. Chatterjee, S.; Majumdar, D.; Misra, S.; Damaševičius, R. Adoption of Mobile Applications for Teaching-Learning Process in Rural Girls’ Schools in India: An Empirical Study. Educ. Inf. Technol. 2020, 25, 4057–4076. [Google Scholar] [CrossRef] [Green Version]
  64. Xie, H.; Chu, H.C.; Hwang, G.J.; Wang, C.C. Trends and Development in Technology-Enhanced Adaptive/Personalized Learning: A Systematic Review of Journal Publications from 2007 to 2017. Comput. Educ. 2019, 140, 103599. [Google Scholar] [CrossRef]
Figure 1. The pictorial annotation created by a participant using an image from Twitter.
Figure 1. The pictorial annotation created by a participant using an image from Twitter.
Sustainability 13 05767 g001
Figure 2. The pictorial annotation created by a participant who concretized an abstract meaning through additional self-generated explanations.
Figure 2. The pictorial annotation created by a participant who concretized an abstract meaning through additional self-generated explanations.
Sustainability 13 05767 g002
Table 1. Participants’ perceptions of Google data and social media data.
Table 1. Participants’ perceptions of Google data and social media data.
Google DataSocial Media Data
The idea of using this data set as resources for language enhancement is interesting/creative/fun.27/30 (90%)28/30 (93%)
The idea of using this data set as resources for language enhancement is feasible/reliable.26/30 (87%)24/30 (80%)
It is easy to search for images that appropriately depict target word meanings from this dataset.29/30 (97%)20/30 (67%)
The images of this dataset are relevant to the keywords for search.26/30 (87%)19/30 (63%)
The images of this dataset are appropriate/make sense.23/30 (76%)12/30 (40%)
Table 2. Overall scores of the learner-generated annotations.
Table 2. Overall scores of the learner-generated annotations.
Google Pictorial AnnotationsSocial Media Pictorial Annotations
NMSDNMSD
Scores of the annotations3004.210.673003.420.70
Table 3. Results of the independent samples t-test of all learner-generated annotations.
Table 3. Results of the independent samples t-test of all learner-generated annotations.
Levene’s Test t-Test
Fp.tdfpMDSEDLower95% CI Upper
Scores of the learner-generated annotations Equal variances assumed3.730.0514.045980.000.790.050.680.90
Equal variances not assumed 14.04596.890.000.790.050.680.90
Table 4. Scores of the learner-generated annotations for verbs and nouns and adjective.
Table 4. Scores of the learner-generated annotations for verbs and nouns and adjective.
Scores of the Annotations of the 6 VerbsScores of the Annotations of the 3 Nouns and 1 Adjective
Google
pictorial annotations
N180120
M4.563.70
SD0.560.47
Social media pictorial annotationsN180120
M3.822.81
SD0.560.38
OverallN360240
M4.193.25
SD0.670.62
Table 5. Results of the independent samples t-test of the learner-generated annotations for different words.
Table 5. Results of the independent samples t-test of the learner-generated annotations for different words.
Levene’s Test t-Test
Fp.tdfpMDSEDLower95% CI Upper
Scores of the learner-generated annotations Equal variances assumed0.420.5117.225980.000.930.050.821.04
Equal variances not assumed 17.50539.680.000.930.050.831.04
Table 6. Descriptive statistics of the participants’ scores in the pre-test and the immediate and delayed post-tests.
Table 6. Descriptive statistics of the participants’ scores in the pre-test and the immediate and delayed post-tests.
NPre-TestImmediate Post-TestDelayed Post-Test
MSDMSDMSD
Learning with Google multimedia320.060.248.031.336.842.35
Learning with social media pictorial annotations 300.060.256.961.625.501.92
Learning with textual annotations310.090.306.001.753.931.93
Table 7. ANOVA results of the immediate post-test scores.
Table 7. ANOVA results of the immediate post-test scores.
SSdfMSFp
Between Groups65.05232.5213.070.00
Within Groups223.93902.48
Total288.9892
Table 8. Results of the multiple comparisons of the immediate post-test scores.
Table 8. Results of the multiple comparisons of the immediate post-test scores.
95% CI
MDSESig.LBUB
Learning with Google pictorial annotations Learning with social media pictorial annotations 1.06 *0.400.020.102.01
Learning with textual annotations 2.03 *0.390.001.082.97
Learning with social media pictorial annotations Learning with textual annotations 0.96 *0.400.040.001.92
*. The mean difference is significant at the 0.05 level.
Table 9. ANOVA results of the delayed post-test scores.
Table 9. ANOVA results of the delayed post-test scores.
SSdfMSFp
Between Groups133.33266.6615.320.00
Within Groups391.59904.35
Total524.9292
Table 10. Results of the multiple comparisons of the delayed post-test scores.
Table 10. Results of the multiple comparisons of the delayed post-test scores.
95% CI
MDSESig.LBUB
Learning with Google pictorial annotations 2Learning with social media pictorial annotations 31.34 *0.530.030.082.60
Learning with textual annotations 12.90 *0.520.001.654.16
Learning with social media pictorial annotations 3Learning with textual annotations 11.56 *0.530.010.292.83
*. The mean difference is significant at the 0.05 level.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zou, D.; Xie, H. Vocabulary Learning Based on Learner-Generated Pictorial Annotations: Using Big Data as Learning Resources. Sustainability 2021, 13, 5767. https://doi.org/10.3390/su13115767

AMA Style

Zou D, Xie H. Vocabulary Learning Based on Learner-Generated Pictorial Annotations: Using Big Data as Learning Resources. Sustainability. 2021; 13(11):5767. https://doi.org/10.3390/su13115767

Chicago/Turabian Style

Zou, Di, and Haoran Xie. 2021. "Vocabulary Learning Based on Learner-Generated Pictorial Annotations: Using Big Data as Learning Resources" Sustainability 13, no. 11: 5767. https://doi.org/10.3390/su13115767

APA Style

Zou, D., & Xie, H. (2021). Vocabulary Learning Based on Learner-Generated Pictorial Annotations: Using Big Data as Learning Resources. Sustainability, 13(11), 5767. https://doi.org/10.3390/su13115767

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop