Next Article in Journal
The Role of Acetyl-Carnitine and Rehabilitation in the Management of Patients with Post-COVID Syndrome: Case-Control Study
Previous Article in Journal
Protonic Transport in Layered Perovskites BaLanInnO3n+1 (n = 1, 2) with Ruddlesden-Popper Structure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Burapha-TH: A Multi-Purpose Character, Digit, and Syllable Handwriting Dataset

1
Faculty of Informatics, Burapha University, Chonburi 20131, Thailand
2
Faculty of Science and Arts, Chanthaburi Campus, Burapha University, Chanthaburi 22170, Thailand
3
AAII, Faculty of Data Science, Musashino University, Tokyo 135-8181, Japan
4
Korea Institute of Oriental Medicine, Daejeon 34054, Korea
5
Korea Institute of Science and Technology Information, Daejeon 34141, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(8), 4083; https://doi.org/10.3390/app12084083
Submission received: 14 March 2022 / Revised: 4 April 2022 / Accepted: 13 April 2022 / Published: 18 April 2022

Abstract

:
In handwriting recognition research, a public image dataset is necessary to evaluate algorithm correctness and runtime performance. Unfortunately, in existing Thai language script image datasets, there is a lack of variety of standard handwriting types. This paper focuses on a new offline Thai handwriting image dataset named Burapha-TH. The dataset has 68 character classes, 10 digit classes, and 320 syllable classes. For constructing the dataset, 1072 Thai native speakers wrote on collection datasheets that were then digitized using a 300 dpi scanner. De-skewing, detection box and segmentation algorithms were applied to the raw scans for image extraction. The experiment used different deep convolutional models with the proposed dataset. The result shows that the VGG-13 model (with batch normalization) achieved accuracy rates of 95.00%, 98.29%, and 96.16% on character, digit, and syllable classes, respectively. The Burapha-TH dataset, unlike all other known Thai handwriting datasets, retains existing noise, the white background, and all artifacts generated by scanning. This comprehensive, raw, and more realistic dataset will be helpful for a variety of research purposes in the future.

1. Introduction

Artificial Intelligence (AI) and related algorithms, especially intense learning algorithms, have become much more popular in the last few years. They have been applied to several research fields. The increase in popularity is mainly attributable to three factors: (1) higher performance hardware; (2) new and better tools, techniques, and libraries to train deeper networks; and (3) availability of more published datasets. While hardware and tools are readily available, the lack of comprehensive public datasets remains a challenge for many domains.
Handwriting recognition is an exciting research topic in the AI domain [1,2,3,4,5,6]. Applications incorporating handwriting recognition are used in the legal industry for postal mail and car plate recognition, invoice imaging, and form data entry. Handwriting recognition software requires a large quantity of high-quality data to train models effectively. Therefore, large, standard datasets are essential for enabling researchers to test, tune, and evaluate each new algorithm’s performance.
Some standard datasets are available, such as the MNITS dataset [7]. MNIST is a handwritten Arabic digit dataset that contains 60,000 images for training and 10,000 images for testing. The database is also widely used for training and testing in machine learning. The EMNIST [8] is an extended MNIST dataset consisting of handwritten digits and letters of the English alphabet. This dataset was introduced in 2017 and is extensively used to improve deep learning algorithms.
There are several scripts that have been proposed as contributions to an international standard handwritten character dataset. The Institute of Automation of the Chinese Academy of Sciences (CASIA) [9] released CASIA-HWDB (offline) and CASIA-OLHWDB (online) in 2011. These datasets contain offline/online handwritten characters and continuous text written by 1020 people using Anoto pens on paper. The datasets of isolated characters contain about 3.9 million samples of 7185 Chinese characters and 171 symbols, and the datasets of handwritten texts contain about 5090 pages with 1.35 million character samples. The National Institute of Japanese Literature released the Kuzushiji dataset in November 2016 [10]. They scanned 35 classical books printed in the 18th century and organized their proposed dataset into three parts: (1) Kuzushiji-MNIST, a drop-in replacement for the MNIST dataset; (2) Kuzushiji-49, a much larger, but imbalanced dataset containing 48 Hiragana characters and one Hiragana iteration mark; and (3) Kuzushiji-Kanji, an imbalanced dataset of 3832 Kanji characters, including rare characters with very few samples. The Kuzushiji dataset currently consists of 3999 character types and 403,242 characters. A public domain handwritten character image dataset for the Malayalam language script contains data provided by 77 native Malayalam writers. It includes independent vowels, consonants, half consonants, vowels, consonant modifiers, and conjunct characters [11]. The glyphs of the Malayalam script have 85 classes that contain 17,236 training images, 5706 validation images, and 6360 testing images. An active contour model-based minimization technique was applied for character segmentation. The dataset was evaluated with different feature extraction techniques. A scattering convolutional network achieves 91.05% recognition accuracy. The PE-92 database project was started in 1992. PE-92 contains 100 image sets of 2350 Korean characters, considered general and in daily use [12,13]. The handwritten Korean language is syllable-based, not alphabet-based like western European languages. One person wrote between 100 and 500 characters for each set in the first 70 images sets. This dataset tries to accumulate as many writing styles as possible. Some problems occurred while developing the database in the data collection process, even though the characters selected to be written were considered general ones. For example, misspelling of complex vowels was often found. In 1997, SERI (System Engineering Research Institute) of Korea University created SERI95, a Hangul dataset. The SERI95 merged with ETRI, which contains 520 sets, one for each of the most frequently used Hangul characters. Each set contains about a thousand samples.
Several datasets of Thai handwriting have been published. Sae-Tang and Methasate introduced a Thai online and offline handwritten character corpus in 2004 [14]. The online handwritten character corpus contains more than 44,000 handwriting samples. The images were collected using a program developed for a WACOM 6 × 8 tablet used by 63 different writers who entered Thai characters, English characters, and special symbols. The characters written include: (1) 79 patterns of Thai consonants, vowels, tones and digit characters; (2) 62 patterns of English uppercase, lowercase, and digit characters; and (3) 15 patterns of special symbols. This offline handwritten character corpus contains handwritten isolated characters, words, and sentences, and 14,000 long samples from 143 different writers. The handwritten isolated character set contains 79 patterns of consonants, vowels, tones, and digits. The word set includes names of 76 Thai provinces and 21 Thai digits. The sentence sets include 16 Thai digits and 3 Thai general articles. A new Thai handwriting dataset, ALICE-THI, was published in 2015 [15]. This dataset was collected to support research on handwritten character recognition using local gradient feature descriptors. ALICE-THI consists of 13,130 training samples and 1360 test samples, 44 consonants, 17 vowels, 4 tones, and 3 symbols. The ALICE-THI handwritten Thai digit dataset contains 8555 training samples and 1000 test samples, for 9555 samples.
It is common knowledge that a large dataset, in terms of the number of images, can help achieve better accuracy when using deep learning techniques [16,17,18]. The construction of a good handwriting dataset, which is comprehensive, has enough variety and is large enough for suitable training of deep learning algorithms. The lack of a diverse dataset of handwritten Thai scripts is problematic for a researcher working on Thai handwriting recognition. Another problem with existing datasets is that they cannot be easily compared. Therefore, Thai handwriting research is not progressing as quickly as it should. These issues inspired us to construct a new dataset to foster more Thai handwriting recognition research.
This paper introduces a new Thai handwriting dataset named Burapha-TH consisting of characters, digits, and syllables. Our dataset is very different from existing standard handwriting datasets. Standard datasets are generally explicit and contain preprocessed images. The preprocessing removes noise, traces image contours, performs smoothing on the images, removes the non-essential background, and performs binarization. In contrast, our new Burapha-TH dataset has only performed de-skewing and segmentation in its preprocessing steps. We did not remove the salt and pepper noise, white background, or artifacts generated by scanning. The objectives when creating this dataset were to provide good opportunities for research on a wide variety of Thai handwriting recognition tasks. The dataset is suitable for research about handwriting recognition, including feature extraction, machine learning [19], deep learning [20], and image processing of handwriting. The dataset contains raw images (without any pre-processing) of each document. We published the original data collection sheets to permit new research on glyph image processing, including alternative preprocessing approaches with more advanced skew correction, line detection, segmentation, image smoothing, and noise and white background removal. The dataset can be used to explore writing patterns related to gender. The researcher can exploit the potential of syllables handwriting style. The syllables can represent a unique style of handwriting. For each syllable, it shows a continuous handwriting style that is suitable for word segmentation or handwriting generation.
Our goal is to provide the dataset necessary to develop more robust and practical real-world applications that make dealing with Thai script images easier. Furthermore, we used a generalized unified framework for constructing the Burapha-TH datasets. It is free for other use. In this paper, we demonstrate the Burapha-TH dataset’s usefulness, an experiment was performed using serval different CNN models, and the results were analyzed.
The rest of the paper is organized as follows. An overview of the Thai language script is provided in Section 2. The construction of the Burapha-TH dataset is described in Section 3. In Section 4, the results from the experiment are discussed. Finally, Section 5 concludes and suggests possible future work.

2. Overview of Thai Language Script

The first inscription of King Ramkamhaeng is historical evidence showing that the Thai script existed and has been in use since 1826. The script has been changed in both form and orthography from time to time. In 1949, Phraya Upakit Silapasarn, an expert in Thai language, Pali language, and Thai literature, presented the patterns of the characteristics of the Thai language as follows: 44 consonants, 4 tone markers, and 21 vowels (form) [21,22]. Around 1991, professionals joined together to form the Thai API Consortium (TAPIC), headed by Thaweesak Koanantakool and sponsored by National Electronics and Computer Technology Center (NECTEC), to draft a proposal for a Thai API standard called WTT 2.0 [23]. Their draft defines two eight-bit character code sets, consisting of 66 control characters (CTRL), 44 consonants (cons), 4 tone markers, 5 diacritics, 10 Thai digits, and 18 vowels (form). Their objective was to make it easier to type Thai on computers.
According to WTT 2.0, Thai words are input and stored letter-by-letter from left to right. These characters are mixed and placed on a line in four zones. All consonant characters are essentially on the baseline. At the same time, vowel characters can be positioned before, after, above or below the consonant characters or in various combinations of these positions. Moreover, tonal characters are located above consonants. If the word contains a vowel character on top of the consonant, a tonal character will be placed above that vowel character. Figure 1 illustrates a sample of word formation in the Thai language, where characters and symbols are located in different zones. ย ก ษ ข ย ว ห ญ ด ม ก are consonant characters located on the mainline, ั ์ เ ี ้ ใ ่ ุ า are vowel characters and symbols located before, after, above and below the consonant characters. Figure 2 displays a pangram of the Thai language, a verse expression that uses almost all of the characters of the Thai language.

2.1. The Characteristics of Thai Characters

2.1.1. Consonants

The structure of a Thai character has the following components: head, tail, mid loop, serration, beak, flag, and pedestal. The head or a single loop is a unique feature of Thai characters, classified into standard single loops such as บ, curly loops such as ข, or serration such as ซ. The tail is a concatenated line above or below zone 2, such as ป and ฤ. The mid loop or second loop looks like the head, but the mid loop is in the middle of the character and occurs as a line connection such as in ห ม. The serration, or a sawtooth, occurs at the characters’ head or tail such as in ต or ฏ. The beak is a line that looks like a bird’s horny projecting jaw in some characters such as ก and ถ. The flag is the end of the line and resembles a flying flag. Examples of this are ธ and ร. Finally, a pedestal can be considered the character’s foot, which occurs in a few characters such as ญ and ฐ.

2.1.2. Vowels

The number of vowels in the Thai language can be counted in different ways, such as 21 forms and 32 or 36 forms and 21 sounds. This is because some vowels are formed from other vowels plus some final consonant sounds. However, the vowel and the sound are not represented differently in the overall picture. In the Thai keyboard system, there are only 17 forms of vowels, but users can still type all Thai vowels by combining them together as shown in Table 1. Vowels in words are placed in five positions: in front of a consonant character, behind a consonant character, above a consonant character, below a consonant character, and surrounding a consonant character. Front vowels are vowels that occur in front of the consonant and include เ แ ไ ใ โ. Back vowels are vowels that occur behind the consonant and include ะ า. Above vowels are vowels that occur above the consonant and include ิ ี ึ ื ั. Below vowels occur below the consonant characters and include ุ ู. Surrounding vowels are vowels whose components surround consonant characters and include เ _ะ แ_ะ โ_ะ เ_าะ เ_อะ เ ียะ เ ีย เือะ เือ ัวะ ัว.

2.1.3. Tone Markers

There are four tones used in the Thai language, and they are written using tonal characters or markers, including  ่ ้ ๊ and ๋. The tone markers are placed above the consonant or vowel characters.

2.1.4. Digits

The Thai language has unique digit characters that are different from Arabic digit characters. These digit characters are often used in official documents. The following characters are the ten Thai digit characters written in a sequence from 0 to 9: ๐ ๑ ๒ ๓ ๔ ๕ ๖ ๗ ๘ ๙.

2.2. How to Write

There are no fixed writing rules for Thai characters in the regular Thai language writing method. Nevertheless, consonant characters are written horizontally from left to right. The writing will mainly start if the character has a head. Thai consonant characters can be categorized differently depending on the criteria used, including character head, a path of the line, and character size.

2.2.1. Character Head

The head is a significant characteristic of Thai consonant characters. A Thai consonant character can be classified according to the characteristics of its head. The head can start from the top line and face out, as seen in the บ and ป characters. Alternatively, sometimes the head faces in such as in the characters ผ and ฝ. Some characters consist of a head starting from the bottom line and facing in or outwards such as ถ, ร, and ภ. The head can also start from the center of the line and turn to the right such as in ค and ศ or turn to the left such as in จ, ด, and ต. The head with serration is also a unique characteristic of Thai consonants and occurs when writing ฑ and ฆ. In addition, some consonants contain two round heads starting from the top line with the head facing out such as ข and ช.

2.2.2. Path of the Line

When considering the line path, the Thai consonant characters are constructed by four different line path types: circle, horizontal line, vertical line, and diagonal line, as displayed in Figure 1.

2.2.3. Character Size

Thai consonant characters can be separated by size, width, and height. Different characters contain small, medium, and large amounts of space between the vertical left and right sides. Examples of each type are ข, ซ, ง, ก and ณ, ฒ, ฌ, respectively. When considering consonant characters on a line, they can be classified into three different groups: above, in, and below the line. Most Thai consonant characters exist in the middle or baseline such as ก, ห, ง, ป, and ฟ, which are examples of characters that have a concatenated line above the middle line. Furthermore, a few characters exist with a concatenated line below the baseline, such as ฤ and ฏ.

3. Burapha-TH Dataset Construction

Our dataset is named Burapha-TH. We created this dataset as a part of research work conducted by Burapha University, Thailand, and the Korea Institute of Science and Technology Information, Korea. In this section, we describe Thai script image data construction and processing.

3.1. Writers

The most crucial aspect is collecting data from as many different styles as possible, especially in the university student group. Figure 3 shows the sex and age distribution of writers who participated in our dataset construction. The number of writers is 1072 (721 female and 351 male). The average age ranges between 17 and 25 years old. The average age is 19.59 and the standard deviation of participants is 1.299. All of the participants were Burapha University undergraduate students.

3.2. Data Collection Sheets

We created data collection sheets, and each sheet was divided into three parts, as shown in Figure 4a. The first part records a participant’s information and displays regulations for writing. The second part is the writing area, which is designed using grids. Each sheet has 100 cells, and the size of each cell is 60 × 60 pixels. We carefully designed cells with equal size, which later helped us to be able to segment each cell automatically. The top cell has a label example, which can help the participant write the blank cell character below. The third part is a blank cell for rewriting any incorrect writing, as shown in Figure 4b.
The participants wrote characters without writing style constraints, had no time limit, and could use any pen of any color. The participants were instructed to write two times on the standard collection sheets. The one regulation is that whenever participants make a writing mistake, they must cross out the wrong image and rewrite it in a blank cell in the last line of the sheet. The regulation allows the writer to write letters within the boxes, and characters must not touch the frame.
In general, Thai people usually use a pen with blue ink in everyday life, followed by black and red. In collecting the handwriting data for this research, we collected as much real data as we could. Therefore, we did not limit the ink color or size of the pens used. In terms of color, we still keep the original ink color because there is research [24] that focuses on general color tuning properties of CNNs trained for object recognition. Such research observes that color images responsible for the activation of color-sensitive kernels were more likely to be misclassified. So, this is the reason that our proposed dataset still retains the original ink color without modification.

3.3. Data Preparation Process

We scanned each data collection sheet using a color document scanner with 300 dpi resolution in the data preparation process. The scanner was fast for scanning the documents, but at the cost of lower resolution. Therefore, the resolution of our dataset is as general as possible. We extract information from low-quality images for handwriting recognition, and they are relatively sound. Next, we describe the algorithms for image deskewing, line detection, and image segmentation.

3.3.1. Image Deskewing

It is challenging to ensure that all paper is in the correct position for the scanning process. When scanning, some images are skewed, as shown in Figure 5a. Algorithm 1 illustrates the pseudo-code for deskewing an image to overcome this problem. We used three OpenCV libraries [25] from a GitHub webpage [26] to implement this part. The libraries are DatasetService, DeskewService, and GraphicsServices. The procedure’s input is an original handwritten form image file (dm), and a deskewed handwritten glyph from an image file (ddm) is the output result of the deskewing process. In this algorithm, the main procedure is deskewing (straightening) text in image form (line 7) by calling the deskew function from DeskewService (DeskewService().deskew) that returns deskewedimage and guessedAngle. The guessedAngle value is used to check for the proper angle at −20 (line 8). We will call the rotate image function from the GraphicsService library in the skewed case, adding 90 to the guessedAngle (line 9). The output of the Image_deskewing algorithm is depicted in Figure 5b.
Algorithm 1: Pseudo-code of Image_deskewing.
Input:  dm #a original handwritten form image
Output:  ddm #a deskewing handwritten from image
Procedure DeskewingImage(dm)
 1: begin
 2:   from services.DatasetService import DatasetService
 3:   from services.DeskewService import DeskewService
 4:   from services.GraphicsService import GraphicsService
 5:   imageCv = GraphicsService().openImageCv(dm)
 6:   imagePath = dm.ImagePath()
 7:   deskewedImage, guessedAngle = DeskewService().deskew(imageCv)
 8:   if (guessedAngle < −20.0):
 9:    dd = GraphicsService().rotateImage(imageCv,(guessedAngle + 90.)* 1.0)
 10:   end if
 11:   DatasetService().saveData(imagePath, deskewedImage)
 12: end

3.3.2. Line Detection

Algorithm 2 is the pseudo-code for the Line detection box. The input is the deskewing handwritten result of Algorithm 1 and the output is in two forms: a matrix and a label matrix. To separate characters in size (60 × 60), we need to detect the box’s size following the cell. The main procedure needs to preprocess the image to obtain a grayscale image (line 2). This grayscale image is used to find the image’s horizontal and vertical scale (line 3–line 5). Then these are combined to form a big picture of an outlier (line 6) before increasing the white region in the final binary image (img_bin_final). The main procedure is used to find a stat matrix (i.e., left, top, width, height, area) with the Perform operation on (line 7).
Algorithm 2: Pseudo-code of Line_detection.
Input: ddm    #a deskewing handwritten from image
Output: stats, labels    # stat matrix and the label matrix
Procedure DetectionBox(ddm, line_min_width = 5)
 1: begin
 2:   gray_scale = convert ddm to gray scale
 3:   img_bin_h = find the outline of the horizontal gray_scale object
 4:   img_bin_v = find the outline of the veritical gray_scale object
 5:   img_bin_final = img_bin_h|img_bin_v
 6:   img_bin_final = increases the white region in the img_bin_final image
 7:   ret, labels, stats, centroids = Perform the operation of img_bin_final
 8: end

3.3.3. Image Segmentation

To provide a set of segmented handwritten images, Algorithm 3 is used. We applied the threshold of the cell’s outlier to segmented images. The input is a set of handwritten form images (D), and the output is a set of segmented handwritten images (DL), as shown in Figure 5d. Then we call the DeskewingImage function in algorithm 1 (line 3), and the detection box function step in algorithm 2 (line 4), respectively. The box detection is shown in Figure 5c. The stat matrix describes the optimal region to cut cells from the grid picture. Finally, each image is appended to the DL set (line 8).
Algorithm 3: Pseudo-code of Image_segmentation.
Input: D = {d1, d2, d3,…, d|D|} #a set of handwritten form image
Output: DL = {dl1, dl2, dl3,…, dl|D|} #a set of segmented handwritten image
Procedure LineDetection(DD)
 1: begin
 2:   for each document dm in D do
 3:     ddm = DeskewingImage(dm)
 4:     stats, labels = DetectionBox(ddm, line_min_width = 5)
 5:     for left, top, width, height, area in stats [2:] do
 6:      if (top > 800 and width > 100 and height > 100)
 7:       dl = ddm [top: top + height, left: left + width]
 8:       append dl in a set of segmented handwritten image DL
 9:      end if
 10:     end for
 11:   end for
 12: end

3.4. Statistical Properties

The raw standard collection sheet images of characters and digits form 1156 sheets, and after segmentation yielded 107,506 images. Simultaneously, there were 1920 sheets of Thai syllables, and those sheets were segmented to produce 279,730 images. We did not resize the images, making the sizes of images non-uniform. For the next step, Thai language experts eliminated some images by using majority voting. The conditions for eliminating images were: (1) incorrect writing, (2) unreadable, (3) heavily distorted, and (4) over edge cut. Some eliminated example images are shown in Figure 6.
After unclear images were eliminated, the number of remaining characters and digits was 87,600 samples (19,906 were removed), and the number of remaining syllables was 268,056 (11,674 were removed). Thus, the total number of proper images is 355,656 samples, and 31,583, or about 8%, were discarded, as shown in Table 2.
Currently, Burapha-TH has three categories: characters, digits, and syllables. The character dataset has 68 classes consisting of 44 Thai characters, 20 Thai vowels, and 4 Thai tone markers. We separated it into 63,327 samples of the training set and 13,600 samples of the test set. The average number of images in the training set is about 931 samples in each class, and the minimum and maximum are 790 and 995. The testing set has 200 samples in each class.
The digit dataset has 10 classes. We sequenced several Thai digits from zero to nine. The images have 10,673 samples, which we divided into training and testing sets with 8673 and 2000 samples, respectively. The average number of images in the training set is about 867 samples in each class, and the minimum and maximum are 772 and 923.
We created a new Thai syllable dataset for extending handwriting recognition research. The number of images in the training and testing sets is 236,056 and 32,000 samples. The average number of images in the training set is about 738 samples, and 503 and 905 are the minimum and maximum counts. The statistics of the proposed Burapha-TH are described in Table 3.
The Burapha-TH handwriting images are written in cursive forms, with or without a head, and they usually have several writing styles. We have tried to gather various collections of Thai scripts in our proposed datasets, as shown in Figure 7. The dataset has example consonants, vowels, and tone markers for characters. The digit dataset depicts Thai digits from zero to nine. Simultaneously, the syllable dataset consists of Thai syllables with different styles of consonants and vowels. The complete version of our proposed dataset includes Table A1 Thai characters with 68 classes, Table A2 Thai digits with 10 classes, and Table A3 Thai syllables with 320 classes.

4. Experiments and Discussion

We performed our study using a desktop computer with an Intel Core i7 3.6 GHz (CPU), 16 GB of memory capacity, and a Nvidia GeForce 1080Ti VGA card. We used Pytorch 1.1.0 with the Ubuntu 16.04 operating system and Python 3.7.
To show the usefulness of our proposed dataset for research in Thai handwriting recognition, we selected three popular CNN architectures: CNN with four convolutional layers, LeNet-5, and VGG-13 with batch normalization. The results are shown in Table 4 to benchmark the proposed dataset. All classifiers were repeated five times by shuffling the training set and averaging accuracy on the test set. The hyper-parameters used to train all the models were batch size 32, dropout 0.5, epoch 100, and optimizer Adam. The testing results show that VGG-13 with BN outperforms the others in terms of accuracy.
To measure the VGG-13 BN model’s performance on the proposed datasets, we used k-fold cross-validation, with three, five, and ten-fold cross-validation. This technique accounts for the model’s variance concerning differences in the training and test datasets and the learning algorithm’s stochastic nature. A model’s performance can be taken as the mean performance across k-folds, given the standard deviation, which could be used to estimate a confidence interval. We used the scikit-learn API to implement the k-fold cross-validation in our experiment. Figure 8 shows the results of percentage accuracy of all data partitions. Syllable and digit datasets show almost 99%, while the character data shows an average value of about 97%.
A statistics overview of the public dataset discussed in the paper is compared with Burapha-TH datasets in Table 5. Our proposed datasets have a wider variety of content when compared with other public datasets. The number of writers of Burapha-TH is similar to CASIA-HWDB, which implies that our dataset has a variety of handwriting styles. The training, validation, and test samples are much more extensive than existing Thai handwriting datasets.

5. Conclusions and Future Work

In this paper, we have presented the Burapha-TH Thai handwriting dataset. The 1072 participants wrote characters, digits, and syllables on standard collection sheets. We extracted dataset samples from 3076 sheets. These were passed through preprocessing consisting of deskewing, line detection, and image segmentation. The expert group eliminated some images.
Our proposed dataset covers 68 consonants and vowels, 10 Thai digits, and 320 Thai syllables. It contains three subsets: 76,927 images in the character dataset, 10,673 images in the digit dataset, and 268,056 images in the syllable dataset. The Burapha-TH dataset images are original, in JPG file format, true color, and without any de-noising or cleansing processing. The best performance shows 95.00%, 98.29%, and 96.16% accuracy using the VGG-13_BN model on the character, digit, and syllable data. Developing this Thai handwriting dataset is essential for improving Thai script recognition research. Our proposed dataset is available for downloading at https://services.informatics.buu.ac.th/datasets/Burapha-TH/ (accessed on 1 April 2022).
In future work, we will be publishing an extension of the present dataset, adding binarization and edge datasets, and expanding it to include more samples and syllable classes. We will also focus on model optimization for Thai handwriting recognition based on the Burapha-TH dataset.

Author Contributions

Conceptualization, H.J. and A.O.; methodology, A.O., T.K. and U.B.; software, T.K.; validation, A.O., H.J. and T.C.; formal analysis, A.O. and U.B.; investigation, A.O., H.J. and T.C.; writing—original draft preparation, A.O.; writing—review and editing, H.J., T.K., U.B. and T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data available in a publicly accessible repository that does not issue DOIs. Publicly available datasets were analyzed in this study. This data can be found here: https://services.informatics.buu.ac.th/datasets/Burapha-TH/ (accessed on 1 April 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Thai character classes considered for the dataset constructions.
Table A1. Thai character classes considered for the dataset constructions.
Consonants (44)
00-161-A1-KO KAI22-183-B7-THO THAHAN
01-162-A2-KHO KHAI23-184-B8-THO THONG
02-163-A3-KHO KHUAT24-185-B9-NO NU
03-164-A4-KHO KHWAI25-186-BA-BO BAIMAI
04-165-A5-KHO KHON26-187-BB-PO PLA
05-166-A6-KHO RAKHANG27-188-BC-PHO PHUNG
06-167-A7-NGO NGU28-189-BD-FO FA
07-168-A8-CHO CHAN29-190-BE-PHO PHAN
08-169-A9-CHO CHING30-191-BF-FO FAN
09-170-AA-CHO CHANG31-192-C0-PHO SAMPHAO
10-171-AB-SO SO32-193-C1-MO MA
11-172-AC-CHO CHOE33-194-C2-YO YAK
12-173-AD-YO YING34-195-C3-RO RUA
13-174-AE-DO CHADA36-197-C5-LO LING
14-175-AF-TO PATAK38-199-C7-WO WAEN
15-176-B0-THO THAN39-200-C8-SO SALA
16-177-B1-THO NANGMONTHO40-201-C9-SO RUSI
17-178-B2-THO PHUTHAO41-202-CA-SO SUA
18-179-B3-NO NEN42-203-CB-HO HIP
19-180-B4-DO DEK43-204-CC-LO CHULA
20-181-B5-TO TAO44-205-CD-O ANG
21-182-B6-THO THUNG45-206-CE-HO NOKHUK
Vowels (20)
46-207-CF-PAIYANNOI56-217-D9-SARA UU
47-208-D0-SARA A57-224-E0-SARA E
48-209-D1-MAI HAN-AKAT58-225-E1-SARA AE
49-210-D2-SARA AA59-226-E2-SARA O
50-211-D3-SARA AM60-227-E3-SARA AI MAIMUAN
51-212-D4-SARA I61-228-E4-SARA AI MAIMALAI
52-213-D5-SARA II62-231-E7-MAITAIKHU
53-214-D6-SARA UE67-236-EC-THANTHAKHAT
54-215-D7-SARA UEE35-196-C4-RU
55-216-D8-SARA U37-198-C6-LU
Tone Markers (4)
63-232-E8-MAI EK65-234-EA-MAI TRI
64-233-E9-MAI THO66-235-EB-MAI CHATTAWA
Table A2. Thai digits classes considered for the dataset constructions.
Table A2. Thai digits classes considered for the dataset constructions.
Digits (10)
68-240-F0-THAI DIGIT ZERO73-245-F5-THAI DIGIT FIVE
69-241-F1-THAI DIGIT ONE74-246-F6-THAI DIGIT SIX
70-242-F2-THAI DIGIT TWO75-247-F7-THAI DIGIT SEVEN
71-243-F3-THAI DIGIT THREE76-248-F8-THAI DIGIT EIGHT
72-244-F4-THAI DIGIT FOUR77-249-F9-THAI DIGIT NINE
Table A3. Thai Syllable classes considered for the dataset constructions.
Table A3. Thai Syllable classes considered for the dataset constructions.
Syllables (320)
เก็บ000-KEP1เมื่อ31-MUEA2แรง62-RAENG0
เกิด001-KOET1เมือง32-MUEANG0แล้ว63-LAEO3
เกิน002-KOEN0เรา33-RAO0และ64-LAE3
เขต003-KHEET1เริ่ม34-ROEM2แห่ง65-HAENG1
เขา004-KHAO4เรียก35-RIAK2โดย66-DOI0
เข้า005-KHAO2เรียน36-RIAN0โต67-TOO0
เขียน006-KHIAN4เรื่อง37-RUEANG2โรค68-ROOK2
เครื่อง007-KHRUEANG2เล็ก38-LEK3โลก69-LOOK2
เงิน008-NGOEN0เล่น39-LEN2ใกล้70-KLAI2
เจอ009-JOOE0เลย40-LOEI0ใคร71-KHRAI0
เจ้า010-JAAO2เล่า41-LAO2ใจ72-JAI0
เช่น011-CHEN2เลือก42-LUEAK2ใช่73-CHAI2
เชิง012-CHOENG0เสีย43-SIA4ใช้74-CHAI3
เชื่อ013-CHUEA2เสียง44-SIANG4ใด75-DAI0
เด็ก014-DEK1เหตุ45-HEET1ใน76-NAI0
เดิน015-DOEN0เห็น46-HEN4ใบ77-BAI0
เดิม016-DOEM0เหมือน47-MUEAN4ใส่78-SAI1
เดียว017-DIAO0เหลือ48-LUEA4ให้79-HAI2
เดือน018-DUEAN0เอง49-EENG0ใหญ่80-YAI1
เท่า019-THAO0เอา50-AO0ใหม่81-MAI1
เธอ20-THOOE0แก51-KAAE0ไง82-NGAI0
เป็น21-PEN0แก่52-KAAE1ได้83-DAI2
เปล่า22-PLAAO1แก้53-KAAE2ไทย84-THAI0
เปลี่ยน23-PLIAN1แดง54-DAENG0ไป85-PAI0
เปิด24-POOET1แต่55-TAAE1ไม่86-MAI2
เพราะ25-PHROR3แทน56-THAAEN0ไม้87-MAI3
เพลง26-PHLEENG0แนว57-NAAEO0ไว้88-WAI3
เพิ่ม27-PHOEM2แบบ58-BAEP0ไหน89-NAI4
เพียง28-PHIANG0แพทย์59-PHAET2ไหม90-MAI4
เพื่อ29-PHUEA2แม่60-MAE2ก็91-KOR2
เพื่อน30-PHUEAN2แรก61-RAEK2กลับ92-KLAP1
Syllables
กลาง93-KLANG0จะ132-JA1ติด171-TIT1
กล่าว94-KLAAO1จัด133-JAT1ถ้า172-THAA2
กลุ่ม95-KLUM1จับ134-JAP1ถาม173-THAAM4
กว่า96-KWAA1จาก135-JAAK1ถึง174-THUENG4
ก่อน97-KORN1จิต136-JIT1ถือ175-THUE4
กัน98-KAN0จีน137-JIIN0ถูก176-THUUK1
กับ99-KAP1จึง138-JUENG0ทรง177-SONG0
การ100-KAAN0จุด139-JUT1ทราบ178-SAAP2
กิน101-KIN0ฉัน140-CHAN4ทั้ง179-THANG3
ขอ102-KHOR4ช่วง141-CHUANG2ทั่ว180-THAW1
ข้อ103-KHOR2ช่วย142-CHUAI2ทาง181-THAANG0
ของ104-KHORNG4ชอบ143-CHORP2ท่าน182-THAAN2
ข้าง105-KHAANG2ชั้น144-CHAN3ทำ183-THAM0
ขาด106-KHAD0ชาติ145-CHAAT2ที184-THII0
ขาย107-KHAAI4ชาย146-CHAAI0ที่185-THII2
ขาว108-KHAAO4ชาว147-CHAAO0นอก186-NORK2
ข่าว109-KHAAO1ชื่อ148-CHUE2น้อง187-NORNG3
ข้าว110-KHAAO2ชุด149-CHUT3นอน188-NORN0
ขึ้น111-KHUEN2ซึ่ง150-SUENG2น้อย189-NOI3
คง112-KHONG0ซื้อ151-SUE3นะ190-NA3
คน113-KHON0ด้วย152-DUAI2นัก191-NAK3
ครั้ง114-KHRANG3ดัง153-DANG0นั่ง192-NANG2
ครับ115-KHRAP3ด้าน154-DAAN2นั้น193-NAN3
ครู116-KHRUU0ดี155-DII0นา194-NAA0
ควร117-KHUAN0ดู156-DUU0น่า195-NAA2
ความ118-KHWAAM0ตน157-TON0นาน196-NAAN0
ค่ะ119-KHA2ต้น158-TON2นาย197-NAAI0
ค่า120-KHAA2ตรง159-TRONG0นำ198-NAM0
คำ121-KHAM0ตรวจ160-TRUAT1น้ำ199-NAAM3
คิด122-KHIT3ต่อ161-TOR1นี่200-NII2
คืน123-KHUEN0ต้อง162-TORNG2นี้201-NII3
คือ124-KHUE0ตอน163-TORN0บท202-BOT1
คุณ125-KHUN0ตอบ164-TORP0บน203-BON0
คู่126-KHUU2ตั้ง165-TANG2บอก204-BORK1
งาน127-NGAAN0ตัว166-TUA0บาง205-BAANG0
ง่าย128-NGAAI2ตา167-TAA0บาท206-BAAT1
จน129-JON0ต่าง168-TAANG1บ้าน207-BAAN2
จบ130-JOP1ตาม169-TAAM0ปรับ208-PRAP1
จริง131-JING0ตาย170-TAAI0ปล่อย209-PLOI1
Syllables
ปลา210-PLAA0ยิ่ง247-YING2สาย284-SAAI4
ปาก211-PAAK1ยิ้ม248-YIM3สาว285-SAAO4
ปี212-PII0ยืน249-YUEN0สิ286-SI1
ผม213-PHOM4ยุค250-YUK3สิ่ง287-SING1
ผล214-PHON4รถ251-ROT3สิบ288-SIP1
ผ่าน215-PHAAN1รวม252-RUAM0สี289-SII4
ผิด216-PHIT1ร่วม253-RUAM2สุข290-SUK1
ผู้217-PHUU2รอ254-ROR0สู่291-SUU1
ฝ่าย218-FAAI1รอบ255-RORP2สูง292-SUUNG4
พบ219-PHOP3รัก256-RAK3หญิง293-YING4
พรรค220-PHAK3รัฐ257-RAT3หน้า294-NAA2
พร้อม221-PHRORM3รับ258-RAP3หนึ่ง295-NUENG1
พระ222-PHRA3ร่าง259-RAANG2หนู296-NUU4
พวก223-PHUAK2ร้าน260-RAAN3หมด297-MOT1
พอ224-PHOR0ราย261-RAAI0หมอ298-MOR4
พ่อ225-PHOR2รีบ262-RIIP2หรือ299-RUE4
พา226-PHAA0รู้263-RUU3หลัก300-LAK1
พี่227-PHII2รูป264-RUUP2หลัง301-LANG4
พูด228-PHUUT2ลง265-LONG0หลาย302-LAAI4
ฟัง229-FANG0ลด266-LOT3ห้อง303-HORNG2
ภาค230-PHAAK2ละ267-LA3หัน304-HAN4
ภาพ231-PHAAP2ล่ะ268-LA4หัว305-HUA4
มอง232-MORNG0ล้าน269-LAAN3หา306-HAA4
มัก233-MAK3ลูก270-LUUK2หาก307-HAAK1
มัน234-MAN0วัด271-WAT3หาย308-HAAI4
มา235-MAA0วัน272-WAN0อย่า309-YAA1
มาก236-MAAK2ว่า273-WAA2อยาก310-YAAK1
มิ237-MI3วาง274-WAANG0อย่าง311-YAANG1
มี238-MII0ส่ง275-SONG1อยู่312-YOO1
มือ239-MUE0สร้าง276-SAANG2ออก313-ORK1
ยก240-YOK3ส่วน277-SUAN1อัน314-AN0
ยอม241-YORM0สวย278-SUAI4อา315-ARE0
ย่อม242-YORM2สอง279-SORNG4อาจ316-AAT1
ยัง243-YANG0สอน280-SORN4อีก317-IIK1
ยา244-YAA0สัก281-SAK1อ่าน318-AAN1
ยาก245-YAAK2สัตว์282-SAT1อื่น319-UEN1
ยาว246-YAAO0สาม283-SAAM4

References

  1. Singh, A.; Bacchuwar, K.; Bhasin, A. A survey of OCR applications. Int. J. Mach. Learn. Comput. 2012, 2, 314. [Google Scholar] [CrossRef] [Green Version]
  2. Jangid, M.; Srivastava, S. Handwritten Devanagari character recognition using layer-wise training of deep convolutional neural networks and adaptive gradient methods. J. Imaging 2018, 4, 41. [Google Scholar] [CrossRef] [Green Version]
  3. Ahlawat, S.; Choudhary, A.; Nayyar, A.; Singh, S.; Yoon, B. Improved handwritten digit recognition using convolutional neural networks (CNN). Sensors 2020, 20, 3344. [Google Scholar] [CrossRef] [PubMed]
  4. Arora, S.; Bhatia, M.S. Handwriting recognition using deep learning in keras. In Proceedings of the 2018 International Conference on Advances in Computing, Communication Control and Networking (ICACCCN), Greater Noida, India, 12–13 October 2018; pp. 142–145. [Google Scholar]
  5. Vaidya, R.; Trivedi, D.; Satra, S.; Pimpale, P.M. Handwritten Character Recognition Using Deep-Learning. In Proceedings of the second International Conference on Inventive Communication and Computational Technologies (ICICCT), Coimbatore, India, 20–21 April 2018; pp. 772–775. [Google Scholar] [CrossRef]
  6. Eltay, M.; Zidouri, A.; Ahmad, I. Exploring deep learning approaches to recognize handwritten Arabic texts. IEEE Access 2020, 8, 89882–89898. [Google Scholar] [CrossRef]
  7. LeCun, Y.A. The MNIST Database of Handwritten Digits. 1998. Available online: http://yann.lecun.com/exdb/mnist/ (accessed on 1 March 2021).
  8. Cohen, G.; Afshar, S.; Tapson, J.; Van Schaik, A. EMNIST: Extending MNIST to handwritten letters. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 2921–2926. [Google Scholar] [CrossRef]
  9. Liu, C.L.; Yin, F.; Wang, D.H.; Wang, Q.F. CASIA online and offline Chinese handwriting databases. In Proceedings of the International Conference on Document Analysis and Recognition, Beijing, China, 18–21 September 2011; pp. 37–41. [Google Scholar] [CrossRef]
  10. Clanuwat, T.; Bober-Irizar, M.; Kitamoto, A.; Lamb, A.; Yamamoto, K.; Ha, D. Deep learning for classical Japanese literature. arXiv 2018, arXiv:1812.01718. [Google Scholar]
  11. Manjusha, K.; Kumar, M.A.; Soman, K.P. On developing handwritten character image database for Malayalam language script. Eng. Sci. Technol. Int. J. 2019, 22, 637–645. [Google Scholar] [CrossRef]
  12. Kim, I.J.; Xie, X. Handwritten Hangul recognition using deep convolutional neural networks. Int. J. Doc. Anal. Recognit. (IJDAR) 2015, 18, 1–13. [Google Scholar] [CrossRef]
  13. KIM, D.H.; Hwang, Y.S.; Park, S.T.; Kim, E.J.; Paek, S.H.; BANG, S.Y. Handwritten Korean character image database PE92. IEICE Trans. Inf. Syst. 1996, 79, 943–950. [Google Scholar]
  14. Sae-Tang, S.; Methasate, L. Thai handwritten character corpus. IEEE Int. Symp. Commun. Inf. Technol. 2004, 1, 486–491. [Google Scholar] [CrossRef]
  15. Surinta, O.; Karaaba, M.F.; Schomaker, L.R.; Wiering, M.A. Recognition of handwritten characters using local gradient feature descriptors. Eng. Appl. Artif. Intell. 2015, 45, 405–414. [Google Scholar] [CrossRef]
  16. Liu, C.L.; Nakagawa, M. Evaluation of prototype learning algorithms for nearest-neighbor classifier in application to handwritten character recognition. Pattern Recognit. 2001, 34, 601–615. [Google Scholar] [CrossRef]
  17. Ciresan, D.C.; Meier, U.; Gambardella, L.M.; Schmidhuber, J. Convolutional neural network committees for handwritten character classification. In Proceedings of the 2011 International Conference on Document Analysis and Recognition, Beijing, China, 18–21 September 2011; pp. 1135–1139. [Google Scholar] [CrossRef] [Green Version]
  18. Pratt, S.; Ochoa, A.; Yadav, M.; Sheta, A.; Eldefrawy, M. Handwritten digits recognition using convolution neural networks. J. Comput. Sci. Coll. 2019, 34, 40–46. [Google Scholar]
  19. Michie, D.; Spiegelhalter, D.J.; Taylor, C.C. Machine Learning, Neural and Statistical Classification; Ellis Horwood Ltd.: Chichester, UK, 1994. [Google Scholar]
  20. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  21. Uppakitsinlapasn. Principles of Thai Language—Akkhawawithi, Wachiwiphak, Wakkayasamphan, Chanthalak. Thai Wattana Phanit Publisher: Bangkok, Thailand, 1931. (In Thailand) [Google Scholar]
  22. Iwasaki, S.; Ingkaphirom, P.; Horie, I.P. A Reference Grammar of Thai; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  23. Koanantakool, H.T.; Karoonboonyanan, T.; Wutiwiwatchai, C. Computers and the Thai language. IEEE Ann. Hist. Comput. 2009, 31, 46–61. [Google Scholar] [CrossRef]
  24. Flachot, A.; Gegenfurtner, K.R. Color for object recognition: Hue and chroma sensitivity in the deep features of convolutional neural networks. Vis. Res. 2021, 182, 89–100. [Google Scholar] [CrossRef] [PubMed]
  25. Bradski, G. The open CV library. Dr. Dobb’s J. Softw. Tools Prof. Program. 2000, 25, 120–123. [Google Scholar]
  26. Ertuna, L. (n.d.) Open CV Library: GitHub—JPLeoRX/Opencv-text-deskew. Available online: https://github.com/JPLeoRX/opencv-text-deskew (accessed on 1 December 2021).
Figure 1. The Thai sentence in four-line zones style.
Figure 1. The Thai sentence in four-line zones style.
Applsci 12 04083 g001
Figure 2. Thai Pangram.
Figure 2. Thai Pangram.
Applsci 12 04083 g002
Figure 3. Statistics of age distribution about the writers.
Figure 3. Statistics of age distribution about the writers.
Applsci 12 04083 g003
Figure 4. Data Collection Sheets: (a) depicts three parts of collected information (b) depicts the case of recording the characters, vowels, and digits and the case of rewriting any incorrect writing.
Figure 4. Data Collection Sheets: (a) depicts three parts of collected information (b) depicts the case of recording the characters, vowels, and digits and the case of rewriting any incorrect writing.
Applsci 12 04083 g004
Figure 5. Overall procedure of data preparation process: (a) is an original image from a document scanner, (b) depicts output images from Algorithm 1, (c) shows line detection process from Algorithm 2, and (d) depicts output from Algorithm 3.
Figure 5. Overall procedure of data preparation process: (a) is an original image from a document scanner, (b) depicts output images from Algorithm 1, (c) shows line detection process from Algorithm 2, and (d) depicts output from Algorithm 3.
Applsci 12 04083 g005
Figure 6. Example of eliminated images: (a) depicts incorrect writing, (b) depicts unreadable images, (c) shows examples of heavily distorted images, and (d) shows over edge cuts.
Figure 6. Example of eliminated images: (a) depicts incorrect writing, (b) depicts unreadable images, (c) shows examples of heavily distorted images, and (d) shows over edge cuts.
Applsci 12 04083 g006
Figure 7. Examples of handwriting styles included in Burapha-TH.
Figure 7. Examples of handwriting styles included in Burapha-TH.
Applsci 12 04083 g007
Figure 8. Result of BURAPHA-TH dataset ranking K-fold cross-validation.
Figure 8. Result of BURAPHA-TH dataset ranking K-fold cross-validation.
Applsci 12 04083 g008
Table 1. The 17 Thai vowel forms based on Thai keyboard.
Table 1. The 17 Thai vowel forms based on Thai keyboard.
Before (5)After (5)Above (5)Below (2)
Table 2. The statistics of raw, segmented, and current data of Burapha-TH Dataset.
Table 2. The statistics of raw, segmented, and current data of Burapha-TH Dataset.
CategoryCollection SheetsSegmented ImagesImage Dataset
Character + Digit1156107,50687,600
Syllable1920279,730268,056
Total3076387,236355,656
Table 3. Number of images in our proposed Burapha-TH datasets.
Table 3. Number of images in our proposed Burapha-TH datasets.
Statistic TopicsCharacter
Dataset
Digit
Dataset
Syllable
Dataset
Number of Class6810320
Number of Train sample63,3278673236,056
Number of Test sample13,600200032,000
Average sample in Train/class931867738
Min—Max sample in Train/class790–995772–923503–905
Number of samples in Test/class200200100
Total3076387,236355,656
Table 4. An overview of the public datasets discussed in the paper and Burapha-TH dataset.
Table 4. An overview of the public datasets discussed in the paper and Burapha-TH dataset.
DatasetRecognizerRecognition Accuracy (%)
TrainingValidationTesting
CNN (4 conv)87.3475.2878.51
CharacterLeNet-588.1675.4278.84
VGG_13_BN97.8892.6495.00
CNN (4 conv)94.0183.0083.93
DigitLeNet-595.4085.6387.26
VGG_13_BN98.5194.0398.29
CNN (4 conv)83.6175.0078.25
SyllableLeNet-576.1571.6474.98
VGG_13_BN98.4194.4096.16
Table 5. A statistics comparison of Burapha-TH datasets and public datasets discussed in the paper.
Table 5. A statistics comparison of Burapha-TH datasets and public datasets discussed in the paper.
DatasetLanguageYearContentClassNo. of WritersStatistics
(Train/Validate/Test)
MNISTEN1998Digit10-60,000/0/10,000
EMNISTEN2017Letter26>500124,800/0/20,800
CASIA-HWDBCN2011Text,
Character
735610203.5 M isolated character, 1.35 M characters in text
Kuzushiji-49JP2016Character49-232,365/0/38,547
MalayalamIN2019Character857717,236/5706/6360
PE92KR1992Syllable2350 ~100 per class
SERI95KR1997Syllable520 465,675/0/51,785
Thai handwritten character corpus 1TH2004Character7914314,000
ALICE-THITH2015Character6815013,138/0/1360
ALICE-THITH2015Digit101508555/0/1000
Burapha-THTH2021Character68107263,327/0/13,600
Burapha-THTH2021Digit108673/0/2000
Burapha-THTH2021Syllable320236,056/0/32,000
1 dataset not available.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Onuean, A.; Buatoom, U.; Charoenporn, T.; Kim, T.; Jung, H. Burapha-TH: A Multi-Purpose Character, Digit, and Syllable Handwriting Dataset. Appl. Sci. 2022, 12, 4083. https://doi.org/10.3390/app12084083

AMA Style

Onuean A, Buatoom U, Charoenporn T, Kim T, Jung H. Burapha-TH: A Multi-Purpose Character, Digit, and Syllable Handwriting Dataset. Applied Sciences. 2022; 12(8):4083. https://doi.org/10.3390/app12084083

Chicago/Turabian Style

Onuean, Athita, Uraiwan Buatoom, Thatsanee Charoenporn, Taehong Kim, and Hanmin Jung. 2022. "Burapha-TH: A Multi-Purpose Character, Digit, and Syllable Handwriting Dataset" Applied Sciences 12, no. 8: 4083. https://doi.org/10.3390/app12084083

APA Style

Onuean, A., Buatoom, U., Charoenporn, T., Kim, T., & Jung, H. (2022). Burapha-TH: A Multi-Purpose Character, Digit, and Syllable Handwriting Dataset. Applied Sciences, 12(8), 4083. https://doi.org/10.3390/app12084083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop