Next Article in Journal
Semi-Analytical Approach in BiER4BP for Exploring the Stable Positioning of the Elements of a Dyson Sphere
Previous Article in Journal
Vortex Dynamics in a Hybrid Aligned Nematic Microvolume with an Orientational Defect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Revival of Classical Algorithms: A Bibliometric Study on the Trends of Neural Networks and Genetic Algorithms

Department of Management Information Systems, National Chengchi University, Taipei 116302, Taiwan
*
Author to whom correspondence should be addressed.
Symmetry 2023, 15(2), 325; https://doi.org/10.3390/sym15020325
Submission received: 25 October 2022 / Revised: 26 December 2022 / Accepted: 17 January 2023 / Published: 23 January 2023

Abstract

:
The purpose of our bibliometric research was to capture and analyze the trends of two types of well-known classical artificial intelligence (AI) algorithms: neural networks (NNs) and genetic algorithms (GAs). Symmetry is a very popular international and interdisciplinary scientific journal that cover six major research subjects of mathematics, computer science, engineering science, physics, biology, and chemistry which are all related to our research on classical AI algorithms; therefore, we referred to the most innovative research articles of classical AI algorithms that have been published in Symmetry, which have also introduced new advanced applications for NNs and Gas. Furthermore, we used the keywords of “neural network algorithm” or “artificial neural network” to search the SSCI database from 2002 to 2021 and obtained 951 NN publications. For comparison purposes, we also analyzed GA trends by using the keywords “genetic algorithm” to search the SSCI database over the same period and we obtained 878 GA publications. All of the NN and GA publication results were categorized into eight groups for deep analyses so as to investigate their current trends and forecasts. Furthermore, we applied the Kolmogorov–Smirnov test (K–S test) to check whether our bibliometric research complied with Lotka’s law. In summary, we found that the number of applications for both NNs and GAs are continuing to grow but the use of NNs is increasing more sharply than the use of GAs due to the boom in deep learning development. We hope that our research can serve as a roadmap for other NN and GA researchers to help them to save time and stay at the cutting edge of AI research trends.

1. Introduction

1.1. Neural Networks

The history of neural networks dates back to the 1940s. The research article of McCulloch and Pitts published in 1943 is often regarded as the beginning of the neural network concept [1]. During the 1950s and 1960s, the first neuro-computer was developed by Frank Rosenblatt and others [2]. Since then, neural networks have continued to gain attention, especially with the rise of deep learning.
Neural networks are one of the many forms of classical artificial intelligence algorithms, also known as artificial neural networks (ANNs). The inspiration for NNs arose from the sophisticated functionality of human brains, with the aim of simulating the way neurons and other nerve cells transmit electrical impulses to each other. NN are composed of three types of layers: an input layer, more than one intermediate layers (hidden layers), and an output layer [3]. Each NN node is connected to another, and they have associated weights and thresholds. When the output of a node is higher than a certain threshold, the specified node is activated and the information is transmitted to the next layer of the NN; otherwise, no information is transmitted to the next layer.
It should be noted that neural network algorithms rely on huge amounts of data to train their learning models and subsequently improve their accuracy. Nevertheless, when these learning models are fine-tuned to achieve high accuracy, they become powerful artificial intelligence tools that can classify and cluster data at extremely high speeds. Taking image recognition as an example, neural network algorithms only take a few minutes to process image recognition data, while human experts need hours for manual image recognition [4].

1.2. Genetic Algorithms

GAs are another form of classical artificial intelligence algorithms which are usually used to solve process optimization problems, such as travel time optimization for the transportation industry and the optimization of computer architectures, computer networks, and investment portfolios. Hybrid solutions involving both NNs and GAs have also been developed, for instance, for parameter selection in machine learning model training [5]. GAs were based on the concept of Charles Darwin’s biological evolution theory of natural selection and apply the methodologies of evolution, including inheritance, natural selection, mutation, and crossover. On the basis of Darwin’s most famous rule of the “survival of the fittest”, GAs simulate the evolutionary mechanisms of natural selection to produce the fittest options and reproduce the best descendants through mutations and crossovers [6]. In other words, the better genes from the fittest organisms are naturally selected and evolve to produce more superior descendants, i.e., organisms evolve to become progressively fitter for their specific environments [7,8].
GAs became more well-known with the assistance of John Holland’s research in 1975. Since then, GAs have been applied in a great deal of research domains, including assembly line balancing, factory layout problems, project scheduling, travel scheduling, and in particular, global optimization problems [9,10].
GAs are heuristic algorithms that simulate the evolutionary mechanism of natural selection to produce the fittest options, replace unfit solutions, and reproduce the best descendants through mutation and crossovers. The process of natural selection is repeated until the best result has been produced [8].

1.3. Genetic Algorithms and Neural Networks

Genetic algorithms and neural networks are both forms of classical algorithms, but GAs have been more popular in the past because they do not require a large amount of data for processing. AI has achieved exceptional accomplishments in recent years, mainly because of the rise of NN-based machine learning. NNs have now become more popular and have become ingrained in our everyday lives. Nevertheless, GAs are still widely used today and the number of applications for GAs has been steadily growing in science, engineering, social sciences, and numerous other fields. An increasing number of AI applications combine GAs and NNs to unlock new solutions. The new combined solutions breathe new life into the field of AI solutions and can speed up the entire process of solving specific problems [11,12].

2. Literature Review

2.1. Applications for Neural Networks

Neural networks are used for a wide range of applications, including applications in computing, science, engineering, technology, environmental, agriculture, mining, climate, business, arts, and more. Some modern applications for neural networks have also been developed, including image recognition, speech recognition, character recognition, machine translation, stock market predictions, medical diagnosis [1], counterfeit products classification [13], sound-absorbing board optimization [14], the optimization of channel equalization in wireless communication [15], the optimization of IoT intrusion detection system performance [16], diabetic retinopathy classification [17], the mitigation of cybersecurity alert fatigue [18], the optimization of hand gesture recognition [19], optimization of indoor human activity recognition [20], the optimization of multi-type object tracking [21], the mitigation of global warming impacts on marine ecosystems [22], the optimization of oil layer prediction [23], the identification of the optimal parameters for recurrent neural network training [24] and the optimal classification of noises in QR codes [25]. Among all these NN applications, the most successful and famous application for neural networks is image recognition, which is based on the convolutional neural network (CNN) architecture. We discuss this in more detail later in the paper.
There are diverse applications for neural networks, but some of them have become more popular in every everyday life, a few examples of which are listed below.
  • Facial Recognition
Facial recognition systems are becoming more popular as intelligent surveillance systems. Facial recognition solutions map out human faces and compare them to previously stored images. These systems are often used for specific entrances in buildings to authenticate human faces and match them up with the data that are stored in their facial image databases.
As stated above, CNNs are often used for image recognition processing. For this, a large number of images of human faces are imported into a database to train the CNNs, although the stored human face images need further processing before the CNN training. CNN models are based on linear algebra principles and matrix multiplication is key for the representation of data and weights. The basic structure of CNNs comprise three layers: a convolutional layer, a pooling layer, and a fully connected layer. The convolutional layer consists of many filters and features and applies those filters to the input to generate feature maps using an activation function. The second layer is the pooling layer, the function of which is to downsample the feature maps. Finally, the fully connected layer, which is also known as the classifier, consists of an activation function to classify the images [26].
Below are some examples of advanced applications for neural networks within facial recognition:
  • The aggregation of the spatial and temporal convolutional features of CNN for video-based facial expression recognition [27].
  • The use of hybrid data augmentation and lightweight convolutional neural network solutions to estimate the age of faces in images [28].
  • Improved accuracy in facial emotion recognition using new hybrid HOG-ESRs (histogram of oriented gradient-ensembles with shared representations) algorithm [29].
2.
Speech recognition
Speech recognition is another example of technology that has now become so common in our daily lives that is used without a second thought. Through popular voice-controlled systems, such as Amazon’s Alexa and Apple’s Siri, automatic speech recognition (ASR) is scattered throughout our computers, tablet devices, home speakers, and cellar phones. ASR technology is now also being used in businesses, for instance, in customer service and many other areas. Enterprises have immediately embraced ASR technology as a means to develop operational efficiencies and simplify processes.
Neural networks have been applied in speech recognition technology for some time. Recurrent neural network (RNNs) are a unique type of NN that have the concept of ”memory” which were originally designed for sequential data processing. RNNs have been employed to develop speech recognition systems with dramatic success. Moreover, a special and powerful RNN architecture with improved memory has been developed which is known as long short-term memory (LSTM). LSTM has proven to be especially effective in speech recognition [30,31].
Below are some examples of advanced applications for neural networks within speech recognition:
  • A convolutional neural network (CNN) structure that can divide continuous speech into syllables by converting sound patterns into the form of RGB (red, green, blue) images [32];
  • A hybrid CNN and bi-directional long short-term memory (BiLSTM) model for a cross-language end-to-end speech recognition system based on transfer learning [33];
  • A hybrid residual time delay neural network (ResTDNN) and BiLSTM model for slot filling in speech recognition [34].

2.2. Applications for Genetic Algorithms

Numerous researchers and scientists have applied genetic algorithms in mathematics, computer science, engineering, business, finance, economics, and social sciences. Many scientists have applied GA concepts to computer science problems, such as the optimization of computer architectures [35,36]. Another important topic of study is the optimization of distributed computer networks. The main objective of this research is to minimize the costs of designing distributed computer network topologies [37]. Other applications have also been developed, such as file allocation methods for distributed systems, signal processing and filtering, hardware debugging, and more. In the finance sector, GAs are used for the automation of sophisticated trading systems, the valuation of real options, and portfolio optimization [38,39]. GAs are also used to solve problems in industry, management, and engineering, for instance, airline revenue management, container loading optimization, automated planning of structural inspection, control engineering, mechanical engineering, marketing mix analysis, optimization of mobile communications infrastructure, plant floor layout design, quality control management, bearing placement optimization, and the identification of optimal routes for multiple vehicles [40,41,42,43,44]. Clearly, GAs can be applied to solve a broad range of general solutions [45,46,47].
Below are some examples of advanced applications for genetic algorithms:
  • A novel aggregated multi-objective genetic algorithm (MOGA) and a variant particle swarm optimization (PSO) solution for 5G software-defined networking (SDN) architecture [48];
  • Combined support vector machine (SVM) and GA solutions for the detection of cyber-attacks and malicious intrusions [49];
  • An integrated auto-generated multi-attribute design structure matrix (DSM) and GA solution for modular design [50];
  • A GA mechanism that can optimize proof-of-work (PoW) blockchains parameters [51].

2.3. Applications for Combination of Genetic Algorithms and Neural Networks

The first attempt at combining genetic algorithms and neural networks was in the late 1980s. Since then, many researchers and scholars have published articles about applications for genetic algorithms and neural networks (GANNs). Various combined genetic algorithm and neural network approaches have been explored, including solar system performance prediction [11], face recognition [52], animation [53], color recipe prediction [54], crude fractional distillation processing [55], and more. In fact, several combinations of genetic algorithms and neural networks have been developed by researchers. Schaffer et al. found that these GANN combinations can be categorized into two different groups: supportive combinations and collaborative combinations. NNs and GAs are applied sequentially in supportive combinations and simultaneously in collaborative combinations [12,56,57]. GAs can also be used in machine learning for feature selection in recurrent neural networks and for training artificial neural networks when pre-classified examples are not available [58].
Below are some examples of advanced applications for NN and GA combinations:
  • A hybrid cascade neural network and metaheuristic optimization genetic algorithm for space–time forecasting [59];
  • A combined GA and back-propagation (BP) neural network for early warning systems in coal mines. [60];
  • An integrated NN and GA framework for sequential semi-supervised classification [61].

2.4. Laws of Bibliometrics

Statistical and mathematical methods are usually applied to analyze information about publications which involve three important laws of bibliometrics: Lotka’s law, Bradford’s law, and Zipf’s law. These three classical laws are the pillars of bibliometrics and are often applied in bibliometric research to test the applicability of publications [62,63,64].

2.4.1. Lotka’s Law

Lotka’s law determines the productivity of authors within specified research areas. The law states that 60% of authors in a specified research area only publish one article, 15% publish two articles (1/2² × 0.60), 7% publish three articles (1/3² × 0.60), etc. When Lotka’s law is applied to several articles over a specified time period, the results are similar but not very accurate. Lotka’s law is used to assess the frequency with which authors publish articles online [65,66,67]. Lotka’s law is also deemed to be quite useful for studying the productivity patterns of authors in bibliographies. In this research, we selected articles published between 2002 and 2021 and performed an author productivity check to observe the NN and GA research trends and forecasts. For verification purposes, we also applied the K–S test to check whether our results complied with Lotka’s inverse square law [68,69,70].

2.4.2. Zipf’s Law

Zipf’s law is regularly described as the word frequency law and is used to identify the frequency distribution of words within articles. It states that the most frequent word in a long article is used twice as often as the second most frequent word and three times as often as the third most frequent word, etc. Zipf’s law is not statistically accurate but it is still quite valuable for bibliometric researchers [71,72].

2.4.3. Bradford’s Law

Bradford’s law is a universal bibliometric principle that determines the number of core journals within specified research area. Bradford’s law separates the core journals within a research area into three categories with an equal number of articles. The first category comprises the core journals on the subject, which tend to be fewer in number but produce approximately one-third of the articles. The second category comprises the same number of articles, but a larger number of journals. The third category comprises the same number of articles but a much larger number of journals. Bradford found that the relationships between the journals in the first category and those in the second and third category were proportional (1: n: n²). Bradford studied 326 geophysics journals and their bibliographies and found that 429 articles were included in 9 journals, 499 articles were included in 59 journals, and 404 articles were included in 258 journals; the proportion was 9: 59: 258, which was similar to 9: 45: 225. Based on this, it can be seen that Bradford’s law is not very precise but it is still deemed to be a general bibliometric principle by many researchers [71,73].

3. Materials and Methods

3.1. Research Methods

In this research, we studied articles published from 2002 to 2021. All of the data that we collected were from the Social Science Citation Index (SSCI) on the Web of Science platform. SSCI indexes over 3400 internationally famous social sciences journals across over 50 social sciences research areas and allows scholars to search articles using different criteria, such as subject, citation, country, territory, document type, institution name, research area, source title, journal, author, etc. [73].
Bibliometric methods are types of study systems that are often used in library science and informatics research and use statistical analyses to identify publication patterns in specific research areas. For instance, scholars can use bibliometric methods to assess the impacts of the relationships between one or more writers or articles. Analytical bibliometric methods mostly involve accessing SSCI and/or the Science Citation Index (SCI) databases to collect data, trace citations, and assess impact. Analytical bibliometric methods can help to identify innovative fields of study and possible research directions [72,73].

3.2. Research Analysis and Discussion

3.2.1. Analysis by Publication Year

We first found an obvious ascending trend in the number of NN articles published since 2015 and a smoother ascending trend in the number of GA articles published since 2016 (Figure 1). The rapid upward trend of NN articles could be explained by the third wave of artificial intelligence development. In 2015, Microsoft used a deep learning technique (i.e., a deep neural network) to outperform humans in image recognition, which sparked the third wave of AI advancement (He et al., 2016) [74]. Based on this, we could also conclude that the smoother ascending trend of GA articles published since 2016 was due to the revolutionary deep learning technique also breathing new life into GA research. Second, we concluded from the results that NN research and development have been rapidly and continuously growing for quite some time. Furthermore, although GAs have been around for a long time, they are still keeping up with the times and developing steadily.

3.2.2. Analysis by Citation

The trends of NN and GA citations both increased rapidly from 2018, which could also be explained by the third wave of AI development (Figure 2). We concluded that the ascending trends of NN and GA citations will continue for some time as long as the AI industry is booming.

3.2.3. Analysis by Country/Territory

From Table 1 and Figure 3, it can be seen that China produced the highest quantity of NN and GA research publications, followed by the USA. Some people believe that whichever country first masters AI technology will rule the world and from our results, the significance of AI within the context of U.S.–China confrontation can clearly be seen. Iran and Taiwan were in third place for the number of NN and GA publications, although the results showed that they are now starting to catch up with other countries. Furthermore, we also found that the NN and GA publications from top three countries made up over 50% of the total number of articles and that AI research was concentrated in a few countries. We discovered that Turkey, South Korea, and Spain are also catching up quickly and becoming major contributors to NN and GA research. Figure 3 shows the top six countries for NN and GA publications.

3.2.4. Analysis by Institution Name

In Table 2, it can be seen that the University of Tehran was the top institution for NN research and the Islamic Azad University was the top institution for GA research. Both are institutions in Iran and both published over 20 articles (2.103% and 2.506%). The Universiti Malaya was the second-place institution for NN research with 17 articles (1.788%) and Hong Kong Polytech University was the second-place institution for GA research with 13 articles (1.481%). The third-place institution for NN research was the Islamic Azad University and the third-place institution for GA research was the Iran University of Science and Technology. Furthermore, we also found that Iran was the most productive country for NN and GA research.

3.2.5. Analysis by Document Type

From Table 3, it can be seen that 92.429% of NN research was in the form of articles (879 articles) and 96.128% of GA research was in the form of articles (844 articles). We concluded that articles were the major and most accepted document type for NN and GA research publications.

3.2.6. Analysis by Language

From Table 4, it can be seen that English was the major language for NN and GA research papers, with 941 and 872 articles (98.948% and 99.317%), respectively. As English is the predominant scientific language worldwide and is the most widely used language among scientists, we could conclude that English was the number one language used in NN and GA research publications.

3.2.7. Analysis by Research Area

From Table 5, the top three NN research areas can be identified: computer science, with 221 articles (23.239%); engineering, with 201 articles (21.136%); and environmental sciences ecology, with 186 articles (19.558%). Table 5 also shows the top three GA research areas: computer science, with 301 articles (34.282%); engineering, with 280 articles (31.891%); and operation research, with 228 articles (25.968%). From this, it can be seen that computer science and engineering were the major topics within NN and GA research and that environmental sciences ecology research is catching up due to the rapid development of ESG. Furthermore, many other important and potential research domains for NN and GA publications were identified, such as business economies, mathematics, psychology, transportation, information science, library science, energy, etc.

3.2.8. Analysis by Source Title

From Table 6, the top three NN research source titles can be identified: Sustainability, with 75 articles (7.886%); Expert Systems with Applications, with 41 articles (4.311%); and the International Journal of Environmental Research and Public Health, with 25 articles (2.629%). Table 6 also shows the top three GA research source titles: Expert Systems with Applications, with 63 articles (7.175%); Sustainability, with 44 articles (5.011%); and the Journal of Operational Research Society, with 25 articles (2.847%). We could use these data to foresee the major NN and GA research trends. Furthermore, many other important research source titles for NN and GA publications were identified, such as Neural Computing Applications, Energy Policy, PLoS ONE, Applied Soft Computing, Computational Economics, Computers Industrial Engineering, etc.

3.2.9. NN Articles Productivity Analysis Using Lotka’s Law

The procedures for certifying the validity of our results were as follows:
  • Gather data from SSCI database;
  • Calculate the distribution of author productivity in NN research;
  • Use the Lotka’s law formulae to calculate the n and c values;
  • Apply the Kolmogorov–Smirnov test to verify whether the results complied with Lotka’s law.

The Distribution of Author Productivity in NN Research

We obtained the author data for 951 NN articles from the SSCI database. Then, we aggregated and accumulated all of the author data to calculate the total number of authors who had published NN research (Table 7). We found that there was a total of 951 articles from 2950 authors, with an average of 3.1 (2950/951) authors per article.

The Use of the Lotka’s Law Formula to Obtain the n Value

Using the data presented in Table 5 and the Lotka’s law formula in Equation (1), we obtained the n value:
n = N X Y X Y N x 2 x 2
Using the data presented in Table 8 and n = 12, we obtained the value of n = −3.313028132:
n = 12 3.298263 86.70113095 12 7.464290 75.348250

The Use of the Lotka’s Law Formula to Obtain the c Value

Then, we also used the Lotka’s law formula in Equation (2) to obtain the c value:
c = 1 1 p 1 1 x n + 1 n 1 p n 1 + 1 2 p n + n 24 p 1 n + 1
p = 20, x = 1, 2, 3, 4, 5, 6,7,8,9,10,11 and 12.
Using the above data, we obtain the value of c = 0.868637581:
c   = 1 1 20 1 1 x 3.313028132 + 1 3.313028132 1 20 3.313028132 1 + 1 2 20 3.313028132 + 3.313028132 24 20 1 3.313028132 + 1
Finally, we substituted n = −3.313028132 and c = 0.868637581 into Equation (3) to obtain:
f x = c x n = 0.868637581 X 3.313028132
From the data presented in Table 7, we calculated that the proportion of authors with one NN article was 90.92%, which was close to the c value (86.86%) that was obtained using Lotka’s law. Moreover, we applied the least squares method to calculate the n and c values and establish whether the results matched those obtained using Lotka’s law [72,73].
On the basis of Pao’s (1989) indication, the absolute value of n should be between 1.2 and 3.8. We found that the absolute value of n = 3.313028132 did match the results obtained using Lotka’s law [70]. The distribution trend is shown in Figure 4.

The Application of the Kolmogorov–Smirnov Test to Verify Whether the Results Complied with Lotka’s Law

Using the Lotka’s law equations, we calculated the following values:
n = −3.313028132 and c = 0.868637581.
We found that the results for the distribution of author productivity in NN research complied with Lotka’s law (Figure 4). We then used the n and c values to calculate the expected and accumulaton numbers of authors and applied the Kolmogorov–Smirnov test to evaluate whether the calculated values matched the theoretical values obtained using Lotka’s law. We also calculated the Dmax value presented in Table 9 as follows:
Dmax = Absolute value of (Fo(x) − Sn(x))
On the basis of the Kolmogorov–Smirnov test, the threshold value was set as follows:
1.63 / 2950 = 0.030010733
We found that the calculated Dmax values were much smaller than the threshold value, which meant that the distribution of author productivity in NN research complied with Lotka’s law.

3.2.10. GA Articles Productivity Analysis Using Lotka’s Law

The procedures for certifying the validity of our results were as follows:
  • Gather data from the SSCI database;
  • Calculate the distribution of author productivity in GA research;
  • Use the Lotka’s law formulae to calculate the n and c value;
  • Apply the Kolmogorov –Smirnov test to verify whether the results complied with Lotka’s law.

The Distribution of Author Productivity in GA Research

We obtained the author data for 878 GA articles from the SSCI database. Then, we aggregated and accumulated all of the author data to calculate the total number of authors who had published GA research (Table 10). We found that there was a total of 878 articles from 2430 authors, with an average of 2.77 (2430/878) authors per article.

The Use of the Lotka’s Law Formula to Obtain The n Value

Using the data presented in Table 10 and Table 11 and N = 7, we obtained the value of n = −3.768656463:
n = 7 3.110087 35.77143977 7 2.489009 13.707992

The Use of the Lotka’s Law Formula to Obtain the c Value

n = −3.768656463, p = 20, x = 1, 2, 3, 4, 5, 6, 7
Using the above data, we obtained the value of c = 0.908883846:
c = 1 1 20 1 1 x 3.768656463 + 1 3.768656463 1 20 3.768656463 1 + 1 2 20 3.768656463 + 3.768656463 24 20 1 3.768656463 + 1
Finally, we substituted n = −3.768656463 and c = 0.909490628 into the below equation to obtain:
f x = 0.909490628 X 3.768656463
From the data presented in Table 10, we found that the proportion of authors with one GA article was 89.51%, which was close to the c value (90.95%) that was obtained using Lotka’s law. Moreover, we applied the least squares method to calculate the n and c values and establish whether the result matched those obtained using Lotka’s law [72,73].
On the basis of Pao’s (1989) indication, the absolute value of n should be between 1.2 and 3.8. We found that the absolute value of n = −3.768656463 did match the results obtained using Lotka’s law [70]. The distribution trend is shown in Figure 5.

The Application of the Kolmogorov–Smirnov Test to Verify Whether the Results Complied with Lotka’s Law

Using the Lotka’s law equations, we calculated the following values:
n = −3.768656463 and c = 0.909490628.
We found that the results for the distribution of author productivity in GA research complied with Lotka’s law (Figure 5). We then used the n and c value to calculate the expected and accumulation numbers of authors and applied the Kolmogorov–Smirnov test to evaluate whether the calculated values matched the theoretical values obtained using Lotka’s law. We also calculated the Dmax value presented in Table 12 as follows:
Dmax = Absolute value of (Fo(x) − Sn(x))
On the basis of the Kolmogorov–Smirnov test, the threshold value was set as follows:
1.63 / 2430 = 0.033066214
We found that the calculated Dmax values were much smaller than the threshold value, which meant that the distribution of author productivity in GA research complied with Lotka’s law.

3.2.11. Advanced H-Index Analysis of NN and GA Research

The h-index is a useful index that characterizes the scientific output of researchers. Hirsch believes that the h-index can solve the shortcomings of measuring an author’s research results using the number of published papers or citation count by synthesizing a new evaluation method that integrates quality and quantity [75]. For example, when the h-index value of researcher A is greater than that of researcher B, it means that the overall quality of the papers from researcher A is better than that of papers from researcher B. The h-index can also be applied to measure the contributions of institutions and countries. H-index value can be calculated using data from the SSCI database on the Web of Science platform.

The H-Index Analysis of Authors of NN and GA Publications

From Table 13, it can be seen that Azadeh had the highest h-index value in GA research (10) and Resende had the highest h-index value in NN research (6). We also found that the average h-index value of authors of GA publications was higher than the average h-index value of authors of NN publications, as shown in Table 13.

The H-Index Analysis of Institutions Producing NN and GA Publications

From Table 14, it can be seen that the University of Tehran was the institution with the highest h-index value in NN research and the Islamic Azad University was the institution with the highest h-index value in GA research. Both are institutions in Iran and their research contributions were the greatest. The Islamic Azad University, the Iran University of Science and Technology, the Sharif University of Technology, and the University of Tehran are also all institutions in Iran, meaning that Iranian institutions contributed the most NN and GA publications, as shown in Table 14. The Universiti Malaya and the Hong Kong Polytech University had the second highest h-index values in NN and GA research, respectively. The Pennsylvania State System of Higher Education had an h-index value of 9 for both NN and GA research. Our results also showed that the Universitat Politècnica de València, the State University System of Florida, and the Indian Institutes of Technology are catching up in terms of their research contributions.

The H-Index Analysis of the Countries Producing NN and GA Publications

From Table 15, it can be seen that the USA contributed the most to NN and GA research as it had the highest h-index values in NN and GA research and produced the second highest quantity of NN and GA publications. In contrast, China had the second highest h-index values in NN and GA research and produced the highest quantity of NN and GA publications. Other countries, including Iran, Taiwan, Turkey, South Korea, and Malaysia, were smaller contributors, but our results showed that they are catching up quickly and becoming potential contributors to NN and GA research.

4. Conclusions

We used a bibliometric analytical technique to examine NN and GA publications in SSCI journals from 2002 to 2021. We found 951 NN publications and 878 GA publications in total. We then analyzed these publications and obtained our final research results by evaluating the following eight criteria: (1) publication year; (2) citation count; (3) country/territory; (4) institution name; (5) document type; (6) language; (7) subject area; and (8) source title. Furthermore, we also applied the Kolmogorov–Smirnov test to verify whether the distributions of author productivity in NN and GA research complied with Lotka’s law and advanced h-index analysis. In sum, this paper provides the following findings and implications:
First, according to our analysis by publication year, we found that the number of NN publications has grown faster than the number of GA publications and NN research has become much more popular. Based on the ascending trends of NN and GA publications, we predict that research in both areas will keep growing continuously. We also predict that NNs and GAs will continue to be important in artificial intelligence research for quite some time due to the third boom in AI development. Second, according to our analysis of the trends of NN and GA citations, we observed upward trends for NN and GA publications and we predict that these trends will continue in the future, as long as the AI industry is still booming.
Third, on the basis of our countries/territories analysis of NN and GA publications, we found that China and the USA contribute the most to NN and GA research and observed the significance of AI within the context of the USA–China confrontation. Iran was in third place for NN publications and Taiwan was in third place for GA publications. We also conclude that contributions from Turkey, South Korea, and Spain are growing fast and they are becoming potential competitors in NN and GA research. Fourth, according to our research institution analysis of NN and GA publications, the top NN research institution was the University of Tehran and the top GA research institution was the Islamic Azad University. Both are institutions in Iran and both published over 20 articles from 2002 to 2021. The second-place institution in NN research was the Universiti Malaya and Hong Kong Polytech University. The third-place institutions in NN and GA research were the Islamic Azad University and the Iran University of Science and Technology, respectively. Furthermore, we also observed that Iran was the most productive country for NN and GA research.
Fifth, from our analysis of document types in NN and GA research, we found that articles are still the major and most popular document type for NN and GA research publications. Sixth, from our analysis of the language used in NN and GA research and as English is the predominant scientific language that is most widely used among scientists worldwide, we concluded that English was still the number one language used in NN and GA research publications. Seventh, from our analysis, we found that computer science and engineering were the major domains of NN and GA research and that environmental sciences ecology research is catching up due to the rapid development of ESG. Furthermore, we also identified many other important and potential research domains for NN and GA publications, including business economies, mathematics, psychology, transportation, information science, library science, energy, etc. We predict that these research areas will become more popular in the future.
Eighth, we discovered that the top three NN research source titles were Sustainability, Expert Systems with Applications, and the International Journal of Environmental Research and Public Health, while the top three GA research source titles were Expert Systems with Applications, Sustainability, and the Journal of Operation Research Society. Additionally, there were many other popular research sources titles for NN and GA publications, including Neural Computing Applications, Energy Policy, PLoS ONE, Applied Soft Computing, Computational Economics, Computers Industrial Engineering, etc. We predict that these source titles will become more attractive in the near future.
Finally, we performed an advanced h-index analyses of NN and GA publications using different criteria, such as author, institution, and country to measure researcher contributions. We obtained the following results:
  • The average h-index value of GA publications was higher than the average h-index value of NN publications, as shown in Table 13.
  • Iranian institutions contributed the most NN and GA publications, as shown in Table 14.
  • The USA contributed the most to NN and GA research, with highest h-index value and the second highest quantity of NN and GA publications. In contrast, China had the second highest h-index values in NN and GA research but produced the second highest quantity of NN and GA publications.
We hope that our bibliometric analysis results will serve as a roadmap for other NN and GA researchers and act as guidelines for further studies. In the future, it is suggested to conduct similar research based on more innovative hybrid classical algorithm models and new bibliometrics methods, such as y-index analysis, frequent keywords analysis, the “life” of references cited in journals, and other advanced data-analysis techniques [75,76].

Author Contributions

Conceptualization, T.-F.L.; methodology, T.-F.L.; formal analysis, T.-F.L.; writing—original draft preparation, T.-F.L.; writing—review and editing, W.-H.H. and T.-F.L.; visualization, T.-F.L.; supervision, W.-H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available in [Web of Science] at [https://clarivate.com/zh-hant/solutions/web-of-science/ (accessed on 1 October 2022)]. These data were derived from the following resources available in the public domain: [https://clarivate.com/zh-hant/solutions/web-of-science/ (accessed on 1 October 2022)].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Abiodun, O.I.; Jantan, A.; Omolara, A.E.; Dada, K.V.; Mohamed, N.A.; Arshad, H. State-of-the-art in artificial Neural Network applications: A survey. Heliyon 2018, 4, e00938. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Yadav, N.; Yadav, A.; Kumar, M. An Introduction to Neural Network Methods for Differential Equations; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  3. Wang, S.C. Interdisciplinary Computing in Java Programming; Springer Science and Business Media: New York, NY, USA, 2003; pp. 81–100. [Google Scholar]
  4. O’Shea, K.; Nash, R. An introduction to convolutional Neural Network. arXiv 2015, arXiv:1511.08458. [Google Scholar]
  5. Holland, J.H. Genetic Algorithm. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  6. Goldberg, D.E.; Holland, J.H. Genetic Algorithm and Machine Learning. Mach. Learn. 1988, 3, 95–99. [Google Scholar] [CrossRef]
  7. Whitley, D. A Genetic Algorithm tutorial. Stat. Comput. 1994, 4, 65–85. [Google Scholar] [CrossRef]
  8. Sivanandam, S.N.; Deepa, S.N. Introduction to Genetic Algorithm; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  9. Yarushkina, N.G. Genetic Algorithm for engineering optimization: Theory and practice. In Proceedings of the 2002 IEEE International Conference on Artificial Intelligence Systems (ICAIS 2002), Divnomorskoe, Russia, 5–10 September 2002; pp. 357–362. [Google Scholar]
  10. Vidal, T.; Crainic, T.G.; Gendreau, M.; Lahrichi, N.; Rei, W. A hybrid Genetic Algorithm for multidepot and periodic vehicle routing problems. Oper. Res. 2012, 60, 611–624. [Google Scholar] [CrossRef]
  11. Garud, K.S.; Jayaraj, S.; Lee, M.Y. A review on modeling of solar photovoltaic systems using artificial Neural Network, fuzzy logic, genetic algorithm and hybrid models. Intern. J. Energy Res. 2021, 45, 6–35. [Google Scholar] [CrossRef]
  12. Rahman, M.M.; Setu, T.A. An implementation for combining Genetic Algorithm and Neural Network. Int. J. Comput. Sci. Technol. 2015, 6, 218–222. [Google Scholar]
  13. Chen, S.W.; Chen, P.H.; Tsai, C.T.; Liu, C.H. An Intelligent Based Symmetrical Classification of Online Shop Selling Counterfeit Products. Symmetry 2022, 14, 2132. [Google Scholar] [CrossRef]
  14. Wang, H.; Zhao, H.; Lian, Z.; Tan, B.; Zheng, Y.; Erdun, E. An Optimized Neural Network Acoustic Model for Porous Hemp Plastic Composite Sound-Absorbing Board. Symmetry 2022, 14, 863. [Google Scholar] [CrossRef]
  15. Kumar Mohapatra, P.; Kumar Rout, S.; Kishoro Bisoy, S.; Kautish, S.; Hamzah, M.; Jasser, M.B.; Mohamed, A.W. Application of Bat Algorithm and Its Modified Form Trained with ANN in Channel Equalization. Symmetry 2022, 14, 2078. [Google Scholar] [CrossRef]
  16. Han, H.; Kim, H.; Kim, Y. Correlation between Deep Neural Network Hidden Layer and Intrusion Detection Performance in IoT Intrusion Detection System. Symmetry 2022, 14, 2077. [Google Scholar] [CrossRef]
  17. Raja Sarobin, M.V.; Panjanathan, R. Diabetic Retinopathy Classification Using CNN and Hybrid Deep Convolutional Neural Networks. Symmetry 2022, 14, 1932. [Google Scholar]
  18. Yang, G.; Tang, C.; Liu, X. DualAC2NN: Revisiting and Alleviating Alert Fatigue from the Detection Perspective. Symmetry 2022, 14, 2138. [Google Scholar] [CrossRef]
  19. Haroon, M.; Altaf, S.; Ahmad, S.; Zaindin, M.; Huda, S.; Iqbal, S. Hand Gesture Recognition with Symmetric Pattern under Diverse Illuminated Conditions Using Artificial Neural Network. Symmetry 2022, 14, 2045. [Google Scholar] [CrossRef]
  20. Cengiz, A.B.; Birant, K.U.; Cengiz, M.; Birant, D.; Baysari, K. Improving the Performance and Explainability of Indoor Human Activity Recognition in the Internet of Things Environment. Symmetry 2022, 14, 2022. [Google Scholar] [CrossRef]
  21. Jiang, T.; Zhang, Q.; Yuan, J.; Wang, C.; Li, C. Multi-Type Object Tracking Based on Residual Neural Network Model. Symmetry 2022, 14, 1689. [Google Scholar] [CrossRef]
  22. Alhakami, H.; Kamal, M.; Sulaiman, M.; Alhakami, W.; Baz, A. A Machine Learning Strategy for the Quantitative Analysis of the Global Warming Impact on Marine Ecosystems. Symmetry 2022, 14, 2023. [Google Scholar] [CrossRef]
  23. Lan, P.; Xia, K.; Pan, Y.; Fan, S. An improved equilibrium optimizer algorithm and its application in LSTM neural network. Symmetry 2021, 13, 1706. [Google Scholar] [CrossRef]
  24. Mikhalev, A.S.; Tynchenko, V.S.; Nelyub, V.A.; Lugovaya, N.M.; Baranov, V.A.; Kukartsev, V.V.; Kurashkin, S.O. The Orb-Weaving Spider Algorithm for Training of Recurrent Neural Networks. Symmetry 2022, 14, 2036. [Google Scholar] [CrossRef]
  25. Rasheed, J.; Wardak, A.B.; Abu-Mahfouz, A.M.; Umer, T.; Yesiltepe, M.; Waziry, S. An Efficient Machine Learning-Based Model to Effectively Classify the Type of Noises in QR Code: A Hybrid Approach. Symmetry 2022, 14, 2098. [Google Scholar] [CrossRef]
  26. Hussain, M.; Bird, J.J.; Faria, D.R. A study on cnn transfer learning for image classification. In Proceedings of the UK Workshop on Computational Intelligence, Aarhus, Denmark, 6–8 July 2018; pp. 191–202. [Google Scholar]
  27. Pan, X.; Guo, W.; Guo, X.; Li, W.; Xu, J.; Wu, J. Deep temporal–spatial aggregation for video-based facial expression recognition. Symmetry 2019, 11, 52. [Google Scholar] [CrossRef] [Green Version]
  28. Liu, X.; Zou, Y.; Kuang, H.; Ma, X. Face image age estimation based on data augmentation and lightweight convolutional neural network. Symmetry 2020, 12, 146. [Google Scholar] [CrossRef] [Green Version]
  29. Zhong, Y.; Sun, L.; Ge, C.; Fan, H. HOG-ESRs Face Emotion Recognition Algorithm Based on HOG Feature and ESRs Method. Symmetry 2021, 13, 228. [Google Scholar] [CrossRef]
  30. Graves, A.; Mohamed, A.R.; Hinton, G. Speech recognition with deep recurrent Neural Network. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  31. Sherstinsky, A. Fundamentals of recurrent Neural Network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
  32. Kubanek, M.; Bobulski, J.; Kulawik, J. A method of speech coding for speech recognition using a convolutional neural network. Symmetry 2019, 11, 1185. [Google Scholar] [CrossRef] [Green Version]
  33. Yu, C.; Chen, Y.; Li, Y.; Kang, M.; Xu, S.; Liu, X. Cross-language end-to-end speech recognition research based on transfer learning for the low-resource Tujia language. Symmetry 2019, 11, 179. [Google Scholar] [CrossRef] [Green Version]
  34. Zhang, Z.; Huang, H.; Wang, K. Using Deep Time Delay Neural Network for Slot Filling in Spoken Language Understanding. Symmetry 2020, 12, 993. [Google Scholar] [CrossRef]
  35. Booker, L.B.; Goldberg, D.E.; Holland, J.H. Classifier Systems and Genetic Algorithm. Artif. Intell. 1989, 40, 235–282. [Google Scholar] [CrossRef]
  36. Chang, T.J.; Meade, N.; Beasley, J.E.; Sharaiha, Y.M. Heuristics for cardinality constrained portfolio optimisation. Comput. Oper. Res. 2000, 27, 1271–1302. [Google Scholar] [CrossRef]
  37. Swift, R.S. Accelerating Customer Relationships: Using GA and Relationship Technologies; Prentice Hall Publisher: Upper Saddle River, NJ, USA, 2001. [Google Scholar]
  38. Sefiane, S.; Benbouziane, M. Portfolio Selection Using Genetic Algorithm Archived 2016-04-29 at the Wayback Machine. J. Appl. Finance Bank. 2012, 2, 143–154. [Google Scholar]
  39. Min, S.H.; Lee, J.; Han, I. Hybrid Genetic Algorithm and support vector machines for bankruptcy prediction. Expert Sys. Appl. 2006, 31, 652–660. [Google Scholar] [CrossRef]
  40. Cheng, R.W.; Gen, M.; Tsujimura, Y. A tutorial survey of job-shop scheduling problems using Genetic Algorithms—I. Representation. Comput. Ind. Eng. 1996, 30, 983–997. [Google Scholar] [CrossRef]
  41. Blazewicz, J.; Domschke, W.; Pesch, E. The job shop scheduling problem: Conventional and new solution techniques. Eur. J. Oper. Res. 1996, 93, 1–33. [Google Scholar] [CrossRef]
  42. Chu, P.C.; Beasley, J.E. A Genetic Algorithm for the generalised assignment problem. Comput. Oper. Res. 1997, 24, 17–23. [Google Scholar]
  43. Jian, W.; Zhu, J.; Zeng, Q. An Optimization Model of Integrated AGVs Scheduling and Container Storage Problems for Automated Container Terminal Considering Uncertainty. Symmetry 2021, 13, 1904. [Google Scholar] [CrossRef]
  44. Hou, D.; Fan, H.; Ren, X. Multi-Depot Joint Distribution Vehicle Routing Problem Considering Energy Consumption with Time-Dependent Networks. Symmetry 2021, 13, 2082. [Google Scholar] [CrossRef]
  45. Ernst, A.T.; Jiang, H.; Krishnamoorthy, M. Staff scheduling and rostering: A review of applications, methods and models. Eur. J. Oper. Res. 2004, 153, 3–27. [Google Scholar] [CrossRef]
  46. Feng, C.W.; Liu, L.; Burns, S.A. Using Genetic Algorithm to solve construction time-cost trade-off problems. J. Comput. Civil Eng. 1997, 11, 184–189. [Google Scholar] [CrossRef]
  47. Hartmann, S. A competitive Genetic Algorithm for resource-constrained project scheduling. Naval Res. Logis. 1998, 45, 733–750. [Google Scholar] [CrossRef]
  48. Liao, L.; Leung, V.C.; Li, Z.; Chao, H.C. Genetic Algorithms with Variant Particle Swarm Optimization Based Mutation for Generic Controller Placement in Software-Defined Networks. Symmetry 2021, 13, 1133. [Google Scholar] [CrossRef]
  49. Aldallal, A.; Alisa, F. Effective Intrusion Detection System to Secure Data in Cloud Using Machine Learning. Symmetry 2021, 13, 2306. [Google Scholar] [CrossRef]
  50. Wang, S.; Li, Z.; He, C.; Liu, D.; Zou, G. An Integrated Method for Modular Design Based on Auto-Generated Multi-Attribute DSM and Improved Genetic Algorithm. Symmetry 2021, 14, 48. [Google Scholar] [CrossRef]
  51. Chin, Z.H.; Yap, T.T.V.; Tan, I.K.T. Genetic-Algorithm-Inspired Difficulty Adjustment for Proof-of-Work Blockchains. Symmetry 2022, 14, 609. [Google Scholar] [CrossRef]
  52. Hancock, P.J.B.; Smith, L.S. Gannet: Genetic design of a neural network for face recognition. Parallel Probl. Solving Nat. 1990, 2, 292–296. [Google Scholar]
  53. Maniezzo, V. Genetic evolution of the topology and weight distribution of Neural Network. IEEE Trans. Neural Netw. 1994, 5, 39–53. [Google Scholar] [CrossRef] [PubMed]
  54. Bishop, J.M.; Bushnell, M.J.; Usher, A.; Westland, S. Genetic optimisation of neural network architectures for colour recipe prediction. In Artificial Neural Networks and Genetic Algorithms; Iowa State University: Ames, IA, USA, 1993; pp. 719–725. [Google Scholar]
  55. Pater, L. Application of artificial Genetic Algorithm and Neural Network for crude fractional distillation process modeling. arXiv 2016, arXiv:1605.00097. [Google Scholar]
  56. Schaffer, J.D.; Whitley, D.; Eshelman, L.J. Combinations of Genetic Algorithm and Neural Network: A survey of the state of the art. In Proceedings of the IEEE COGANN-92: International Workshop on Combinations of Genetic Algorithm and Neural Networks, Baltimore, MD, USA, 6 June 1992; pp. 1–37. [Google Scholar]
  57. Hussain, T. Methods of Combining Neural Network and Genetic Algorithm, Tutorial Presentation. In Proceedings of the 1997 ITRC/TRIO Researcher Retreat, Kingston, ON, Canada, 6–8 May 1997. [Google Scholar]
  58. Syed, O. Applying Genetic Algorithm to Recurrent Neural Network for Learning Network Parameters and Architecture. Master’s Thesis, Case Western Reserve University, Cleveland, OH, USA, 1995. [Google Scholar]
  59. Caraka, R.E.; Yasin, H.; Chen, R.C.; Goldameir, N.E.; Supatmanto, B.D.; Toharudin, T.; Pardamean, B. Evolving Hybrid Cascade Neural Network Genetic Algorithm Space–Time Forecasting. Symmetry 2021, 13, 1158. [Google Scholar] [CrossRef]
  60. Meng, F. Safety warning model of coal face based on FCM fuzzy clustering and GA-BP neural network. Symmetry 2021, 13, 1082. [Google Scholar] [CrossRef]
  61. Adnan, A.; Muhammed, A.; Abd Ghani, A.A.; Abdullah, A.; Hakim, F. Hyper-heuristic framework for sequential semi-supervised classification based on core clustering. Symmetry 2020, 12, 1292. [Google Scholar] [CrossRef]
  62. Potter, W.G. Of Making Many Books There is No End: Bibliometrics and Libraries. J. Acad. Librariansh. 1988, 14. 238a–238c. [Google Scholar]
  63. Pao, M.L. Concept of Information Retrieve; Libraries Unlimited: Englewood, CO, USA, 1989. [Google Scholar]
  64. Ehrgott, M.; Gandibleux, X. A survey and annotated bibliography of multiobjective combinatorial optimization. Or Spektrum 2000, 22, 425–460. [Google Scholar] [CrossRef]
  65. Potter, W.G. Lotka’s Law Revisited. Libr. Trends 1981, 30, 21–39. [Google Scholar]
  66. Coille, R.C. Lotka’s frequency distribution of scientific productivity. J. Am. Soc. Inform. Sci. 1977, 28, 366–370. [Google Scholar] [CrossRef]
  67. Vlachy, J. Frequency distribution of scientific performance: A bibliography of Lotka’s law and related phenomena. Scientometrics 1978, 1, 109–130. [Google Scholar] [CrossRef]
  68. Gupta, D.K. Lotka’s law and productivity of entomological research in Nigeria for the period 1900–1973. Scientometrics 1987, 12, 33–46. [Google Scholar] [CrossRef]
  69. Nicholls, P.T. Bibliometric modeling processes and empirical validity of Lotka’s law. J. Am. Soc. Inform. Sci. 1989, 40, 379–385. [Google Scholar] [CrossRef]
  70. Pao, M.L. Lotka’s law: A testing procedure. Inform. Process. Manag. 1985, 21, 305–320. [Google Scholar] [CrossRef]
  71. Tsay, M.Y. The Characteristic of Informetric and Bibliometric; Hwa-Tai Publisher: Taipei, Taiwan, 2003. [Google Scholar]
  72. Tsai, H.H.; Yang, J.M. Analysis of Knowledge Management Trend by Bibliometric Approach. Int. J. Ind. Manuf. Eng. 2010, 4, 75–79. [Google Scholar]
  73. Yin, C.Y.; Yang, J.M. Assessment of Home Networking Research Trends, 1981–2009. J. Converg. Inform. Technol. 2010, 5, 17–27. [Google Scholar]
  74. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  75. Hirsch, J.E. An index to quantify an individual’s scientific research output. Proc. Natl. Acad. Sci. USA 2005, 102, 16569–16572. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  76. Matricciani, E. The probability distribution of the age of references in engineering papers. IEEE Trans. Prof. Comm. 1991, 34, 7–12. [Google Scholar] [CrossRef]
Figure 1. The trend of NN and GA literature publication (source: SSCI database).
Figure 1. The trend of NN and GA literature publication (source: SSCI database).
Symmetry 15 00325 g001
Figure 2. Trends of NN and GA citations per year (source: SSCI database).
Figure 2. Trends of NN and GA citations per year (source: SSCI database).
Symmetry 15 00325 g002
Figure 3. The top six countries/territories for NN and GA publications (source: SSCI database).
Figure 3. The top six countries/territories for NN and GA publications (source: SSCI database).
Symmetry 15 00325 g003
Figure 4. The author productivity in NN research.
Figure 4. The author productivity in NN research.
Symmetry 15 00325 g004
Figure 5. The author productivity in GA research.
Figure 5. The author productivity in GA research.
Symmetry 15 00325 g005
Table 1. A list of top 25 countries/territories producing NN and GA articles (from 2002 to 2021, SSCI).
Table 1. A list of top 25 countries/territories producing NN and GA articles (from 2002 to 2021, SSCI).
RankingNeural NetworkGenetic Algorithm
Countries/TerritoriesNP% of 951Countries/TerritoriesNP% of 878
1PEOPLES R CHINA17718.612PEOPLES R CHINA20923.804
2USA16917.771USA16318.565
3IRAN899.359TAIWAN819.226
4TURKEY666.94IRAN728.2
5TAIWAN565.889SOUTH KOREA546.15
6SPAIN474.942SPAIN525.923
7ITALY464.837ENGLAND475.353
8CANADA454.732INDIA455.125
9ENGLAND434.522TURKEY343.872
10MALAYSIA424.416CANADA323.645
11SOUTH KOREA394.101AUSTRALIA293.303
12INDIA353.68ITALY252.847
13GERMANY313.26FRANCE212.392
14POLAND252.629GERMANY192.164
15AUSTRALIA242.524JAPAN192.164
16FRANCE242.524BRAZIL182.05
17BRAZIL212.208PORTUGAL171.936
18NETHERLANDS212.208GREECE141.595
19JAPAN181.893NETHERLANDS141.595
20PAKISTAN171.788COLOMBIA131.481
21SAUDI ARABIA171.788MEXICO121.367
22PORTUGAL151.577MALAYSIA111.253
23SOUTH AFRICA141.472SINGAPORE111.253
24GREECE131.367BELGIUM91.025
25EGYPT111.157CHILE80.911
NP = number of publications.
Table 2. A list of the top 25 institutions producing NN and GA publications (from 2002 to 2021, SSCI).
Table 2. A list of the top 25 institutions producing NN and GA publications (from 2002 to 2021, SSCI).
RankingNeural Network Genetic Algorithm
Institution NamesNP% of 951CountryInstitution NamesNP% of 878Country
1UNIV OF TEHRAN202.103IranISLAMIC AZAD UNIV222.506Iran
2UNIVERSITI MALAYA171.788MalaysiaHONG KONG POLYTECHNIC UNIV131.481China
3ISLAMIC AZAD UNIV161.682IranIRAN UNIV SCIENCE TECHNOLOGY121.367Iran
4STATE UNIV SYSTEM OF FLORIDA151.577USANATIONAL INSTITUTE OF TECHNOLOGY121.367India
5EGYPTIAN KNOWLEDGE BANK EKB111.157EgyptPENNSYLVANIA COMMONWEALTH SYSTEM OF HIGHER EDUCATION111.253USA
6IRAN UNIV SCIENCE TECHNOLOGY111.157IranSHARIF UNIV OF TECHNOLOGY111.253Iran
7PENNSYLVANIA COMMONWEALTH SYSTEM OF HIGHER EDUCATION101.052USAUNIVERSITAT POLITECNICA DE VALENCIA111.253Spain
8CHINESE ACADEMY OF SCIENCES90.946ChinaUNIV OF TEHRAN111.253Iran
9INDIAN INSTITUTE OF TECHNOLOGY SYSTEM90.946 IndiaBEIJING JIAOTONG UNIV101.139China
10UCSI UNIV90.946MalaysiaINDIAN INSTITUTE OF TECHNOLOGY SYSTEM101.139 India
11UNIVERSITI TEKNOLOGI MALAYSIA90.946MalaysiaPENNSYLVANIA STATE UNIV101.139USA
12UNIV OF ALBERTA90.946CanadaNATIONAL CHENG KUNG UNIV91.025Taiwan
13UNIV OF LONDON90.946UKSTATE UNIV SYSTEM OF FLORIDA91.025USA
14UNIV OF TUNKU ABDUL RAHMAN90.946MalaysiaYONSEI UNIV91.025Korea
15CENTRAL SOUTH UNIV80.841ChinaCHINESE ACADEMY OF SCIENCES80.911China
16DELFT UNIV OF TECHNOLOGY80.841NetherlandsNATIONAL UNIV OF SINGAPORE80.911Singapore
17HONG KONG POLYTECHNIC UNIV80.841ChinaTSINGHUA UNIV80.911China
18PENNSYLVANIA STATE UNIV80.841USAUNIV OF MICHIGAN80.911USA
19SHARIF UNIV OF TECHNOLOGY80.841IranUNIV OF MICHIGAN SYSTEM80.911USA
20UDICE FRENCH RESEARCH UNIVERSITIES80.841FranceWUHAN UNIV80.911China
21CENTRE NATIONAL DE LA RECHERCHE SCIENTIFIQUE CNRS70.736FranceAMIRKABIR UNIV OF TECHNOLOGY70.797Iran
22ISFAHAN UNIV OF TECHNOLOGY70.736IranNATIONAL KAOHSIUNG UNIV OF SCIENCE TECHNOLOGY70.797Taiwan
23ISTANBUL TECHNICAL UNIV70.736TurkeySEOUL NATIONAL UNIV SNU70.797Korea
24SANTA CLARA UNIV70.736USATARBIAT MODARES UNIV70.797Iran
25TARBIAT MODARES UNIV70.736IranUNIVERSIDADE DO PORTO70.797Portugal
NP = number of publications.
Table 3. The different document types in NN and GA research (from 2002 to 2021, SSCI).
Table 3. The different document types in NN and GA research (from 2002 to 2021, SSCI).
Neural NetworkGenetic Algorithm
Document TypeNP% of 951Document TypeNP% of 878
Article87992.429Article84496.128
Meeting Abstract383.996Proceeding Paper364.1
Review Article242.524Meeting Abstract131.481
Proceeding Paper131.367Book Review111.253
Early Access111.157Review Article80.911
Correction40.421Early Access70.797
Editorial Material30.315Correction20.228
Book Review20.21
Letter10.105
Retracted Publication10.105
NP = number of publications.
Table 4. The languages used for NN and GA research (from 2002 to 2021, SSCI).
Table 4. The languages used for NN and GA research (from 2002 to 2021, SSCI).
Neural NetworkGenetic Algorithm
LanguageNP% of 951LanguageNP% of 878
English94198.948English87299.317
Portuguese30.315Spanish30.342
Spanish20.21Czech20.228
Czech10.105Turkish10.114
French10.105
German10.105
Russian10.105
Turkish10.105
NP = number of publications.
Table 5. An analysis of NN and GA research areas (from 2002 to 2021, SSCI).
Table 5. An analysis of NN and GA research areas (from 2002 to 2021, SSCI).
RankingNeural NetworkGenetic Algorithm
Research AreasNP% of 951Research AreasNP% of 878
1Computer Science22123.239Computer Science30134.282
2Engineering20121.136Engineering28031.891
3Environmental Sciences Ecology18619.558Operations Research Management Science22825.968
4Business Economics15816.614Business Economics22125.171
5Science Technology Other Topics12413.039Environmental Sciences Ecology11112.642
6Operations Research Management Science848.833Mathematics9110.364
7Psychology848.833Science Technology Other Topics677.631
8Public Environmental Occupational Health677.045Transportation627.062
9Energy Fuels555.783Information Science Library Science374.214
10Mathematics505.258Geography364.1
11Transportation474.942Psychology232.62
12Neurosciences Neurology434.522Physical Geography212.392
13Social Sciences Other Topics333.47Public Administration192.164
14Physics272.839Energy Fuels182.05
15Health Care Sciences Services252.629Public Environmental Occupational Health171.936
16Information Science Library Science202.103Social Sciences Other Topics171.936
17Materials Science191.998Automation Control Systems151.708
18Chemistry181.893Mathematical Methods In Social Sciences151.708
19Geography181.893Telecommunications131.481
20Telecommunications181.893Education Educational Research111.253
21Education Educational Research161.682Physics111.253
22Psychiatry141.472Urban Studies101.139
23Mathematical Methods in Social Sciences131.367Chemistry80.911
24Physical Geography131.367Geology70.797
25Geology121.262Health Care Sciences Services70.797
NP = number of publications.
Table 6. An analysis of the source titles in NN and GA research (from 2002 to 2021, SSCI).
Table 6. An analysis of the source titles in NN and GA research (from 2002 to 2021, SSCI).
RankingNeural NetworkGenetic Algorithm
Source TitleNP% of 951Source TitleNP% of 878
1SUSTAINABILITY757.886EXPERT SYSTEMS WITH APPLICATIONS637.175
2EXPERT SYSTEMS WITH APPLICATIONS414.311SUSTAINABILITY445.011
3INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH252.629JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY252.847
4NEURAL COMPUTING APPLICATIONS191.998APPLIED SOFT COMPUTING242.733
5ENERGY POLICY161.682COMPUTATIONAL ECONOMICS171.936
6PLoS ONE161.682COMPUTERS INDUSTRIAL ENGINEERING161.822
7APPLIED SCIENCES BASEL131.367INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE151.708
8NEUROCOMPUTING111.157OMEGA INTERNATIONAL JOURNAL OF MANAGEMENT SCIENCE151.708
9COMPUTATIONAL ECONOMICS101.052INTERNATIONAL TRANSACTIONS IN OPERATIONAL RESEARCH131.481
10ENERGIES101.052COMPUTERS ENVIRONMENT AND URBAN SYSTEMS111.253
11ENERGY101.052INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH111.253
12IEEE ACCESS90.946COMPUTERS OPERATIONS RESEARCH101.139
13ACCIDENT ANALYSIS AND PREVENTION80.841EUROPEAN JOURNAL OF OPERATIONAL RESEARCH101.139
14APPLIED SOFT COMPUTING80.841INFORMATION PROCESSING MANAGEMENT101.139
15JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY80.841SOFT COMPUTING91.025
16FRONTIERS IN PSYCHOLOGY70.736ECONOMIC COMPUTATION AND ECONOMIC CYBERNETICS STUDIES AND RESEARCH80.911
17IRANIAN JOURNAL OF PUBLIC HEALTH70.736TRANSPORTATION RESEARCH PART E LOGISTICS AND TRANSPORTATION REVIEW80.911
18JOURNAL OF INTELLIGENT FUZZY SYSTEMS70.736IEEE ACCESS70.797
19TRANSPORTATION RESEARCH PART D TRANSPORT AND ENVIRONMENT70.736INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH70.797
20BEHAVIOURAL PROCESSES60.631INTERNATIONAL JOURNAL OF PRODUCTION ECONOMICS70.797
21ENGINEERING CONSTRUCTION AND ARCHITECTURAL MANAGEMENT60.631JOURNAL OF INTELLIGENT MANUFACTURING70.797
22JOURNAL OF CLEANER PRODUCTION60.631MATHEMATICAL PROBLEMS IN ENGINEERING70.797
23JOURNAL OF FORECASTING60.631PLoS ONE70.797
24MATHEMATICS60.631TRANSPORTATION RESEARCH PART B METHODOLOGICAL70.797
25PHYSICA A STATISTICAL MECHANICS AND ITS APPLICATIONS60.631ANNALS OF OPERATIONS RESEARCH60.683
NP = number of publications.
Table 7. An analysis of author productivity in NN research.
Table 7. An analysis of author productivity in NN research.
NPAuthor(s)NP × AuthorAccumulated RecordAccumulated Record (%)Accumulated Author (s)Accumulated Author(s) (%)
12112120.36%10.03%
1100120.36%10.03%
1000120.36%10.03%
919210.63%20.07%
818290.87%30.10%
7321501.49%60.20%
6212621.85%80.27%
5630922.75%140.47%
412481404.18%260.88%
3421262667.95%682.31%
220040066619.89%2689.08%
1268226823348100.00%2950100.00%
NP = number of publications.
Table 8. The calculation of the exponent n for NN publications.
Table 8. The calculation of the exponent n for NN publications.
NP (x)Author (y)X = log(x)Y = log(y)XYXX
126820.003.430.000.00
22000.302.300.690.09
3420.481.620.770.23
4120.601.080.650.36
560.700.780.540.49
620.780.300.230.61
730.850.480.400.71
810.900.000.000.82
910.950.000.000.91
1001.00----1.00
1101.04----1.08
1211.080.000.001.16
Total29508.689.993.307.46
NP = number of publications; x = number of publications; y = author; X = logarithm of x; Y = logarithm of y.
Table 9. The results of the Kolmogorov–Smirnov test for NN research.
Table 9. The results of the Kolmogorov–Smirnov test for NN research.
NPObservation by Author(s)Accumulated Value Sn(x)Expected Value by Author(s) Accumulated Value Fo(x)ABS Value Fo(x) − Sn(x)
Dmax
10.909150.909150.880760.880760.028390763
20.067800.976950.081830.962590.014357179
30.014240.991190.020380.982970.008211635
40.004070.995250.007600.990580.004676717
50.002030.997290.003540.994120.003172625
60.000680.997970.001890.996010.001956854
70.001020.998980.001120.997130.001857395
80.000340.999320.000710.997830.001490022
90.000340.999660.000470.998300.0013573
100.000000.999660.000330.998630.001028591
110.000000.999660.000240.998870.0007915
120.000341.000000.000180.999050.000954539
NP = number of publications; Sn(x) = observed cumulative frequency; Fo(x) = theoretical cumulative frequency; D = maximum deviation.
Table 10. An analysis of author productivity in GA research.
Table 10. An analysis of author productivity in GA research.
NPAuthor(s)NP × AuthorAccumulated RecordAccumulated Record (%)Accumulated Author (s)Accumulated Author(s) (%)
71770.25%10.04%
6636431.54%70.29%
5420632.26%110.45%
410401033.69%210.86%
3471412448.74%682.80%
218737461822.13%25510.49%
1217521752793100.00%2430100.00%
NP = number of publications.
Table 11. The calculation of the exponent n for GAs.
Table 11. The calculation of the exponent n for GAs.
NP (x)Author (y)X = log(x)Y = log(y)XYXX
121750.003.340.000.00
21870.302.270.680.09
3470.481.670.800.23
4100.601.000.600.36
540.700.600.420.49
660.780.780.610.61
710.850.000.000.71
Total24303.709.663.112.49
NP = number of publications; x = number of publications; y = author; X = logarithm of x; Y = logarithm of y.
Table 12. The results of the Kolmogorov–Smirnov test for GA research.
Table 12. The results of the Kolmogorov–Smirnov test for GA research.
NPObservation by Author(s)Accumulated Value Sn(x)Expected Value by Author(s) Accumulated Value Fo(x)ABS Value Fo(x) − Sn(x)
Dmax
10.895060.895060.909490.909490.0144289
20.076950.972020.066730.976220.004203931
30.019340.991360.014480.990700.000660181
40.004120.995470.004900.995590.000120588
50.001650.997120.002110.997710.00058614
60.002470.999590.001060.998770.000820778
70.000411.000000.000590.999360.000638127
NP = number of publications; Sn(x) = observed cumulative frequency; Fo(x) = theoretical cumulative frequency; D = maximum deviation.
Table 13. The top 10 authors in our h-index analysis (from 2002 to 2021, SSCI).
Table 13. The top 10 authors in our h-index analysis (from 2002 to 2021, SSCI).
RankingNeural NetworksGenetic Algorithm
AuthorH-IndexAuthorH-Index
1Resende, Mauricio G. C.6Azadeh, Ali10
2Niaki, Seyed Taghi Akhavan5OOI, Keng-Boon8
3Pendharkar, Parag C.5Saberi, Morteza7
4Oh, Kyong Joo4Ghiassi, M.7
5Ruiz, Rubén4Leong, Lai-Ying L.Y.6
6Chung, Sai-Ho4Tan Wei Han, Garry5
7Azadeh, Ali4Maurelli, Guido2
8Huang, Bo4Dawson, Michael RW2
9Ceylan, Halim4Palmer, Alfonso2
10Dung-Ying Lin3Buscema, Paolo Massimo2
Table 14. The top 10 institutions in our h-index analysis (from 2002 to 2021, SSCI).
Table 14. The top 10 institutions in our h-index analysis (from 2002 to 2021, SSCI).
RankingNeural NetworksGenetic Algorithm
InstitutionH-IndexInstitutionH-Index
1UNIVERSITY OF TEHRAN12ISLAMIC AZAD UNIVERSITY10
2UNIVERSITI MALAYA12HONG KONG POLYTECHNIC UNIVERSITY9
3ISLAMIC AZAD UNIVERSITY10PENNSYLVANIA COMMONWEALTH SYSTEM OF HIGHER EDUCATION PCSHE9
4PENNSYLVANIA COMMONWEALTH SYSTEM OF HIGHER EDUCATION PCSHE9UNIVERSITAT POLITECNICA DE VALENCIA9
5STATE UNIVERSITY SYSTEM OF FLORIDA8INDIAN INSTITUTE OF TECHNOLOGY SYSTEM IIT SYSTEM9
6IRAN UNIVERSITY SCIENCE TECHNOLOGY8IRAN UNIVERSITY SCIENCE TECHNOLOGY8
7EGYPTIAN KNOWLEDGE BANK EKB7NATIONAL INSTITUTE OF TECHNOLOGY NIT SYSTEM8
8INDIAN INSTITUTE OF TECHNOLOGY SYSTEM IIT SYSTEM7SHARIF UNIVERSITY OF TECHNOLOGY8
9UCSI UNIVERSITY7UNIVERSITY OF TEHRAN8
10CHINESE ACADEMY OF SCIENCES6BEIJING JIAOTONG UNIVERSITY6
Table 15. The top 10 countries in our h-index analysis (from 2002 to 2021, SSCI).
Table 15. The top 10 countries in our h-index analysis (from 2002 to 2021, SSCI).
RankingNeural NetworksGenetic Algorithm
Countries/TerritoriesH-IndexCountries/TerritoriesH-Index
1USA42USA40
2PEOPLES R CHINA32PEOPLES R CHINA37
3IRAN25TAIWAN26
4TURKEY21SOUTH KOREA22
5MALAYSIA21IRAN21
6TAIWAN19INDIA19
7CANADA16SPAIN18
8SPAIN15ENGLAND18
9INDIA15TURKEY16
10AUSTRALIA15CANADA15
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lou, T.-F.; Hung, W.-H. Revival of Classical Algorithms: A Bibliometric Study on the Trends of Neural Networks and Genetic Algorithms. Symmetry 2023, 15, 325. https://doi.org/10.3390/sym15020325

AMA Style

Lou T-F, Hung W-H. Revival of Classical Algorithms: A Bibliometric Study on the Trends of Neural Networks and Genetic Algorithms. Symmetry. 2023; 15(2):325. https://doi.org/10.3390/sym15020325

Chicago/Turabian Style

Lou, Ta-Feng, and Wei-Hsi Hung. 2023. "Revival of Classical Algorithms: A Bibliometric Study on the Trends of Neural Networks and Genetic Algorithms" Symmetry 15, no. 2: 325. https://doi.org/10.3390/sym15020325

APA Style

Lou, T. -F., & Hung, W. -H. (2023). Revival of Classical Algorithms: A Bibliometric Study on the Trends of Neural Networks and Genetic Algorithms. Symmetry, 15(2), 325. https://doi.org/10.3390/sym15020325

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop