Next Article in Journal
Green Hydrogen Production—Fidelity in Simulation Models for Technical–Economic Analysis
Previous Article in Journal
Vehicle Motion Control for Overactuated Vehicles to Enhance Controllability and Path Tracking
Previous Article in Special Issue
Understanding the Flows of Signals and Gradients: A Tutorial on Algorithms Needed to Implement a Deep Neural Network from Scratch
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast Type-II Hartley Transform Algorithms for Short-Length Input Sequences

by
Marina Polyakova
1,* and
Aleksandr Cariow
2,*
1
Institute of Computer Systems, Odesa Polytechnic National University, Shevchenko Ave., 1, 65044 Odesa, Ukraine
2
Faculty of Computer Science and Information Technology, West Pomeranian University of Technology in Szczecin, Żołnierska 49, 71-210 Szczecin, Poland
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2024, 14(22), 10719; https://doi.org/10.3390/app142210719
Submission received: 7 October 2024 / Revised: 7 November 2024 / Accepted: 15 November 2024 / Published: 19 November 2024
(This article belongs to the Special Issue Advanced Digital Signal Processing and Its Applications)

Abstract

:
This paper presents the type-II fast discrete Hartley transform (DHT-II) algorithms for input data sequences of lengths from 2 to 8. The starting point for developing the eight algorithms is the representation of DHT-II as a matrix–vector product. The underlying matrices usually have a good block structure. These matrices must then be successfully factorized to obtain a computational procedure that reduces the number of operations in computing the matrix–vector product. In some cases, it is necessary to pre-decompose the original matrices into submatrices and rearrange the rows and/or columns of the resulting matrices to find the factorizations that would substantially save the arithmetic operations. As a result of applying the pointed transformations, we synthesized the final algorithms with reduced computational complexity. The correctness of the obtained algorithmic solutions was theoretically justified using the rigorous mathematical background of each of them. Then, the complex algorithms were further tested using the MATLAB R2023b software to confirm their performance. Finally, an evaluation of the computational complexity for each obtained solution was compared with the computational complexity of the direct calculation of the matrix–vector product and existing fast DHT-II algorithms. The obtained factorizations of the DHT-II transformation matrices on average reduce the number of additions by 5% and the number of multiplications by 73% compared with the direct calculation of the matrix–vector product.

1. Introduction

Among other orthogonal transforms, the discrete Hartley transform (DHT) is a well-known and well-tested tool in digital signal and image processing [1,2,3]. This transform is used in smart grid applications [4], signal processing [5], digital image watermarking [6,7,8], audio watermarking [9], image compression and encryption [10,11], image enhancement [12], OFDM communication systems [13,14,15], face recognition [16], audio classification [17,18], etc. Its advantage is primarily that it can successfully supplement the discrete Fourier transform in many applications [1,2,3]. For example, in audio classification applications, features extracted from the representation of signals in the frequency domain are typically focused on the magnitude spectral content, while the phase spectral content is ignored [19]. The conventional Fourier phase spectrum is a highly discontinuous function. To extract the features in classification applications, function continuity is required. Thus, the Hartley phase spectrum was introduced as an alternative to the Fourier phase spectrum. The former encapsulates the phase content of the signal more efficiently if the sources of phase discontinuities are detected and overcome [20]. The latter is because the Hartley phase spectrum does not suffer from the phase “wrapping ambiguities” introduced due to the inverse tangent function employed in the Fourier phase spectrum computation [19].
Compared to wavelet analysis, the Hartley transform provides only global frequency information, although the ability to analyze signals in terms of both time and frequency may be obtained with windowing applied. As a result of windowing, a short DHT is obtained [21]. The DHT allows for efficient signal denoising, being a fixed global transform. Moreover, the traditional DHT can be adapted to detect fine signal details by using windows of different sizes [8,22]. Although DHT [10,11] is exploited in image compression infrequently, it is widely applied in audio and image watermarking [6,7,8,22] to enhance robustness against different noise attacks. Also, a significant increase in computing efficiency by saving arithmetic operations and memory storage is achieved [22].
To date, four types of discrete Hartley transforms have been formulated [23]. However, the most popular and most developed transform is still DHT-I [1,10,13,14,15]. The other transforms from this family have been studied much less. One important, but not the main, advantage of discrete orthogonal transforms is the huge number of so-called fast algorithms that have been designed for them, allowing for the acceleration of their implementation [23]. The direct computation of the DHT-II of an N-point sequence requires N2 arithmetic operations. Therefore, fast DHT-II algorithms have been introduced to reduce the computational complexity and implementation cost [24,25,26,27,28,29].
It should be noted that the papers dealing with the efficient implementation of the DHT-II transform considered the case of large lengths of input data sequences [24,25,26,27,28,29]. However, short-length DHT-II algorithms are of particular interest, since these transforms are considered typical modules in the synthesis of more complex algorithms [23]. Once constructed, short-length DHT-II algorithms can be successfully applied in various projects to unify the development process of the final product. There are many papers devoted to the efficient implementation of traditional short-length trigonometric transforms, but they mainly concern cosine or sine transforms [30,31,32,33]. Then, the development of fast DHT-II algorithms for short-length input sequences becomes relevant. Below, the current state of the problem is analyzed to select an approach to the development of such fast DHT-II algorithms.

1.1. Related Papers

To develop fast algorithms for discrete orthogonal transforms the strategy of “divide and conquer” is applied [23,34,35,36,37]. This means dividing a large problem into smaller ones. Then, the solution to the big problem can be obtained by collecting the solutions of smaller problems, which are organized according to the dividing procedures. Often, an obtained solution to the big problem is modular and systematic, which is desirable for implementation. Based on the divide and conquer strategy, fast algorithms for discrete orthogonal transforms have been developed as split-radix algorithms and prime factor algorithms [23].
Split-radix algorithms assume that the length of the input data sequence can be expressed as N = q × 2m, where m is a positive integer and q is an odd integer. Various sequence length values can be represented by setting different values of m and q. For example, the radix-2 algorithm recursively decomposes the length N DHT-II into two N/2-length DHT-IIs. In contrast, the split-radix algorithm decomposes the N-length DHT-II using different radices. In the literature, split-radix algorithms are developed based on decimation-in-time (DIT) and decimation-in-frequency decomposition [23]. The advantages of split-radix algorithms are substantial, saving the number of arithmetic operations and supporting a wide range of sequence lengths.
Despite the advantages of the cost-effective radix algorithms, computational overheads usually occur when the input sequence length is not matched to the assumptions of the available algorithm. Hence, prime factor algorithms were introduced for various sequence lengths, in particular, for sequence lengths other than a power of two. However, prime factor algorithms generally have a more complex computational structure. Specifically, they require irregular index mapping to reorder the input data sequence while computing. Therefore additional computation time and memory space are generally needed. Also, the desirable property of in-place computation, which is easily achieved by split-radix algorithms, is difficult to implement.
In the literature, the fast algorithms of both categories are described for DHT-II [24,25,26,27,28,29]. Thus, in [23], the prime factor algorithm initially decomposed N-point DHT-II into q p-point discrete cosine transforms and p q-point DHTs. Then, to reduce the required number of arithmetic operations and eliminate the irregular index mapping, the resulting decomposition was combined with the split-radix algorithm. A regular computational structure was achieved, providing flexibility for composite sequence lengths and substantially saving the computational complexity.
In [24], an algorithm of the DHT-II computation was presented when two sequences of N/2-length DHT-II adjacent coefficients were given. In this way, two new radix-2 algorithms were developed for computing the N-point DHT-II. Next, obtaining the N-point DHT-II was described based on the two adjacent N/2-point DHT-IIs. Compared with the traditional Hu’s algorithm, the arithmetic operations could be saved from 16% to 24% for N varying from 16 to 64 [24]. If N = 8, the number of arithmetic operations slightly increased.
In [25], the algorithm from [24] was further developed. The obtained radix-2 recursive algorithm required a reduced number of arithmetic operations compared with existing algorithms and could be easily implemented.
The algorithms proposed in [24,25] were based on the decimation-in-frequency approach. But, if DHT-II fast algorithms need to be developed for real-time applications, then the DIT approach is preferable. That is why in [26], the radix-2 DIT fast algorithm for DHT-II calculation was presented. It was described by a regular signal flow graph, which provided flexibility to different transform lengths, substantially reducing the arithmetic complexity compared with Hu’s algorithms.
In [27], a split-radix DIT algorithm for the fast calculation of the DHT-II was proposed. The splitting of DHT-II of length N into one DHT-II of length N/2 was implemented for even-indexed samples. The splitting of DHT-II of length N into two DHTs-II of length N/4 was applied for odd-indexed samples. The proposed algorithm considerably reduced the arithmetic complexity compared to the existing DHT-II algorithms. In particular, the number of multiplications could be decreased by 29% compared to the radix-2 DHT-II algorithm [26], and the number of additions was reduced by 10%.
In [28], a modular VLSI algorithm for the fast calculation of DHT-II was introduced. The author efficiently incorporated the obfuscation technique with low overheads, which presented the additional advantage of hardware security. The proposed VLSI algorithm applied modular and regular computational structures named quasi-band correlations. In addition, the use of the systolic arrays architectural paradigm allowed for the efficient implementation of this algorithm on a VLSI chip.
The aforementioned split-radix algorithms were focused on input sequences, with length N being the power of two. Fast algorithms for calculating DHT-II with length N being the power of three were also developed. For example, in [29], a fast radix-3 algorithm for the computation of the DHT-II was developed for input sequences of length N, where N = 3m, m ≥ 2. Then, this algorithm was applied to the direct computation of DHT-II when three adjacent sequences of length N/3 DHT-II were given. The arithmetic operations could be saved by up to 29%. Furthermore, the new algorithm could be easily implemented in VLSI.

1.2. The Main Contributions of This Paper

As a result of the analysis of articles devoted to developing the fast DHT-II algorithms, the following should be noted [5]. Firstly, the mathematical notations used by the authors to describe the proposed fast algorithms are often quite complicated. As a result, the understanding of the essence of the suggested solutions may be difficult for practitioners unfamiliar with the polynomial theory, number theory, Galois field theory, etc. Secondly, the mentioned theories and the developed algorithms concern specialized tasks. However, reducing the computing operations is desirable when a wide range of problems are solved. To lessen these disadvantages, in [5], a structural approach was proposed to develop fast algorithms for matrix–vector multiplication. It was based on a deep analysis of the structure of base transform matrices, identifying individual features of the arrangement of identical entries. Then, if necessary, the matrix structure was changed for subsequent use of the submatrix factorization [5]. The structural approach did not limit the length of the original data sequence, for example, to a power of two or three [5]. The number of multiplication and addition operations necessary for matrix–vector product computation could be reduced using this incredibly simple and effective approach. Since DHT-II coefficient matrices are structured, this research aimed to develop reduced-complexity DHT-II algorithms for input sequences of lengths N = 2, 3, 4, 5, 6, 7, and 8 based on a structural approach.
This paper is structured as follows. Section 1 introduces the problem and defines the subject of this study. Section 2 contains a short mathematical background and defines the system of used notations and symbols. Section 3 contains the derivation of algorithms for DHT-II transform sizes from 2 to 8. Section 4 and Section 5 contain a discussion of this study’s results and a comparative evaluation of the computational costs required to implement the developed and known algorithms. Conclusions are presented in Section 6.

2. Preliminary Remarks

The discrete Hartley transform (DHT) is one of the orthogonal transforms used, among others, for the analysis and processing of sounds and signals. The DHT advantages are primarily because it can successfully replace the discrete Fourier transform in many applications. If we apply the DHT to real-valued signals, then real-valued results are obtained in contrast to the Fourier transform. The use of the latter to real-valued signals in general gives complex-valued results. The real-data processing significantly simplifies the implementation of calculations [1,2,3].
For example, the direct generalized DHT can be represented by the following expression [23]:
X k = n = 0 N 1 x n   cas ( n + n 0 ) ( k + k 0 ) 2 π N ,   k = 0 ,   1 ,   ,   N 1 ,
where
cas θ = cos θ + sin θ = 2 cos ( θ π / 4 ) ;
X k is the output sequence after the direct generalized DHT is performed, x n is the sequence of input data, and N is the number of signal samples. By expression (1), we have the DHT-I if n 0 = 0 ,   k 0 = 0 ; the DHT-II if n 0 = 1 / 2 ,   k 0 = 0 ; the DHT-III if n 0 = 0 ,   k 0 = 1 / 2 ; and the DHT-IV if n 0 = 1 / 2 ,   k 0 = 1 / 2 .
Therefore the DHT-II can be obtained as follows [23]:
y k = n = 0 N 1 x n   cas π 2 n + 1 k N ,   k = 0 ,   1 ,   ,   N 1 ,
where y k is the output sequence after the DHT-II is performed.
In matrix notation, DHT-II can be represented as follows:
Y N × 1 = C N X N × 1 ,
where Y N × 1 = y 0 ,   y 1 ,     ,   y N 1 T , X N × 1 = x 0 ,   x 1 ,   ,   x N 1 T , c kl = cas π 2 n + 1 k N , k, l = 0, …, N − 1.
In this paper, we use the following markings and signs [5]:
  • I N is an order N identity matrix;
  • H 2 is a 2 × 2 Hadamard matrix;
  • 1 N × M   is a N × M matrix of ones (a matrix where every element is equal to one);
  • ⊗ is the Kronecker product of two matrices;
  • ⊕ is the direct sum of two matrices.
DHT-II in matrix notation is as follows:
[ y 0 y 1 y N 1 ] = [ c 0 , 0 c 0 , 1 c 1 , 0 c 1 , 1 c 0 , N 1 c 1 , N 1 c N 1 ,   0 c N 1 ,   1   c N 1 , N 1 ] [ x 0 x 1 x N 1 ]
An empty cell in a matrix means that it contains zero. The multipliers are marked as s m ( N ) .

3. Fast Algorithms for Small-Size DHT-II

3.1. Algorithm for 2-Point DHT-II

Let us elaborate the algorithm for two-point DHT-II. The expression for two-point DHT-II is as follows:
Y 2 × 1 = C 2 X 2 × 1 ,
where Y 2 × 1 =   y 0 ,   y 1 T , X 2 × 1 =   x 0 ,   x 1 T , C 2 = a 2 a 2 a 2 a 2 , a 2 = 1.
Then, the expression for DHT-II for N = 2 can be presented as follows:
Y 2 × 1 = H 2 X 2 × 1 .
Figure 1 shows a data flow graph of the synthesized algorithm for the two-point DHT-II. As can be seen, the number of addition operations is 2, which is the same as when using the direct method.

3.2. Algorithm for 3-Point DHT-II

Now, the algorithm for the three-point DHT-II can be developed. The expression for the three-point DHT-II is as follows:
Y 3 × 1 = C 3 X 3 × 1 ,
where Y 3 × 1 = y 0 ,   y 1 ,   y 2 T , X 3 × 1 =   x 0 ,   x 1 ,   x 2 T , C 3 = 1 c 3 b 3 1 1 1 1 b 3 c 3 with b 3 = 0.366, c 3 = 1.366.
We denote c 3 = 1 + b 3 and decompose the matrix C 3 into two components:
C 3 = C 3 ( a ) + C 3 ( b ) ,
where C 3 ( a ) = 1 b 3 b 3 1 1 1 1 b 3 b 3 , C 3 ( b ) = 1 1 .
Taking into account properties of structural matrices [5,32], the computational procedure for the three-point DHT-II is represented by the expression
Y 3 × 1 = W 3 × 5 D 5 W 5 × 3 X 3 × 1 ,
where W 5 × 3 = 1 1 1 1 1 1 1 1 , D 5 = diag 1 ,   1 ,   1 ,   s 0 ( 3 ) ,   1 , s 0 ( 3 ) = b 3 ,   W 3 × 5 = 1 1 1 1 1 1 1 .
A data flow graph of the proposed algorithm for the three-point DHT-II is shown in Figure 2. As can be seen, we are able to reduce the number of multiplication operations from 4 to 1. However, the number of addition operations is increased from 6 to 7.

3.3. Algorithm for 4-Point DHT-II

Let us design the algorithm for the four-point DHT-II. The four-point DHT-II is expressed as follows:
Y 4 × 1 = C 4 X 4 × 1 ,
where Y 4 × 1 = y 0 ,   y 1 ,   y 2 ,   y 3 T , X 4 × 1 = x 0 ,   x 1 ,   x 2 ,   x 3 T , C 4 = 1 a 4 1 0 1 0 1 a 4 1 a 4 1 0 1 0 1 a 4 with a 4   = 1.4142.
To change the order of the columns of C 4 , we define the permutation π 4 in the following form: π 4 = 1 2 1 3 3 4 2 4   .
After permutation of the columns of C 4 according to π 4 , we obtain the matrix C 4 ( A ) = 1 1 a 4 0 1 1 0 a 4 1 1 a 4 0 1 1 0 a 4   with the permutation matrix P 4 = 1 1 1 1 .
Taking into account the properties of structural matrices [5,30], the computational procedure for the four-point DHT-II is represented by the expression
Y 4 × 1 = W 4 ( 1 ) D 4 W 4 ( 0 ) P 4 X 4 × 1 ,
where W 4 ( 0 ) = H 2 I 2 , D 4 = diag 1 ,   1 ,   s 0 ( 4 ) ,   s 0 ( 4 ) , s 0 ( 4 ) = a 4 ,   W 4 ( 1 ) = H 2 I 2 .
A data flow graph of the proposed algorithm of the four-point DHT-II is presented in Figure 3. In particular, the number of multiplication operations may be reduced from 4 to 2 and the number of addition operations is decreased from 8 to 6.

3.4. Algorithm for 5-Point DHT-II

Let us obtain the algorithm for the five-point DHT-II. The five-point DHT-II is expressed as follows:
Y 5 × 1 = C 5 X 5 × 1 ,
where Y 5 × 1 = y 0 ,   y 1 ,   y 2 ,   y 3 ,   y 4 T , X 5 × 1 =   x 0 ,   x 1 ,   x 2 ,   x 3 ,   x 4 T , b 5 = 1.3968, c 5   = 0.6420, d 5 = 1.2601, e 5   = 0.2212,
C 5 = 1 b 5 d 5 c 5 e 5 1 c 5 b 5 e 5 d 5 1 1 1 1 1 1 d 5 e 5 b 5 c 5 1 e 5 c 5 d 5 b 5 .
To change the order of rows and columns of C 5 , the permutations π 5 and π 6 are defined in the following form:
π 5 = 1 2 3 1   3 4 5 2 5 4   ,   π 6 = 1 2 1 2   3 4 5 3 5 4   .
The rows of C 5 are permutated according to π 5 and the columns of C 5 are permutated according to π 6 . After permutations, the matrix acquires the following structure:
C 5 ( a ) = 1 1 1 1 1 1 b 5 d 5 e 5 c 5 1 c 5 b 5 d 5 e 5 1 e 5 c 5 b 5 d 5 1 d 5 e 5 c 5 b 5 .
Then, the matrix C 5 ( a ) is decomposed into two components:
C 5 ( a ) = C 5 ( b ) + C 5 ( c ) ,
where C 5 ( b ) = 1 1 1 1 1 1 1 1 1 and C 5 ( c ) =             b 5 d 5 e 5 c 5   c 5 b 5 d 5 e 5   e 5 c 5 b 5 d 5   d 5 e 5 c 5 b 5 .
Matrix C 5 ( b ) has the same entries except on the sign in the first column and first row, which allows for reducing the number of operations without the need for further transformations [32]. After eliminating rows and columns containing only zero entries in matrix C 5 ( c ) , we obtain matrix C 4 ( d ) :
C 4 ( d ) = b 5 d 5 e 5 c 5 c 5 b 5 d 5 e 5 e 5 c 5 b 5 d 5 d 5 e 5 c 5 b 5
which matches the matrix pattern A 2 B 2 B 2 A 2 , where A 2 = b 5 d 5 c 5 b 5 , B 2 = e 5 c 5 d 5 e 5 .
Hence, the matrix C 4 ( d ) can be represented as [5,30]
C 4 ( d ) = ( I ¯ 2 I 2 ) ( H 2 I 2 ) 1 / 2 [ ( A 2 + B 2 ) ( A 2 B 2 ) ] ( H 2 I 2 ) ,
where I ¯ 2 = 1 0 0 1 . Considering the structures of the resulting matrices A 2 + B 2 and ( A 2 B 2 ), we note that these matrices match the patterns a b b a and c d d c , respectively. Here, a = b 5     e 5 ;   b = c 5 + d 5 ;   c = b 5 + e 5 ;   d =   c 5 + d 5 . Then [5,31],
A 2 + B 2 = T 2 × 3 ( 4 )   ×   diag ( a b ,   a b ,   b ) T 3 × 2 ( 3 ) A 2 B 2 =   I ¯ 2 H 2 × diag ( ( c + d ) / 2 ,   ( c d ) / 2 )   H 2
where T 2 × 3 ( 4 ) = 1 0 1 0 1 1 , T 3 × 2 ( 3 ) = 1 0 0 1 1 1 .
We define s 0 ( 5 ) = ( a b ) / 2 ;   s 1 ( 5 ) =   ( a + b ) / 2 ;   s 2 ( 5 ) = b / 2 ;   s 3 ( 5 ) = ( c + d ) / 4 ;   s 4 ( 5 ) = ( c d ) / 4 . Then
C 4 ( d ) I ¯ 2 I 2 ) ( H 2 I 2 ) ( T 2 × 3 ( 4 ) ( I ¯ 2 H 2 ) ) × diag ( s 0 ( 5 ) ,   s 1 ( 5 ) ,   s 2 ( 5 ) ,   s 3 ( 5 ) ,   s 4 ( 5 ) )   ( T 3 × 2 ( 3 ) H 2 ) ( H 2 I 2 ) .
Taking into account the properties of structural matrices [5,32], the computational procedure for the five-point DHT-II is represented by the expression
Y 5 × 1 = P 5 ( 0 ) W 5 × 7 W 7 ( 0 ) W 7 × 8 D 8 W 8 × 7 W 7 ( 1 ) W 7 × 10 W 10 × 5 P 5 ( 1 ) X 5 × 1
where
P 5 ( 0 ) = 1 1 1 1 1 ,   P 5 ( 1 ) = 1 1 1 1 1
  W 8 × 7 = 1 T 3 × 2 ( 3 ) H 2 I 2 ,   W 7 ( 1 ) = 1 H 2 I 2 I 2 ,   W 7 × 10 = I 5 P 2 × 5 ( 2 ) ,   W 10 × 5 = 1 2 × 1 I 5 ,   W 7 × 8 = 1 T 2 × 3 ( 4 ) ( I ¯ 2   H 2 ) I 2 ,   D 8 = diag ( 1 ,   s 0 5 ,   s 1 5 ,   s 2 5 ,   s 3 5 ,   s 4 5 ,   1 ,   1 ) , P 2 × 5 ( 2 ) = 1 1 1 1 1 , W 7 ( 0 ) = 1 1 1 1 1 1 1 1 1 1 1 ,   W 5 × 7 = 1 1 1 1 1 1 1 1 1 1
A data flow graph of the proposed algorithm of the five-point DHT-II is presented in Figure 4. In particular, the number of multiplication operations may be reduced from 16 to 5, but the number of addition operations is increased from 20 to 23.

3.5. Algorithm for 6-Point DHT-II

We now propose an algorithm for the six-point DHT-II. The six-point DHT-II is expressed as follows:
Y 6 × 1 = C 6 X 6 × 1 ,
where Y 6 × 1 = y 0 ,   y 1 ,   y 2 ,   y 3 ,   y 4 ,   y 5 T , X 6 × 1 = x 0 ,   x 1 ,   x 2 ,   x 3 ,   x 4 ,   x 5 T , C 6 = 1 c 6 c 6 1 b 6 b 6 1 1 1 1 1 1 1 b 6 b 6 1 c 6 c 6 1 c 6 c 6 1 b 6 b 6 1 1 1 1 1 1 1 b 6 b 6 1 c 6 c 6 with b 6 = 0.366, c 6 = 1.366.
As can be seen, it is c 6 = 1 +   b 6 ; then, we decompose
C 6 = C 6 ( a ) + C 6 ( b ) ,
where
C 6 ( a ) = 1 1 1 1 0 0 1 1 1 1 1 1 1 0 0 1 1 1 1 1 1 1 0 0 1 1 1 1 1 1 1 0 0 1 1 1 ,   C 6 ( b ) = b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 b 6 .
Based on the properties of structural matrices [5,32] and denoted s 0 6 = b 6 , the computational procedure for the six-point DHT-II is represented by the expression
Y 6 × 1 = W 6 × 8 D 8 W 8 × 10 W 10 × 6 X 6 × 1 ,
where
W 10 × 6 = 1 1 1 1 1 1 1 1 1 1 ,   W 6 × 8 = 1 1 1 1 1 1 1 1 1 1 ,   W 8 × 10 = C 6 ( a ) B 2 × 4 ,   B 2 × 4 = 1 1 1 1 1 1 1 1 ,   D 8 = diag ( 1 ,   1 ,   1 ,   1 ,   1 ,   1 ,   s 0 6 ,   s 0 6 ) .
A data flow graph of the proposed algorithm of the six-point DHT-II is presented in Figure 5. We are able to reduce the number of multiplications from 16 to 2 and the number of addition operations from 30 to 26.

3.6. Algorithm for 7-Point DHT-II

To elaborate the algorithm for the seven-point DHT-II, the matrix notation of this transform is expressed as follows:
Y 7 × 1 = C 7 X 7 × 1 ,
where Y 7 × 1 = y 0 ,   y 1 ,   y 2 ,   y 3 ,   y 4 ,   y 5 ,   y 6 T , X 7 × 1 = x 0 ,   x 1 ,   x 2 ,   x 3 ,   x 4 ,   x 5 ,   x 6 T , b 7 =1.3349, c 7 =1.1974, d 7 =0.1583, e 7 =1.4053, g 7 =0.4671, f 7 =0.7524,
C 7 = 1 b 7 e 7 c 7 f 7 d 7 g 7 1 c 7 g 7 e 7 d 7 b 7 f 7 1 d 7 c 7 b 7 g 7 f 7 e 7 1 1 1 1 1 1 1 1 e 7 f 7 g 7 b 7 c 7 d 7 1 f 7 b 7 d 7 e 7 g 7 c 7 1 g 7 d 7 f 7 c 7 e 7 b 7 .
Now, we need to change the order of rows and columns. Let us define the permutations π 7 ,   π 8 in the following form:
π 7 = 1 2 4 1   3 4 5 6 7 2 3 7 6 5   ,   π 8 = 1 2 3 4 5 6 7 1 2 3 4 7 6 5 .
Then, we permute rows and columns of C 7 according to π 7 ,   π 8 . After permutations, the matrix acquires the following structure:
C 7 ( a ) = 1 1 1 1 1 1 1 1 b 7 e 7 c 7 g 7 d 7 f 7 1 c 7 g 7 e 7 f 7 b 7 d 7 1 d 7 c 7 b 7 e 7 f 7 g 7 1 g 7 d 7 f 7 b 7 e 7 c 7 1 f 7 b 7 d 7 c 7 g 7 e 7 1 e 7 f 7 g 7 d 7 c 7 b 7 .
Further, the matrix C 7 ( a ) is decomposed into two components:
C 7 ( a ) = C 7 ( b ) + C 7 ( c ) ,
where
C 7 ( b ) = 1 1 1 1 1 1 1 1 1 1 1 1 1 ,   C 7 ( c ) = b 7 e 7 c 7 g 7 d 7 f 7 c 7 g 7 e 7 f 7 b 7 d 7 d 7 c 7 b 7 e 7 f 7 g 7 g 7 d 7 f 7 b 7 e 7 c 7 f 7 b 7 d 7 c 7 g 7 e 7 e 7 f 7 g 7 d 7 c 7 b 7 .
Matrix C 7 ( B ) has the same entries except on the sign in the first column and first row, which allows us to reduce the number of operations without the need for further transformations [31]. After eliminating the rows and columns containing only zero entries in matrix C 7 ( c ) , we obtain matrix C 6 ( d ) :
C 6 ( d ) = b 7 e 7 c 7 g 7 d 7 f 7 c 7 g 7 e 7 f 7 b 7 d 7 d 7 c 7 b 7 e 7 f 7 g 7 g 7 d 7 f 7 b 7 e 7 c 7 f 7 b 7 d 7 c 7 g 7 e 7 e 7 f 7 g 7 d 7 c 7 b 7 .
The obtained matrix acquires the structure A 3 B 3 B 3 A 3 , with A 3 = b 7 e 7 c 7 c 7 g 7 e 7 d 7 c 7 b 7 and B 3 = g 7 d 7 f 7 f 7 b 7 d 7 e 7 f 7 g 7 . Then [5,32],
C 6 ( d ) = ( I ¯ 2 I 3 ) ( H 2 I 3 ) 1 / 2 [ ( A 3 + B 3 ) ( A 3 B 3 ) ] ( H 2 I 3 ) .
Submatrices ( A 3 + B 3 ) , ( A 3 B 3 ) of quasi-diagonal matrix ( A 3 + B 3 ) ( A 3 B 3 ) are represented as circular convolution matrices [32]:
A 3 + B 3 =   b 7 g 7   e 7 + d 7 c 7 + f 7   c 7 + f 7 b 7 g 7 ( e 7 + d 7 )   e 7 + d 7 ( c 7 + f 7 ) b 7 g 7 ,   A 3 B 3 =   b 7 + g 7   e 7 d 7 c 7 f 7   c 7 f 7 ( b 7 + g 7 ) ( e 7 d 7 ) ( e 7 d 7 ) ( c 7 f 7 ) b 7 + g 7 .
Now, we will move onto dealing with matrices ( A 3 + B 3 ) and ( A 3 B 3 ) . In this case, the expressions for calculating the entries of a circular convolution matrix H 3 =   h 0   h 2   h 1   h 1   h 0   h 2   h 2   h 1   h 0   for N = 3 are as follows [32]:
H 3 = T 3 ( 1 ) T 3 × 4 × diag ( s 0 ,   s 1 ,   s 2 ,   s 3 ) × T 4 × 3 T 3 ( 0 ) ,
where s 0 = ( h 0 + h 1 + h 2 ) / 3 ,   s 1 = h 0 h 2 ,   s 2 = h 1 h 2 ,   s 3 = ( h 0 + h 1 2 h 2 ) / 3 ; T 3 ( 1 ) = 1 1 1 1 1 1 1 , T 3 × 4 = 1 1 1 1 1 , T 4 × 3 = 1 1 1 1 1 , T 3 ( 0 ) = 1 1 1 1 1 1 1 .
To make the ( A 3 + B 3 ) , ( A 3 B 3 ) matrices consistent with the circular convolution expression, we need to modify them. In the ( A 3 + B 3 ) matrix, we change the sign of all terms in the first column and first row. In the ( A 3 B 3 ) matrix, we change the sign in the second column. In this way, we obtain the matrices:
( A 3 + B 3 ) ( 0 ) =   b 7 g 7   ( e 7 + d 7 ) ( c 7 + f 7 ) ( c 7 + f 7 ) b 7 g 7 ( e 7 + d 7 ) (   e 7 + d 7 ) ( c 7 + f 7 ) b 7 g 7 , ( A 3 B 3 ) ( 0 ) =   b 7 + g 7   ( e 7 d 7 ) c 7 f 7   c 7 f 7 b 7 + g 7 ( e 7 d 7 )   ( e 7 d 7 ) c 7 f 7 b 7 + g 7 .
Using the three-point convolution algorithm, the values s i , I = 0, 1, 2, 3 for matrices ( A 3 + B 3 ) ( 0 ) and ( A 3 B 3 ) ( 0 ) take the following form:
s 0 ( 7 ) = ( b 7   g 7   ( e 7 + d 7 )     ( c 7 + f 7 ) ) / 6 ;   s 1 ( 7 ) = ( b 7   g 7 + e 7 + d 7 ) / 2 ;   s 2 ( 7 ) = ( e 7 + d 7     c 7     f 7 ) / 2 ; s 3 ( 7 ) = ( b 7     g 7   c 7     f 7 + 2 ( e 7 + d 7 ) ) / 6 ;   s 4 ( 7 ) = ( b 7   + g 7 + c 7     f 7 e 7 + d 7 ) / 6 ;   s 5 ( 7 ) = ( b 7   + g 7 + e 7   d 7 ) / 2 ; s 6 ( 7 ) = ( c 7     f 7 + e 7   d 7 ) / 2 ;   s 7 ( 7 ) = ( b 7   + g 7 + c 7     f 7 + 2 ( e 7   d 7 ) ) / 6 .
Then,
C 6 ( d ) = ( I ¯ 2 I 3 ) ( H 2 I 3 ) ( T 3 ( 1 ) T 3 × 4 T 3 ( 1 ) T 3 × 4 ) × diag ( s 0 ( 7 ) ,   s 1 ( 7 ) ,   s 2 ( 7 ) ,   s 3 ( 7 ) ,   s 4 ( 7 ) ,   s 5 ( 7 ) ,   s 6 ( 7 ) ,   s 7 ( 7 ) ) × × ( T 4 × 3 T 3 ( 0 ) T 4 × 3 T 3 ( 0 ) ) ( H 2 I 3 ) .
Based on the properties of structural matrices [5,32], the computational procedure for the seven-point DHT-II is represented by the expression
Y 7 × 1 = P 7 ( 1 ) W 7 × 9 W 9 W 9 × 11 D 11 W 11 × 9 W 9 × 14 W 14 × 7 P 7 ( 0 ) X 7 × 1 ,
where
P 7 ( 0 ) = 1 1 1 1 1 1 1 , P 7 ( 1 ) = 1 1 1 1 1 1 1 ,   W 11 × 9 = 1 ( T 4 × 3 T 3 ( 0 ) ) ( T 4 × 3 T 3 ( 0 ) ) I 2 ,     W 9 × 11 = 1 ( T 3 ( 1 ) T 3 × 4 ) ( T 3 ( 1 ) T 3 × 4 ) I 2 , W 9 × 14 = 1 H 2 I 3 P 2 × 7 ,   W 9 = 1 ( I ¯ 2 I 3 ) ( H 2 I 3 ) I 2 ,   W 14 × 7 = 1 2 × 1 I 7 , D 11 = diag ( 1 ,   s 0 7 ,   s 1 7 ,   s 2 7 ,   s 3 7 ,   s 4 7 ,   s 5 7 ,   s 6 7 ,   s 7 ( 7 ) ,   1 ,   1 ) , P 2 × 7 = 1 1 1 1 1 1 1 , W 7 × 9 = 1 1 1 1 1 1 1 1 1 1 1 1 1 1 .
Figure 6 shows a data flow graph of the synthesized algorithm for the seven-point DHT-II. As can be seen, we are able to reduce the number of multiplications from 36 to 8, but the number of addition operations is increased from 42 to 46.

3.7. Algorithm for 8-Point DHT-II

Let us design the algorithm for the eight-point DHT-II. The eight-point DHT-II is expressed as follows:
Y 8 × 1 = C 8 X 8 × 1 ,
where Y 8 × 1 = y 0 ,   y 1 ,   y 2 ,   y 3 ,   y 4 ,   y 5 ,   y 6 ,   y 7 T , X 8 × 1 = x 0 ,   x 1 ,   x 2 ,   x 3 ,   x 4 ,   x 5 ,   x 6 ,   x 7 T , C 8 = 1 b 8 d 8 b 8 1 c 8 0 c 8 1 b 8 0 b 8 1 c 8 d 8 c 8 1 c 8 d 8 c 8 1 b 8 0 b 8 1 c 8 0 c 8 1 b 8 d 8 b 8 1 b 8 d 8 b 8 1 c 8 0 c 8 1 b 8 0 b 8 1 c 8 d 8 c 8 1 c 8 d 8 c 8 1 b 8 0 b 8 1 c 8 0 c 8 1 b 8 d 8 b 8 with b 8 = 1.3066, c 8 = 0.5412, d 8 = 1.4142.
Let us define the permutation
π 8 = 1 2 1 5   3 4 5 6 7 8 3 7 2 4 6 8  
to change the order of columns of C 8 :
C 8 ( a ) = 1 1 d 8 0 b 8 b 8 c 8 c 8 1 1 0 d 8 b 8 b 8 c 8 c 8 1 1 d 8 0 c 8 c 8 b 8 b 8 1 1 0 d 8 c 8 c 8 b 8 b 8 1 1 d 8 0 b 8 b 8 c 8 c 8 1 1 0 d 8 b 8 b 8 c 8 c 8 1 1 d 8 0 c 8 c 8 b 8 b 8 1 1 0 d 8 c 8 c 8 b 8 b 8
After permutations, the matrix C 8 ( a ) acquires the structure
A 4 B 4 A 4 B 4   with   A 4 = 1 1 d 8 0 1 1 0 d 8 1 1 d 8 0 1 1 0 d 8   and   B 4 = b 8 b 8 c 8 c 8 b 8 b 8 c 8 c 8 c 8 c 8 b 8 b 8 c 8 c 8 b 8 b 8 .
Then, C 8 ( a ) = ( H 2 I 4 ) ( A 4 B 4 ) . Based on the properties of structural matrices [30,31], the matrix A 4 has the following structure:
A 2 ( a ) B 2 ( A ) A 2 ( A ) B 2 ( A )   with   A 2 ( A ) = H 2   and   B 2 ( a ) = d 8 I 2 .
Hence, the matrix A 4 is represented as
A 4 = ( H 2 I 2 ) ( H 2   diag ( s 0 ( 8 ) ,   s 1 ( 8 ) ) )
where s 0 ( 8 ) =   s 1 ( 8 ) =   d 8 .
In the B 4 matrix, we change the sign of all terms in the fourth column and fourth row. The obtained matrix B 4 ( a )   matches the pattern A 2 ( b ) B 2 ( b ) B 2 ( b ) A 2 ( b ) . Therefore, we decompose the obtained matrix B 4 ( a ) as
B 4 ( a ) = ( T 23 ( 4 ) I 2 ) [ ( A 2 ( b )   B 2 ( b ) )   ( A 2 ( b )   B 2 ( b ) )   B 2 ( B ) ] ( T 32 ( 3 ) I 2 ) ,
where T 23 ( 4 ) and T 32 ( 3 ) are defined as in (16), A 2 ( b )   B 2 ( b ) = b 8 c 8 b 8 c 8 b 8 c 8 b 8 + c 8 , A 2 ( b )   B 2 ( b ) = b 8 c 8 b 8 c 8 b 8 c 8 b 8 + c 8 , B 2 ( b ) = c 8 c 8 c 8 c 8 .
Next, we present A 2 ( b )   B 2 ( b ) , A 2 ( b )   B 2 ( b ) , and B 2 ( b ) as [5]
A 2 ( b )   B 2 ( b ) = H 2 × diag ( s 2 ( 8 ) ,   s 3 ( 8 ) ) ,   s 2 ( 8 ) = s 3 ( 8 ) = b 8 c 8 , A 2 ( b )   B 2 ( b ) = H 2 × diag ( s 4 ( 8 ) ,   s 5 ( 8 ) ) ,   s 4 ( 8 ) = s 5 ( 8 ) = b 8 c 8 , B 2 ( b ) = H 2 × diag ( s 6 ( 8 ) ,   s 7 ( 8 ) ) ,   s 6 ( 8 ) = s 7 ( 8 ) = c 8 .
Relying on the properties of structural matrices [5,32], the computational procedure for the eight-point DHT-II is represented by the expression
Y 8 × 1 = W 8 W 8 × 10 W 10 D 10 W 10 × 8 P 8 X 8 × 1 ,
where
W 8 = H 2 I 4 ,   W 8 × 10 = ( H 2 I 2 ) ( T 23 ( 4 ) I 2 ) ,   W 10 × 8 = I 4 ( T 3 × 2 ( 3 ) I 2 ) , W 10 = H 2 I 2 H 2 H 2 H 2 ,   D 10 = diag ( 1 ,   1 ,   s 0 ( 8 ) ,   s 1 ( 8 ) ,   s 2 ( 8 ) ,   s 3 ( 8 ) ,   s 4 ( 8 ) ,   s 5 ( 8 ) ,   s 7 ( 8 ) ) , W 8 = 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 , P 8 = 1 1 1 1 1 1 1 1 .
A data flow graph of the proposed algorithm of the eight-point DHT-II is presented in Figure 7. Based on this algorithm, the number of multiplications can be reduced from 40 to 8 and the number of additions from 48 to 26.

4. Results

The experiment confirmed the correctness of the DHT-II matrix factorizations obtained in Section 3. Namely, the coinciding results of the multiplication of factorized matrices from Section 3 (Equations (7), (10), (12), (18), (21), (27), and (30)) with the original DHT-II matrix was evaluated using the MATLAB R2023b software. For this, the DHT-II matrices were directly calculated for different input sequence sizes N using the expression (3) from Section 3. Next, the factorized matrices were obtained using the expressions (7), (10), (12), (18), (21), (27), and (30) for N = 2, 3, 4, 5, 6, 7, and 8, respectively. Further, the results of the direct calculation of the DHT-II matrix were compared to the results of the multiplication of matrices included in the DHT-II factorizations. The results’ consistency was evaluated, indicating the correctness of the developed factorizations.
Based on the DHT-II matrix factorization, the proposed fast algorithms of the DHT-II calculation were represented in Section 3 by the data flow graphs in Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7. To test the correctness of the constructed data flow graphs for the proposed algorithms, 210 sequences of random normally distributed numbers of length N from 2 to 8 were generated. Each sequence was transformed by a direct matrix–vector product using a data flow graph. The coinciding results obtained for each N indicated the correctness of the constructed data flow graphs for the proposed algorithms.
Next, the numbers of multiplications and additions for the direct DHT-II calculation method and the proposed algorithms were estimated and compared. To ensure the accuracy of the estimates, zero elements in the base DHT-II matrices were not taken into account. This meant that the estimate of the multiplicative complexity of the calculations was calculated using the formula N2-Q and the additive complexity was calculated using the formula N(N − 1)-Q, where Q is the number of positions with zero elements in the base transform matrix. So, in the matrix C 4 , four zeros appeared; in the matrix C 8 ,   eight zeros were found. Overall, the number of multiplications on the data flow graphs was determined by the number of graph vertices, indicated by circles with factors. The number of additions was determined by the number of occurrences of two graph edges in one vertex. The proposed algorithms significantly reduced the number of multiplications in the DHT-II algorithms for a 1D signal, with the number of samples in the range of 3 to 8. At the same time, the number of additions was slightly reduced or even increased. The number of additions was reduced by an average of 5% and the number of multiplications was reduced by an average of 73% with the number of samples in the range of 3 to 8. The obtained results are shown in Table 1. The percentage difference in the number of operations is indicated in parentheses, a plus means that the number of operations increased compared to the direct method, and a minus means it was reduced.
The numbers of multiplications and additions for the algorithms known from the literature are shown in Table 2. Also, the percentage difference in the number of operations is indicated in parentheses.

5. Discussion of Computational Complexity

Analyzing the obtained results, we note that for N = 4, the proposed algorithm did not differ from the existing algorithms by the number of multiplications. If N = 8, the number of multiplications may be reduced with the proposed solution by 33% compared to Hu’s algorithm [24] and Hamood’s algorithms [26,27]. Compared to the prime factor algorithm [23], the number of multiplications for N = 8 was not reduced. The number of additions for N = 8 was reduced by 7–46% compared to Hu’s algorithm [24], the prime factor algorithm [23], and the radix-2 algorithms [24,25]. Compared to Hamood’s algorithms [26,27], the number of additions for N = 4 was not reduced and the number of additions for N = 8 was increased by 8%. For N = 4, the number of additions was reduced by 40% and 67% compared to the algorithms from [23,24], respectively.
The constructive difference between the proposed DHT-II algorithms and the existing algorithms is that the critical path in the signal flow graph of any of the presented algorithms contains only one multiplication. If there is more than one multiplication on the critical path of the algorithm, then the multiplication of two n-bit operands results in a 2n-bit product. Repeated multiplication requires more manipulation of the operands and, therefore, requires more time and effort compared to multiplication alone. In fixed-point devices, this may result in overflow–underflow processing. To maintain accuracy, double access to memory is required both when writing and when reading. Using floating point arithmetic requires exponent alignment, mantissa addition, etc. Thus, the presence of more than one multiplication on the critical path for the known algorithms is a noticeable drawback.
Through this research on the proposed DHT-II algorithms, the following shortcomings of the traditional algorithms were determined. The number of additions and multiplications for the algorithms described in the literature, in general, is greater compared to the proposed algorithms. In addition, traditional algorithms often limit the length of the input data sequence to powers of two or three. However, the application of the proposed algorithms may be limited by the following. First, the structural approach is more suitable for constructing fast algorithms for short data sequences. For long data sequences, the structure of the transformation matrices is more difficult to identify. Second, the efficiency of synthesizing fast algorithms based on the structural approach depends on the structural properties of the transformation matrices, namely, how efficiently these matrices can be transformed so that their block structure corresponds to the matrix patterns formulated in [5].

6. Conclusions

As a result of research, the problem of designing the fast algorithms of type-II DHT was solved for short-length input signals based on a structural approach. The algorithms for short-length input signals are of special interest because they are subsequently used as building blocks of algorithms for long-length sequences.
After a literature review, it was decided to use the structural properties of the type-II DHT coefficients matrix to develop fast algorithms for this transform. The proposed algorithms differ from existing solutions by applying pre-decomposition of the original matrices into submatrices, as well as rearranging the rows and/or columns of the resulting matrices to find the factorizations that would reduce the computational complexity. The relations between the values of the matrix entries were also considered. In an experiment, the computational complexity of the proposed algorithms was compared with the direct calculation of matrix–vector products. The resulting DHT-II matrix factorizations reduced the number of additions by 5% on average and the number of multiplications by 73% for DHT-II lengths in the range of 3 to 8.
As a result, the amount of resources used on the signal processor was significantly reduced while allowing for easier operation in real-time. A significant contribution to the speedup of signal processing came from the reduction in multiplications since multiplications are more labor-intensive and time-expensive to use than additions. To achieve computational efficiency, the structure of the transform matrix and not the parameters of DHT-II were optimized.
A comparison of the proposed algorithms based on the structural approach with algorithms known from the literature showed that the efficiency of the proposed algorithms significantly exceeded the efficiency of radix or split-radix algorithms and was close to the efficiency of prime factor algorithms.
As for prospects for further research, we note the following. In [38], as an example, the application of the Hartley transform in wavelet-like decomposition was shown. The application of known orthogonal transformations in multiscale decomposition and reconstruction significantly reduced computational costs and gave the same effect as wavelet decomposition and reconstruction [39,40]. Thus, we assume that the wavelet-like decomposition can be applied to increase the performance of texture image segmentation for skin disease diagnosis [41].

Author Contributions

Conceptualization, A.C.; methodology, A.C., M.P.; software, M.P.; validation, A.C., M.P.; formal analysis, A.C., M.P.; investigation, M.P.; writing—original draft preparation, M.P.; writing—review and editing, A.C.; supervision, A.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Patra, N.; Nayak, S.S. Discrete Hartley transform and its applications—A review. J. Integr. Sci. Technol. 2022, 10, 173–179. [Google Scholar]
  2. Parsai, S.; Jain, S.; Dangi, J. Review paper on fast DHT algorithm using Vedic mathematics. Int. J. Comput. Appl. 2015, 120, 32–35. [Google Scholar] [CrossRef]
  3. Britanak, V.; Yip, P.; Rao, K. Discrete Cosine and Sine Transforms: General Properties, Fast Algorithms and Integer Approximations, 3rd ed.; Academic Press: New York, NY, USA, 2007. [Google Scholar]
  4. de Oliveira Nascimento, F.A. Hartley transform signal compression and fast power quality measurements for smart grid application. IEEE Trans. Power Deliv. 2023, 99, 4134–4144. [Google Scholar] [CrossRef]
  5. Cariow, A. Strategies for the synthesis of fast algorithms for the computation of the matrix-vector product. J. Signal Process. Theory Appl. 2014, 3, 1–19. [Google Scholar] [CrossRef]
  6. Lipinski, P.; Puchala, D. Digital image watermarking using fast parametric transforms. Bull. Pol. Acad. Sci. Tech. Sci. 2019, 67, 463–477. [Google Scholar] [CrossRef]
  7. Kasban, H. A spiral-based image watermarking scheme using Karhunen–Loeve and discrete hartley transforms. Multidimens. Syst. Signal Process. 2017, 28, 573–595. [Google Scholar] [CrossRef]
  8. Zhang, X.; Su, Q. A spatial domain-based colour image blind watermarking scheme integrating multilevel discrete Hartley transform. Int. J. Intell. Syst. 2021, 36, 4321–4345. [Google Scholar] [CrossRef]
  9. Budiman, G.; Suksmono, A.B.; Danudirdjo, D.; Pawellang, S. QIM-based audio watermarking with combined techniques of SWT-DST-QR-CPT using SS-based synchronization. In Proceedings of the 6th International Conference on Information and Communication Technology (ICoICT), Bandung, Indonesia, 3–4 May 2018. [Google Scholar]
  10. Coutinho, V.A.; Cintra, R.J.; Bayer, F.M. Low-complexity three-dimensional discrete Hartley transform approximations for medical image compression. Comput. Biol. Med. 2021, 139, 105018. [Google Scholar] [CrossRef] [PubMed]
  11. Ye, H.S.; Zhou, N.R.; Gong, L.H. Multi-image compression-encryption scheme based on quaternion discrete fractional Hartley transform and improved pixel adaptive diffusion. Signal Process. 2020, 175, 107652. [Google Scholar] [CrossRef]
  12. Ma, M.; Sun, Q.; Gao, X.; Wang, G.; Deng, H.; Zhang, Y.; Guan, Q.; Zhong, X. High-efficiency single-pixel imaging using discrete Hartley transform. AIP Adv. 2021, 11, 075211. [Google Scholar] [CrossRef]
  13. Al-Gharabally, M.; Almutairi, A.F. Frequency-domain subcarrier diversity receiver for discrete Hartley transform OFDM systems. EURASIP J. Wirel. Commun. Netw. 2019, 2019, 78. [Google Scholar] [CrossRef]
  14. Ouyang, X.; Jin, J.; Jin, G.; Li, P. Low complexity discrete Hartley transforms pre-coded OFDM system over frequency-selective fading channel. ETRI J. 2015, 37, 32–42. [Google Scholar] [CrossRef]
  15. Wu, M.; Xie, Y.; Shi, Y.; Zhang, J.; Yao, T.; Liu, W. An adaptively biased OFDM based on Hartley transform for visible light communication systems. IEICE Trans. Fundam. 2024, E107-A, 928–931. [Google Scholar] [CrossRef]
  16. Narendra, K.C.; Satyanarayana, S. Hartley transform based correlation filters for face recognition. In Proceedings of the International Conference on Signal Processing and Communications (SPCOM), Noida, India, 26–28 December 2016. [Google Scholar]
  17. Muliono, R.; Khairina, N.; Harahap, M.K.; Putri, S.M. Analysis discrete Hartley transform for the recognition of female voice based on voice register in singing techniques. J. Phys. Conf. Ser. IOP Publ. 2019, 1361, 012039. [Google Scholar]
  18. Jleed, H.; Bouchard, M. Acoustic environment classification using discrete Hartley transform features. In Proceedings of the IEEE 30th Canadian Conference on Electrical and Computer Engineering (CCECE), Windsor, ON, Canada, 30 April–3 May 2017. [Google Scholar]
  19. Paraskevas, I.; Chilton, E.; Rangoussi, M. Audio classification using features derived from the Hartley transform. In Proceedings of the International Conference on Systems, Signals and Image Processing (IWSSIP’06), Budapest, Hungary, 21–23 September 2006. [Google Scholar]
  20. Paraskevas, I.; Rangoussi, M. The Hartley phase spectrum as an assistive feature for classification. In Advances in Nonlinear Speech Processing. NOLISP 2009. Lecture Notes in Computer Science; Solé-Casals, J., Zaiats, V., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; Volume 5933. [Google Scholar]
  21. Sundararajan, N.; Ebrahimi, A.; Vasudha, N. Two dimensional short time Hartley transforms. Sultan Qaboos Univ. J. Sci. 2016, 21, 4. [Google Scholar] [CrossRef]
  22. Zhu, L.; Zhao, Y.; Fang, Y.; Wang, J. A novel robust digital image watermarking scheme based on attention U-Net++ structure. Vis. Comput. 2024, 40, 8791–8807. [Google Scholar] [CrossRef]
  23. Bi, G.; Zeng, Y. Transforms and Fast Algorithms for Signal Analysis and Representations; Birkhäuser: Boston, MA, USA, 2004. [Google Scholar]
  24. Shu, H.; Wang, Y.; Senhadji, L.; Luo, L. Direct computation of type-II discrete Hartley transform. IEEE Signal Process. Lett. 2007, 14, 329–332. [Google Scholar] [CrossRef]
  25. Chiper, D.F. Fast radix-2 algorithm for the discrete Hartley transform of type II. IEEE Signal Process. Lett. 2011, 18, 687–689. [Google Scholar] [CrossRef]
  26. Hamood, M.T. Fast algorithm for computing the discrete Hartley transform of type-II. Indones. J. Electr. Eng. Inform. 2016, 4, 120–125. [Google Scholar] [CrossRef]
  27. Hamood, M.T. Direct split-radix algorithm for fast computation of type-II discrete Hartley transform. Telecommun. Comput. Electron. Control 2020, 18, 3067–3072. [Google Scholar] [CrossRef]
  28. Chiper, D.F. An improved algorithm for the VLSI implementation of type II generalized DHT that allows an efficient incorporation of obfuscation technique. In Proceedings of the 45th International Conference on Telecommunications and Signal Processing (TSP), Prague, Czech Republic, 13–15 July 2022. [Google Scholar]
  29. Shu, H.; Wu, J.; Yang, C.; Senhadji, L. Fast radix-3 algorithm for the generalized discrete Hartley transform of type II. IEEE Signal Process. Lett. 2012, 19, 348–351. [Google Scholar] [CrossRef]
  30. Cariow, A.; Makowska, M.; Strzelec, P. Small-size FDCT/IDCT algorithms with reduced multiplicative complexity. Radioelectron. Commun. Syst. 2019, 62, 559–576. [Google Scholar] [CrossRef]
  31. Cariow, A.; Lesiecki, Ł. Small-size algorithms for type-IV discrete cosine transform with reduced multiplicative complexity. Radioelectron. Commun. Syst. 2020, 63, 465–487. [Google Scholar] [CrossRef]
  32. Bielak, K.; Cariow, A.; Raciborski, M. The development of fast DST-II algorithms for short-length input sequences. Electronics 2024, 13, 2301. [Google Scholar] [CrossRef]
  33. Polyakova, M.; Witenberg, A.; Cariow, A. The design of fast type-V discrete cosine transform algorithms for short-length input sequences. Electronics 2024, 13, 4165. [Google Scholar] [CrossRef]
  34. Gautam, D. Resourceful fast discrete Hartley transform to replace discrete Fourier transform with implementation of DHT algorithm for VLSI architecture. Turk. J. Comput. Math. Educ. 2021, 12, 5290–5298. [Google Scholar]
  35. Jain, R.; Jain, P. FPGA implementation of DHT through parallel and pipeline structure. In Proceedings of the International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 27–29 January 2021. [Google Scholar]
  36. Chiper, D.F. A structured dual split-radix algorithm for the discrete Hartley transform of length 2N. Circuits Syst. Signal Process. 2018, 37, 290–304. [Google Scholar] [CrossRef]
  37. Chiper, D.F. A systolic array algorithm based on band-convolution structure for an efficient VLSI implementation of the odd-time generalized discrete Hartley transform. In Proceedings of the International Symposium on Signals, Circuits and Systems (ISSCS), Iasi, Romania, 11–12 July 2019. [Google Scholar]
  38. Korohoda, P.; Dabrowski, A. Generalized convolution as a tool for the multi-dimensional filtering tasks. Multidimens. Syst. Signal Process. 2008, 19, 361–377. [Google Scholar] [CrossRef]
  39. Majorkowska, D.; Ţariov, A. Procedures of multilevel 2D data decomposition and reconstruction with wavelet-like packets. Elektron. Konstr. Technol. Zastos. 2008, 48, 48–52. (In Polish) [Google Scholar]
  40. Majorkowska, D.; Ţariov, A. Multilevel signal representation with wavelet-like packets with use of various orthogonal transform matrices. Elektron. Konstr. Technol. Zastos. 2008, 48, 135–140. (In Polish) [Google Scholar]
  41. Polyakova, M.V. Image segmentation with a convolutional neural network without pooling layers in dermatological disease diagnostics systems. Radio Electron. Comput. Sci. Control 2023, 1, 51–61. [Google Scholar] [CrossRef]
Figure 1. Data flow graph of the proposed algorithm for the computation of two-point DHT-II.
Figure 1. Data flow graph of the proposed algorithm for the computation of two-point DHT-II.
Applsci 14 10719 g001
Figure 2. Data flow graph of the algorithm for the computation of three-point DHT-II.
Figure 2. Data flow graph of the algorithm for the computation of three-point DHT-II.
Applsci 14 10719 g002
Figure 3. Data flow graph of the algorithm for the computation of four-point DHT-II.
Figure 3. Data flow graph of the algorithm for the computation of four-point DHT-II.
Applsci 14 10719 g003
Figure 4. Data flow graph of the algorithm for the computation of five-point DHT-II.
Figure 4. Data flow graph of the algorithm for the computation of five-point DHT-II.
Applsci 14 10719 g004
Figure 5. Data flow graph of the algorithm for the computation of six-point DHT-II.
Figure 5. Data flow graph of the algorithm for the computation of six-point DHT-II.
Applsci 14 10719 g005
Figure 6. Data flow graph of the algorithm for the computation of seven-point DHT-II.
Figure 6. Data flow graph of the algorithm for the computation of seven-point DHT-II.
Applsci 14 10719 g006
Figure 7. Data flow graph of the algorithm for the computation of eight-point DHT-II.
Figure 7. Data flow graph of the algorithm for the computation of eight-point DHT-II.
Applsci 14 10719 g007
Table 1. The number of operations of the direct method and the proposed algorithms.
Table 1. The number of operations of the direct method and the proposed algorithms.
NDirect MethodProposed Algorithms
AdditionsMultiplicationsAdditionsMultiplications
2202 (0%)0 (0%)
3647 (+16%)1 (−75%)
4846 (−25%)2 (−50%)
5201623 (+15%)5 (−69%)
6301626 (−13%)2 (−88%)
7423646 (+10%)8 (−78%)
8484026 (−35%)8 (−80%)
Table 2. The number of operations for the algorithms known from the literature.
Table 2. The number of operations for the algorithms known from the literature.
AlgorithmReference, Year of PublicationN = 4N = 8
MultiplicationsAdditionsMultiplicationsAdditions
Hu’s radix-2 DIT algorithm[24], 20072 (0%)18 (−67%)12 (−33%)48 (−46%)
Prime factor algorithm[23], 20042 (0%)10 (−40%)8 (0%)28 (−7%)
Radix-2 algorithm[24], 20074 (−50%)18 (−67%)14 (−43%)42 (−38%)
Radix-2 algorithm[25], 201110 (−20%)28 (−7%)
Radix-2 DIT algorithm[26], 20162 (0%)6 (0%)12 (−33%)24 (+8%)
Split-radix algorithm[27], 20202 (0%)6 (0%)12 (−33%)24 (+8%)
Proposed algorithm26826
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Polyakova, M.; Cariow, A. Fast Type-II Hartley Transform Algorithms for Short-Length Input Sequences. Appl. Sci. 2024, 14, 10719. https://doi.org/10.3390/app142210719

AMA Style

Polyakova M, Cariow A. Fast Type-II Hartley Transform Algorithms for Short-Length Input Sequences. Applied Sciences. 2024; 14(22):10719. https://doi.org/10.3390/app142210719

Chicago/Turabian Style

Polyakova, Marina, and Aleksandr Cariow. 2024. "Fast Type-II Hartley Transform Algorithms for Short-Length Input Sequences" Applied Sciences 14, no. 22: 10719. https://doi.org/10.3390/app142210719

APA Style

Polyakova, M., & Cariow, A. (2024). Fast Type-II Hartley Transform Algorithms for Short-Length Input Sequences. Applied Sciences, 14(22), 10719. https://doi.org/10.3390/app142210719

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop