Next Article in Journal
Asymptotic Properties and Application of GSB Process: A Case Study of the COVID-19 Dynamics in Serbia
Next Article in Special Issue
Smart Lithium-Ion Battery Monitoring in Electric Vehicles: An AI-Empowered Digital Twin Approach
Previous Article in Journal
Hydrogen Storage Prediction in Dibenzyltoluene as Liquid Organic Hydrogen Carrier Empowered with Weighted Federated Machine Learning
Previous Article in Special Issue
Meta-Heuristic Optimization of LSTM-Based Deep Network for Boosting the Prediction of Monkeypox Cases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Probabilistic Models for Data Compression

1
Department of Mathematics and Computer Science, Technical University of Civil Engineering, 020396 Bucharest, Romania
2
Faculty of Mathematics and Computer Science, University of Bucharest, 010014 Bucharest, Romania
3
Department of Applied Mathematics, Bucharest University of Economic Studies, 010734 Bucharest, Romania
4
“Gheorghe Mihoc-Caius Iacob” Institute of Mathematical Statistics and Applied Mathematics, 050711 Bucharest, Romania
5
“Costin C. Kiriţescu” National Institute of Economic Research, 050711 Bucharest, Romania
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(20), 3847; https://doi.org/10.3390/math10203847
Submission received: 1 September 2022 / Revised: 4 October 2022 / Accepted: 12 October 2022 / Published: 17 October 2022

Abstract

:
Our research objective is to improve the Huffman coding efficiency by adjusting the data using a Poisson distribution, which avoids the undefined entropies too. The scientific value added by our paper consists in the fact of minimizing the average length of the code words, which is greater in the absence of applying the Poisson distribution. Huffman Coding is an error-free compression method, designed to remove the coding redundancy, by yielding the smallest number of code symbols per source symbol, which in practice can be represented by the intensity of an image or the output of a mapping operation. We shall use the images from the PASCAL Visual Object Classes (VOC) to evaluate our methods. In our work we use 10,102 randomly chosen images, such that half of them are for training, while the other half is for testing. The VOC data sets display significant variability regarding object size, orientation, pose, illumination, position and occlusion. The data sets are composed by 20 object classes, respectively: aeroplane, bicycle, bird, boat, bottle, bus, car, motorbike, train, sofa, table, chair, tv/monitor, potted plant, person, cat, cow, dog, horse and sheep. The descriptors of different objects can be compared to give a measurement of their similarity. Image similarity is an important concept in many applications. This paper is focused on the measure of similarity in the computer science domain, more specifically information retrieval and data mining. Our approach uses 64 descriptors for each image belonging to the training and test set, therefore the number of symbols is 64. The data of our information source are different from a finite memory source (Markov), where its output depends on a finite number of previous outputs. When dealing with large volumes of data, an effective approach to increase the Information Retrieval speed is based on using Neural Networks as an artificial intelligent technique.

1. Introduction

The assessment of similarity or distance between two information entities is crucial for all information discovery tasks (whether Information Retrieval or Data mining). Appropriate measures are required for improving the quality of information selection and also for reducing the time and processing costs [1].
Even if the concept of similarity originates from philosophy and psychology, its relevance arises in almost every scientific [2]. This paper is focused on measuring similarity in the computer science domain, i.e., Information Retrieval, in images, video, and to some extent audio, respectively. In the paper domain, the similarity measure “is an algorithm that determines the degree of agreement between entities.” [1].
The approaches for computing similarity or dissimilarity between various object representations can be classified [2] in:
(1)
distance-based similarity measures
This class includes the following models: Euclidean distance, Minkowski distance, Mahalanobis distance, Hamming distance, Manhattan/City block distance, Chebyshev distance, Jaccard distance, Levenshtein distance, Dice’s coefficient, cosine similarity, soundex distance.
(2)
feature-based similarity measures (contrast model)
This method, proposed by Tversky in 1977, represents an alternate to distance-based similarity measures, based on computing similarity by common features of compared entities. Entities are more similar if they share more common features, while they are dissimilar in the case of more distinctive features. The similarity between entities A and B can be determined using the formula:
s ( A , B ) = α g ( A B ) β g ( A B ) γ g ( B A ) ,
where:
α , β , γ are used to define the respective weights of associated values,
g ( A B ) describes the common features in A and B,
g ( A B ) stands for the distinctive features of A
g ( B A ) means the distinctive features of the entity B.
(3)
probabilistic similarity measures
For assessing the relevance among some complex data types, the following probabilistic similarity measures are necessary: maximum likelihood estimation, maximum a posteriori estimation.
(4)
extended/ additional measures
Includes similarity measures based on fuzzy set theory [3], similarity measures based on graph theory, similarity-based weighted nearest-neighbors [4], similarity-based neural networks [5].
A lot of algorithms and techniques like Classification [6], Clustering, Regression [7], Artificial Intelligence(AI), Neural Networks (NNs), Association Rules, Decision Trees, Genetic Algorithm, Nearest Neighbor method etc., are attempted [2] for knowledge discovery from databases.
The Artificial Neural Networks (ANNs) are widely used in various fields of engineering and science. They generate useful tools in quantitative analysis, due to their unique feature of approximating complex and non-linear equations. Their performance and advantage consists in their ability to model both linear and non-linear relationships.
The Artificial Neural Networks are well-suited for a very broad class of nonlinear approximations and mappings. An Artificial Neural Network (ANN) with nonlinear activation functions is more effective than linear regression models in dealing with nonlinear relationships.
ANNs are regarded as one of the important components in AI. They have been studied [8] for many years with the goal of achieving human-like performance in many branches, such as classification, clustering, and pattern recognition, speech and image recognition and Information Retrieval [9] by modelling the human neural system.
IR “is different from data retrieval in databases using SQL queries because the data in databases are highly structured and stored in relational tables, while information in text is unstructured. There is no structured query language like SQL for text retrieval.” [10].
Gonzales and Woods [11] have developed Huffman Coding to remove coding redundancy, by yielding the smallest number of code symbols per source symbol.
Burgerr and Burge [12] as well as Webb [13] have developed several algorithms for Image Compression using Discrete Cosine Transformation.
The main objective of this paper consists in improving the performance of the Huffman Coding algorithm by achieving a minimum average length of the code word. Our new approach is important for removing more effectively the Coding Redundancy in Digital Image Processing.
The remainder of the paper is organized as follows. In Section 2 we discuss some general aspects about the Poisson distribution and the data compression.
Then, in Section 3 we introduce and analyze an approach to reduce the documents, entitled Discrete Cosine Transformation, in order to achieve the Latent Semantic Model.
We follow with the Fourier descriptor method in Section 4 to describe the shape of an object by considering its boundaries.
We define some notions from the Information Theory in Section 5 as they are useful to model the information generation like a probabilistic process.
The Section 6 presents the Huffman Coding, which is built to remove the coding redundancy and to find the optimal code for an alphabet of symbols.
We introduce an experimental evaluation of the new model on the task of computing image entropy, the average length of the code words and the Huffman coding efficiency too in Section 7.
We conclude in Section 8 by highlighting that the scientific value added by our paper consists in the fact of computing the average length of the code words in the case of applying of the Poisson distribution.

2. Preliminaries

2.1. Discrete Random Variables and Distributions

By definition [14], a random variable X and its distribution are discrete if the possible values of X denoted x 1 , x 2 , x 3 , are finitely many or, at most, countably many values, with the probabilities p 1 = P ( X = x 1 ) , p 2 = P ( X = x 2 ) , p 3 = P ( X = x 3 ) , .

The Poisson Distribution

The Poisson distribution (named after S.D. Poisson) is the discrete distribution which has infinite possible values and the probability function
P ( k ) = λ k k ! e λ , λ > 0 .
The Poisson distribution is similar with the binomial distribution in the mean in that it is achieved as a limiting case of this distribution, for n and p 0 , where the product n p = λ > 0 is kept constant. As it is used for a rare occurrence of an event, the Poisson distribution is also called the distribution of the rare events that occur in order to achieve success in a sequence of some independent Bernoulli samples.
This distribution is frequently encountered in the study of some phenomena in biology, telecommunications, statistical quality control (when the probability of obtaining a defect is very small), in the study of phenomena that present some agglomerations (in the theory of threads of waiting).
The probability that of the n drawn balls, k are white is: [15]
P ( k ) = lim n P n ( k ) = lim n C n k p k q n k = lim n C n k λ n k 1 λ n n k = = λ k k ! · lim n n ! n k ( n k ) ! = 1 · lim n 1 λ n n k e λ ,
namely it results in the Equation (2).
The simulation of X P o ( λ ) can be achieved [16,17]: using the Matlab generator, using the inverse transform method, using a binomial distribution.
In [17] we generated a selection of volume 1000 on the random variable X, having some continuous distributions (such as normal or exponential distribution) or discrete distributions (as the geometric or Poisson distribution). The methods used in the generation of the random variables X are illustrated in the Table 1.
The means and the variances corresponding to the resulting selections will be compared with the theoretical ones. In each case we build both the graphical histogram and the probability density of X too. For example, the Figure 1 shows the histograms built for X G e o m ( p ) , using three methods (Matlab generator, inverse transform method, by counting the failures) and the probability density of X.
The statistics associated to the concordance tests are also computed in the Table 1.
Example 1.
Adjust the following empirical distribution, which is a discrete one, using the Poisson distribution:
x i 0123456789
P ( x i ) 0.160.2170.270.180.110.030.10.0020.0010.002
The resulting Poisson distribution will be:
x0123456789
P ( x ) 0.07990.20190.25510.21490.13580.06860.02890.01040.00330.0009

2.2. Data Compression

The concept of data compression [11,18,19,20,21,22,23,24,25,26,27,28] means the reduction of that amount of data, which is used to represent a given quantity of information.
The data represent the way in which the information is transmitted, such that different amounts of data can be used to represent the same quantity of information.
For example, if n 1 and n 2 are the number of bits in two representations of the same information, then the relative data redundancy R D ( , 1 ) of the representation with n 1 bits can be defined as:
R D = 1 1 C R ,
where C R ( 0 , ) signifies the compression ratio and has the expression:
C R = n 1 n 2 .
There are the following three types of redundancies [11] in the Digital Image Processing:
(A)
Coding Redundancy- it is necessary to evaluate the optimal information coding by the average length of the code words ( L a v g ) to remove this kind of redundancy:
L a v g = k = 0 L 1 l ( r k ) p r ( r k ) ,
where:
L being the number of the intensity values associated to a an M × N image;
M N L a v g bits are necessary to represent the respective M × N image;
the discrete random variable r k [ 0 , L 1 ] represents the intensities of that M × N image;
n k is the absolute frequency of the k- th intensity r k ;
l ( r k ) , k 0 , L 1 ¯ means the number of bits that are used to represent each value of r k ;
p r ( r k ) = n k M N is the probability of the occurrence of the r k value;
(B)
Interpixel Redundancy one refers to a reduction of the redundancy associated with spatially and temporally correlated pixels through mapping such as the run-lengths, differences between adjacent pixels and so on; reversible mapping implies the reconstruction without error;
(C)
Psychovisual Redundancy is when certain information which has relatively less importance for the perception of the image quality; it is different from the Coding Redundancy and the Interpixel Redundancy by the fact that it is associated with the real information, which can be removed by a quantization method.

3. Discrete Cosine Transformation

The appearance based approach constitutes [2] one of the various approaches used to select the features from an image, by retaining the most important information of the image and rejecting the redundant information. This class includes Principal Component Analysis (PCA), Discrete Cosine Transformation (DCT) [29], Independent Component Analysis (ICA) and Linear Discriminant Analysis (LDA).
In case of large document collections, the high dimension of the vector space matrix F generates problems in the text document set representation and induces high computing complexity in Information Retrieval.
The most often used methods for reducing the text document space dimension which have been applied in Information Retrieval are: Singular Value Decomposition (SVD) and PCA.
Our approach is based on using the Discrete Cosine Transformation for reducing the text documents. Thus, the set of keywords is reduced to the much smaller feature set. The resulting model represents the Latent Semantic Model.
The DCT [30] represents an orthogonal transformation, similar with the PCA. The elements of the transformation matrix are obtained using the following formula:
t m i = 2 δ m 1 n cos π n ( i 1 2 ( m 1 ) ) , ( ) i , m = 1 , n ¯ ,
n being the size of the transformation and
δ m = 1 if m = 1 , 0 otherwise .
The DCT requires the transformation of the n dimensional vectors X p , p = 1 , N ¯ , where N denotes the number of vectors that must be transformed), to the vectors Y p , ( ) p = 1 , N ¯ , using the formula:
Y p = T · X p , ( ) p = 1 , N ¯ ,
T = { t m i } i , m = 1 , n ¯ meaning the transformation matrix.
We have to choose, between all the components of the vectors Y p , p = 1 , N ¯ , a number of m components, corresponding to the positions which conduct to a mean square belonging to the first m mean squares, in descending order, while the other n m components will be cancelled.
The vector Y p , p = 1 , N ¯ , is defined through the formula (8):
Y p = y p 1 y p n .
The mean square of the transformed vectors is given by:
E ( Y p 2 ) = E ( y p 1 2 ) E ( y p n 2 ) ,
where
E ( y p j 2 ) = y p j 2 ¯ = 1 N p = 1 N y p j 2 , ( ) j = 1 , n ¯ .
The DCT application consists in determining the vectors Y ^ p , p = 1 , N ¯ corresponding to the m components of the vectors Y p , p = 1 , N ¯ that do not cancel.

Image Compression Algorithm Using Discrete Cosine Transformation

Digital image processing represents [11,12,13,31,32] a succession of hardware and software processing steps, as well as the implementation of several theoretical methods.
The first step of this process involves the image acquisition. It requires an image sensor for achieving a two-dimensional image, such as a video camera (for example, the Pinhole Camera Model, one of the simplest camera models).
The analog signal (which is continuous in time and values) [33] resulted at the output of the video camera, must be converted into a digital signal, for its processing using the computer. This transformation involves the following steps [12]:  
Step 1
(Spatial sampling). This step aims to achieve the spatial sampling of the continuous light distribution. The spatial sampling of an image represents the conversion of the continuous signal to its discrete representation and it depends on the geometry of the sensor elements associated to the acquisition device.
Step 2
(Temporal sampling). In this stage, the resulting discrete function is sampled in the time domain to create a single image. The temporal sampling is achieved by measuring at regular intervals the amount of light incident on each individual sensor element.
Step 3
(Quantization of pixel values). This step aims to quantize the resulting values of the image to a finite set of numeric values for storing and processing the image values on the computer.
Definition 1
 ([12]). A digital image I represents a two-dimensional function of natural coordinates ( u , v ) N × N , which maps to a range of possible image values P, with the property that I ( u , v ) P .
The pixel values are described by binary words of length k (which define the depth of the image); therefore, a pixel can represent any of 2 k different values.
As an illustration, the pixels of the grayscale images:
  • are represented by using k = 8 bits (1 byte) per pixel;
  • have the intensity values belonging to the set { 0 , 1 , , 255 } , where the value 0 corresponds to the minimum brightness (black) and 255 represents the maximum brightness (white).
The result of performing the three steps Step 1Step 3 is highlighted in a “description of the image in the form of a two-dimensional, ordered matrix of integers” [12], illustrated in the Figure 2.
CASIA Iris Image Database Ver 3.0 (or CASIA-IrisV3 for short) contains three subsets, totally 22,051 iris images of more than 700 subjects.
Figure 3 displays a coordinate system for image processing, which is flipped in the vertical direction, provided that the origin, defined by u = 0 and v = 0 lies in the upper left corner.
Another advantage of DCT in image compression consists in not depending on the input data.
The coordinates u , v represent the columns and rows of the image, respectively. For an image with the resolution M × N , the maximum column number is u max = M 1 , while the maximum row number is v max = N 1 .
After achieving the digital image, it is necessary to preprocess it in order to improve it; we can mention some examples of preprocessing image techniques:
  • image enhancement, which assumes the transformation of the images for highlighting some hidden or obscure details, interest features, etc.;
  • image compression, performed for reducing the amount of data needed to represent a given amount of information;
  • image restoration aims to correct those errors that appear at the image capture.
Among different methods for image compression, the DCT “achieves a good compromise between the ability of information compacting and the computational complexity” [12]. Another advantage of using DCT in image compression consists in not depending on the input data.
The DCT algorithm is used for the compression of 256 × 256 matrix of integers X = ( x i j ) i , j = 1 , 256 ¯ , where x i j { 0 , 1 , , 256 } means the original pixel values. The Algorithm 1 consists in performing the following steps [2]:
We have performed the compression algorithm based on the DCT, using the Lena.bmp image [2], which has 256 × 256 pixels and 256 levels of grey; it is represented in the Figure 4.
The Table 2 and Figure 5 display the experimental results obtained by implementing the DCT compression algorithm, in Matlab.
Algorithm 1: DCT compression algorithm.
Step 1
Split the initial image into 8 × 8 pixel blocks (1024 image blocks).
Step 2
Process each block by applying the DCT, using relation (8).
Step 3
The first nine coefficients for each transformed block are retained in a zigzag fashion and the rest of
( 64 9 ) the coefficients are cancelled (by making them equal to 0). This stage is illustrated in Figure 6.
Step 4
The inverse DCT is applied for each of the 1024 blocks resulted in the previous step.
Step 5
The compressed image represented by the matrix X ^ = ( x ^ i j ) i , j = 1 , 256 ¯ is achieved,
where x ^ i j denotes the encoded pixel values. Then, the pixel values are converted into integer values.
Step 6
The performances of the DCT compression algorithm is evaluated in terms of the
Peak Signal-to-Noise Ratio (PSNR), given by [34,35]:
P S N R = 20 log 10 255 M S E ,
where the Mean Squared Error (MSE) is defined as follows [34,35]:
M S E = 1 N × N i = 1 N j = 1 N ( x i j x ^ i j ) 2 ,
where N × N means the total number of pixels in the image (in our case N = 256 ).
Figure 5. Visual evaluation of the performances corresponding to the DCT compression algorithm.
Figure 5. Visual evaluation of the performances corresponding to the DCT compression algorithm.
Mathematics 10 03847 g005
Figure 6. Zigzag fashion to retain the first nine coefficients.
Figure 6. Zigzag fashion to retain the first nine coefficients.
Mathematics 10 03847 g006

4. Fourier Descriptors

The descriptors of different objects can be compared for achieving [36] a measurement of their similarity [2].
The Fourier descriptors [37,38] show interesting properties in terms of the shape of an object, by considering its boundaries.
Let γ be [2,35] a closed pattern, oriented counterclockwise, described by the parametric representation: z ( l ) = ( x ( l ) , y ( l ) ) , where l denotes the length of a circular arc, along the curve γ , starting from an origin and 0 l < L , where L means the length of the boundary.
A point lying on the boundary generates the complex function u ( l ) = x ( l ) + i y ( l ) . We note that u ( l ) is a periodic function, with period L.
Definition 2.
The Fourier descriptors are the coefficients associated to the decomposition of the function u ( l ) in a complex Fourier series.
By using a similar approach of implementing a Fourier series for building a specific time signal, which consists of cosine/sine waves of different amplitude and frequency, “the Fourier descriptor method uses a series of circles with different sizes and frequencies to build up a two dimensional plot of a boundary” [36] corresponding to an object.
The Fourier descriptors are computed using the formula [2,35]:
a n = 1 L 0 L u ( l ) e i 2 π L n l d l
such that
u ( l ) = n = a n · e i 2 π L n l .
In the case of a polygonal contour, depicted in the Figure 7, we will derive [2,35] an equivalent formula to (12).
Denoting
λ = V k 1 M ¯ M V k ¯ ,
namely (see the Figure 8):
λ = l l k 1 l k l ,
where M ( x M , y M ) , V k 1 ( x k 1 , y k 1 ) , V k ( x k , y k ) and
l 0 = 0 , l k = i = 1 k | V i V i 1 | ;
from (12) it results:
a n = 1 L k = 1 m l k 1 l k ( x M ( l ) + i y M ( l ) ) · e i 2 π L n l d l .
From (14) we deduce:
x M = x k 1 + λ x k 1 + λ , y M = y k 1 + λ y k 1 + λ .
We will regard each coordinate pair as a complex number, as illustrated in the Figure 9:
Taking into consideration the previous assumption and relationships (18) and (15) we get:
x M ( l ) + i y M ( l ) = x k 1 + l l k 1 l k l · x k 1 + l l k 1 l k l + i y k 1 + l l k 1 l k l · y k 1 + l l k 1 l k l
= ( x k 1 + i y k 1 ) l k + l [ ( x k x k 1 ) + i ( y k y k 1 ) ] ( x k + i y k ) l k 1 l k l k 1
= V k 1 l k + l ( V k V k 1 ) V k l k 1 l k l k 1 ;
hence, the formula (17) will become:
a n = 1 L k = 1 m 1 l k l k 1 l k 1 l k [ V k 1 l k + l ( V k V k 1 ) V k l k 1 ] · e i 2 π L n l d l = 1 L k = 1 m 1 l k l k 1 l k 1 l k [ V k 1 l k + l ( V k V k 1 ) V k l k 1 ] · e i 2 π L n l d l =
a n = 1 L k = 1 m 1 l k l k 1 ( V k 1 l k V k l k 1 ) · l k 1 l k e i 2 π L n l d l + ( V k V k 1 ) · l k 1 l k l · e i 2 π L n l d l I .
By computing
I = l k 1 l k l · e i 2 π L n l d l = 1 i 2 π L n l k · e i 2 π L n l k l k 1 · e i 2 π L n l k 1 + 1 2 π L n 2 e i 2 π L n l k e i 2 π L n l k 1
and substituting it into (19) we shall achieve:
a n = 1 L k = 1 m 1 l k l k 1 · T k ,
where
T k = 1 i 2 π L n ( V k 1 l k V k l k 1 ) · e i 2 π L n l k e i 2 π L n l k 1 1 i 2 π L n ( V k V k 1 ) · l k e i 2 π L n l k l k 1 e i 2 π L n l k 1
+ 1 2 π L n 2 ( V k V k 1 ) · e i 2 π L n l k e i 2 π L n l k 1 ,
namely
T k = 1 i 2 π L n V k ( l k l k 1 ) · e i 2 π L n l k V k 1 ( l k l k 1 ) · e i 2 π L n l k 1 + 1 2 π L n 2 ( V k V k 1 ) · e i 2 π L n l k e i 2 π L n l k 1 .
Therefore, on the basis of the relations (21) and (16) from which it results
l k l k 1 = | V k V k 1 | , ( ) k = 1 , m ¯ ,
the formula (20), which allows us the computation of the Fourier descriptors will be:
a n = 1 2 π i n k = 1 m V k · e i 2 π L n l k V k 1 · e i 2 π L n l k 1 + L ( 2 π n ) 2 k = 1 m V k V k 1 | V k V k 1 | · e i 2 π L n l k e i 2 π L n l k 1 .
The principal advantages of the Fourier Descriptor method for object recognition consists in the invariance to translation, rotation and scale displayed by the Fourier descriptors are [36].

5. Measuring Image Information

5.1. Entropy and Information

The fundamental premise in Information Theory is that the information generation can be modeled like a probabilistic process [39,40,41,42,43,44,45,46,47]
Hence, an event E, which occurs with probability P ( E ) will contain I ( E ) information units, where [11]:
I ( E ) = log 1 P ( E ) = log P ( E ) .
There is the convention that the basis of the logarithm determines the unit used to measure the information. In the case when the basis is equal with two, then the information unit it is called a bit (binary digit).
Assuming a discrete set of symbols { a 1 , , a m } , having the associated probabilities { P ( a 1 ) , , P ( a m ) } we see that the entropy of the discrete distribution is [11]:
H ( z ) = j = 1 m P ( a j ) log P ( a j ) ,
where z = ( P ( a 1 ) , , P ( a m ) ) t .
We can note that the entropy from the Equation (23) just depends on the probabilities of the symbols and it measures the randomness or unpredictability of the respective symbols drawn from a given sequence; in fact, the entropy defines the average amount of information that can be obtained by observing a single output of a primary source.
The information which is transferred to the receiver through an information transmission system is a random discrete set of symbols { b 1 , , b n } too, with the corresponding probabilities { P ( b 1 ) , , P ( b n ) } , where [11]:
P ( b k ) = j = 1 m P ( b k | a j ) · P ( a j ) , ( ) k = 1 , n ¯ .
The Equation (24) can be written in the matrix form [11]:
v = Q z ,
where
  • v = ( P ( b 1 ) , , P ( b n ) ) t is the probability distribution of the output alphabet { b 1 , , b n } ;
  • the matrix Q = ( q k , j ) 1 k n 1 j m means is associated with the information transmission system:
    Q = P ( b 1 | a 1 ) P ( b 1 | a 2 ) P ( b 1 | a m ) P ( b 2 | a 1 ) P ( b 2 | a 2 ) P ( b 2 | a m ) P ( b n | a 1 ) P ( b n | a 2 ) P ( b n | a m ) ,
    where
    q k , j = P ( b k | a j ) , ( ) j = 1 , m ¯ , k = 1 , n ¯ .
One can define [11]:
(1)
the conditional entropy:
H ( z | v ) = j = 1 m k = 1 n P ( a j , b k ) log P ( a j | b k ) ,
where P ( a j , b k ) means the joint probability, namely the probability of b k occurring at the same time that a j occurs:
P ( a j , b k ) = P ( b k ) · P ( a j | b k ) .
(2)
the mutual information between z and v, which expresses the reduction of uncertainty about z because of the knowledge of v:
I ( z ; v ) = H ( z ) H ( z | v ) ,
namely
I ( z ; v ) = j = 1 m P ( a j ) log P ( a j ) + j = 1 m k = 1 n P ( a j , b k ) log P ( a j | b k ) =
= j = 1 m k = 1 n P ( a j , b k ) log P ( a j ) + j = 1 m k = 1 n P ( a j , b k ) log P ( a j | b k ) =
= j = 1 m k = 1 n P ( a j , b k ) log P ( a j ) log P ( a j | b k ) =
= j = 1 m k = 1 n P ( a j , b k ) log P ( a j | b k ) P ( a j ) ;
therefore
I ( z ; v ) = j = 1 m k = 1 n P ( a j , b k ) log P ( a j , b k ) P ( a j ) · P ( b k ) .
Taking into account the Bayes Rule [48]:
P ( a j , b k ) = P ( a j ) · P ( b k | a j ) = P ( b k ) · P ( a j | b k )
and from the Equation (24) it will result that:
I ( z ; v ) = j = 1 m k = 1 n P ( a j ) · q k j log q k j j = 1 m P ( a j ) q k j .
From the Equation (30) we can deduce that the minimum value of I ( z ; v ) is 0 and can be achieved in the case when the input and the output alphabet are mutually independent [48], i.e.,
P ( a j , b k ) = P ( a j ) · P ( b k ) .

5.2. The Case of the Binary Information Sources

Let a binary information source with the source alphabet A = { a 1 , a 2 } = { 0 , 1 } and let P ( a 1 ) , P ( a 2 ) be the probabilities that the source to produce the symbols a 1 and a 2 , such that [11]:
P ( a 1 ) = p b s
P ( a 2 ) = 1 p b s = p ¯ b s .
By using the Equation (23) it will result that the following entropy of the binary source [11]:
H ( z ) = p b s log 2 p b s ( 1 p b s ) log 2 ( 1 p b s ) ,
where z = ( p b s , 1 p b s ) t , i.e., one achieves the binary entropy function [11]:
H ( z ) = p b s log 2 p b s p ¯ b s log 2 p ¯ b s ,
having its maximum value (of 1 bit) when p b s = 1 2 .
If there is some noise during the data transmission then the matrix Q from the Equation (26) can be defined as [11]:
Q = 1 p e p e p e 1 p e = p ¯ e p e p e p ¯ e ,
p e being the probability of an error during the transmission of any symbol.
In the case when the output alphabet delivered at the receiver is the the set of symbols B = { b 1 , b 2 } = { 0 , 1 } one achieves that [11]:
v = p ¯ b s p b s + p e p ¯ b s p e p b s + p ¯ e p ¯ b s ,
i.e., one can compute:
P ( b 1 ) = p ¯ e p b s + p e p ¯ b s
P ( b 2 ) = p e p b s + p ¯ e p ¯ b s .
Hence, on the basis of the formula (31), the mutual information between z and v will be:
I ( z ; v ) = p b s p ¯ e log 2 p ¯ e p ¯ e p b s + p e p ¯ b s + p b s p e log 2 p e p e p b s + p ¯ e p ¯ b s + + p ¯ b s p e log 2 p e p ¯ e p b s + p e p ¯ b s + p ¯ e p ¯ b s log 2 p ¯ e p e p b s + p ¯ e p ¯ b s = = ( p b s p ¯ e + p ¯ b s p e ) log 2 ( p ¯ e p b s + p e p ¯ b s ) ( p b s p e + p ¯ e p ¯ b s ) log 2 ( p b s p e + p ¯ e p ¯ b s ) + + p ¯ e ( p b s + p ¯ b s ) log 2 p ¯ e + p e ( p b s + p ¯ b s ) log 2 p e ;
therefore [11]:
I ( z ; v ) = H b s ( p b s p e + p ¯ e p ¯ b s ) H b s ( p e ) .
From the Equation (39) one can notice that:
  • I ( z ; v ) = 0 when p b s has the value 0 or 1;
  • the maximum value of I ( z ; v ) is C = 1 H b s ( p e ) (it means the maximum amount of information that can be transferred, i.e., the capacity of the binary transmission system = BST) and can be achieved when the symbols of the binary source are equally likely to occur, namely p b s = 1 2 . For p e = 1 one obtains the maximum capacity of the BST: C = 1 H b s = 1 , while for p e = 1 / 2 one deduces that C = 0 , i.e., any information can not be transferred through the BST.

5.3. The Noiseless Coding Theorem

Let be a zero-memory source  ( A , z ) , from the Equation (23), which is an information source, having only statistically independent symbols. We suppose that the output of this source is a n- tuple of symbols, then the output of the given zero-memory source is the set A = { α 1 , , α m n } has m n possible values α i , each of them consisting in n symbols from the alphabet A.
As there are the inequalities [11]:
log 1 P ( α i ) l ( α i ) < log 1 P ( α i ) + 1 ,
where l ( α i ) , introduced by the Equation (5) is the length of the coding word used to represent α i , one achieves:
i = 1 m n P ( α i ) log 1 P ( α i ) i = 1 m n P ( α i ) l ( α i ) < i = 1 m n P ( α i ) log 1 P ( α i ) + 1 ;
therefore [11]:
H ( z ) L a v g < H ( z ) + 1 ,
where L a v g means the average number of code symbols required to represent all n-symbol groups.
From the Equation (42) one deduces the Shannon’s first theorem (the noiseless coding theorem) [11], which claims that the output of a zero-memory source can be represented with an average of H ( z ) information units per source symbol:
H ( z ) n L a v g n < H ( z ) n + 1 n ,
namely:
H ( z ) L a v g n < H ( z ) + 1 n
or:
lim n L a v g n = H ( z ) .
The Equation (45) proves that the expression L a v g n can be approximated with H ( z ) by encoding infinitely long extensions corresponding to the single-symbol source.
The efficiency of the coding strategy is given by [11]:
η = n · H ( z ) L a v g .

6. Huffman Coding

Huffman Coding [11] is an error-free compression method, designed to remove the coding redundancy, by yielding the smallest number of code symbols per source symbol, which in practice can be represented by the intensities of an image or the output of a mapping operation.
The Huffman algorithm finds the optimal code for an alphabet of symbols, taking into account the condition that the probabilities associated with the symbols have to be coded one at a time.
The approach of Huffman consists in following:
Step 1
Approximate the given data with a Poisson distribution to avoid the undefined entropies.
Step 2
Create a series of source reductions by sorting the probabilities of the respective symbols in a descending order in order to combine the lowest probability symbols into a single symbol, which replace them in the next source reduction. This process can be repeated as long as a source with two symbols is not reached.
Step 3
Code each reduced source, by starting with the smallest source and then go back to the original source of this work, taking into account that the symbols 0 and 1 are the binary codes with minimal length for a two-symbol source.
The Huffman coding efficiency can be computed using the formula [11]:
η = L a v g H ( z ) ,
L a v g being the average length of the code words, defined in the relation (5) and H ( z ) being the entropy of the discrete distribution, introduced by the Equation (23).

7. Experimental Evaluation

7.1. Data Sets

We will use [2] the images from the PASCAL dataset for evaluating the performance of our method using Matlab.
In this paper we have used 10,102 images from the VOC data sets, which contain significant variability in terms of object size, orientation, pose, illumination, position and occlusion.
The database consists of 20 object classes: aeroplane, bicycle, bird, boat, bottle, bus, car, motorbike, train, sofa, table, chair, tv/monitor, potted plant, person, cat, cow, dog, horse and sheep.
We have used The ColorDescriptor engine [49] for extracting the image descriptors from all the images.
The Figure 10 shows 50 images from the VOC data base.
The performance of our method has been assessed using images from the PASCAL dataset The Pascal VOC challenge represents [34] a benchmark in visual object category recognition and detection, as it provides the vision and machine learning communities with a standard data set of images and annotation.
Our approach uses 64 descriptors for each image belonging to the training and test set, therefore the number of symbols is 64.

7.2. Experimental Results

For our experiments, we used the descriptors corresponding to some images from the VOC data sets and after we approximated our data with a Poisson distribution, we computed the image entropy, the average length of the code words and the Huffman coding efficiency too. We have applied the following Algorithm:
Step 1
Approximate the given data with a Poisson distribution.
Step 2
Create a series of source reductions by sorting the probabilities of the respective symbols in a descending order in order to combine the lowest probability symbols into a single symbol, which replace them in the next source reduction. This process can be repeated as long as a source with two symbols is not reached.
Step 3
Code each reduced source, by starting with the smallest source and then go back the original source of this work, taking into account that the symbols 0 and 1 are the binary codes with minimal length for a two-symbol source.
The achieved results are illustrated in the Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10 and Table 11.

8. Conclusions

This paper proposes to improve the Huffman coding efficiency by adjusting the data using a Poisson distribution, which avoids the undefined entropies too. The performance of our method has been assessed in Matlab, by using a set of images from the PASCAL dataset.
The scientific value added of our paper consists in applying the Poisson distribution in order to minimize the average length of the Huffman code words.
The PASCAL VOC challenge represents [34] a benchmark in visual object category recognition and detection as it provides the vision and machine learning communities with a standard data set of images and annotation.
In this paper we have used 10,102 images from the VOC data sets, which contain significant variability in terms of object size, orientation, pose, illumination, position and occlusion.
The database consists of 20 object classes: aeroplane, bicycle, bird, boat, bottle, bus, car, motorbike, train, sofa, table, chair, tv/monitor, potted plant, person, cat, cow, dog, horse and sheep.
The data of our information source are different from a finite memory source (Markov), where its output depends on a finite number of previous outputs.

Author Contributions

Conceptualization, I.I., M.D., S.D. and V.P.; Data curation, I.I. and M.D.; Formal analysis, M.D., S.D. and V.P.; Investigation, I.I., M.D., S.D. and V.P.; Methodology, I.I., M.D., S.D. and V.P.; Project administration, I.I. and V.P.; Resources, I.I.; Software, I.I. and M.D.; Supervision, I.I. and V.P.; Validation, I.I., M.D., S.D. and V.P.; Writing—original draft, I.I. and S.D.; Writing—review & editing, I.I., M.D., S.D. and V.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a grant of the Romanian Ministery of Education and Research, CNCS—UEFISCDI, project number PN-III-P4-ID-PCE-2020-1112, within PNCDI III.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zaka, B. Theory and Applications of Similarity Detection Techniques. 2009. Available online: http://www.iicm.tugraz.at/thesis/bilal_dissertation.pdf (accessed on 14 July 2022).
  2. Iatan, I.F. Issues in the Use of Neural Networks in Information Retrieval; Springer: Cham, Switzerland, 2017. [Google Scholar]
  3. Hwang, C.M.; Yang, M.S.; Hung, W.L.; Lee, M.L. A similarity measure of intuitionistic fuzzy sets based on the Sugeno integral with its application to pattern recognition. Inf. Sci. 2012, 189, 93–109. [Google Scholar] [CrossRef]
  4. Chen, Y.; Garcia, E.K.; Gupta, M.R.; Rahimi, A.; Cazzanti, L. Similarity-based Classification: Concepts and Algorithms. J. Mach. Learn. Res. 2009, 10, 747–776. [Google Scholar]
  5. Suzuki, K.; Yamada, H.; Hashimoto, S. A similarity-based neural network for facial expression analysis. Pattern Recognit. Lett. 2007, 28, 1104–1111. [Google Scholar] [CrossRef]
  6. Duda, D.O.; Hart, P.E.; Stork, D.G. Pattern Classification, 2nd ed.; John Wiley: New York, NY, USA, 2001. [Google Scholar]
  7. Andersson, J. Statistical Analysis with Swift; Apress: New York, NY, USA, 2021. [Google Scholar]
  8. Reshadat, V.; Feizi-Derakhshi, M.R. Neural network-based methods in information retrieval. Am. J. Sci. Res. 2012, 58, 33–43. [Google Scholar]
  9. Cai, F.; China, C.; de Rijke, M. A Survey of Query Auto Completion in Information Retrieval. Found. Trends R Signal Process. 2016, 10, 273–363. [Google Scholar]
  10. Liu, B. Web DataMining; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  11. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 4th ed.; Pearson: New York, NY, USA, 2018. [Google Scholar]
  12. Burgerr, W.; Burge, M.J. Principles of Digital Image Processing; Fundamental Techniques; Springer: London, UK, 2009. [Google Scholar]
  13. Webb, A. Statistical Pattern Recognition, 2nd ed.; John Wiley and Sons: New York, NY, USA, 2002. [Google Scholar]
  14. Kreyszig, E. Advanced Engineering Mathematics; John Wiley and Sons: New York, NY, USA, 2006. [Google Scholar]
  15. Trandafir, R.; Iatan, I.F. Modelling and Simulation: Theoretical Notions and Applications; Conspress: Bucharest, Romania, 2013. [Google Scholar]
  16. Anastassiou, G.; Iatan, I. Modern Algorithms of Simulation for Getting Some Random Numbers. J. Comput. Anal. Appl. 2013, 15, 1211–1222. [Google Scholar]
  17. Iatan, I.F.; Trandafir, R. Validating in Matlab of some Algorithms to Simulate some Continuous and Discrete Random Variables. In Proceedings of the Mathematics and Educational Symposium of Department of Mathematics and Computer Science; Technical University of Civil Engineering Bucharest: București, Romania; MatrixRom: Bucharest, Romania, 2014; pp. 67–72. [Google Scholar]
  18. Kumar, P.; Parmar, A. Versatile Approaches for Medical Image Compression. Procedia Comput. Sci. 2020, 167, 1380–1389. [Google Scholar] [CrossRef]
  19. Wilhelmsson, D.; Mikkelsen, L.P.; Fæster, S.; Asp, L.E. X-ray tomography data of compression tested unidirectional fibre composites with different off-axis angles. Data Brief 2019, 25, 104263. [Google Scholar] [CrossRef]
  20. Wu, F.Y.; Yang, K.; Sheng, X. Optimized compression and recovery of electrocardiographic signal for IoT platform. Appl. Soft Comput. J. 2020, 96, 106659. [Google Scholar] [CrossRef]
  21. Norris, D.; Kalm, K. Chunking and data compression in verbal short-term memory. Cognition 2021, 208, 104534. [Google Scholar] [CrossRef]
  22. Peralta, M.; Jannin, P.; Haegelen, C.; Baxter, J.S.H. Data imputation and compression for Parkinson’s disease clinical questionnaires. Artif. Intell. Med. 2021, 114, 102051. [Google Scholar] [CrossRef]
  23. Calderoni, L.; Magnani, A. The impact of face image compression in future generation electronic identity documents. Forensic Sci. Int. Digit. Investig. 2022, 40, 301345. [Google Scholar] [CrossRef]
  24. Coutinho, V.A.; Cintra, R.J.; Bayer, F.B. Low-complexity three-dimensional discrete Hartley transform approximations for medical image compression. Comput. Biol. Med. 2021, 139, 3105018. [Google Scholar] [CrossRef]
  25. Ettaouil, M.; Ghanou, Y.; El Moutaouakil, K.; Lazaar, M. Image Medical Compression by a new Architecture Optimization Model for the Kohonen Networks. Int. J. Comput. Theory Eng. 2011, 3, 204–210. [Google Scholar] [CrossRef]
  26. Dokuchaev, N.G. On Data Compression and Recovery for Sequences Using Constraints on the Spectrum Range. Probl. Inf. Transm. 2021, 57, 368–372. [Google Scholar] [CrossRef]
  27. Du, Y.; Yu, H. Medical Data Compression and Sharing Technology Based on Blockchain. In International Conference on Algorithmic Applications in Management; Lecture Notes in Computer Science Book Series (LNTCS); Springer: Cham, Switzerland, 2020; Volume 12290, pp. 581–592. [Google Scholar]
  28. Ishikawa, M.; Kawakami, H. Compression-based distance between string data and its application to literary work classification based on authorship. Comput. Stat. 2013, 28, 851–878. [Google Scholar] [CrossRef]
  29. Jha, C.K.; Kolekar, M.H. Electrocardiogram data compression using DCT based discrete orthogonal Stockwell transform. Biomed. Signal Process. Control 2018, 46, 174–181. [Google Scholar] [CrossRef]
  30. Netravali, A.N.; Haskell, B.G. Digital Pictures: Representation and Compression; Springer: Cham, Switzerland, 2012. [Google Scholar]
  31. Vlaicu, A. Digital Image Processing; Microinformatica Group: Cluj-Napoca, Romania, 1997. (In Romanian) [Google Scholar]
  32. Shih, F.Y. Image Processing and Pattern Recognition; Fundamentals and Techniques; John Wiley and Sons: New York, NY, USA, 2010. [Google Scholar]
  33. Tuduce, R.A. Signal Theory; Bren: Bucharest, Romania, 1998. [Google Scholar]
  34. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. A Fuzzy Neural Network and its Application to Pattern Recognition. IEEE Trans. Fuzzy Syst. 2010, 88, 303–338. [Google Scholar]
  35. Neagoe, V.E.; Stǎnǎşilǎ, O. Pattern Recognition and Neural Networks; Matrix Rom: Bucharest, Romania, 1999. (In Romanian) [Google Scholar]
  36. Janse van Rensburg, F.J.; Treurnicht, J.; Fourie, C.J. The Use of Fourier Descriptors for Object Recogntion in Robotic Assembly. In Proceedings of the 5th CIRP International Seminar on Intelligent Computation in Manufacturing Engineering, Ischia, Italy, 25–28 July 2006. [Google Scholar]
  37. Yang, C.; Yu, Q. Multiscale Fourier descriptor based on triangular features for shape retrieval. Signal Process. Image Commun. 2019, 71, 110–119. [Google Scholar] [CrossRef]
  38. De, P.; Ghoshal, D. Recognition of Non Circular Iris Pattern of the Goat by Structural, Statistical and Fourier Descriptors. Procedia Comput. Sci. 2016, 89, 845–849. [Google Scholar] [CrossRef] [Green Version]
  39. Preda, V. Statistical Decision Theory; Romanian Academy: Bucharest, Romania, 1992. [Google Scholar]
  40. Preda, V.C. The Student distribution and the principle of maximum entropy. Ann. Inst. Stat. Math. 1982, 34, 335–338. [Google Scholar] [CrossRef]
  41. Preda, V.; Balcau, C.; Niculescu, C. Entropy optimization in phase determination with linear inequality constraints. Rev. Roum. Math. Pures Appl. 2010, 55, 327–340. [Google Scholar]
  42. Preda, V.; Dedu, S.; Sheraz, M. Second order entropy approach for risk models involving truncation and censoring. Proc. Rom.-Acad. Ser. Math. Phys. Tech. Sci. Inf. Sci. 2016, 17, 195–202. [Google Scholar]
  43. Preda, V.; Băncescu, I. Evolution of non-stationary processes and some maximum entropy principles. Ann. West Univ.-Timis.-Math. Comput. Sci. 2018, 56, 43–70. [Google Scholar] [CrossRef]
  44. Barbu, V.S.; Karagrigoriou, A.; Preda, V. Entropy and divergence rates for Markov chains: II. The weighted case. Proc. Rom.-Acad.-Ser. A 2018, 19, 3–10. [Google Scholar]
  45. Abdul-Sathar, E.I.; Sathyareji, G.S. Estimation of Dynamic Cumulative Past Entropy for Power Function Distribution. Statistica 2018, 78, 319–334. [Google Scholar]
  46. Sachlas, A.; Papaioannou, T. Residual and Past Entropy in Actuarial Science and Survival Models. Methodol. Comput. Appl. Probab. 2014, 16, 79–99. [Google Scholar] [CrossRef]
  47. Sheraz, M.; Dedu, S.; Preda, V. Entropy measures for assessing volatile markets. Procedia Econ. Financ. 2015, 22, 655–662. [Google Scholar] [CrossRef] [Green Version]
  48. Lehman, E.; Leighton, F.T.; Meyer, A.R. Mathematics for Computer Science; 12th Media Services: Suwanee, GA, USA, 2017. [Google Scholar]
  49. Van de Sande, K.E.A.; Gevers, T.; Snoek, C.G.M. Evaluating color descriptors for object and scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1582–1596. [Google Scholar] [CrossRef]
Figure 1. The histograms built for X G e o m ( p ) , using three methods.
Figure 1. The histograms built for X G e o m ( p ) , using three methods.
Mathematics 10 03847 g001
Figure 2. Transformation of a continuous intensity function F ( x , y ) to a discrete digital image F ( u , v ) .
Figure 2. Transformation of a continuous intensity function F ( x , y ) to a discrete digital image F ( u , v ) .
Mathematics 10 03847 g002
Figure 3. Image coordinates.
Figure 3. Image coordinates.
Mathematics 10 03847 g003
Figure 4. Image Lena.bmp (see https://www.cosy.sbg.ac.at/∼pmeerw/Watermarking/lena.html, accessed on 14 July 2022).
Figure 4. Image Lena.bmp (see https://www.cosy.sbg.ac.at/∼pmeerw/Watermarking/lena.html, accessed on 14 July 2022).
Mathematics 10 03847 g004
Figure 7. Polygonal contour.
Figure 7. Polygonal contour.
Mathematics 10 03847 g007
Figure 8. Parametric representation of a segment.
Figure 8. Parametric representation of a segment.
Mathematics 10 03847 g008
Figure 9. Treating a point M as a complex number.
Figure 9. Treating a point M as a complex number.
Mathematics 10 03847 g009
Figure 10. 50 images from the VOC data base.
Figure 10. 50 images from the VOC data base.
Mathematics 10 03847 g010
Table 1. The methods used to generate a selection of volume 1000 on the random variable X, having some continuous and discrete distribution.
Table 1. The methods used to generate a selection of volume 1000 on the random variable X, having some continuous and discrete distribution.
Random
Variable
Simulation
Method
MeanVarianceStatistics
χ 2 Kolmogorov
NormalMatlab21.939243.947254.39860.0143
Box-Muller22.021644.162542.74150.0101
Rejection polar22.070644.083028.43960.0176
ExponentialMatlab11.03611.03136.30550.0249
Inverse method10.965710.9786102.82230.0193
Gamma distribution11.0111.01108.65510.0253
GeometricMatlab2.33332.3277.77787.453556.91550.05
Inverse method2.33332.7637.77787.682534.66670.0408
Counting the failures2.33332.3617.77787.766439.60330.0454
PoissonMatlab10.99910.96421.14090.049
Binomial distribution11.03810.097739.09640.039
Table 2. Experimental results achieved by implementing the DCT compression algorithm.
Table 2. Experimental results achieved by implementing the DCT compression algorithm.
Number of the Retained CoefficientsPerformances of the DCTCompression lgorithm
PSNRMSE
m 1 = 15 27.605410.6232
m 1 = 9 25.343613.7836
m 1 = 2 21.224222.1477
Table 3. The Huffman coding efficiency.
Table 3. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0014.63244.65340.9955
Mathematics 10 03847 i0024.54014.56090.9955
Mathematics 10 03847 i0034.58024.60050.9955
Mathematics 10 03847 i0044.60374.62330.9957
Mathematics 10 03847 i0054.52774.54610.9959
Table 4. The Huffman coding efficiency.
Table 4. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0064.58394.60390.9957
Mathematics 10 03847 i0074.56814.58690.9959
Mathematics 10 03847 i0084.45874.48230.9947
Mathematics 10 03847 i0094.52214.54110.9958
Mathematics 10 03847 i0104.42574.45270.9939
Table 5. The Huffman coding efficiency.
Table 5. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0114.55094.56930.9959
Mathematics 10 03847 i0124.57374.59270.9958
Mathematics 10 03847 i0134.54844.56680.9959
Mathematics 10 03847 i0144.46964.49280.9948
Mathematics 10 03847 i0154.55674.57660.9956
Table 6. The Huffman coding efficiency.
Table 6. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0164.4624.48560.9947
Mathematics 10 03847 i0174.57424.59340.9958
Mathematics 10 03847 i0184.44074.46470.9946
Mathematics 10 03847 i0194.53624.55630.9956
Mathematics 10 03847 i0204.45664.47990.9948
Table 7. The Huffman coding efficiency.
Table 7. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0214.51564.53680.9953
Mathematics 10 03847 i0224.63504.65590.9955
Mathematics 10 03847 i0234.61244.63310.9955
Mathematics 10 03847 i0244.45974.48330.9947
Mathematics 10 03847 i0254.49244.5150.9949
Table 8. The Huffman coding efficiency.
Table 8. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0264.49304.51530.9951
Mathematics 10 03847 i0274.55414.57320.9958
Mathematics 10 03847 i0284.50264.52190.9957
Mathematics 10 03847 i0294.56914.58770.9959
Mathematics 10 03847 i0304.5414.56120.9955
Table 9. The Huffman coding efficiency.
Table 9. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0314.52334.54220.9958
Mathematics 10 03847 i0324.39614.42470.9935
Mathematics 10 03847 i0334.63184.65290.9954
Mathematics 10 03847 i0344.58374.60370.9957
Mathematics 10 03847 i0354.56614.58530.9958
Table 10. The Huffman coding efficiency.
Table 10. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0364.59214.61150.9957
Mathematics 10 03847 i0374.60934.62990.9955
Mathematics 10 03847 i0384.50634.52570.9957
Mathematics 10 03847 i0394.51164.53220.9954
Mathematics 10 03847 i0404.58254.60250.9957
Table 11. The Huffman coding efficiency.
Table 11. The Huffman coding efficiency.
ImageEntropy L avg Coding Efficiency
Mathematics 10 03847 i0414.35544.38800.9926
Mathematics 10 03847 i0424.59114.61050.9958
Mathematics 10 03847 i0434.49444.51590.9952
Mathematics 10 03847 i0444.45164.47520.9947
Mathematics 10 03847 i0454.36464.39580.9929
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Iatan, I.; Drăgan, M.; Dedu, S.; Preda, V. Using Probabilistic Models for Data Compression. Mathematics 2022, 10, 3847. https://doi.org/10.3390/math10203847

AMA Style

Iatan I, Drăgan M, Dedu S, Preda V. Using Probabilistic Models for Data Compression. Mathematics. 2022; 10(20):3847. https://doi.org/10.3390/math10203847

Chicago/Turabian Style

Iatan, Iuliana, Mihăiţă Drăgan, Silvia Dedu, and Vasile Preda. 2022. "Using Probabilistic Models for Data Compression" Mathematics 10, no. 20: 3847. https://doi.org/10.3390/math10203847

APA Style

Iatan, I., Drăgan, M., Dedu, S., & Preda, V. (2022). Using Probabilistic Models for Data Compression. Mathematics, 10(20), 3847. https://doi.org/10.3390/math10203847

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop