1. Introduction
Information theory has developed greatly in recent years, and has found broader applications in various research fields [
1]. Shannon developed the information theory of entropy [
2], and there have been extensive studies of entropy generalizations. On the other hand, the theory of Tsallis statistics originated in the 1980s, and the principle of entropy maximization was a means of expanding on the basic statistics theory. It was based on the fact that Boltzmann–Gibbs statistical mechanics could be reconstructed by the entropy maximization principle, a development beginning in 1957 with the work of Jaynes [
3,
4]; the framework has been developed primarily for extending statistical mechanics. Tsallis entropy was introduced in 1988 by Constantino Tsallis as a basis for generalizing the standard statistical mechanics on the basis of
q-statistics [
5]. It is possible to derive Tsallis distributions from the optimization of Tsallis entropy. For example, the
q-Gaussian is one of the probability distributions that arise from the maximization of Tsallis entropy.
In this short paper, we aim to define the channel capacity of the coding system for transmission of messages on the basis of Tsallis entropy and aim to understand Tsallis entropy theory from the viewpoint of coding theory and seek a way to maximize the entropy corresponding to the signal event number. For the purpose, we introduce the coding symbols, the appearance probability of the symbols, and the message duration. Through this theory-based development, we reconsider the significance of Tsallis entropy in coding theory. Especially with respect to source coding, the objective is to determine the relational expression when applying the entropy maximization principle to establishing the relationship between code length and the appearance probability of the code used to calculate the channel capacity of a coding system. This theory regarding the relationship between code length and the appearance probability of the code was developed by Brillouin and the simple formula was obtained by maximizing information in the coding system, on the basis of information theory [
6,
7,
8]. Several research progressions have been achieved regarding mutual information [
7]; however, the relational expression between source code length and the appearance probability of the code remains to be improved in the theory.
2. Source Coding for Tsallis Entropy Formulation
Consider all the possible distinct messages that correspond to all the possible combinations of the code
Aj, whose code length is
τj. For instance, a message is described using
n types of code symbols,
Sj (1 ≤
j ≤
n) as follows:
Our aim is to identify the way of coding in which the total information within a given duration can be maximized. The messages, which consist of symbols
Aj with numbers
Nj (1 ≤
j ≤
n), will correspond to all the possible combinations of symbols
Aj. Therefore,
N1 = 2,
N2 = 1,
N3 = 6, and
n = 5 in the message (1). Here, we consider
Ψ, the total number of such distinct messages, in the selection of
n symbols. We assume absolutely no restrictions, constraints, or correlations in using various symbols. We obtain information
I derived from the above messages consisting of
Nj,
Here,
K is an arbitrary constant. If we use entropy unit, we take
K =
kB, Boltzmann’s constant. On the other hand, in information science,
K is equivalent to log
2e. Shannon defines the channel capacity as follows [
2]:
Here,
Ψ is signified as a function of
τ, a message of total duration. The unit of channel capacity is given by bits per second, if the total duration is measured in seconds. We define the total number of code symbols
N in a given message as:
For example,
N = 11 in (1). Thus,
N is variable in different individual messages. Next,
pj is the appearance probability of the
Sj symbol in the messages consisting of a total of
N symbols. In the following summations over
j:
and
Using (5), we can rewrite (3) as follows:
Here,
S represents the Shannon entropy,
S. In this study, we investigate the channel capacity when the entropy is given by Tsallis entropy. Here, we introduce the
q-duration of the message, as follows:
τj signifies the
jth code length. Here, we used escort probability
φj:
In actuality,
For simplification, we use general notation in Tsallis statistics:
Tsallis entropy is given by [
6] (
http://www.tsakkus,cat.cbpf.br/TEMUCO.pdf):
The theory of Tsallis statistics, based on the generalized form of entropy
Sq (
q ∈
R), when
q→1, recovers the Shannon entropy:
We aimed to maximize Tsallis entropy (12) [
8],
Sq, instead of Shannon entropy. Then, we introduced a function
G, using non-determined parameters
β and
γ, in reference to (8), (9), and (10):
Then
For calculation of (15), we used
For maximization of
, setting the right sides of (15) and (16) equal to zero, we have:
and solving the above equations with respect to undetermined coefficients β and γ, we have:
and
Rewriting (21), using the
q-logarithm function,
and we obtain from (8), (21) and (22):
Equation (23) implies that most probable code symbols must be short, while the improbable code may be long. In fact, when
q approaches 1, (23) gives the logarithm according to Brillouin’s work using another constant
γ’:
Thus, the probability of symbol appearance can be described using the Tsallis duration in (22), which is similar to the Shannon entropy coding in (24) [
9]. The above result is explicitly a natural extension of Brillouin’s theory, regarding the relationship between coding and Shannon entropy, to the concept of Tsallis entropy.
4. Conclusions
In this short article, we achieved an important formulation between code length and appearance probability in Equation (26). In a similar way to how Shannon’s entropy was extended to Tsallis entropy, the source-coding theory based on the former entropy by Brillouin [
9,
10] was theoretically generalized for the theory based on Tsallis entropy [
5,
7,
11]. On the other hand, it remains to be discussed how one should interpret
q-duration in Equations (7) and (19). The duration of a given message event is generally shorter than the actual message duration; however, when the appearance probability distribution obeys a
q-Gaussian, the
q-duration is an indicator available for use in the analysis of the coding background, in place of the actual duration. From this perspective, we will further investigate the definition of the
q-duration and the interpretation of the limitation of our calculation in future.
My theoretical attempt to generalize source coding can be applied to data management within the areas of data communications, processing, and conversion, particularly in the development of imaging applications. Further investigation is needed with regards to the tractability of Tsallis entropy and
q-statistics in evaluating actual experimental data. We actually applied the medical imaging technique, following previous reports [
12,
13]; in future work, we look to investigate systems in which entropy is measurable and/or definable.