• Nem Talált Eredményt

A parametric mean length is defined as the quantity αβLu= α α−1  1−X Piβ ui Puipβi !α1 D−ni(α−1α

N/A
N/A
Protected

Academic year: 2022

Ossza meg "A parametric mean length is defined as the quantity αβLu= α α−1  1−X Piβ ui Puipβi !α1 D−ni(α−1α"

Copied!
5
0
0

Teljes szövegt

(1)

http://jipam.vu.edu.au/

Volume 6, Issue 4, Article 117, 2005

SOME RESULTS ON A GENERALIZED USEFUL INFORMATION MEASURE

1ABUL BASAR KHAN,1BILAL AHMAD BHAT, AND2S. PIRZADA

1DIVISION OFAGRICULTURALECONOMICS ANDSTATISTICS, SHER-E-KASHMIR

UNIVERSITY OFAGRICULTURALSCIENCES ANDTECHNOLOGYJAMMU

FACULTY OFAGRICULTURE

MAINCAMPUSCHATHA-180009 INDIA

bhat_bilal@rediffmail.com

2DEPARTMENT OFMATHEMATICS

UNIVERSITY OFKASHMIR

SRINAGAR-190006, INDIA

sdpirzada@yahoo.co.in

Received 01 June, 2005; accepted 23 September, 2005 Communicated by N.S. Barnett

ABSTRACT. A parametric mean length is defined as the quantity

αβLu= α α1

1X

Piβ ui Puipβi

!α1

D−ni(α−1α )

,

where α 6= 1, X pi= 1

this being the useful mean length of code words weighted by utilities,ui. Lower and upper bounds forαβLuare derived in terms of useful information for the incomplete power distribution, pβ.

Key words and phrases: Entropy, Useful Information, Utilities, Power probabilities.

2000 Mathematics Subject Classification. 94A24, 94A15, 94A17, 26D15.

1. INTRODUCTION

Consider the following model for a random experimentS, SN = [E;P;U]

whereE = (E1, E2, . . . , En)is a finite system of events happening with respective probabilities P = (p1, p2, . . . , pN), pi ≥ 0, andP

pi = 1and credited with utilitiesU = (u1, u2, . . . , uN),

ISSN (electronic): 1443-5756

c 2005 Victoria University. All rights reserved.

The authors wish to thank the anonymous referee for his valuable suggestions, which improved the presentation of the paper.

176-05

(2)

ui >0,i= 1,2, . . . , N. Denote the model byE, where

(1.1) E =

E1 E2 · · · EN p1 p2 · · · pN u1 u2 · · · uN

We call (1.1) a Utility Information Scheme (UIS). Belis and Guiasu [3] proposed a measure of information called ‘useful information’ for this scheme, given by

(1.2) H(U;P) =−X

uipilogpi,

whereH(U;P) reduces to Shannon’s [8] entropy when the utility aspect of the scheme is ig- nored i.e., when ui = 1 for each i. Throughout the paper, P

will stand for PN

i=1 unless otherwise stated and logarithms are taken to baseD(D >1).

Guiasu and Picard [5] considered the problem of encoding the outcomes in (1.1) by means of a prefix code with codewords w1, w2, . . . , wN having lengths n1, n2, . . . , nN and satisfying Kraft’s inequality [4]

(1.3)

N

X

i=1

D−ni ≤1,

whereDis the size of the code alphabet. The useful mean lengthLu of code was defined as

(1.4) Lu =

Puinipi

Puipi and the authors obtained bounds for it in terms ofH(U;P).

Longo [8], Gurdial and Pessoa [6], Khan and Autar [7], Autar and Khan [2] have studied generalized coding theorems by considering different generalized measures of (1.2) and (1.4) under condition (1.3) of unique decipherability.

In this paper, we study some coding theorems by considering a new function depending on the parametersαandβ and a utility function. Our motivation for studying this new function is that it generalizes some entropy functions already existing in the literature (see C. Arndt [1]).

The function under study is closely related to Tsallis entropy which is used in physics.

2. CODINGTHEOREMS

Consider a function

(2.1) αβH(U;P) = α

α−1

1−

Puipαβi Puipβi

!α1

, whereα >0 (6= 1),β >0,pi ≥0,i= 1,2, . . . , N andP

pi ≤1.

(i) When β = 1 and α → 1, (2.1) reduces to a measure of useful information for the incomplete distribution due to Belis and Guiasu [3].

(ii) When ui = 1 for eachi i.e., when the utility aspect is ignored,P

pi = 1, β = 1and α →1, the measure (2.1) reduces to Shannon’s entropy [10].

(iii) Whenui = 1for eachi, the measure (2.1) becomes entropy for theβ-power distribution derived fromP studied by Roy [9]. We callαβH(U;P)in (2.1) the generalized useful measure of information for the incomplete power distributionPβ.

(3)

Further consider,

(2.2) αβLu = α

α−1

1−X

Piβ ui Puipβi

!α1

D−ni(α−1α )

, whereα >0 (6= 1),P

pi ≤1.

(i) Forβ = 1,ui = 1for eachiandα→1,αβLuin (2.2) reduces to the useful mean length Lu of the code given in (1.4).

(ii) Forβ = 1,ui = 1for eachiandα→1,αβLubecomes the optimal code length defined by Shannon [10].

We establish a result, that in a sense, provides a characterization of αβH(U;P) under the condition of unique decipherability.

Theorem 2.1. For all integersD >1

(2.3) αβLuαβH(U;P)

under the condition (1.3). Equality holds if and only if

(2.4) ni =−log uiPiαβ

Puipαβi

! . Proof. We use Hölder’s [11] inequality

(2.5) X

xiyi ≥X xpi

1pX yiq

1q

for allxi ≥ 0, yi ≥ 0, i = 1,2, . . . , N whenP < 1 (6= 1)andp−1+q−1 = 1,with equality if and only if there exists a positive numbercsuch that

(2.6) xpi =cyiq.

Setting

xi =p

αβ α−1

i

ui Puipβi

!α−11 D−ni,

yi =p

αβ 1−α

i

ui

Puipβi

!1−α1 ,

p= 1−1/αandq= 1−αin (2.5) and using (1.3) we obtain the result (2.3) after simplification

for α−1α >0asα >1.

Theorem 2.2. For every code with lengths{ni},i= 1,2, ..., N,αβLu can be made to satisfy, (2.7) αβLuαβH(U;P)D(1−αα )+ α

1−α h

1−D(1−αα )i . Proof. Letnibe the positive integer satisfying, the inequality

(2.8) −log uiPiαβ

Puipαβi

!

≤ni <−log uiPiαβ Puipαβi

! + 1.

Consider the intervals

(2.9) δi =

"

−log uiPiαβ Puipαβi

!

,−log uiPiαβ Puipαβi

! + 1

#

(4)

of length 1. In everyδi, there lies exactly one positive numbernisuch that (2.10) 0<−log uiPiαβ

Puipαβi

!

≤ni <−log uiPiαβ Puipαβi

! + 1.

It can be shown that the sequence {ni}, i = 1,2, . . . , N thus defined, satisfies (1.3). From (2.10) we have

ni <−log uiPiαβ Puipαβi

! (2.11) + 1

⇒D−ni < uiPiαβ Puipαβi

! D

⇒D−ni(α−1α ) < uiPiαβ Puipαβi

!1−αα Dα−1α

Multiplying both sides of (2.11) bypβi

ui

Puipαβi

α1

,summing overi = 1,2, . . . , N and simpli-

fying, gives (2.7).

Theorem 2.3. For every code with lengths{ni}, i = 1,2, ..., N, of Theorem 2.1,αβLu can be made to satisfy

(2.12) αβH(U;P)≤ αβLu < αβH(U;P) + α

α−1(1−D) Proof. Suppose

(2.13) ni =−log uiPiαβ

Puipαβi

!

Clearly ni and ni + 1 satisfy ‘equality’ in Hölder’s inequality (2.5). Moreover, ni satisfies Kraft’s inequality (1.3).

Supposeniis the unique integer betweenni andni+ 1, then obviously,ni satisfies (1.3).

Sinceα >0 (6= 1), we have Xpβi ui

Puipβi

!α1

Dni(α−1)/α (2.14)

≤X

pβi ui

Puipβi

!α1

Dni(α−1)/α

< D

Xpβi ui Puipβi

!α1

Dni(α−1)/α

Since,

Xpβi ui Puipβi

!

1

αDni(α−1)/α=

Puipαβi Puipβi

!α1

(5)

Hence, (2.14) becomes Puipαβi

Puipβi

!α1

≤X

pβi ui Puipβi

!α1

D−ni(α−1)/α< D

Puipαβi Puipβi

!α1

which gives the result (2.12).

REFERENCES

[1] C. ARNDT, Information Measures- Information and its Description in Science and Engineering, Springer, (2001) Berlin.

[2] R. AUTARANDA.B. KHAN, On generalized useful information for incomplete distribution, J. of Comb. Information and Syst. Sci., 14(4) (1989), 187–191.

[3] M. BELIS AND S. GUIASU, A qualitative-quantitative measure of information in Cybernetics Systems, IEEE Trans. Information Theory, IT-14 (1968), 593–594.

[4] A. FEINSTEIN, Foundation of Information Theory, McGraw Hill, New York, (1958).

[5] S. GUIASU AND C.F. PICARD, Borne infericutre de la Longuerur utile de certain codes, C.R.

Acad. Sci, Paris, 273A (1971), 248–251.

[6] GURDIAL AND F. PESSOA, On useful information of orderα, J. Comb. Information and Syst.

Sci., 2 (1977), 158–162.

[7] A.B. KHANANDR. AUTAR, On useful information of orderαandβ, Soochow J. Math., 5 (1979), 93–99.

[8] G. LONGO, A noiseless coding theorem for sources having utilities, SIAM J. Appl. Math., 30(4) (1976), 739–748.

[9] L.K. ROY, Comparison of Renyi entropies of power distribution, ZAMM, 56 (1976), 217–218.

[10] C.E. SHANNON, A Mathematical Theory of Communication, Bell System Tech-J., 27 (1948), 394–423, 623–656.

[11] O. SHISHA, Inequalities, Academic Press, New York, (1967).

Hivatkozások

KAPCSOLÓDÓ DOKUMENTUMOK

Abstract: We present bounds and approximations for the smallest positive zero of the Laguerre polynomial L (α) n (x) which are sharp as α → −1 +. We indicate the applicability of

The various properties and characteristics for functions belonging to the class U (λ, α, β, k) derived here include (for example) a characterization theorem, coefficient

It follows that the equality holds if and only if the vectors α and β are linearly dependent; or the vector γ is a linear combination of the vector α and β, and xy = 0 but x and y

Abstract: For functions f (z) which are starlike of order α, convex of order α, and λ-spiral- like of order α in the open unit disk U , some interesting sufficient conditions

For functions f (z) which are starlike of order α, convex of order α, and λ-spiral- like of order α in the open unit disk U , some interesting sufficient conditions

Second order lower bounds for the entropy function expressed in terms of the index of coincidence are derived.. Equivalently, these bounds involve entropy and Rényi entropy of

Second order lower bounds for the entropy function expressed in terms of the index of coincidence are derived.. Equivalently, these bounds involve entropy and Rényi entropy of

Key words and phrases: Univalent functions, Starlike functions of order α, Convex functions of order α, Inverse functions, Coefficient estimates.. 2000 Mathematics