Schramm 1955
Schramm 1955
SPRING 1955
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
Information Theory and Mass Communication 133
end of the line, we do not have com- transmitted signal succeeds in reducing
munication. Unless the concept in the the number of equally probable out-
semantic system of Mr. A. is repro- comes at the receiving end by one-half,
duced in the semantic system of Mr. B., one bit of information is said to have
communication has not taken place. been transferred. (Bit comes from bi-
Begging the question of whether a nary digit.) Thus, when you reduce the
meaning as seen by one individual can two equally probable outcomes of a
ever be reproduced exactly by another coin toss to one, you are using one bit
individual-or whether we can test it of information. You can see that the
accurately enough to be sure-we have computing of this information readily
no great difficulty in adapting this defi- lends itself to using logarithms to the
nition to our common understanding of base 2, rather than our common base
the term communication. 10. In the case of the coin toss, log,2
But when we define information in = 1 bit. But it would take log,42 or
terms of information theory, then we about 5.4 bits of information to predict
have to get used to a somewhat differ- which typewriter key would be struck
ent approach. We can, of course, meas- at random, or log,26 (4.7 bits) to pre-
ure the “information” transmitted along dict which letter of the alphabet will
a communication chain in terms of come up, if one is chosen at random.
many kinds of units-letters, mor- This brings us to the basic terms of
phemes, phonemes, facts (if we can information theory, entropy and redun-
satisfactorily define a fact). But none dancy. Entropy simply means the un-
of these is satisfactory for the precise certainty or disorganization of a sys-
needs of information theory. Informa- tem; redundancy is the opposite. En-
tion is there defined in terms of its abil- tropy is, of course, a famous term
ity to reduce the uncertainty or dis- derived from mathematical physics,
organization of a situation at the re- where it has been used to talk about
ceiving end. Newton’s second law of thermodynam-
Let us take an example. Suppose I ics. The law that “entropy always in-
tell you one “fact” about a coin toss creases,” said Eddington, “holds, I
and one “fact” about typewriter keys. think, the supreme position among the
I tell you that tails will not come up laws of Nature.” It is this law, he also
when the coin is next tossed, and that said-the tendency of physical systems
the letter G will not be struck when the to become always more shuffled, less
next key is depressed on the typewriter. organized-which is the only way we
Now it is obvious that the information could tell whether a movie of the phys-
about the coin is more useful to you ical world were being run backward or
than the information about the type- forward. It is not surprising that Shan-
writer in predicting what will happen. non, trying to describe information in
You will have no remaining doubt as to terms of the reduction of uncertainty,
which side of the coin will come up, should use the term entropy and the tra-
whereas you will still be uncertain ditional mathematical symbol for that
which of the remaining 41 keys of the term, H.
typewriter will be struck. In terms of Entropy is measured in terms of the
information theory, more information information required to eliminate the
has been transferred about the coin uncertainty or randomness from a situ-
than about the typewriter. When a ation within a system or involving two
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
134 JOURNALISM QUARTERLY
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
Information Theory and Mass Communication 135
amination of the constraints on these The amount of redundancy-using
events, and discovering the dependency that term freely-is therefore one of
of events in one of these systems on the great strategy questions confronting
events in another system. mass communication. The most eco-
For example, a large part of what we nomical writing is not always the most
call “effects” study is the comparison of effective writing. We could write this
events in one system with events in an- entire paper in the terse, economical
other. A readership study compares the language of mathematics, but that
events in a newspaper with the events would not necessarily communicate
in an individual‘s reading behavior. A most to the people who will read this
retention study compares the events in paper. A newspaper reporter may
a medium with the events in an indi- choose to explain the term photosyn-
vidual’s recall. And so forth. We have thesis in twenty words, which is redun-
every reason to suspect, therefore, that dancy unnecessary to a scientist but
a mathematical theory for studying elec- highly necessary to a layman. There is
tronic communication systems ought t o a kind of rule of thumb, in preparing
have some carry-over to human com- technical training materials, that two or
munication systems. more examples or illustrations should
be given for each important rule or
ENTROPY AND REDUNDANCY
term. There is another rule of thumb,
The term entropy is still strange to in broadcast commercials, that product
students of social communication, but names should be repeated three times.
redundancy is an old and familiar idea. All these are strategy decisions, aimed
The redundancy concept of informa- at using the optimum of redundancy.
tion theory gives us no great trouble. And indeed, finding the optimum de-
Kedundancy is a measure of certainty gree of redundancy for any given com-
or predictability. In information theory, munication purpose is one of the chief
as in social communication, the more problems of encoding.
redundant a system is, the less infor-
Relative entropy, as we have pointed
mation it is carrying in a given time.
out, is merely the other side of the coin
On the other hand, any language or any
from redundancy. The lower the redun-
code without redundancy would be
dancy, the higher the relative entropy.
chaos. In many cases, increasing the
redundancy will make for more effi- One of the aspects of human com-
cient communication. munication where entropy and redun-
For example, on a noisy teletype dancy measures have already proved
line, it helps considerably to have cer- their usefulness is in the study of lan-
tain probabilities controlling what let- guage. Morphemes, phonemes, letters
ters follow others. If a q (in English) and other linguistic units obviously do
occurs followed by two obvious errors, not occur in a language completely at
the operator at the receiving end can random; they are bound by certain se-
be quite sure that the q is followed by quential relationships, and therefore
a u and that the next letter will be an- subject to measures of entropy and re-
other vowel. When a circuit is bad, dundancy. We know, among other
operators arbitrarily repeat key words other things, that the relative entropy
twice. Remember the familiar cable of English is slightly less than 50%.2
language-THIS IS NOT-REPEAT,
NOT. . . ’This is calculated BS follows: The maximum
entropy of 26 English letters 15 log26 or about
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
136 JOURNALISM QUARTERLY
Shannon has estimated, incidentally, -if we get this result, it is clear that
that if the relative entropy of the lan- the uncertainty or relative entropy of
guage was only 20%-if the next let- Paragraph B is considerably greater for
ter in a sequence were, on the average, this audience than is that of Paragraph
SO% predictable-then it would be im- A. Paragraph A is apparently more re-
possible to construct interesting cross- dundant than B. Taylor has gone into
word puzzles. But if the relative en- this use of information theory in his
tropy were 70%-if the structure were doctoral dissertation, and it is clear
only 30% redundant-then it would that the redundancy or relative entropy
bc easily possible to construct three- of a passage is closely related to its
dimensional crossword puzzles ( 1). readability.
This information about crossword puz- If we consider an entire mass me-
zles, of course, is not intended to rep- dium as a system, then it is evident
resent the results of modern linguistic that the maximum entropy of a news-
scholarship. For a more representative paper o r a broadcasting station is im-
example, see Jacobson, Halle and mensely greater than that of a sema-
Cherry on the entropy of Russian pho- phor, a calling card, a personal letter or
nemes (9). a sermon. The paper or the station has
Wilson Taylor’s “Cloze” procedures a very great freedom to do different
is one of the interesting ways we have things and produce strikingly different
available for use in estimating the en- products. A large newspaper, like the
tropy or redundancy of prose. Taylor New York Times,has higher maximum
deletes every nth word in a passage, entropy than a smaller newspaper. If
and asks readers to supply the missing we could devise any way to make a
words. The scatter of different words valid comparison, I think we should
suggested for each of the missing terms find that the relative entropy of radio
provides a measure of the predictabil- and television would be less than that
ity of the passage to that particular of newspapers. If this is indeed the
audience. For example, if we present case, it may be that the tremendous
two paragraphs to the same group of wordage of broadcasting puts a burden
20 readers, and on the average this is on originality, and the scant feedback
the score they make: to a broadcasting station puts a pre-
mium on any formula which has proved
Paragraph A
popular. A successful formula is soon
16 specify word A (correct) imitated. A popular program promptly
2 B
2 C spawns a whole family that look like
it. A joke passes quickly from come-
Paragraph B dian to comedian. We might say that
6 specify word A (correct) for comedians, joint and conditional
4 B entropy are quite low. For comic strips,
4 C
3 D relative entropy is obviously very low,
1 E and redundancy very high.
1 F But it is also evident that no medium
1 G
would be lower if we figured sequential entropy
4.7 bitr per letter. The sequential entropy of for sequences longer than eight ktten.
groups of eight letters as they occur in English 1% Wilson L. Taylor, “Cloze Procedure: A
usage is about 2.35 bits per letter. Therefore, the New Tool for Measuring Readability,” J O ~ N A L -
relaclve entropy is 2.3V4.7 or about 3. This ISM Q U A I ~ L Y , 30:415-33 (Fall 1953).
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
Information Theory and Mass Communication 137
FIGURE A
-
I 8 a ma
I20
SOURCES OF NEWS
B Y CITIES
l8
o0o
ao
PLLY N E W YORK
11/8/54
TIMES
uses as much of the available freedom Times for that day was about 52%.
as it could. Complete freedom would Throughout that week it hung around
mean random content. The art of being SO%, minus or plus 5. This seems quite
an editor or a program manager con- typical of large newspapers. Four Chi-
sists in no small degree of striking the cago papers, two other New York pa-
right belance between predictability pers, and the Washington Post and
and uncertainty-right balance being Times Herald, all were between 41 and
defined as the best combination of sat- 57% for the same period. The London
isfied anticipation and surprise. From Times and Paris Figaro were a little
time to time we have tried to quantify over 40%. During the same period, a
this amount of organization or predict- radio news wire averaged about 45%
ability in a mass medium. One of the relative entropy.
simpler ways to approach it is to tabu- This rather remarkable order of
late the news sources in a paper. ~greementrepresents a pattern of con-
For example, Figure A is a typical straint which, if we understood it com-
distribution of news items by source in pletely, would tell us a great deal about
a metropolitan newspaper for one day. mass media. Why do large papers, on
The usual way we handle figures like the average, use about half the freedom
this is by means of the statistics of cen- open to them to represent different
tral tendency-mean, standard devia- news sources? Availability is one rea-
tion, etc. Suppose we were to handle it son, but the chief reason is simply that
by information theory mathematics. If this is the editors' definition of what
relative entropy were at a maximum, their clientele want, and can absorb,
each of these news sources would be and should have, and can be given
represented equally. Actually, the rela- within the bounds of physical limits
tive entropy of news sources in the and custom. Information theory ap-
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
138 JOURNALISM QUARTERLY
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
Information Theory and Mass Communication 139
ronment to represent, and others. All Or think of the receiver at the end of
these are subject to the stability and the mass communication chain. What
fidelity measures of information theory: stories from the Reader‘s Digest, what
how likely are they to pass on the in- items from the newspaper, does he pass
formation that comes to them? How on to his friends? And how accurately
faithfully are they likely to reproduce does he represent the content? Does he
it? reproduce the part of the content which
Even the terms used to talk about reinforces his previous attitudes? Does
fidelity in electronic systems sound fa- he get the point of an article?
miliar to us in light of our experience Rumor analysis is a fascinating use
with mass communication. How much for the coupling concepts of informa-
of the information do the gatekeepers tion theory. What kinds of rumors en-
filter out? How much fading and boom- courage the stability of the chain-that
ing do they introduce (by changing the is, what kinds of rumors will tend to be
emphasis of the message)? How much passed on? And what factors govern
systematic distortion are they respon- how faithfully a rumor is passed on?
sible for (through bias)? How much Content analysis codes are subject to
random distortion (through carelessness study for stability and fidelity. How
or ignorance) ? much of the information in the meas-
The newspaper itself-if we want to ured content do they respond to? How
consider it as a system-is a gatekeeper faithfully do they reproduce it? As a
of the greatest importance. The daily matter of fact, many of the concepts of
life of a city presents itself to the pa- information theory are stimulating to
per’s input. Selected bits of the daily content study. For example, the heavy
life of the rest of the world enter the redundancy of Communist propaganda
input through the telegraph desk. What shows up from almost any content
comes out? What is the stability of the study, as does the relatively low en-
paper for reproducing local news as tropy of the semantic systems within
compared with national news, civic which the Communist propagandist
news as compared with crime news, works. The close coupling of units in
news of one presidential candidate as the Communist propaganda chain is
compared with news of another? And striking. And the stability and fidelity
what about fidelity? T o what extent of the Communist gatekeepers, trans-
does the paper change its input by cut- mitting the official line, are very high.
ting, by rewriting, by choosing a head- If they are not, the Party gets a new
line which affects the meaning, by giv- gatekeeper.
ing one story better position than an- Measures of stability and fidelity are
other? available, in information theory, and
Think of the reporter walking his relatively easy to use. When they are
beat. Everything he sees and hears is applied to a long chain-such as the
news for someone, but he must make a one, previously referred to, which car-
selection in term of what his editors ries news from “India to Indiana” and
and-supposedly-his readers want. back-it becomes apparent that the sta-
His stability is necessarily low. But how bility of the main points along the
is his fidelity? Does he get the quotes chain is quite high: that is, a bureau
from a speech right? Does he report an like London is quite likely to pass along
accident accurately? a story that comes from New Delhi.
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
140 JOURNALISM QUARTERLY
The closer one gets to the source of faster than most mechanical systems
news, the lower the stability, because (such as smoke signals and flags), but
the input is large, the output capacity far slower than that of most electronic
relatively small. Bloomington, for ex- devices (e.g., the electronic comput-
ample, regularly publishes about 65 lo- ers). We still have a great deal to find
cal stories, but can only put two or out about man’s capacity for handling
three on the wire. Delhi, likewise, can language and pictorial information.
send London only a small part of the Many of the capacity problems of
Indian news. Chicago, on the other mass communications, of course, find
hand, can send out more than half the man at the mercy of his works. The re-
stories available. The problem in meas- porter who has only 30 minutes to
uring the fidelity of this kind of chain write his story before deadline, the edi-
is to define measurable units. Using tor who is permitted to file only 200
length as one criterion, it becomes ap- words on the wire, the radio news bu-
parent that the greatest loss is near the reau desk which has room for only 13
source of news. Using rewriting as a minutes of copy and must select from
criterion, it seems that the chief re- 300 stories, the editor who finds a big
writing is done at the first wire points advertising day crowding out his news
and the chief national bureaus. -all these are communicators suffer-
CHANNEL CAPACITY ing from capacity problems they have
Channel capacity is another impor- helped to make. It is also obvious that
tant concept which is common both to the channel capacity of the New York
information theory and to mass com- Times is greater than that of a small
munication. All channels, human, elec- daily. But for the Times and its smaller
tronic or mechanical, have an upper brothers there is an even greater chan-
limit on their ability to assume differ- nel restriction: the reader. The reader
ent states or carry different events. We 01 a daily can spend, on the average,
can estimate, for example, the amount about 40 minutes on his paper. And he
of information the eye is capable of reads rather slowly. Even so, he can
transmitting to the optic nerve, and it read faster than he can listen, so to
is less than the information available to speak. A radio speaker usually stays
the eye, although apparently more than under 150 words a minute, not be-
the semantic system can handle. We cause he cannot talk faster, but because
can estimate the capacity of a t e l e he fears he will crowd the channel ca-
phone line or a microphone, and have pacity of his listeners.
very good formulas for doing so. But Shannon has developed a theorem
when we consider the characteristics of for a channel with noise which is both
a chain and recall that the chain is no remarkable in itself and highly mean-
stronger than its weakest link, then our ingful for persons concerned with mass
chief interest turns to the channel ca- communication ( 1 ) . His theorem says,
pacity of man, who is the weakest link in effect, that for rates of transmission
in most communication chains. less than the capacity of a channel it is
Perceptual experiments have told us possible to reduce noise to any desired
a great deal about the ability of man to level by improving the coding of the in-
transmit information through some of formation to be transmitted; but that
his systems. In general, we can say that for rates of transmission greater than
man’s ability to handle information is channel capacity it is never possible to
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
Information Theory and Mass Communication 141
FIGURE B
Ability t o Repeat Information from Newscasts
I n t o r r n o t l o n i n p u t
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
142 JOURNALISM QUARTERLY
would be required, but some striking dundancy, noise, fidelity and capacity
confirmation of it comes out of experi- measures, they suggest trafic (what
ments with retention of newscasts. members do the most talking, and how
Subjects were presented newscasts of much talking is done?), closure (to
increasing density but constant length what extent is the group a closed cor-
-5, 10, 20, 30, 40. 50 items. The aver- poration?), and congruence (to what
age subject’s ability to recall the subject extent do members participate equally
of these items leveled off vary sharply in the communication of the group, or
between 10 and 20. There was practi- to what extent are there members who
cally no additional learning between 20 are chiefly talkers and others who are
and 30. After 20, the number of errors chiefly listeners?). All these formula-
began to increase rather sharply. In tions can be dealt with mathematically.
other words, the amount of information Measures like these suggest a quite dif-
transmitted behaved about as hypothe- ferent and stimulating way of studying
sized above, and the resulting curve small groups, and in particular they
was strikingly like those typically re- commend themselves for use in study-
sulting from experiments on the capac- ing the important groups within mass
ity of subjects to discriminate among communication.
stimuli-as shown in Figure B. Suppose, for example, we want to
study some part of the world news net-
NETWORKS
work. Suppose that we take the chief
Of all the potential contributions of newspapers of the leading cities in half
information theory to mass communi- a dozen countries-for example, the
cation, perhaps the most promising is United States, Great Britain, France,
in the study of communication net- Germany, Italy and the Soviet Union-
works. Networks are as important in and tabulate for one week the stories
mass communication as in electronic which the papers in each city carry
communication. Every functional group from the other cities in the network.
is a communication network. The staff This has been done in a small way,
of a newspaper or a broadcasting sta- with interesting results. Washington has
tion, a film production crew, the group the greatest output traffic, New York
with which a member of the mass com- the greatest input traffic. Moscow has
munication audience talks over what he the greatest degree of individual clo-
reads, hears and sees-all these are sure: that is, it is most likely to talk, if
communication networks. The inter- at all, to itself. Within a country, there
communication within the network is are startling differences in the amount
measurable, whether it consists of con- and distribution of input. In general
versation, print, gestures or electronic there appears to be a little more organi-
currents. zation (redundancy) in the pattern of
Osgood and Wilson, in a mimeo- input than in the pattern of output:
graphed publicati~n,~ have suggested a that is, source entropy is higher than
series of measures derived from infor- destination entropy. And the congru-
mation theory, for dealing with groups. ence (the correlation between source
In addition to the common entropy, re- and destination frequencies of points
in the network) varies markedly with
8 “A,,Vocabulary for Talking about Communi- political conditions and cultural rela-
cation, colloquium paper. Institute of Communi-
cations Research, University of Illinob. tionships at a particular time.
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
Information Theory and Mass Communication 143
Let us take a simpler example of that simulated the work of an actual
group communication. Here is a record newspaper staff, including reporting,
of telephone calls amongst four boys reference, editing, copyreading and set-
(who telephoned incessantly). The calls ting in type. All their intercommunica-
were tabulated at periods two months tions were recorded. Not enough
apart-20 calls while the boys were or- groups have yet been put through the
ganizing a school newspaper, and 20 procedure to reveal all the variables,
calls two months later after the paper but the pattern so far is very clear and
was well launched. interesting. Some of the groups were
started on their assignments entirely
Twenty Telephone Calls by Four Boys unstructured-that is, no roles were as-
A. In process of organizing a school signed. In others a leader was appoint-
newspaper: ed. In still others, every person was
Mike Bud MikeT. John
Mike .... 4 4 2 1 0 assigned a job. Inasmuch as some
Bud ..... 3 1 2 6 measure of leadership almost always
Mike T. . . 1 1 0 2 appeared, regardless of assignment,
John .... 1 1 0 2
5 6 5 4 participants were asked at the end
B. After school newspaper had been pub- whether they perceived a leader or
lished 2 months: leaders, and if so, whom? This, in gen-
Mike Bud MikcT. John eral, seems to be the pattern:
Mike . . . . 3 1 1 5 (a) As the perception of leadership
Bud ..... 7 1 0 8
Mike T. . . 5 1 0 6 increases, the relative transitional en-
John .... 1 0 0 1 tropy of communication in the group
1 3 4 2 1 decreases-that is, it becomes easier to
It is clear that the relative transi- predict who will talk to whom.
tional entropy of this group became less (b) As the degree of initial organi-
in the two months-that is, it became zation is increased, the total amount of
better organized-and also that the communication decreases and the total
congruence had changed so that in- time required to do the job decreases.
creasingly one pattern could be pre- (c) However, between the group in
dicted: i.e., the boys would call Mike. which a leader is appointed and the
It seems that whereas Mike must have group in which all members are as-
been the organizer at fist, he became signed roles, these measures change
the leader later, and the other boys much less than between the other
turned to him for advice or instruc- groups and the unstructured group. In
tions. some cases, the group in which a leader
This kind of result suggests the hy- only was appointed actually finished
pothesis that the entropy of communi- the job more quickly than the group in
cation within a functional group de- which all roles were assigned. This sug-
creases as the group becomes more gests that there may be a stage in which
fully organized into work roles and bet- increasing organization does not con-
ter perceives the existence of leadership. tribute to efficiency; and also, that it
By way of testing this and preparing must make a difference who is appoint-
the way for studying actual media ed leader, even in these previously un-
staffs, some experiments have been acquainted groups.
done with groups of five journalism These results are presented only to
students who were given assignments suggest that the approach is a promis-
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
144 JOURNALISM QUARTERLY
ing one for group study, and especially only analogic value, and that the con-
for the study of the kind of functional tribution of its mathematical tools is
groups that play such an important necessarily small. These tools seem to
part in mass communication. me to be extremely promising in the
study of language, channel capacities,
FINALLY couplings, and network groups, if no-
How can we sum up the import of where else. It will be to our advantage
all this for the study of mass commu- to explore these uses and others.
nication?
Appendix
Even such a brief overview as this
THE BASIC FORMULAS
must make it clear that information
It may be helpful to explain the basic
theory is suggestive and stimulating for entropy formula here in order to give a
students of human communication. It better idea of what information theory has
must be equally clear that the power of to offer mathematically.
the theory and its stimulating analogic Let us begin with an event which we
quality are greatly at variance with the call i within a system which we can call
puny quality of the mathematical ex- I.” (For example i may be the yellow light
on a traffic light I.) Then let us call p(i)
amples I have been able to cite-that the probability of event i occumng within
is, examples of the use so far made of the system. This is equivalent to saying
information theory mathematics in that p(i) equals l/a, in which a is a cer-
studying mass communication. Why tain number of equally probable classes.
(For example, the yellow light in a traffic
should this be? light occurs two times in four events, so
The theory is now-1948, as I have that its probability is 1/2.) The informa-
said, for most of us. Its application is tion we need to predict the Occurrence of
fringed with dangers. One of these has event i is therefore loaa. By algebraic
transformation, we can say that, since p(i)
been indicated-the danger of working equals I/a, a equals l/p(i). Therefore the
with stochastic processes in functional information necessary to specify the one
systems which may learn and thereby event i is lo&(l/p(i)). Since the logarithm
change the probabilities. It should also of x/y always equals log x - log y, we
have the information necessary to specify
be said that we do not as yet know event i equal to log21- log?p(i). The log
much about the sampling distributions of 1 is always zero, and therefore we ar-
of these entropy formulas, and it is rive at an equation which states the
therefore not always wise to use them amount of information necessary to spe-
for hypothesis testing and statistical in- cify one event in a system (let us call this
information h(i)),
ference. Finally, we must admit frank-
ly the difficulty of bridging the gap be- h (i) = -lo&p(i)
tween the formula’s concept of infor- Now what we need is an estimate of
the average amount of entropy associated
mation (which is concerned only with with all the states of a system. The aver-
the number of binary choices necessary age of a sample of numbers can be ex-
to specify an event in a system) and pressed as
our concept of information in human 2 i f(i)
communication (which is concerned i n
with the relation of a fact to outside where i is the numerical value of any
events+.g., how “informative” is it?). QThisexplanation of the formula for observed
entropy in general follows the approach of Wil-
This is not to say that the transfer son in bibliography item (10). Wilson’s treatment
cannot be made. Certainly I have no in- of the subject is easy to read and still both solid
and stimulating, and is recommended to begin-
tention of saying that the theory has ners in this field.
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
Information T h e o r y and Mass Communication 145
class of numbers, f ( i ) is the frequency of currence of events n and j together. We
occurrence of that class, n is the sample also get a formula for conditional entropy,
size, and p is the term for sum of all the which deals with the occurrence of two
i events in sequence (for example, the oc-
currence of u after q in a sample of Eng-
i's. But f(i)/n is the same as an estimate lish words). This is written,
of probability, which we called p (i), and
which we can here substitute in the term HtJ= 2 p(i.j)log?pl(j)
for an average as follows: 1J
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015
146 JOURNALISM QUARTERLY
“The duty of the press at this moment is to show the way to recovery
from the blight of fear and cautious conformity.
“There has been far too much defeatism. . . . I believe that the press in
America today has the greatest opportunity it has ever had to help point
the way back to sanity.
“What is more, I believe it has begun to live up to this challenging op-
portunity, and on this 50th anniversary we can look forward to another
half century in which this School of Journalism and the University will play
their part in educating other generations for the grave responsibility o f
freedom. . . .
“The press of America cut its teeth on controversy and controversy has
been its lifeblood.
“What is most extraordinary is to find in press and radio those who seem
so frightened and insecure that they would drive out every opinion that
does not conform to their own narrowly reactionary standard. They spend
a great deal of time vilifying those who do not agree with them as though
this were in itself a crime.
‘‘I would not want to deprive the radicals of either the right or the left
of their privileges of saying within the bounds o f constitutional freedom
what they please. And I might say in passing that we sometimes forget
what we should have learned from the example of Italy and Germany: that
there are radicals of the extreme right just as ready to destroy existing
institutions as are the extremists of the left. The Jacobin can wear black as
well as red.’’--hf.mQuis CHILDS, at University of Wisconsin.
Downloaded from jmq.sagepub.com at UCSF LIBRARY & CKM on April 11, 2015