RE: Information: Brad's reply (was Information: a very

Brad Jones (bjones@tartarus.uwa.edu.au)
Wed, 1 Jul 1998 18:34:02 +0800

>
> At 01:19 PM 6/30/98 +0800, Brad Jones wrote:
> >
> >> >*Biology question*
> >> >1. If the mutation does not occur in DNA replication then all
> it would do
> >> >is create a different protein every now and then. How can this lead to
> >> >cumulative mutations that in turn lead to macro evolution?
> >
> >Greg answered this quite well I thought.
>
> Greg expressed his opinion that macroevolution can't occur. He
> didn't cite
> andy data, I cited the last chapter of Gilbert's Developmental Biology.
> Why wouldn't you be curious enough to at least go look at that book rather
> than accept an opinion that agrees with your own as authoritative?
>

I appreciated Greg's opinion because it was clear, concise and seemed to be
reasonable. I do not have time to research every point. Anyway it does not
have specific bearing on whether random mutations add information so I'll
leave it there.

> >
> >But I am a bit confused as to what you think create sthe
> information, is it
> >generational or sequential. Firstly we were arguing
> generational, then you
> >changed to sequential which doesn't seem to have much effect on the whole
> >organism in terms of major change.
>
> The information of a DNA molecule is sequential. One must measure
> H via the
> nucleotides in the sequential axis NOT the generational axis. The
> generational axis was the only way I could explain your statement in your
> first post when you said on Tue, 23 Jun 1998 22:17:39 +0800:
>
> >A DNA sequence of AAAAATAAAA will output this each and every
> >time eg: AAAAATAAAA AAAAATAAAA AAAAATAAAA
>
> If I take the above statement as being sequential then it is an assertion
> that DNA is repeated segments of
> AAAAATAAAAAAAAATAAAAAAAAATAAAAAAAAATAAAA..., which is clearly nonsense and
> observationally wrong. But if placed in a generational axis, the above
> statement makes sense. I was merely trying to give you the benefit of the
> doubt.
>

You are correct, that would be nonsense. But modeling DNA as a source is
nonsense anyway so whats the difference? No finite sequence of symbols is a
source.

by sequential I was considering the proteins that are created by the DNA, ie
DNA specifies proteins and does so repeatedly. If this does not happen or I
am totally on the wrong track then I apologise, I am no biology expert as I
have said.

I am here to debate information theory and your comments regarding the
creation of information. I disagree that "information can be created by
random mutations" independent of the biological mechanisms.

> >
> >ie. one cell with a faulty protein isn't going to be passed to the
> >offspring. In fact wouldn't it have to be DNA replicating in a sex cell
> >which can create the necessary information that you require?
>
> Absolutely. The germ cells are what is passed on and the DNA contained in
> them. Thus when a body has a really great advantageous mutation, the body
> is more likely to pass on more germ cells to many more offspring thus
> multiplying the advantageous germ cell.
>
> >
> >I would like a clarification of the position you are arguing, it seems to
> >vary a bit as the discussion continues.
>
> I try to stay put but I am responding to various issues and you may
> percieve them as different, they aren't. I also think that part of the
> problem is that you haven't read Yockey and can't quit understand the
> connotation of the words used as Yockey uses them and I follow Yockey's
> work, which I find quite credible and consistent.

umm, this doesn't clarify your position on much....

If Yockey has relevant points then I assume you will bring them out. I am
looking for Yockey's book but it is not readily available.

>
> >> Once again you are equivocating on the terms information as used
> >> for knowledge or intelligence rather than information as a
> >> mathematically defined concept. See Greg Billock's post
> >> yesterday or see Yockey's "Application of Information Theory to
> >> the Central Dogma and the
> >> sequence hypothesis" Journal of Theoretical biology 46:369-406
> >>
> >
> >I disagree here. Information is related to and requires meaning.
> You cannot
> >have information without meaning. You CAN have information
> carrying capacity
> >without meaning, but no information.
>
> Then you will be disagreeing with what is becoming the norm in information
> science. And on this we will continue to disagree. Yockey notes:
>
> "Accordingly, we will keep in mind that the word *information* is the name
> of a mathematical function. We must not allow ourselves to ascribe any
> other meaning associated with the word *information* as used in ordinary
> language. We are, after all, constructing a mathematical
> formalism to avoid
> the lack of exactness characteristics of ordinary discourse. We will see
> later that a plausible argument in ordinary language may lead to
> conclusions which are more than just inexact; they are false." ~H. P.
> Yockey, "An Application of Information Theory to the Central
> Dogma and the
> Sequence Hypothesis," Journal of Theoretical Biology
> 46(1974):369-406, p. 375
>
> and
>
> "One must further remember that entropy is a name of a mathematical
> function, nothing more. One must not ascribe meaning to the function that
> is not in the mathematics. For example, the word 'information' in this
> book is never used to connote knowledge or any other dictionary meaning of
> the word not specifically stated here. It will be helpful to
> remember that
> a communication system, including the genetic information system, is
> essentially a data recording and processing system. Thus one avoids much
> of the anthropomorpyhism that words convey."~Hubert Yockey, Information
> Theory and Molecular Biology, (Cambridge: Cambridge University Press,
> 1992), p. 68.
>

I am not sure what this proves.

I agree that information theory does not have to know the meaning. It sure
helps knowing that there is meaning though.

>
>
> [of DNA brad wrote]
> >I thought the error rate was pretty small, around 10E-9 if I remember
> >correctly, that is easily small enough to compare to a CD.
>
> the error rate only applies to the generational axis. There is NO error
> rate in the sequential axis of a non-reproducing DNA sequence.

if there is no error rate then there is nothing to discuss on the sequential
axis as far as information creation is concerned.

We are discussing the information created by random mutations, if there are
no mutations then there can be no information created. Read quote below:

***********************************************************************
>But in our original notes on information theory, both Brian and I were
>talking about the Sequence axis. Information is measured along the
>sequence axis, not per se the generational axis.

>Thus when I pointed out that the sequence AAAAAAAAAA had zero information
>content, and the mutation to AAAAATAAAA represented an increase in
>information it does because we are not talking about the generation axis.
>But even putting it into your terminology, the output(generational axis) of
>the DNA sequence AAAAAAAAAA is not always AAAAAAAAAA but occasionally is
>AACAAAAAAA or AAAAATAAAA. There is a Generational markov matrix that is
>something like:
--Glenn, 28/06/98

**************************************************************************

So you are directly contradicting yourself here. This is why I have
previously (and still do) ask for you to clarify your position.

>
> >
> >Either DNA stores information or it does not, which one is it? If it does
> >then it can be compared to other information storage devices. I
> do not know
> >what you mean about treating it as source and reciever, I was
> treating it as
> >neither, rather the channel.
>
> We may be having a semantic dispute here in which the different
> applications of information theory use the word channel
> differently. Yockey
> separated source from channel and reciever. If your texts don't make that
> distinction then that would explain the difference. Let me quote Yockey.
>
> "A discrete memoryless source is defined as one in which there are no
> restrictions or intersymbol influence between letters of the alphabet. If
> the response of the channel to the input letters is independent
> of previous
> letters, the channel is called a discrete memoryless channel. A channel
> that allows input symbols to be transmitted in any sequence is called an
> unconstrained channel. A source that transmits messages written
> in natural
> languages is not a memoryless source, since natural languages do have
> intersymbol influence. Monod in his philosophy of chance and necessity,
> says that protein sequences are due to chance and justifies this on the
> ground that one cannot predict missing amino acids from the properties of
> their neighbors. That reflects only the fact that there is no intersymbol
> influence in proteins as there is in natural languages....
> "Thus the DNA-mRNA-protein system is discrete, memoryless and
> unconstrained. It transmits the same message to the destination
> repeatedly, as in a tape recorder.

****

Here your own quote compares it to a tape recorder yet you continually
complain when I compare it to a CD or book etc....

This does not make sense Glenn.

Therefore your own quote refutes your argument as to why DNA cannot be
modeled as a CD. So now that the comparison is valid according to Yockey I
can use it.

***

> The particular message recorded in the
> DNA is independent of the genetic information apparatus.
> "In communication theory the messages that have meaning, or in
> molecular biology what is called specificity, are imbedded in the ensemble
> of random sequences that have the same statistical structure, that is, the
> same entropy. We know the statistical structure of the ensemble but not
> that of the individual sequences. For that reason, the output of any
> information source, and, in particular, DNA in molecular biology, is
> regarded as a random process that is completely characterized by the
> probability spaces [Omega, A, p],[Omega, B, p]." H. P. Yockey, Information
> Theory and Molecular Biology, Cambridge, 1992, p. 114
>
> As you can see, Yockey separates the channel from the DNA.

The DNA cannot be a source, it MUST be a channel. Any type of storage device
does not create information and so is not a source. YOU say that random
mutations create information so THAT must be the source, not the DNA.

>
> >
> >ANY information storage device IS a channel. Do not make the mistake of
> >thinking anything that has an output is a source. This is not correct.
> >
>
> I think there is some difference in terminology as noted above.

I think not

>
> >>
> >
> >I would like to read information theory text to correct your
> application of
> >information theory. It was you who started using info theory
> (incorrectly)
> >and that is what I am debating.
>
> That is still a matter of dispute. Two people have told you that meaning
> is not part of information theory and even one of your own quotes
> equivocates on that difference.
>
>
> >> Because you use the CD analogy in a equivocation between
> >> information used as 'knowledge' or 'intelligence' rather than
> >> information as defined by H =-k sum(-p[I]log[p[I]). My son told
> > ^^^^^^^^^^^^^^^^^^^^^^^^^
> >This is the most inaccurate method of anaysing anything but the
> most trivial
> >examples of sources. DNA is certainly not acurately modeled by this. I
> >showed this in great length in earlier posts and you have never
> refuted my
> >arguments. (or even acknowedged them).
>
> I did acknowledge them and replied to them. You should go look again
> before you make such charges.

You never refuted my mathematics with any reasonable objections other than
"you are wrong"

>
>
> >White noise does not have any information. White noise has the
> capacity to
> >carry a great deal of information but it isn't carrying it because nobody
> >put any into it. do you get this yet?
>
> here you are once again equivocating on information as used in ordinary
> language rather than as defined by mathematics. See the Yockey quote
> above. I fully agree that white noise doesn't have information in the
> ordinary, colloquial usage of the term information. But we are dealing
> with a mathematical NOT DICTIONARY definition of information.

I use information as meaning REAL information as used in REAL systems. Not
pretend information used to test systems or to MODEL real systems.

DNA is a REAL system and I treat it as such.

>
> >
> >Now mutations in electronic devices are EXACTLY what I study and you are
> >totally wrong on this one. If you think that random changes add
> information
> >to a CD then you are sadly mistaken and should go ask anyone involved in
> >electronics.
>
> As I keep repeating, mathematically information is not the
> english language
> or intelligence or knowledge. And I agree that white noise will
> not add the
> dictionary definiton of information to anything. But that is not what we
> are talking about. If this we become stymied at this point there is not
> much use in going on. Information is defined as I noted above.

I agree, Information does not have to be english language or anything else
like that. In fact information can be almost anything that is created to
convey some kind of meaning. Random noise is not information.

>
> "For this reason we will call H the message enropy of the ensemble of
> messages. It is also called the information content since the number of
> messages of length N is 2^NH. Each message can be assigned a meaning or
> specificity and thus carries *information*[italicized to differentiate it
> from the previous use of information--grm], knwoledge or intelligence. H
> reflects the character of the probability distribution P[i]. A broad
> distribution provides a maximum choice at the source or uncertainty at the
> receiver. H is a maximum if all p[i]=1/n." ~H. P. Yockey, "An Application
> of Information Theory to the Central Dogma and the Sequence Hypothesis,"
> Journal of Theoretical Biology 46(1974):369-406, p. 373
>
> I repeat the last sentence.
>
> " H is a maximum if all p[i]=1/n" This means that the probabilities of the
> characters are RANDOM, RANDOM RANDOM. H is maximum if the
> sequence is RANDOM.

NO NO NO.

equiprobable symbols does NOT mean random data. it means what it says.
source coding specifically tries to get equiprobable symbols as a means of
maximising data throughput. This does not mean that random data is sent.

You are confusing the MODEL with the REAL SOURCE. the model is similar to
the real source but is NOT IT.

I can easily find the probabilties of the letters appearing in my post, does
that make my post random? NO (well you may think it is :P )

>
> >
> >Random errors in a CD will ALWAYS reduce the information, and the meaning
> >and everything else.
>
> I agree that they will reduce info in the dictionary sense, but not in the
> mathematical sense.
>
> >> Once again you are making a fundamental error in information
> >> theory. See Greg Billock's post if you won't believe me.
> >
> >Hmm, amazing how I can make such fundamental errors in info
> theory yet still
> >achieve high marks while studying it at university level....
>
> Interesting that you won't listen to Yockey, or Greg or me on the
> idea that
> information is not what you think it is in information theory.

Not that interesting, I still think you are wrong.

>
>
> >It is entirely possible to know that you have the best
> compression possible,
> >that is what info theory is all about. This has NOTHING to do with
> >algoirthmic complexity at all.
> >
> >By finding the true information content of a signal you can devise
> >compression to compress it to as close to 100% efficiency as you like, it
> >just takes more processing power. This is known as "Shannon's noiseless
> >coding theorum" and involves taking extensions of a source to
> find the true
> >information content and then coding using this.
> >
> >This was what I did to refute your first post about the addition of
> >information. You have never shown any error I made in doing this so the
> >result stands.
> >
>
> I have answered it. You haven't understood the answer. You wrote:
>
> >>The general rule is that any information source that repeats a
> >>sequence will have zero information. In plain english this is:
> >>
> >>"If we already know what is being sent then there is no information
> >>being sent"
> >>
> >>btw this is the basic principal that all compression software relies on.
>
> I answered you by showing that DNA is not a set of repeats. That
> invalidates your analysis.

If DNA creates proteins then it is repeating.

If DNA replicates then it is repeating.

either way my analysis is valid. But if you want state your new position
fully and I'll analyse that...

>
> >> That is why you can't tell whether I am writing real mandarin
> >> chinese (pinyin) below or real gibberish.
> >>
> >> Ni xue yao xuexi hen duo!
> >
> >No I cannot tell, but with a bit of investigation I could.
>
> Do it without consulting a chinese student or a chinese textbook.
> Go ahead?

Why on earth would I do it without consulting a textbook?!?

Read my other post where I talk about the difference between investigating
the source as opposed to the message. I have no doubt that I cannot tell if
it is random from the message, BUT I have never stated I could.

I have stated that I could investigate the source to find out. see the
difference?

> And let me give you two others to tell me which is and isn't information.
>
> Zhe ge mao you mao.
>
> xi gong zuo chi xiao xue.
>
>
> When you have the answers I will tell you which is and isn't Mandarin,
> which I do speak 'ee diar diar".

I'm not going to bother as it has no bearing on the debate.

>
>
> >Once I know which
> >one it is I would be able to find the true information content.
> If you were
> >speaking mandarin I would be able to find the information content of that
> >language, if it was gibberish I would just igore you. You see If
> you speak
> >gibberish and I ignore you I gain the same information as if I listen to
> >you, this is therefore the ultimate compression.
>
> There is real mandarin in the above. Tell me which!
>
> >
> >You are not helping your case with examples like this.
>
> Actually I am. I am showing you that you cannot use information theory to
> determine meaning. Which of the following is a terrible curse in Mandarin?
> If information theory is about meaning tell me the bad word!

I NEVER stated that I could determine meaning with information theory. That
is clearly absurd and has no bearing on this debate.

>
> lao wu gui
>
> hen xiao chun
>
> jiu dian zhong
>
> >
> >Hehe, well it is also a real exam question on information theory. Tell us
> >Glenn, is noise the most informative source?
> >
> >just post a yes/no answer, its not hard.
>
> By the dictionary definition it is not. By the mathematical definition it
> is!!! See Yockey above. H is maximized when the probability for each
> character is the same.

Read my other post on the exam question.

> I wrote:
> >> Now, let me get this straight. You didn't know who Yockey was;
> >> originally you said you hadn't read any of his articles; you
> >> didn't understand that a Markov transition matrix could be
> >> constructed which didn't rely on the previous character (you
> >> admitted that I was correct );
> >
> >You were correct, but I knew it was possible. I just have never heard of
> >anyone using it in that way.
>
> So, knowing that is was possible is why you wrote on Sat, 27 Jun 1998
> 17:06:50 +0800:
>
> >It is therefore impossible by definition to have a markov source where
> >the next symbol is not dependent on the previous.
>
> That is an awfully funny way to say you knew it was possible. There must
> be some mutations in Australian English that I am unaware of.

Ok, I am guilty of some bad grammar?!?

your point?

>
> I wrote:
> >> you didn't know ...
> >> using mathematics for a non-degenerate code;
> >
> >I know that the code is not important in the maths I used.
>
> I cite Yockey again. "The third term in equation (7) is one of the aspects
> of information theory in biology which differes from infomration theory in
> elelctrical engineering. This is because there is no degeneracy in the
> codes used in communications." ~H. P. Yockey, "An Application of
> Information Theory to the Central Dogma and the Sequence Hypothesis,"
> Journal of Theoretical Biology 46(1974):369-406, p. 37
>
> Can you cite the 3rd term in equation 7 from electrical engineering texts?

umm, what are you talking about here? what is the "3rd term in equation 7"?

I have not been talking about codes and both channels and sources can be
analysed without any knowledge of the codes that may be used.

>
>
> I would suggest that if we are not going to make any progress on the issue
> of meaning vs. information then we are at a standstill. I have
> cited Yockey
> you have cited your opinion that information means 'meaning'. If all we
> are going to do is have you say (after I have cited a textbook)
> "information theory deals in meaning" and I repeat the citeation, then we
> are at an impasse. I have enjoyed the debate, but I see no way to make
> progress beyond this point if you won't believe a textbook.

I believe you misunderstand the textbook and are confusing the use of
mathematical models with the actual information source.

>
>
> But if you can tell me which of the chinese statements are meaningful, I
> will tell you their meaning. I don't expect to hear from you on this.
> glenn

I'm not going to bother because it is not relevant. I never stated that I
could tell information from gibberish from the message alone. I stated that
I could investigate the source to tell is it is information. Big differences
here.

--------------------------------------------
Brad Jones
3rd Year BE(IT)
Electrical & Electronic Engineering
University of Western Australia