Search This Blog

Monday, 29 June 2020

On the origin of information.

The Information Enigma: A Closer Look

 Brian Miller

 

 

The first video in our Intelligent Design YouTube Festival, “The Information Enigma,” consolidates the research and writing of Stephen Meyer and Douglas Axe into a 21-minute video. The presentation demonstrates how the information present in life points unambiguously to intelligent design. This topic is central to intelligent design arguments and the ID research program. Here I will flesh out in greater detail the concept of biological information, and I will explain why significant quantities of it cannot be generated through natural processes. 

A Primer on Information

A pioneer in the field of information theory was Claude Shannon who connected the concept of information to the reduction in uncertainty and to probability. As an example, knowing the five-digit ZIP Code for an address eliminates uncertainty about a building’s location. And, the four-digit extension to the ZIP Code provides additional information that reduces the uncertainty even further. Conversely, randomly generating the correct five-digit ZIP Code corresponds to a probability of 1 in 100,000, while generating the correct nine-digit ZIP Code corresponds to a probability of 1 in a billion. The latter is much less probable, so the nine-digit code contains more information. 

Shannon quantified the amount of information in a pattern in what he defined as the Shannon measure of information. In the simplest case, the quantity is proportional to the log of 1/p, where p is the probability of a pattern occurring by chance. For the five-digit code, p would be 1/100,000, and 1/p would be 100,000. This measure can be thought of as the minimal number of answers to Yes-No questions that would be required to identify 1 out of N choices. To illustrate, imagine attempting to identify a pre-chosen famous actor out of eight possible people. If the answer to each question about the mystery individual eliminated half of the options, the correct answer could be determined with three questions. Therefore, learning the answer corresponds to acquiring 3 bits of information. Note that 2 to the power of 3 is 8, or conversely, log (base 2) of 8 is 3. 

Information and Biology

Information theory has been applied to biology by such figures as Hubert Yockey. In this context, Shannon’s definition had to be modified to distinguish between arbitrary patterns and those that performed some function. Shannon’s measure was modified to quantify “functional information.” The measure of functional information corresponds to the probability of a random pattern achieving some target goal. For instance, if 1 in 1024 amino acid sequences formed a structure that accelerated a specific reaction, the functional information associated with that sequence would equate to 10 bits since 10 Yes-No questions would have to be asked to select 1 entity out of 1024 possibilities. 

Mathematically, 2 to the power of 10 is 1024, or log (base 2) of 1024 is 10. More advanced measures for functional information have been developed including algorithmic specified complexity and the more generalized canonical specified complexity. They follow the same basic logic. These measures help relate the information content of biological molecules and structures to their functional capacities. 

Information and Proteins

The information content of a protein’s amino acid sequence directly relates to its ability to control chemical reactions or other processes. In general, the higher the information content, the higher the level of fine-grained control over outcomes and the greater the capacity for elaborate molecular manipulations. For amino acid sequences with higher information content are more specified, so they can fold into three-dimensional shapes of greater precision and complexity. In turn, the higher specificity requirement corresponds to proteins being more susceptible to mutations — a few amino acid changes will often completely disable them. Functional sequences are consequently less probable, which is another signature of greater information content. This connection between information, sequence rarity, and complexity of function has profound implications for Doug Axe’s protein research. 
Axe demonstrated that the probability for a random amino acid sequence to fold into one section (domain) of a functional β-lactamase protein is far too small for it to ever occur by chance. Therefore, the information content is too great to originate through a random search. Yet β-lactamase performs the relatively simple task of breaking apart an antibiotic molecule. In contrast, many of the proteins required for the origin of life perform much more complex operations (see here, here, and here). 

 The same holds true for many proteins required in the construction of new groups of organisms (e.g., animal phyla). Therefore, these proteins’ information content must be even greater. So the probability of their originating through a random search is even smaller. 

Information and Design

This conclusion is deeply problematic for evolutionary theory since no natural process can generate quantities of information substantially larger than what could result from a random search. 
The limitation results from No Free Lunch theorems as demonstrated by the research of Robert J. Marks, Winston Ewert, and William Dembski (see here and here). It is further supported by theorems derived from research in computer science. For instance, computer scientist Leonid Levin demonstrated the “conservation of independence” in information-bearing systems. He stated the following
The information I(x:y) has a remarkable invariance; it cannot be increased by random or deterministic (recursive) processing of x or y. This is natural, since if x contains no information about y then there is little hope to find out something about y by processing x. (Torturing an uninformed witness cannot give information about the crime!)
The conservation law simply means that the information, I(x:y), present in one system, x, coinciding with another, y, cannot increase through any natural process. A nearly identical conclusion comes from information theory in what is known as the data processing inequality. It states that the information content of a signal cannot be increased by any local physical operation. 

In terms of evolution, the first system (the signal) could be a duplicated gene or a nonfunctional section of DNA freely mutating, and the second could be any functional protein sequence into which the gene/section could potentially evolve. The theorems mandate that a DNA sequence (x) could never appreciably increase in functional information, such as more greatly resembling a new enzyme (y). This constraint makes the evolution of most novel proteins entirely implausible. 

In the video, Stephen Meyer 6explains how information points to intelligent design by the same logic used in the historical sciences. In addition, the information processing machinery of life demonstrates unmistakable evidence of foresight, coordination, and goal direction. And, these signatures unambiguously point to intelligent agency. The same arguments hold true to an even greater degree for the origin of life. 
The only pressing question is to what extent critics can continue to allow their philosophical bias to override biological information’s clear design implications.

No comments:

Post a Comment