Search This Blog

Sunday 30 June 2024

Yet more preDarwinian design vs. Darwinism.

 “Irreducible Complexity” May Be Part of the Definition of Life


There are many bad counter-arguments to Michael Behe’s famous irreducible complexity conundrum, and (in my opinion) one pretty good one. 

For those unfamiliar with Behe’s argument, it goes like this: 
                        Darwinian evolution is supposed to build complex systems gradually, overcoming vast improbabilities in tiny steps over billions of years. But, strangely, many systems in living organisms are “irreducibly complex” — they contain a core set of key elements that are all absolutely necessary for the system to function at all. Gradual evolution through random variation and natural selection could never build such a system, because the system would have no adaptive function until it was already completely finished.
                 After Behe made this case in his 1996 book Darwin’s Black Box, scientists (and non-scientists) scrambled to rebut him. Some argued that the systems in question weren’t really irreducibly complex; others, that they could have arisen through cooption of parts from other systems; others, that they emerged as reductions from larger complex systems that were not irreducibly complex… and so on. 

None of those arguments have held up to logical or empirical scrutiny. But I don’t think those argument are the real reason that most who find Behe’s argument unpersuasive find it unpersuasive. I suspect that the real objection for most people is something more gut-level and foundational, which might be expressed something like this: 
                               Okay, so maybe it’s hard to see how gradual, blind processes could produce a few special systems like the bacterial flagella. Because of irreducible complexity — got it. But Darwin’s theory still makes sense for everything else. So are we really going to throw out the whole theory on the basis of a few things we can’t explain? Isn’t it more likely that there’s some explanation for these things, and we just have to wait for it?1
                                   After all, if Darwinian evolution works in theory, then it seems to follow that Darwinian evolution should have happened. And then, if living organism don’t look like they were made by Darwinian evolution, the question just becomes, “So where the heck are the things that were made by Darwinian evolution?” Even if the presence of irreducible complexity shows that all the organisms we study didn’t arise by Darwinian evolution, it doesn’t explain why they didn’t arise by Darwinian evolution.
       
Confusion and Mystery

In other words, for the irreducible complexity argument to persuade someone away from Darwinism, it’s not enough to show that some structures in living organisms don’t look like they were made from unguided Darwinian processes. As long as unguided Darwinian processes work in theory, the existence of irreducibly complexity in life may add confusion and mystery, but it doesn’t do away with the theory. For the argument to be really convincing, you need to also show that Darwinism doesn’t actually work to construct living organisms even in theory. 

Might this be the case? Well, it would be the case if irreducible complexity is actually necessary for living systems. If something needs to be irreducibly complex in order to achieve the characteristics that would make us call it “alive,” then Darwin’s theory doesn’t even work in theory, and the mystery is solved — we see features that Darwinian evolution can’t explain simply because Darwinian evolution didn’t actually happen, and can’t happen. 

Behe argued something like this in response to the criticisms of his first book. But it was at first an open question — there is no quick-and-easy way to tell if irreducible complexity is intrinsic and necessary to life, or not. 

It’s extremely interesting, then, that the prominent theoretical biologist Stuart Kauffman has been promoting a definition of life that entails irreducible complexity — though Kauffman (who is unsympathetic towards ID) doesn’t use that term.

A Definition of Life 

Kauffman has been arguing that what sets living organisms apart from non-living things, and what makes them able to function and to evolve, is that in living organisms the parts exist for and by means of the whole. Kauffman calls such systems “Kantian wholes” (because the idea comes from Immanuel Kant’s Critique of Judgement). A Kantian whole, to put it another way, is a self-creating system in which everything supports and depends upon everything else. 

It’s easy to see how living organisms fit this definition. Your various parts can’t exist without you — you’ll never find a brain or a spleen lying around on its own (at least, not for very long). Likewise, you wouldn’t exist if you didn’t have those parts (at least, not for very long). 

It’s also easy to see that such a system is by definition irreducible complex. The “whole” — by definition — encompasses all of the parts. So, if the whole is necessary for the continued existence of the parts, then all of the parts are necessary for the continued existence of the parts — which is the definition of irreducible complexity. Not all irreducibly complex systems are necessarily Kantian wholes, but Kantian wholes are necessarily irreducibly complex. 

Irreducible Complexity in LUCA

Of  course, someone will probably point out that this is all very interesting philosophizing, but science is about empirical evidence. And Kauffman, as a scientist, is eager to provide it. To this end, he co-authored (with the up-and-coming origin of life researcher Joana Xavier and others) a paper published in Proceedings of the Royal Society B which seemed to show that life has existed in the form of Kantian wholes as far back in evolutionary history as we can see. 

Xavier et al. took a database of metabolic reactions in bacteria and archaea (the two domains of the simplest lifeforms) and looked at which reactions they had in common. They found in the intersection of bacteria and archaea a collectively autocatalytic set of 172 reactions. (“Collectively autocatalytic” means that the set of reactions is self-creating — all the catalysts of the reactions in the set are created by other reactions in the same set; e.g. A creates B, B creates C, C creates A.) From a phylogenetic perspective, this implies that the common ancestor of bacteria and archaea — and thus presumably the “last universal common ancestor” (LUCA) itself — was characterized by complex autocatalytic metabolic cycles. In a paper in the volume Evolution “On Purpose”: Teleonomy in Living Systems, Kauffman and his colleague Andrea Roli write that these findings “very strongly suggest that life arose as small-molecule collectively autocatalytic sets.” 

Kauffman and his co-theorists believe that collectively auto-catalytic sets are Kantian wholes. Therefore, they argue that life has been characterized by Kantian whole-ness from the very beginning, in accordance with Kauffman’s contention that living things are Kantian wholes by their very nature. If that’s true, then — as we have seen — that means that life, by its very nature, is irreducibly complex. 

What Was the Question?  

If irreducible complexity really is part of the definition of life, this solves the problem raised in the response to Behe’s irreducible complexity argument. 

It all comes down to What is it that we’re trying to explain? when we invoke evolution or design. Why does life need an explanation at all? What is it that makes people, cows, mushroom, pine trees, bacteria, and so forth, so very perplexing to us?

Darwin seemed to think the problem was mere complexity, or the adapted-ness of organisms to their environment. That seems plausible at first glance, but in retrospect we should have known that it isn’t the case. A pile of sand is complex — the odds of obtaining that exact same arrangement of grains of sand a second time are almost nil — but nobody thinks that the existence of piles of sand is some big mystery. 

No, the thing that makes living organisms so mysterious (one of the things that makes them mysterious, anyway) is that they are irreducibly complex: they move, act, reproduce, and grow by means of an elaborate system of interconnected, interworking parts. It’s obvious (with 20-20 hindsight) that this is the real mystery in need of explanation, and it’s equally obvious that the ability of natural selection to pile up tiny, individually useful random variations in no way explains (or even attempts to explain) how such an intricate network could come to be.

So when Behe pointed to irreducible complexity, he wasn’t noticing some random, inexplicable feature of certain biological systems and using it to attack Darwin’s theory. Rather, he was putting his finger on what exactly it is about life that makes us feel it needs explaining. And that turned out to be something about which Darwin’s insights, brilliant though they were, had nothing to say.

Notes

For example, this line of thinking has got to be why evolutionary biologist Bret Weinstein feels that “if we pursue that question [a particular problem raised by ID proponents], what we’re going to find is, oh, there’s a layer of Darwinism we didn’t get and it’s going to turn out that the intelligent design folks are going to be wrong” — even though he admits that ID proponents are pointing to genuine holes in the current theory of evolution. 

Saturday 29 June 2024

Yet another of the fossil records many big bangs.

 Fossil Friday: Snake Origins —Yet Another Biological Big Bang


This Fossil Friday features the “legged” snake Najash rionegrina from the Late Cretaceous of Patagonia, which is one of the oldest fossil snakes known to science. It was found in terrestrial sediments and shows a well-defined sacrum with pelvis connected to the spine and functional hind legs. Therefore it was considered as supporting an origin of snakes from burrowing rather than aquatic ancestors (Groshong 2006). I had reported about the highly controversial and hotly debated topic of snake origins in a previous article (Bechly 2023), where you can find links to all the relevant scientific literature.

Another Open Question

But  there was another open question concerning the origin of snakes: Did their distinct body plan evolve gradually as predicted by Darwinian evolution, or did snakes appear abruptly on the scene as predicted by intelligent design theory? Earlier this year a seminal new study was published by a team of researchers from the University of Michigan and Stony Brook University in the prestigious journal Science (Title et al. 2024). This study brought important new insights with the mathematical and statistical modelling of the most comprehensive evolutionary tree of snakes and lizards, based on a comparative analysis of the traits of 60,000 museum specimens and the partial sequencing of the genomes of 1,000 species (SBU 2024, Osborne 2024). The study found that all the characteristic traits of the snake body plan, such as the flexible skull with articulated jaws, the loss of limbs, and the elongated body with hundreds of vertebrae, all appeared in a short window of time about 100-110 million years ago (Rapp Learn 2024).

The authors commented in the press releases that this burst of biological novelty suggests that “snakes are like the Big Bang ‘singularity’ in cosmology” (SBU 2024; also see Cosmos 2024,Osborne 2024, Sivasubbu & Scaria 2024, Wilcox 2024). This arguably would imply that snakes became “evolutionary winners” because they evolved “in breakneck pace” (Wilcox 2024), which the senior author of the study explained with the ad hoc hypothesis that “snakes have an evolutionary clock that ticks a lot faster than many other groups of animals, allowing them to diversify and evolve at super quick speeds” (Osborne 2024). Well, that is not an explanation at all, but just a rephrasing of the problem. How could such a super quick evolution be accommodated within the framework of Darwinian population genetics and thus overcome the waiting time problem? After all, the complex re-engineering of a body plan requires coordinated mutations that need time to occur and spread in an ancestral population. Did anybody bother to do the actual math to check if such a proposed supercharged evolution is even feasible, given the available window of time and reasonable parameters for mutation rates, effective population sizes, and generation turnover rates? Of course not. We just have the usual sweeping generalizations and fancy just-so stories.

The Fatal Waiting Time Problem

My prediction is that this will prove to be another good example of the fatal waiting time problem for neo-Darwinism. In any case we can add the origin of snakes to the large number of abrupt appearances in the history of life (Bechly 2024), and I am happy to embrace the name coined by the authors of the new study for this remarkable event: The macroevolutionary singularity of snakes. This does not sound very Darwinian, does it? So what do the authors suggest as causal explanation? They have none and the press release from Stony Brook University (SBU 2024) therefore concludes with this remarkable admission: “The authors note that the ultimate causes, or triggers, of adaptive radiations is a major mystery in biology. In the case of snakes, it’s likely there were multiple contributing factors, and it may never be possible to fully define each factor and their role in this unique evolutionary process.” It other words, it was a biological Big Bang and they have no clue what caused it. But of course it must have been unguided evolution, no intelligence allowed!

References

Friday 28 June 2024

AI is on the verge of becoming a grey goo problem?

 

JEHOVAH'S folly trumps man's wisdom ?

 New Paper on the Panda’s Thumb: “Striking Imperfection or Masterpiece of Engineering?”


Readers are invited to consider my new paper,   “The Panda’s Thumb: Striking Imperfection or Masterpiece of Engineering?The abstract is below.

Abstract: Key Points of the Contents 

Before going further, a brief note on the synonyms that I’m using here such as the “double/dual/complementary function” of the panda’s thumb. Each of the synonyms has its own subtly different overtones. With this in mind, I hope the basic points discussed below may be better understood. 

Above: “Some Key Points in a Long-Lasting Controversy”: Different views of evolutionary biologists on the panda’s thumb. Some assessments of the panda’s dexterity by intelligent design theorists.
Introduction: The panda’s thumb has become a paradigm for evolution in general. Links to articles by Stephen Dilley, and notes on the recent controversy between Nathan Lents and Stuart Burgess.
If the panda’s thumb is an embodiment of bad design, where are the evolutionists’ proposals indicating how they could have done better?
Some citations from a public talk by Stuart Burgess on the ingenious design of the wrist.
A massive contradiction within the theory of evolution itself.
Double/dual/complementary function is often overlooked.
“What makes the modern human thumb myology special within the primate clade is … [the appearance of] two extrinsic muscles, extensor pollicis brevis and flexor pollicis longus.”
It is a fundamental mistake to use the human thumb as a yardstick for the perfection or imperfection of the panda’s thumb.
A closer look at the differences of the radial sesamoid in a basal ursoid in comparison to that of the panda (Ailuropoda) for gripping and walking and the grasping hand of Homo sapiens according to Xiaoming Wang et al. (2022).
In comparison to other bear species, “only in A. melanoleuca can it be considered to be hyper-developed, reaching a similar size to that of the first metacarpal.”
Doubts concerning a simple homology of different sesamoid bones in various species.
Radial sesamoid as the ideal starting point to develop a thumb-like digit in pandas.
Natural selection of the radial sesamoid according to Wang et al. as well as Barrette in contrast to Stanley.
Implications of the ruling neo-Darwinian paradigm (gradualism plus natural selection) for the origin of the panda’s thumb.
Further discussion of Barrette’s points as “the length of the radial sesamoid, and therefore that of the false thumb, is limited firstly by its location under the hand,” etc.
Less efficient feeding would emphasize the enormous problem involved in the theory of natural selection.
The panda’s ecological impact and the “Optimal Panda Principle” in contrast to the evolutionary “Panda Principle” of Gould and his followers.
How to pick up little Necco candy wafers with thumbless mittens?
When directly observing pandas in zoos, Gould and Davis marveled at the dexterity/competence/virtuosity of the panda’s hand. I have done so, too. The panda’s hand is not “clumsy” at all.
Key question from two PhD students at the Max Planck Institute of Plant Breeding Research (Cologne) who came to my office and asked: Wouldn’t it be much more economical for an intelligent designer to modify, as far as possible, an already existing structure for some new functions than to create a totally new structure for similar roles/purposes/tasks from scratch?
Some comments on Barette’s statement that “We owe this metaphor [of approximate tinkering/bricolage] to François Jacob, a French biologist and recipient of the Nobel Prize. Far from being perfect, such approximate tinkering is a trace left by evolutionary history,” and thus a proof of it.
Davis on the enlarged radial sesamoid as “unquestionably” a direct product of natural selection.
Possible number of genes involved in the origin of pandas according to Davis and some others.
What do we know in the interim about panda genetics?
SNPs in the Ursidae including our beloved pandas.
As already mentioned in other articles of mine (for example: https://www.weloennig.de/Hippo.pdf): Note please that virtually all highlighting/emphasis is by W.-E. L. (except italics for genera and species as well as adding a note when the cited authors themselves have emphasized certain points). Why so often? Well, since many people do not have the time to study a more extensive work in detail, these highlights can serve as keywords to get a first impression of what is being discussed. 

Concerning the key points enumerated above: Page numbers may change in a future update, and so are not presented here. Incidentally, citations do not imply the agreement of the authors quoted with my overall views nor vice versa. Moreover, I alone am responsible for any mistakes.

On questions concerning absolute dating methods, see http://www.weloennig.de/HumanEvolution.pdf, p. 28. 


Thursday 27 June 2024

technology of the zygote vs. Darwin.

 Let’s Think About a Zygote Like an Engineer


Having read Evolution News for years, contributing an occasional article or two, in addition to my 81-part series on “The Designed Body,” I’ve noticed that there’s a certain way we proponents of intelligent design tend to frame our arguments. We usually provide information on what it takes for life to work, rather than just how it looks (per much of neo-Darwinism). Then we look for reasonable explanations of causation which must include where the information came from to properly produce, assemble, and coordinate all the necessary parts of a given system that we know is absolutely needed for survival (most of that is absent from neo-Darwinism).

But in my collaboration with Steve Laufmann to produce our book Your Designed Body, we came to the conclusion that a different style may be more useful. What we propose is that, in addition to what’s described above, we also engage readers with examples of “problem-solving” just like engineers do it. After all, it takes one to know one. If you’ve never used mental energy to try to solve any one of these hard problems of life, then how can you appreciate what it took to come up with and apply the solution? 

Let’s try the following as an exercise. Once you’ve gone through it, you’ll be better prepared to understand all the causal hurdles that had to have been surmounted. And this will allow you to ask better questions and not be as vulnerable to many of the “just so” stories of neo-Darwinism. 

“Separation of Concerns”

Recently, there was an article in The Scientist, “The First Two Cells in a Human Embryo Contribute Disproportionately to Fetal Development.” It noted a study published in Cell, “The first two blastomeres contribute unequally to the human embryo,” indicating that “a research team showed that, contrary to current models, one early embryonic cell dominates lineages that will become the fetus.” 

The gist of the article was that the current thinking — that it’s at the eight-cell stage where totipotent embryonic cells take the first “fork in the road” of commitment to developing into the fetus or the placenta — may be incorrect. It would now seem that this first “separation of concerns” (as Laufmann and I call it) may take place earlier on, when the zygote divides into the first two blastomeres. 

Ingenious methods were used to label and track the cell lineage from the two-cell to the blastocyst stage:“Thus, they could determine the contribution of each cell to the development of two early structures: the trophectoderm (TE) that becomes the placenta and the inner cell mass (ICM) that eventually produces the fetal tissue.” 

“They are not identical,” explains Magdalena Zernicka-Goetz, a developmental and stem cell biologist at Caltech and the University of Cambridge who is a study co-author. “Only one of the two cells is truly totipotent, meaning it can give rise to body and placenta, and the second cell gives rise mainly to placenta.” She adds, “I was always interested in how cells decide their fate.” The article in The Scientist concludes by telling us that “next, Zernicka-Goetz aims to investigate the features and origins of the differences between clones at the two-cell stage.”

Points to Ponder

It is clear that scientists still do not fully understand how human life develops from the zygote to a newborn and then into a mature fertile adult. One has to wonder what signaling and communication must take place at exactly the right times and in the right orders for all of this to happen properly, never mind where the information and instructions came from. Despite this self-acknowledged lack of understanding, we are told by evolutionary science that it certainly was an unguided and undirected natural process that brought it into being, and not a mind at work, as intelligent design contends. 

What do you think? If you took your car to a mechanic and he told you that he has no idea what’s wrong with it but he’s sure he can fix it, would you engage his services? Just because the scientist is smarter than you about what parts do what and how, that doesn’t necessarily mean that her conclusions about causation are true. After all, in saying that “I was always interested in how cells decide their fate,” she’s attributing agency, a mind at work, to the zygote. So don’t be misled.

Human Life Is a Hard Problem

Actually, life is a series of millions of hard problems that have to be solved all the time, or else. I’m talking about, among many other things, the cellular, metabolic, anatomical, and neuromuscular problems of human life. Let’s start from square one — the human zygote — how each of us began after the sperm of our father joined with the egg of our mother within her womb. That is the one cell from which, within nine months, we developed into a three-trillion-cell newborn with all the equipment we needed to survive. 

If you could go back in time to that moment in your life, be nanosized and micro-pipetted into your own first cell, what would be the first problem you’d have to solve? In other words, once the zygote comes into being, what’s the first thing it has to do? 

Well, if it’s going to become a newborn in nine months or so, it’s got to start dividing. But that won’t happen for at least 24 hours, so you have to consider what else may be more important as the zygote floats within the fluid of your mom’s uterus.

The chemical content of the fluid inside the zygote (high potassium, low sodium) is the opposite of what’s in the fluid surrounding it (low potassium, high sodium). And because these ions can cross the cell membrane, diffusion would naturally make them try to equalize on both sides (inside and outside the zygote) which would spell disaster. So, the sodium/potassium pumps in the zygote’s cell membrane have to kick in right away to keep pushing sodium out and bringing potassium back in, right?

Yes, the action of the million or so sodium/potassium pumps in the zygote’s cell membrane are needed for it to stay alive. But what do they need to do their work?

All work requires energy. So, as with all of life, the first priority of the zygote is to generate enough energy through glycolysis (without oxygen) and cellular respiration (with oxygen). The zygote needs oxygen and glucose (or other substances) to metabolize to get the energy it needs.

And if the zygote’s going to divide into two cells, then four, eight, sixteen, and more, then it’s also going to need nutrients to be able to make more copies of itself. Where does the new human life get the oxygen and nutrients it needs, and how does it make sure of its supply until it becomes a newborn? 

The Engineering Problem

This is how Steve Laufmann and I framed this engineering problem in our book:

All cells need oxygen and nutrients. Early life is no exception. Fertilization results in a zygote, which multiplies through cell division to become an embryo. In the early phase, the embryo gets what is needed by diffusion from the surrounding fluid. This works when there’s only a few dozen cells. But within several weeks the embryo will grow into a fetus, and in a few months into a newborn with trillions of specialized cells organized into coherent, interdependent, finely tuned organ systems. For this to be possible, the embryo needs a better way to get oxygen and nutrients, and to get rid of carbon dioxide and waste materials. If he cannot meet this challenge, he will not survive. But he’s in a special situation, dwelling inside his mother, so he’ll need a solution altogether different from anything else in the body’s inventory — a distinct yet temporary system that can meet this need while he’s developing his permanent internal systems.

We go on to ask a very important engineering question:

How do you build a series of finely tuned, coherent interdependent systems, each necessary for life, and stay alive the whole time? It just wouldn’t do if the body needed to go dead for a while, build some stuff, then come back to life when everything was ready to go. What the child in the womb needs is a complete set of temporary systems to meet the needs of his rapidly growing body, to keep it alive until its own systems are ready to take over. Then at birth, when they are no longer needed, these systems must be discarded as the child transitions to long-term systems.

The Solution Is the Placenta

The  answer to the very hard engineering problem we asked above is the placenta. Somehow or other the zygote has the foresight to know that down the road it will develop into a fetus that requires the placenta for its metabolic and nutritional needs. 

This is how we explain the solution in our book:
           Tissues of the embryo (TE) combine with tissues of the mother (lining of the uterus) to make the placenta — a totally separate organ that provides the scaffolding needed to keep the developing child alive. The placenta enables the mother to sustain the developing child while his internal organ systems and tissues are being fabricated, integrated, and launched. The developing child is, quite literally, on life support between the zygote phase and birth, when his body is finally ready to take over the job.

Up until this recent study it was thought that it’s not until the embryo consists of at least eight cells that some of them start to commit to being part of the placenta (TE). But now it seems that it takes place at the two-cell stage. If your nanosized self is inside the zygote, which lever do you pull to make sure that one of the two forming blastomeres goes down the TE-track? And even more important, where did the lever come from? 

It appears that, based on the findings of this study, the answer to the first question will be the concern of future research. But since, as we are regularly assured, we all know that life came about from the unguided and undirected processes of natural selection acting on random variation, the second question is assumed to have already been answered back in 1859, before we knew any of these intricate particulars and when biological systems were assumed to be vastly simpler than they turned out to be. What do you think? Any questions?

More light ,less heat re:dark matter?

 

The plague of plagues?

 

Predarwinian design vs. Darwinism

 Life Can’t Exist Without Repair Mechanisms, and That’s a Problem for Origin-of-Life Theories


A cell is often described as a factory — a quite extraordinary factory that can run autonomously and reproduce itself. The first cell required a lengthy list of components, layers of organization, and a large quantity of complex specified information, as described by previous episodes of Long Story Short. The latest entry in the series emphasizes yet another requirement for life: an abundance of specific repair mechanisms. 

Damage to the “factory” of the cell occurs on two levels: damage to the stored information (either during replication or by natural degradation over time) and damage to the manufacturing machinery (either from faulty production of new machinery or damage incurred during use). Each type of damage requires specific repair mechanisms that demonstrate foresight — the expectation that damage will occur and the ability to recognize, repair and/or recycle only those components that are damaged. All known life requires these mechanisms. 

Damage to Stored Information

The initial process of DNA replication is facilitated by a polymerase enzyme which results in approximately one error for every 10,000 to 100,000 added nucleotides.1 However, no known life can persist with such a high rate of error, if left uncorrected.2 Fortunately, DNA replication in all life includes a subsequent proofreading step — a type of damage repair — that enhances the accuracy by a factor of 100 to 1,000. The current record holder for the sloppiest DNA replication of a living organism, under normal conditions, is Mycoplasma mycoides (and its human-modified relative, JVCI-syn 3A), where only 1 in 33,000,000 nucleotides are incorrectly copied.3

Following the replication of DNA, a daily barrage of DNA damage occurs during normal operating conditions. Life therefore requires sophisticated and highly specific DNA repair mechanisms. In humans, DNA damage response is estimated to involve a hierarchical organization of 605 proteins in 109 assemblies.4 Efforts to make the simplest possible cell by stripping out all non-essential genes has successfully reduced DNA repair to a minimal set of six genes.5 But, these six genes are encoded in thousands of base pairs of DNA, and the machinery to transcribe and translate those genes into the repair enzymes requires a minimum of 149 genes.6 Thus, the DNA code that is required to make DNA repair mechanisms easily exceeds 100,000 base pairs. Here, we encounter a great paradox, first identified in 1971 by Manfred Eigen7: DNA repair is essential to maintain DNA but the genes that code for DNA repair could not have evolved unless the repair mechanisms were already present to protect the DNA. 

Faulty Production of New Machinery

We  used to think that the metabolic machinery in a cell always produced perfect products. But the reality is that faulty products are unavoidable, resulting in the production of interfering or toxic garbage. All living organisms must therefore have machinery that identifies problems and either repairs or recycles the faulty products. 

The cell’s central manufacturing machine is the ribosome, a marvel that produces functional proteins from strands of mRNA (with the help of many supporting molecules). Unfortunately, about 2-4 percent of mRNA strands get stuck in the ribosome during translation into a protein.8 Not only does this halt production, but it could result in production of a toxic, half-finished protein. 

If the mitochondria could not get “unstuck,” life as we know it would end. In the process of self-replication, a single cell must produce an entire library of proteins, placing a heavy burden on the cell’s mitochondria. But with a 2-4 percent rate of stuck mRNA strands, the average cell would have each of its mitochondria get stuck at least five times before the cell could replicate.9 Therefore, life could never replicate and metabolism would cease unless this problem was solved.

Fortunately, all forms of life, even the simplest,9 are capable of trans–translation, typically involving a three-step process. First, a molecule combining transfer and messenger RNA and two helper molecules (SymB and EF-Tu) recognizes that mRNA is stuck in the ribosome and attaches a label to the half-formed protein. This label, called a degron, is essentially a polyalanine peptide. The condemned protein is recognized, degraded, and recycled by one of many proteases. Finally, the mRNA must also be labeled and recycled to keep it from clogging other ribosomes. In some bacteria,10 a pyrophosphohydrolase enzyme modifies the end of the mRNA, labeling it for destruction. An RNAse (another enzyme) then recognizes this label, grabs hold of the mRNA, and draws it close to its magnesium ion, which causes cleavage of the RNA. Another RNAse then finishes the job, breaking the mRNA up into single nucleotides which can be re-used.

The required presence of tools that can destroy proteins and RNA also comes with a requirement that those tools are highly selective. If these tools evolved, one would expect the initial versions to be non-selective, destroying any proteins or RNA within reach, extinguishing life and blocking the process of evolution.11

Note that the set of tools for trans-translation and protein and RNA recycling are all stored in DNA, which must be protected by repair mechanisms. And, these tools cannot be produced without mitochondria, but the mitochondria cannot be unstuck without the action of trans-translation. Thus, we encounter another case of circular causality. 

Damage Incurred During Use

The normal operation of enzymes or metabolites like co-enzymes or cofactors involves chemical reactions that follow specific paths. Deviations from the desired paths can occur from interferences like radiation, oxidative stress, or encountering the wrong “promiscuous” enzyme. These deviations result in rogue molecules that interfere with metabolism or are toxic to the cell. As a result, even the simplest forms of life require several metabolic repair mechanisms: 

“[T]here can be little room left to doubt that metabolite damage and the systems that counter it are mainstream metabolic processes that cannot be separated from life itself.”12
“It is increasingly evident that metabolites suffer various kinds of damage, that such damage happens in all organisms and that cells have dedicated systems for damage repair and containment.”13
As a relatively simple example of a required repair mechanism, even the simplest known cell (JVCI Syn 3A) has to deal with a sticky situation involving sulfur. Several metabolic reactions require molecules with a thiol group — sulfur bonded to hydrogen and to an organic molecule. The organism needs to maintain its thiol groups, but they have an annoying tendency to cross link (i.e., two thiol groups create a disulfide bond, fusing the two molecules together). Constant maintenance is required to break up this undesired linking. Even the simplest known cell requires two proteins (TrxB/JCVISYN3A_0819 and TrxA/JCVISYN3A_0065) to restore thiol groups and maintain metabolism.12 Because the repair proteins are themselves a product of the cell’s metabolism, this creates another path of circular causality: You can’t have prolonged metabolism without the repair mechanisms but you can’t make the repair mechanisms without metabolism.

An Ounce of Prevention is Worth a Pound of Cure

In addition to life’s required repair mechanisms, all forms of life include damage prevention mechanisms. These mechanisms can destroy rogue molecules, stabilize molecules that are prone to going rogue, or guide chemical reactions toward less harmful outcomes. As an example, when DNA is replicated, available monomers of the four canonical nucleotides (G, C, T, and A) are incorporated into the new strand. Some of the cell’s normal metabolites, like dUTP (deoxyuridine triphosphate), are similar to a canonical nucleotide and can be erroneously incorporated into DNA. Even the simplest cell (once again, JVCI-syn3A) includes an enzyme (deoxyuridine triphosphate pyrophosphatase) to hydrolyze dUTP and prevent formation of corrupted DNA.6

Summing Up the Evidence

Those who promote unguided abiogenesis simply brush off all of these required mechanisms, claiming that life started as simplified “proto-cells” that didn’t need repair. But there is no evidence that any form of life could persist or replicate without these repair mechanisms. And the presence of the repair mechanisms invokes several examples of circular causality — quite a conundrum for unintelligent, natural processes alone. Belief that simpler “proto-cells” didn’t require repair mechanisms requires blind faith, set against the prevailing scientific evidence.   

Notes

Babenek A, and Zuizia-Graczyk I. Fidelity of DNA replication — a matter of proofreading. Curr Genet. 2018; 54: 985-996.
Some viruses have high error rates when replicating, but viruses cannot replicate without the help of cellular life, which requires very low error rates. Some specialized DNA polymerases intentionally operate with lower fidelity on a temporary basis for purposes such as antibody diversity. 
Moger-Reischer RZ, et al. Evolution of a Minimal Cell. Nature. 2023; 620: 122-127.
Kratz A, et al. A multi-scale map of protein assemblies in the DNA damage response. Cell Systems 2023; 14: 447-463.
Hutchison CA, et al. Design and synthesis of a minimal bacterial genome. Science 2016; 351: aad6253. 
Breuer M, et al. Essential Metabolism for a Minimal Cell. eLife 2019;8:e36842 DOI: 10.7554/eLife.36842. 
Eigen, M. Self-organization of matter and evolution of biological macromolecules. Naturwissenschaften, 1971; 58: 465–523.
Ito K, et al. Nascentome analysis uncovers futile protein synthesis in Escherichia coli. PLoS One 2011; 6: e28413
Keiler KC, Feaga HA. Resolving nonstop translation complexes is a matter of life or death. Journal of Bacteriology 2014; 196: 2123-2130.
Mackie GA. RNase E: at the interface of bacterial RNA processing and decay. Nature Reviews Microbiology 2013; 11: 45-57.
“Because RNA degradation is ubiquitous in all cells, it is clear that it must be carefully controlled to accurately recognize target RNAs.” Houseley J and Tollervey D. The many pathways of RNA degradation. Cell 2009; 136: 763-776.
Hass D, et al. Metabolite damage and damage control in a minimal genome. American Society for Microbiology 2022; 13: 1-16.
Linster CL, et al. Metabolite damage and its repair or pre-emption. Nature Chemical Biology 2013; 9: 72-80.


Tuesday 25 June 2024

Common design vs. Common descent.

 New Paper Argues that Variant Genetic Codes Are Best Explained by Common Design


A popular argument for a universal common ancestor is the near-universality of the conventional genetic code. Critics of common descent often point to deviations from the standard code as evidence against it. A recent paper published in the journal BIO-Complexity, by Winston Ewert, reviews the character and distribution of genetic code variants and the implications these have for common ancestry, and “develops a framework for understanding codes within a common design framework, based crucially on the premise that some genetic code variants are designed and others are the result of mutations to translation machinery.” Ewert explains that,

Upon first investigation, evolutionary theory appears to have a compelling account of the character and distribution of variant codes. Evolutionary theory suggests that if genetic code evolution is possible it should be very rare. This would explain why most genomes follow the standard code and why the exceptions only vary in a few codons. It would also explain the following details about the variant codes. Most variations are found in mitochondria, whose very small genomes would make code evolution easier. Many variations are also found in highly reduced genomes, such as those of endosymbiotic bacteria. No variations are found in the nuclear genomes of complex multicellular organisms like plants and animals. The distribution of many codes can be easily explained by identifying certain points on the tree of life where codons were reassigned and then inherited by all of their descendants.

EWERT W (2024) ON THE ORIGIN OF THE CODES: THE CHARACTER AND DISTRIBUTION OF VARIANT GENETIC CODES IS BETTER EXPLAINED BY COMMON DESIGN THAN EVOLUTIONARY THEORY. BIO-COMPLEXITY 2024 (1):1-25.

Three Tenets

The paper proposes “a framework that seeks to explain the character and distribution of variant genetic codes within a common design framework.” Ewert’s framework has three tenets: First, “the canonical genetic code has been well optimized and is thus an ideal choice for most genomes.” There are multiple optimized parameters and thus “A designer must identify the best trade-offs to select the ideal genetic code.” The second tenet of Ewert’s framework is that a minor variation on the standard code is better suited to some organisms, since those organisms may acquire an advantage by a different set of trade-offs with respect to genetic code optimization. The third tenet is that the translation machinery has been damaged by mutations in some organisms, and that this has resulted in their misinterpreting the code they were initially designed to employ. These are examples of genetic code variants that have evolved naturally.

Five Criteria

Ewert offers five criteria that may be used to distinguish genetic codes that are evolved from those that are designed. First, evolved codes are expected to be found in taxonomic groups below the family level, whereas those that are designed are predicted to be above the level of family. Second, evolved codes should be readily “explicable in terms of some simple mutation to the translation machinery of the cell.” Third, it is predicted that codes that are evolved will be limited to the genomes of endosymbionts. Fourth, it is expected that codes that are evolved utilize a small number of codons so that the variation does not cause the organism too much harm. Fifth, it is predicted that evolved codes will fall into a simple nested hierarchical (phylogenetic) distribution. By contrast, 

[D]esigned codons are found in high-level taxa of at least genus-level but typically higher. They involve many reassignments that are difficult to explain with any sort of simple mutation. They are found in free-living organisms. They sometimes reassign codons that are expected to be rare. They are often distributed in a complex fashion that does not fit phylogenetic expectations.

EWERT W (2024) ON THE ORIGIN OF THE CODES: THE CHARACTER AND DISTRIBUTION OF VARIANT GENETIC CODES IS BETTER EXPLAINED BY COMMON DESIGN THAN EVOLUTIONARY THEORY. BIO-COMPLEXITY 2024 (1):1-25.

It has been conventionally thought that evolution provides a good explanation for the character and distribution of genetic code variants — in particular, the near-universality of the standard code; the prevalence of variant codes in simple genomes such as those of mitochondria; and the phylogenetic distribution of variant codes. Ewert notes that, in light of evolutionary theory, it would in fact be expected that there would be variant codes found at the higher taxonomic levels, which would be consistent with the genetic code still being variable at the time of the last common ancestor. However, “What we observe instead are modifications of the standard code. They are not associated with the high-level taxa…” 

Furthermore, though we should expect on evolutionary theory that it would be exponentially harder to reassign a code as the number of genes increases, “variant codes are found in nuclear genomes that are not particularly small. They are found in ciliates, which have comparable numbers of genes to the human genome. Additionally, we find them in some multicellular green algae. In fact, we find more code variation in eukaryotic nuclear genomes than in bacterial genomes, despite eukaryotes having much larger genomes.” Thus, Ewert concludes, “despite the initial impression, evolutionary theory does not account well for the kinds of genomes with variant codes.”

Invoking “Inexplicable Events”

Finally, though evolutionary theory would predict that the distribution of variant codes would be consistent with the standard phylogeny, Ewert observes that, “In many cases, the distribution of a code is complex, defying evolutionary explanations. Codes recur in closely related groups in a way not explained by common descent. Evolutionary theory has to invoke inexplicable events such as reversions to the standard code.”

Thus, Ewert concludes, “Initially, evolutionary theory appeared to have some explanatory power. However, upon closer inspection, the features of the variant codes that seemed well explained by evolutionary theory turned out to either be inaccurate or to not follow from evolutionary theory.” Instead, he argues that the character and distribution of variant codes is often better explained under a framework of common design.

The paper is well worth a careful read. It can be accessed here.


Yet another house divided?

 

The future we were promised is finally here?

 

Friday 21 June 2024

Every silver lining has a cloud?

 

Settled science is the real science stopper?

 

Thoughtful Darwinism to ID : lets be frenemies II

 Evolutionary Biologist Concedes Intelligent Design Is the Cutting Edge


Bret Weinstein and Heather Heying are well-known evolutionary biologists (and husband and wife) with a podcast, the DarkHorse Podcast. Recently Weinstein posed a provocative question, “Is intelligent design a competitor to Darwinian evolution?” His answer may surprise you: Yes.  

No, he’s not about to come over to the dark side. Weinstein is confident that Darwinism will meet the challenges that ID has set, about the Cambrian Explosion and more, but he concedes that it hasn’t done so yet.

He describes conversations with Richard Dawkins and Jerry Coyne, asking them why evolutionary biology hasn’t had a breakthrough since 1976 when Dawkins published The Selfish Gene. Dawkins and Coyne, separately, both answered that it was because all the big questions had already been answered, and all that was left was a “clean up” operation. Weinstein recognizes that that is just so much blowing smoke, and the work of Stephen Meyer and his “high-quality colleagues” in the ID research community has exposed the problems that Darwinists need to be working on to solve. 

Settled Science?

About Dr. Meyer, he says:
           I encountered people like Stephen Meyer, who were not phony scientists, pretending to do the work. They were actually very good at what they did. And I believe Stephen Meyer is motivated by a religious motivation, but we don’t generally ask the question when somebody takes up science, “What are you really in it for? Are you in it for the fame?” That’s not a legitimate challenge to somebody’s work. 

And the fact is, Stephen Meyer is very good at what he does. He may be motivated by the thought that at the end of the search he’s going to find Jesus. But in terms of the quality of his arguments, I was very impressed when I met him: his love for biology, his love for creatures, the weirder the better, he likes them, right? So that looked very familiar to me. 

No Mind Readers Here

In other words, motivation should be irrelevant. The quality of the science is what counts. I would add, none of us is a mind reader and we can never know what someone else’s motivation really is. In any event, says Weinstein, ID clearly is about science, not religion:

And it all also became obvious to me in interacting with Stephen Meyer and many of his high-quality colleagues that they’re actually motivated, for whatever reason, to do the job that we are supposed to be motivated to do inside of biology. They’re looking for cracks in the theory. Things that we haven’t yet explained. And they’re looking for those things for their own reasons, but the point is we’re supposed to be figuring out what parts of the stories we tell ourselves aren’t true, because that’s how we get smarter over time. 

Shrinking from a Fight 

Darwinists, say Weinstein, are shrinking from a fight they wrongly feel they shouldn’t have to bother with:

If you decide… that your challengers aren’t entitled to a hearing because they’re motivated by the wrong stuff, then you do two things. One, you artificially stunt the growth of your field, and you create a more vibrant realm where your competitors have a better field to play in because you’ve left a lot of holes in the theory ready to be identified, which I think is what’s going on. The better intelligent design folks are finding real questions raised by Darwinism, and the Darwinists, instead of answering those questions, [are] deciding it’s not worthy of their time. And that is it is putting us on a collision course.

“Giving Up Darwin”

Heying cites the 2019 public defection from Darwinism of Yale computer scientist David Gelernter, who pointed to Meyer’s writing as his primary reason for “Giving Up Darwin.” She admits she hasn’t kept up with the challenges from ID, but agrees that she should keep up, and that’s because challenges like those from ID can make the evolutionary establishment “smarter.” Ignoring the challenges makes the establishment dumber — stagnant and self-satisfied.

I’m not familiar with most of the arguments that are coming out of the intelligent design movement. It hasn’t felt like it was my obligation to be familiar with them. Perhaps what you’re arguing is it is our responsibility.

Weinstein, unlike Coyne or Dawkins, is up for talking and debating with ID proponents:

I’m open to that battle and I expect that if we pursue that question, what we’re going to find is, oh, there’s a layer of Darwinism we didn’t get and it’s going to turn out that the intelligent design folks are going to be wrong. But they will have played a very noble and important role in the process of us getting smarter. And look, I think Stephen Meyer at the end of the day, I don’t think he’s going to surrender to the idea that there’s no God at the end of this process. But if we find a layer of Darwinism that hasn’t been spotted, that answers his question, I think he’s going to be delighted with it the same way he’s delighted by the prospect of seeing whale sharks.

Again, these are remarkable concessions from a couple of scientists who are not at all looking to make the leap to ID, but who understand that intelligent design, not Darwinism, is currently at biology’s cutting edge.

Yet more on the fossil records anti Darwinian bias

 Fossil Friday: Ediacaran Animal Embryos Put to Test and Put to Rest


The Weng’an biota of the Doushantuo Formation in South China is a famous fossil Lagerstätte, which is of particular importance, because it is dated to an Early-Middle Ediacaran age (590-575 million years ago), right in the time when molecular clock estimates place the origin of crown group metazoan animal phyla. The absence of actual unequivocal fossil animals from this period has often been explained away as an artifact of an incomplete fossil record. However, the discovery of various Ediacaran localities of the Burgess-Shale-Type decisively refuted this artifact hypothesis (Bechly 2020b), because those localities would have easily preserved any early small and soft-bodied animals, but only yielded macroalgae and a few problematic forms of uncertain affinity.

The Last Straw

The phosphatized microfossils of the Doushantou Formation, which are three-dimensionally preserved down to the cellular level, represent the last straw to somehow align the molecular clock expectations with the actual fossil record. This explains the urge by paleontologists to readily interpret some of the Doushantuo fossils as metazoan embryos. To the great frustration of evolutionists, all these attempts proved to be highly contentious and ultimately failed to provide any convincing evidence for Ediacaran metazoans. I discussed the dubious nature of these alleged animal embryos in several previous articles (Bechly 2020a, 2020b, 2022a; also see Evolution News 2016), where you can also find all the references to the peer-reviewed literature.

Now a new study by Sun et al. (2024), published in the Proceedings of the Royal Society, looks at the developmental biology of the genus Spiralicellula that was previously proposed as a potential metazoan embryonic stage, a kind of planktonic larva. The authors found a very different developmental mode exhibited by Spiralicellula and all the other alleged metazoan embryos, lacking any cell differentiation, compared to any crown group metazoans. They therefore explicitly “reject a crown-metazoan affinity for Spiralicellula and all other components of the Weng’an biota, diminishing the probability of crown-metazoan diversification before the early Ediacaran.” The authors do not even consider Spiralicellula as a plausible stem-animal but conclude “that Spiralicellula is more likely affiliated with non-metazoan holozoans than with stem metazoans.” So the highly critical view that I elaborated in my previous articles is still well in line with the most recent mainstream research by the leading experts on these fossils.

A New Study

Therefore, it is well worth quoting at some length from the conclusions of this interesting and important new study:

While the embryo-like fossils from the Weng’an biota were once thought to represent the earliest metazoans, recent research has suggested that some species, including Caveasphaera [40], Helicoforamina [29], Ostiosphaera [47], Sporosphaera [33] and now Spiralicellula, have affinities that lie outside crown Metazoa. Interpretations of soma–germ cell differentiation in Megasphaera [21] have been used to support their interpretation as stem metazoans, at best. The primary evidence for a stem-metazoan affinity of Megasphaera derives from ‘matryoshka’ structures that are interpreted to reflect cell differentiation. Nevertheless, it remains unclear whether these structures are endogenous or exogenous in origin [7,48]….

The Weng’an biota, renowned for its exceptional preservation fidelity, is considered a distinctive taphonomic window that holds great potential for documenting the earliest metazoans. The absence of definitive evidence of crown metazoans in the biota is inconsistent with the expectations of the molecular clock estimates which posit a Tonian or Cryogenian origin for the clade [1,2]. It remains formally possible that the absence of crown-group animals from the Weng’an biota and earlier strata reflects the incompleteness of the fossil record, and the discovery of unequivocal metazoans from the Weng’an biota or older strata remains a viable possibility, not least given the discovery of crown metazoans, including cnidarians and bilaterians, within the later Ediacaran [50–54]. However, claims of crown metazoans from the Cryogenian [55,56] and Tonian [57,58] are all highly contested [59–62] and intense exploration of the Weng’an biota, the most exceptional of all sites of fossil preservation, has failed to yield the anticipated evidence of early crown metazoans, instead yielding only evidence of non-metazoan holozoans or possible stem metazoans. Alongside the Weng’an biota, the Doushantuo silicified Lagerstätte in South China serves as its lateral counterpart [28,63–65]. Despite differing preservational settings, this Lagerstätte remarkably preserves fossil structures down to a subcellular level. It contains a diverse array of microfossils, including cyanobacteria, acritarchs, multicellular algae, and embryo- like fossils [63,66,67], all found in the Weng’an biota. Notably absent, however, are fossils of crown-group metazoans. As such, the available fossil evidence suggests a relatively low probability of crown metazoans diversifying in the early Ediacaran, rather than ecological constraints within the Weng’an biota’s preservational setting. Such insights prompt a recalibration of molecular timescales in light of these discoveries.

In short: There are no fossil animals in the Ediacaran, when they should be found according to the gradualistic predictions of Darwinian evolution and according to molecular clock datings. The fossil record does not agree with either of these predictions, so that the theory fails the empirical test.

Time for a Better Theory!

In his New York Times bestseller book Darwin’s Doubt, Stephen Meyer (2013) considered that the Ediacaran Doushantuo fossils may indeed include actual animal embryos of sponges. As I have already indicated (Bechly 2020b, 2022a, 2022b), this concession may have been far too generous. Even mainstream evolutionist science more and more recognizes that there simply are no metazoan animal embryos in the Doushantuo Formation. So, where are all the postulated ancestors of the more than twenty different animal phyla appearing abruptly in the Cambrian Explosion? Even Richard Dawkins (2009) has admitted that the Cambrian shows us a substantial number of major animal phyla “already in an advanced state of evolution, the very first time they appear. It is as though they were just planted there, without any evolutionary history. Needless to say, this appearance of sudden planting has delighted creationists.” Well, maybe nature is telling you something.

References

Bechly G 2020a. The Myth of Precambrian Sponges. Evolution News May 12, 2020. https://evolutionnews.org/2020/05/the-myth-of-precambrian-sponges/
Bechly G 2020b. The Demise of the Artifact Hypothesis. Evolution News June 6, 2020. https://evolutionnews.org/2020/07/demise-of-the-artifact-hypothesis-aggravates-the-problem-of-the-cambrian-explosion/
Bechly G 2022a. “Lying on the Internet”? Debunking Dave Farina on Stephen Meyer. Evolution News December 1, 2022. https://evolutionnews.org/2022/12/lying-on-the-internet-debunking-dave-farina-on-stephen-meyer/
Bechly G 2022b. Let’s Help “Professor Dave” Understand the Precambrian. Evolution News December 2, 2022. https://evolutionnews.org/2022/12/lets-help-professor-dave-understand-the-precambrian/
Dawkins R 2009. The Greatest Show on Earth. Free Press, New York (NY), 470 pp.
Evolution News 2016. New Precambrian Embryos Are equivocal at Best. Evolution News August 18, 2016. https://evolutionnews.org/2016/08/new_precambrian_1/
Meyer SC 2013. Darwin’s Doubt: The Explosive Origin of Animal Life and the Case for Intelligent Design. HarperOne, New York (NY), viii+498 pp.
Sun W, Yin Z, Liu P, Zhu M & Donoghue P 2024. Developmental biology of Spiralicellula and the Ediacaran origin of crown metazoans. Proceedings of the Royal Society B 291: 20240101, 1–10. DOI: https://doi.org/10.1098/rspb.2024.0101


The carbon atom vs. Darwin.

The Remarkable Carbon Atom


In an article yesterday, I discussed the incredible design of the nonmetal atoms, and the striking coincidence that the very atoms from which one can build stable, defined shapes also give us the hydrophobic force, which is the key to arranging them into higher-level structures. Here, I will discuss the fitness of carbon for life, and the incredibly fortuitous circumstances that promote its abundance in the universe.

The Fitness of Carbon for Life

The carbon atom, the primary constituent of organic molecules, is, in several respects, uniquely fit for the assembly of the complex macromolecules found in the cell. First, due to the stability of carbon-carbon bonding, only carbon can form long polymers of itself, forming long chains or rings, while also bonding to other kinds of atoms. Though silicon can also form long chains by bonding with itself, these bonds are significantly less stable than carbon-carbon bonds. Plaxco and Gross note that “while silicon-silicon, silicon-hydrogen, and silicon-nitrogen bonds are similar in energy, the silicon-oxygen bond is far more stable than any of the other three types. As a consequence, silicon readily oxidizes to silicon dioxide, limiting the chemistry available to this atom whenever oxygen is present. And oxygen is the third most common atom in the Universe.”1 As Primo Levi explains, carbon “is the only element that can bind itself in long stable chains without a great expense of energy, and for life on earth (the only one we know so far) precisely long chains are required. Therefore carbon is the key element of living substance.”2

Second, carbon is tetravalent — that is, each atom can form four covalent bonds with other atoms. Third, carbon possesses a relatively small atomic nucleus, entailing short bond distances, thereby allowing it to form stable bonds with itself as well as other atoms. This property is also possessed by the other small, non-metal atoms in period two. Carbon is able to form single, double, and triple bonds with other atoms. Nitrogen can also form single, double, or triple bonds and oxygen can form single and double bonds. Contrast this with the nonmetal atoms directly beneath them in the periodic table — silicon, phosphorus, and sulfur — which possess larger atomic radii and therefore form such bonds less easily due to multiple bonds having reduced stability. 

Another property of organic bonds is that their strength sits within a Goldilocks zone, being neither too strong nor too weak for biochemical manipulations in the cell. If the strength of those bonds were to be altered by a single order of magnitude, it would render impossible numerous biochemical reactions that take place in the cell. If it were too strong, the activation energy needed to break bonds could not be sufficiently reduced by enzymatic activity (enzymes strain chemical bonds by engaging in specific conformational movements while bound to a substrate). Conversely, if organic bonds were much weaker, bonds would be frequently disrupted by molecular collisions, rendering controlled chemistry impossible. 

Another special characteristic of carbon is that there is not much variation in energy levels of carbon bonds from one atom to the next. Robert E. D. Clark explains that carbon “is a friend of all. Its bond energies with hydrogen, chlorine, nitrogen, oxygen, or even another carbon differ little. No other atom is like it.”3 Kevin W. Plaxco and Michael Gross further comment, “Carbon presents a fairly level playing field in which nature can shuffle around carbon-carbon, carbon-nitrogen, and carbon-oxygen single and double bonds without playing too great a cost to convert any one of these into another… Given all this, it’s no wonder that on the order of ten million unique carbon compounds have been described by chemists, which is as many as all of the described non-carbon-containing compounds put together.”4

Carbon Resonance

As we have seen, carbon is absolutely fundamental to life. It also happens to be — after hydrogen, helium, and oxygen — the fourth most abundant element in our galaxy. A carbon nucleus can be generated by smashing together two nuclei of helium-4 to make beryllium-8 (containing four protons and four neutrons) and then adding a further nucleus of helium to generate carbon-12 (containing six protons and six neutrons). However, beryllium is quite unstable, and can be expected to break apart into two nuclei of helium in 10-16seconds. On occasion, prior to the breaking apart of beryllium, a third helium nucleus collides with beryllium, resulting in a carbon nucleus. As it happens, the carbon atom possesses a special quantum property called a resonance, which facilitates this process. A resonance describes the discrete energy levels at which protons and neutrons in the nucleus can exist. Indeed, it turns out that the resonance of the carbon atom just so happens to correspond to the combined energy of the beryllium atom and a colliding nucleus of helium.

As Geraint Lewis and Luke Barnes explain, “if there were a resonance at just the right place in carbon, the combined energy of the beryllium and helium nuclei would result in a carbon nucleus in one of its excited states. The excited carbon nucleus knows how to handle the excess energy without simply falling apart. It is less likely to disintegrate, and more likely to decay to the ground state with the emission of a gamma-ray photon. Carbon formed, energy released… success!”5 Without this specific resonance level, the universe would contain relatively few carbon atoms — in 1953, this specific resonance that had previously been predicted by Fred Hoyle was discovered by William Fowler, precisely where Hoyle had predicted it would be.

A Remarkable Coincidence

This special carbon resonance (known as the Hoyle state), which corresponds to the energy levels of the combined beryllium-8 nucleus and a helium-4 nucleus, renders the otherwise improbable process of carbon-12 formation feasible and efficient in the high-temperature environments of stellar cores. This delicate balance of energy levels is a remarkable aspect of nuclear astrophysics that allows for the creation of the elements necessary for life. If it were not for this special resonance, life very probably would not exist in our universe. This is another one of many countless features of our universe that have to be “just right” for life — in particular, advanced life — to exist.

Notes

Kevin W. Plaxco and Michael Gross. Astrobiology: A Brief Introduction, 2nd edition (The John Hopkins University Press, 2011), chapter 1.
Primo Levi, The Periodic Table (Abacus, 1990), 226-227.
Robert E.D. Clark, The Universe: Plan or Accident? 3rd edition, (Zondervan, 1972), 97.
Kevin W. Plaxco and Michael Gross. Astrobiology: A Brief Introduction, 2nd edition (The John Hopkins University Press, 2011), chapter 1.
Geraint F. Lewis and Luke A. Barnes, A Fortunate Universe: Life in a Finely Tuned Cosmos (Cambridge University Press, 2017), 116-117.