More Recent Comments

Tuesday, June 19, 2007

What is a gene, post-ENCODE?

Back in January we had a discussion about the definition of a gene [What is a gene?]. At that time I presented my personal preference for the best definition of a gene.
A gene is a DNA sequence that is transcribed to produce a functional product.
This is a definition that's widely shared among biochemists and molecular biologists but there are competing definitions.

Now, there's a new kid on the block. The recent publication of a slew of papers from the ENCODE project has prompted many of the people involved to proclaim that a revolution is under way. Part of the revolution includes redefining a gene. I'd like to discuss the paper by Mark Gerstein et al. (2007) [What is a gene, post-ENCODE? History and updated definition] to see what this revolution is all about.

The ENCODE project is a large scale attempt to analyze and annotate the human genome. The first results focus on about 1% of the genome spread out over 44 segments. These results have been summarized in an extraordinarily complex Nature paper with massive amounts of supplementary material (The Encode Project Consortium, 2007). The Nature paper is supported by dozens of other papers in various journals. Ryan Gregory has a list of blog references to these papers at ENCODE links.

I haven't yet digested the published results. I suspect that like most bloggers there's just too much there to comment on without investing a great deal of time and effort. I'm going to give it a try but it will require a lot of introductory material, beginning with the concept of alternative splicing, which is this week's theme.

The most widely publicized result is that most of the human genome is transcribed. It might be more correct to say that the ENCODE Project detected RNA's that are either complimentary to much of the human genome or lead to the inference that much of it is transcribed.

This is not news. We've known about this kind of data for 15 years and it's one of the reasons why many scientists over-estimated the number of humans genes in the decade leading up to the publication of the human genome sequence. The importance of the ENCODE project is that a significant fraction of the human genome has been analyzed in detail (1%) and that the group made some serious attempts to find out whether the transcripts really represent functional RNAs.

My initial impression is that they have failed to demonstrate that the rare transcripts of junk DNA are anything other than artifacts or accidents. It's still an open question as far as I'm concerned.

It's not an open question as far as the members of the ENCODE Project are concerned and that brings us to the new definition of a gene. Here's how Gerstein et al. (2007) define the problem.
The ENCODE consortium recently completed its characterization of 1% of the human genome by various high-throughput experimental and computational techniques designed to characterize functional elements (The ENCODE Project Consortium 2007). This project represents a major milestone in the characterization of the human genome, and the current findings show a striking picture of complex molecular activity. While the landmark human genome sequencing surprised many with the small number (relative to simpler organisms) of protein-coding genes that sequence annotators could identify (~21,000, according to the latest estimate [see www.ensembl.org]), ENCODE highlighted the number and complexity of the RNA transcripts that the genome produces. In this regard, ENCODE has changed our view of "what is a gene" considerably more than the sequencing of the Haemophilus influenza and human genomes did (Fleischmann et al. 1995; Lander et al. 2001; Venter et al. 2001). The discrepancy between our previous protein-centric view of the gene and one that is revealed by the extensive transcriptional activity of the genome prompts us to reconsider now what a gene is.
Keep in mind that I personally reject the premise and I don't think I'm alone. As far as I'm concerned, the "extensive transcriptional activity" could be artifact and I haven't had a "protein-centric" view of a gene since I learned about tRNA and ribosomal RNA genes as an undergraduate in 1967. Even if the ENCODE results are correct my preferred definition of a gene is not threatened. So, what's the fuss all about?

Regulatory Sequences
Gerstein et al. are worried because many definitions of a gene include regulatory sequences. Their results suggest that many genes have multiple large regions that control transcription and these may be located at some distance from the transcription start site. This isn't a problem if regulatory sequences are not part of the gene, as in the definition quoted above (a gene is a transcribed region). As a mater of fact, the fuzziness of control regions is one reason why most modern definitions of a gene don't include them.
Overlapping Genes
According to Gerstein et al.
As genes, mRNAs, and eventually complete genomes were sequenced, the simple operon model turned out to be applicable only to genes of prokaryotes and their phages. Eukaryotes were different in many respects, including genetic organization and information flow. The model of genes as hereditary units that are nonoverlapping and continuous was shown to be incorrect by the precise mapping of the coding sequences of genes. In fact, some genes have been found to overlap one another, sharing the same DNA sequence in a different reading frame or on the opposite strand. The discontinuous structure of genes potentially allows one gene to be completely contained inside another one’s intron, or one gene to overlap with another on the same strand without sharing any exons or regulatory elements.
We've known about overlapping genes ever since the sequences of the first bacterial operons and the first phage genomes were published. We've known about all the other problems for 20 years. There's nothing new here. No definition of a gene is perfect—all of them have exceptions that are difficult to squeeze into a one-size-fits-all definition of a gene. The problem with the ENCODE data is not that they've just discovered overlapping genes, it's that their data suggests that overlapping genes in the human genome are more the rule than the exception. We need more information before accepting this conclusion and redefining the concept of a gene based on analysis of the human genome.
Splicing
Splicing was discovered in 1977 (Berget et al. 1977; Chow et al. 1977; Gelinas and Roberts 1977). It soon became clear that the gene was not a simple unit of heredity or function, but rather a series of exons, coding for, in some cases, discrete protein domains, and separated by long noncoding stretches called introns. With alternative splicing, one genetic locus could code for multiple different mRNA transcripts. This discovery complicated the concept of the gene radically.
Perhaps back in 1978 the discovery of splicing prompted a re-evaluation of the concept of a gene. That was almost 30 years ago and we've moved on. Now, many of us think of a gene as a region of DNA that's transcribed and this includes exons and introns. In fact, the modern definition doesn't have anything to do with proteins.

Alternative splicing does present a problem if you want a rigorous definition with no fuzziness. But biology isn't like that. It's messy and you can't get rid of fuzziness. I think of a gene as the region of DNA that includes the longest transcript. Genes can produce multiple protein products by alternative splicing. (The fact that the definition above says "a" functional product shouldn't mislead anyone. That was not meant to exclude multiple products.)

The real problem here is that the ENCODE project predicts that alternative splicing is abundant and complex. They claim to have discovered many examples of splice variants that include exons from adjacent genes as shown in the figure from their paper. Each of the lines below the genome represents a different kind of transcript. You can see that there are many transcripts that include exons from "gene 1" and "gene 2" and another that include exons from "gene 1" and "gene 4." The combinations and permutations are extraordinarily complex.

If this represents the true picture of gene expression in the human genome, then it would require a radical rethinking of what we know about molecular biology and evolution. On the other hand, if it's mostly artifact then there's no revolution under way. The issue has been fought out in the scientific literature over the past 20 years and it hasn't been resolved to anyone's satisfaction. As far as I'm concerned the data overwhelmingly suggests that very little of that complexity is real. Alternative splicing exists but not the kind of alternative splicing shown in the figure. In my opinion, that kind of complexity is mostly an artifact due to spurious transcription and splicing errors.
Trans-splicing
Trans-splicing refers to a phenomenon where the transcript from one part of the genome is attached to the transcript from another part of the genome. The phenomenon has been known for over 20 years—it's especially common in C. elegans. It's another exception to the rule. No simple definition of a gene can handle it.
Parasitic and mobile genes
This refers mostly to transposons. Gerstein et al say, "Transposons have altered our view of the gene by demonstrating that a gene is not fixed in its location." This isn't true. Nobody has claimed that the location of genes is fixed.
The large amount of "junk DNA" under selection
If a large amount of what we now think of as junk DNA turns out to be transcribed to produce functional RNA (or proteins) then that will be a genuine surprise to some of us. It won't change the definition of a gene as far as I can see.
The paper goes on for many more pages but the essential points are covered above. What's the bottom line? The new definition of an ENCODE gene is:
There are three aspects to the definition that we will list below, before providing the succinct definition:
  1. A gene is a genomic sequence (DNA or RNA) directly encoding functional product molecules, either RNA or protein.
  2. In the case that there are several functional products sharing overlapping regions, one takes the union of all overlapping genomic sequences coding for them.
  3. This union must be coherent—i.e., done separately for final protein and RNA products—but does not require that all products necessarily share a common subsequence.
This can be concisely summarized as:
The gene is a union of genomic sequences encoding a coherent set of potentially overlapping functional products.
On the surface this doesn't seem to be much different from the definition of a gene as a transcribed region but there are subtle differences. The authors describe how their new definition works using a hypothetical example.

How the proposed definition of the gene can be applied to a sample case. A genomic region produces three primary transcripts. After alternative splicing, products of two of these encode five protein products, while the third encodes for a noncoding RNA (ncRNA) product. The protein products are encoded by three clusters of DNA sequence segments (A, B, and C; D; and E). In the case of the three-segment cluster (A, B, C), each DNA sequence segment is shared by at least two of the products. Two primary transcripts share a 5' untranslated region, but their translated regions D and E do not overlap. There is also one noncoding RNA product, and because its sequence is of RNA, not protein, the fact that it shares its genomic sequences (X and Y) with the protein-coding genomic segments A and E does not make it a co-product of these protein-coding genes. In summary, there are four genes in this region, and they are the sets of sequences shown inside the orange dashed lines: Gene 1 consists of the sequence segments A, B, and C; gene 2 consists of D; gene 3 of E; and gene 4 of X and Y. In the diagram, for clarity, the exonic and protein sequences A and E have been lined up vertically, so the dashed lines for the spliced transcripts and functional products indicate connectivity between the proteins sequences (ovals) and RNA sequences (boxes). (Solid boxes on transcripts) Untranslated sequences, (open boxes) translated sequences.
This isn't much different from my preferred definition except that I would have called the region containing exons C and D a single gene with two different protein products. Gerstein et al (2007) split it into two different genes.

The bottom line is that in spite of all the rhetoric the "new" definition of a gene isn't much different from the old one that some of us have been using for a couple of decades. It's different from some old definitions that other scientists still prefer but this isn't revolutionary. That discussion has already been going on since 1980.

Let me close by making one further point. The "data" produced by the ENCODE consortium is intriguing but it would be a big mistake to conclude that everything they say is a proven fact. Skepticism about the relevance of those extra transcripts is quite justified as is skepticism about the frequency of alternative splicing.


Gerstein, M.B., Bruce, C., Rozowsky, J.S., Zheng, D., Du, J., Korbel, J.O., Emanuelsson, O., Zhang, Z.D., Weissman, S. and Snyder, M. (2007) What is a gene, post-ENCODE? History and updated definition. Genome Res. 17:669-681.

The ENCODE Project Consortium (2007) Nature 447:799-816. [PDF]

[Hat Tip: Michael White at Adaptive Complexity]

Monday, June 18, 2007

Gene Genie #9

 
The latest issue of the carnival Gene Genie has just been posted on DNAdirect talk.

Skepticism About "Out-of-Africa"

 
Alan R. Templeton has long been a critic of those who would over-interpret the genetic data on human origins. He's not alone. There are a surprisingly large number of biologists who refuse to jump on the "Out-of-Africa" bandwagon. This group does not get the same amount of publicity as the advocates of a recent (<100,000 years) wave of migration out of Africa. I think it's because skepticism of a new theory is seen as sour grapes. That, plus the fact that it's hard to publish criticisms of work that's already in the scientific literature.

An upcoming issue of the journal Evolution will contain a review by Templeton on human origins and the Out-of-Africa theory. Right now it's only available online [GENETICS AND RECENT HUMAN EVOLUTION]. Here's the abstract,
Starting with "mitochondrial Eve" in 1987, genetics has played an increasingly important role in studies of the last two million years of human evolution. It initially appeared that genetic data resolved the basic models of recent human evolution in favor of the "out-of-Africa replacement" hypothesis in which anatomically modern humans evolved in Africa about 150,000 years ago, started to spread throughout the world about 100,000 years ago, and subsequently drove to complete genetic extinction (replacement) all other human populations in Eurasia. Unfortunately, many of the genetic studies on recent human evolution have suffered from scientific flaws, including misrepresenting the models of recent human evolution, focusing upon hypothesis compatibility rather than hypothesis testing, committing the ecological fallacy, and failing to consider a broader array of alternative hypotheses. Once these flaws are corrected, there is actually little genetic support for the out-of-Africa replacement hypothesis. Indeed, when genetic data are used in a hypothesis-testing framework, the out-of-Africa replacement hypothesis is strongly rejected. The model of recent human evolution that emerges from a statistical hypothesis-testing framework does not correspond to any of the traditional models of human evolution, but it is compatible with fossil and archaeological data. These studies also reveal that any one gene or DNA region captures only a small part of human evolutionary history, so multilocus studies are essential. As more and more loci became available, genetics will undoubtedly offer additional insights and resolutions of human evolution.

[Hat Tip: Gene Expression]

Skepticism About Evo-Devo

 
The May issue of Evolution contains an article by Hoekstra and Coyne on Evolutionary-Developmental Biology or Evo-Devo. There are many evolutionary biologists who have serious doubts about the claims of evo-devo but these doubts don't often make it into the scientific literature because it's very hard to publish critiques. The Hoekstra and Coyne (2007) article is a welcome contribution to the debate.

Here's the abstract,
An important tenet of evolutionary developmental biology (“evo devo”) is that adaptive mutations affecting morphology are more likely to occur in the cis-regulatory regions than in the protein-coding regions of genes. This argument rests on two claims: (1) the modular nature of cis-regulatory elements largely frees them from deleterious pleiotropic effects, and (2) a growing body of empirical evidence appears to support the predominant role of gene regulatory change in adaptation, especially morphological adaptation. Here we discuss and critique these assertions. We first show that there is no theoretical or empirical basis for the evo devo contention that adaptations involving morphology evolve by genetic mechanisms different from those involving physiology and other traits. In addition, some forms of protein evolution can avoid the negative consequences of pleiotropy, most notably via gene duplication. In light of evo devo claims, we then examine the substantial data on the genetic basis of adaptation from both genome-wide surveys and single-locus studies. Genomic studies lend little support to the cis-regulatory theory: many of these have detected adaptation in protein-coding regions, including transcription factors, whereas few have examined regulatory regions. Turning to single-locus studies, we note that the most widely cited examples of adaptive cis-regulatory mutations focus on trait loss rather than gain, and none have yet pinpointed an evolved regulatory site. In contrast, there are many studies that have both identified structural mutations and functionally verified their contribution to adaptation and speciation. Neither the theoretical arguments nor the data from nature, then, support the claim for a predominance of cis-regulatory mutations in evolution. Although this claim may be true, it is at best premature. Adaptation and speciation probably proceed through a combination of cis-regulatory and structural mutations, with a substantial contribution of the latter.

Hoekstra, Hopi, E. and Coyne, Jerry, A. (2007) THE LOCUS OF EVOLUTION: EVO DEVO AND THE GENETICS OF ADAPTATION. Evolution 65:995–1016.

Monday's Molecule #31

 
Today's molecule is complicated but it makes a lot of sense if you know your basic biochemistry. We don't need a long complicated name this time. It's sufficient to simply describe what you're looking at and why it's significant. You have to identify the key residue to get credit for the answer.

As usual, there's a connection between Monday's molecule and this Wednesday's Nobel Laureate(s). This one is an obvious direct connection. Once you have identified the molecule you should be able to name the Nobel Laureate(s).

The reward (free lunch) goes to the person who correctly identifies the molecule and the reaction and the Nobel Laureate(s). Previous free lunch winners are ineligible for one month from the time they first collected the prize. There are no ineligible candidates for this Wednesday's reward since recent winners (including last week's winner, "Kyo") have declined the prize on the grounds that they live in another country and can't make it for lunch on Thursday (a feeble excuse, in my opinion, haven't you heard of airplanes?).

Comments will be blocked for 24 hours. Comments are now open.

UPDATE: The molecule is the lariat structure of the RNA splicing intermediate. The key residue is the adenylate residue that's joined through its 2′ hydroxyl group to the 5′ end of the intron [see RNA Splicing: Introns and Exons]. The Nobel Laureates are Rich Roberts and Phil Sharp. (See the comments for an interesting anecdote concerning the discovery of this molecule.)

Saturday, June 16, 2007

Cellular Respiration Ninja Enzymes

 
Here's a student video presentation of glycolysis and respiration. It's much better than most [e.g., An Example of High School Biochemistry]. However, there are two errors in the video. The first one is fairly serious. The second one is less serious but it's something we cover in my class and it helps illustrate a fundamental concept about how certain reactions work. The second error was very common in most biochemistry textbooks in the past but it's been eliminated from the majority of 21st century textbooks. Can you spot both errors?



[Hat Tip: Greg Laden]
[Hint: What are the products produced by glycolysis and by the Krebs cycle?]

Friday, June 15, 2007

Penicillin Resistance in Bacteria: After 1960

 
The widespread appearance of penicillin-resistant bacteria by 1960 prompted the introduction of new drugs that could not be degraded by newly evolved β-lactamases [see Penicillin Resistance in Bacteria: Before 1960].

The most important of these new drugs are the cephalosporins, modified β-lactams with bulky side chains at two different positions. These drugs still inhibit the transpeptidases and prevent cell wall formation but because of the bulky side chains they cannot be hydrolyzed by β-lactamases. Thus, they are effective against most of the penicillin-resistant strains that arose before 1960.

Other drugs, such as methicillin, were modified penicillins. They also had modified side chains that prevented degradation by the β-lactamases.

It wasn't long before cephalosporin- and methicillin-resistant strains began to appear in hospitals. As a general rule, these strains were not completely resistant to high doses of the new class of drugs but as time went on the resistant strains became more and more immune to the drugs.

The new version of drug resistance also involves the transpeptidase target but instead of developing into β-lactamases they evolve into enzymes that can no longer bind the cephalosporins. Usually the development of resistance takes place in several stages.

There are many different transpeptidases in most species of bacteria. The are usually referred to as penicillin-binding proteins or PBP's. Often the first sign of non-lactamase drug resistance is a mutant version of one PDP (e.g., PDP1a) and subsequent development of greater resistance requires the evolution of other PDB's that don't bind the drug. In the most resistant strains there will be one particular PDB (e.g., PDB2a) that is still active at high drug concentrations while the other transpeptidases will be inhibited.

Resistant enzymes have multiple mutations, which explains the slow, stepwise acquisition of drug resistance. An example is shown in the figure. This is PDP1a from Streptococcus pneumoniae (Contreras, et al. 2006) and the mutant amino acids are displayed as gold spheres. Most of the mutations do not affect the binding of the drug but those surrounding the entry to the active site are crucial. The necessary amino acid substitutions are numbered in the figure. You can see that they line the groove where the cephalosporin drug (purple) is bound. The effect of the mutations is to prevent the bulky β-lactam from inhibiting the enzyme. This is a very different form of drug resistance than the evolution of degradation enzymes that characterized the first stage of penicillin resistant bacteria.


Chambers, H.F. (2003) Solving staphylococcal resistance to beta-lactams. Trends Microbiol. 11:145-148.

Contreras-Martel, C., Job, V., Di Guilmi, A.M., Vernet, T., Dideberg, O. and Dessen, A. (2006) Crystal structure of penicillin-binding protein 1a (PBP1a) reveals a mutational hotspot implicated in beta-lactam resistance in Streptococcus pneumoniae. J. Mol. Biol. 355:684-696.

Livermore, D.M. (2000) Antibiotic resistance in staphylococci. Int. J. Antimicrob. Agents 16:s3-s10.

Penicillin Resistance in Bacteria: Before 1960

 
The Nobel Prize for the discovery and analysis of penicillin was awarded in 1945 [Nobel Laureates: Sir Alexander Fleming, Ernst Boris Chain, Sir Howard Walter Florey]. It was about this time that penicillin became widely available in Europe and North America.

By 1946 6% of Staphylococcus aureus strains were resistant to penicillin. Resistance in other species of bacteria was also detected in the 1940's. By 1960 up to 60% of Staphylococcus aureus strains were resistant with similar levels of resistance reported in other clinically relevant strains causing a wide variety of diseases (Livermore, 2000).

Penicillins are a class of antibiotics with a core structure called a β-lactam. The different types of penicillin have different R groups on one end of the core structure. A typical examples of a penicillin is penicillin G [Monday's Molecule #30]. Others common derivatives are ampicillin and amoxicillin.

The original resistance to this entire class of drugs was caused mostly by the evolution of bacterial enzymes that could degrade them before they could block cell wall synthesis. (Recall that bacteria have cell walls and penicillin blocks cell wall synthesis [How Penicillin Works to Kill Bacteria].)
It seems strange that the evolution of penicillin resistance would require a totally new enzyme for degrading the drug. Where did this enzyme come from? And how did it arise so quickly in so many different species?

The degrading enzyme is called penicillinase, β-lactamase, or oxacillinase. They all refer to the same class of enzyme that binds penicillins and then cleaves the β-lactam unit releasing fragments that are inactive. The enzymes are related to the cell wall transpeptidase that is the target of the drug. The inhibition of the transpeptidase is effective because penicillin resembles the natural substrate of the reaction: the dipeptide, D-alanine-D-alanine.

In the normal reaction, D-Ala-D-Ala binds to the enzyme and the peptide bond is cleaved causing release one of the D-Ala residues. The other one, which is part of the cell wall peptidoglycan, remains bound to the enzyme. In the second part of the reaction, the peptidoglycan product is transferred from the enzyme to a cell wall crosslinking molecule. This frees the enzyme for further reactions (see How Penicillin Works to Kill Bacteria for more information).

Penicillin binds to the peptidase as well and the β-lactam bond is cleaved resulting in the covalent attachment of the drug to the enzyme. However, unlike the normal substrate, the drug moiety cannot be released from the transpeptidase so the enzyme is permanently inactivated. This leads to disruption of cell wall synthesis and death.

Resistant strains have acquired mutations in the transpeptidase gene that allow the release of the cleaved drug. Thus, the mutant enzyme acts like a β-lactamase by binding penicillins, cleaving them, and releasing the products. Although the β-lactamases evolved from the transpeptidase target enzymes, the sequence similarity between them is often quite low in any given species. This is one of the cases where structural similarity reveals the common ancestry [see the SCOP Family beta-Lactamase/D-ala carboxypeptidase]. It's clear that several different β-lactamases have evolved independently but, in many cases, a particular species of bacteria seems to have licked picked up a β-lactamase gene by horizontal transfer from another species. The transfer can be mediated by bacteriophage or plasmids.


Livermore, D.M. (2000) Antibiotic resistance in staphylococci. Int. J. Antimicrob. Agents 16:s3-s10.

Thursday, June 14, 2007

Catherine Shaffer Responds to My Comments About Her WIRED Article

 
Over on the WIRED website there's a discussion about the article on junk DNA [One Scientist's Junk Is a Creationist's Treasure]. In the comments section, the author Catherine Shaffer responds to my recent posting about her qualifications [see WIRED on Junk DNA]. She says,
You might be interested to learn that I contacted Larry Moran while working on this article and after reading the archives of his blog. I wanted to ask him to expand upon his assertion that junk DNA disproves intelligent design. His response was fairly brief, did not provide any references, and did not invite further discussion. It's interesting that he's now willing to write a thousand words or so about how wrong I am publicly, but was not able to engage this subject privately with me.
Catherine Shaffer sent me a brief email message where she mentioned that she had read my article on Junk DNA Disproves Intelligent Design Creationism. She wanted to know more about this argument and she wanted references to those scientists who were making this argument. Ms. Shaffer mentioned that she was working on an article about intelligent design creationism and junk DNA.

I responded by saying that the presence of junk DNA was expected according to evolution and that it was not consistent with intelligent design. I also said that, "The presence of large amounts of junk DNA in our genome is a well established fact in spite of anything you might have heard in the popular press, which includes press releases." She did not follow up on my response.
His blog post is inaccurate in a couple of ways. First, I did not make the claim, and was very careful to avoid doing so, that “most” DNA is not junk. No one knows how much is functional and how much is not, and none of my sources would even venture to speculate upon this, not even to the extent of “some” or “most.”
Her article says, "Since the early '70s, many scientists have believed that a large amount of many organisms' DNA is useless junk. But recently, genome researchers are finding that these "noncoding" genome regions are responsible for important biological functions." Technically she did not say that most DNA is not junk. She just strongly implied it.

I find it difficult to believe that Ryan Gregory would not venture to speculate on the amount of junk DNA but I'll let him address the validity of Ms. Shaffer's statement.
Moran also mistakenly attributed a statement to Steven Meyer that Meyer did not make.
I can see why someone might have "misunderstood" my reference to what Myer said so I've edited my posting to make it clear.
Judmarc and RickRadditz—Here is a link to the full text of the genome biology article on the opossum genome: Regulatory conservation of protein coding and microRNA genes in vertebrates: lessons from the opossum genome. We didn't have space to cover this in detail, but in essence what the researchers found was that upstream intergenic regions were more highly conserved in the possum compared to coding regions, but also represented a greater area of difference between possums and humans.
This appears to be a reference to the paper she was discussing in her article. It wasn't at all clear to me that this was the article she was thinking about in the first few paragraphs of her WIRED article.

Interested readers might want to read the comment by "Andrea" over on the WIRED site. She He doesn't pull any punches in demonstrating that Catherine Shaffer failed to understand what the scientific paper was saying. Why am I not surprised? (Recall that this is a science writer who prides herself on being accurate.)
So, yes, this does run counter to the received wisdom, which makes it fascinating. You are right that the discussion of junk vs. nonjunk and conserved vs. nonconserved is much more nuanced, and we really couldn't do it justice in this space. Here is another reference you might enjoy that begins to deconstruct even our idea of what conservation means: “Conservation of RET regulatory function from human to zebrafish without sequence similarity.” Science. 2006 Apr 14;312(5771):276-9. Epub 2006 Mar 23. Revjim—If you have found typographical errors in the copy, please do point them out to us. The advantage of online publication is that we do get a chance to correct these after publication.
Sounds to me like Catherine Shaffer is grasping at straws (or strawmen).
For Katharos and others—I interviewed five scientists for this article. Dr. Francis Collins, Dr. Michael Behe, Dr. Steve Meyers, Dr. T. Ryan Gregory, and Dr. Gill Bejerano. Each one is a gentleman and a credentialed expert either in biology or genetics. I am grateful to all of them for their time and kindness.
I think we all know just how "credentialed" Stephen Meyer is. He has a Ph.D. in the history and philosophy of science. Most of us are familiar with the main areas of expertise of Michael Behe and none of them appear to be science.

Wednesday, June 13, 2007

WIRED on Junk DNA

Junk DNA is the DNA in your genome that has no function. Much of it accumulates mutations in a pattern that's consistent with random genetic drift implying strongly that the sequences in junk DNA are unimportant. In fact, the high frequency of sequence change (mutation plus fixation) is one of the most powerful bits of evidence for lack of function.

Catherine Shaffer is a science writer who describes herself like this on her website,
I am a writer specializing in biotechnology, genetics, genomics, and other molecular, biological sciences. I have experience with news and features. My strengths include a meticulous attention to detail, an absolutely fanatical devotion to scientific accuracy, and enthusiasm. Readers appreciate my clean, uncluttered prose; my crisp, novelistic style; and (sometimes) my zany sense of humor. I am a writer who always meets deadlines and is organized and dependable.

I studied biochemistry at the graduate level at the University of Michigan, and worked in the pharmaceutical industry for several years. I am especially knowledgeable about genomics, proteomics, biotechnology, drug discovery, and chromatographic separations.
She has written an article for WIRED on junk DNA [One Scientist's Junk Is a Creationist's Treasure]. Here's how the article begins,

Without your "junk DNA" you might be reading this article while hanging upside down by your tail.

That's one of the key findings of the opossum genome-sequencing project, and a surprising group is embracing the results: intelligent-design advocates. Since the early '70s, many scientists have believed that a large amount of many organisms' DNA is useless junk. But recently, genome researchers are finding that these "noncoding" genome regions are responsible for important biological functions.

The opossum data revealed that more than 95 percent of the evolutionary genetic changes in humans since the split with a common human-possum ancestor occurred in the "junk" regions of the genome. Creationists say it's also evidence that God created all life, because God does not create junk. Nothing in creation, they say, was left to chance.

"It is a confirmation of a natural empirical prediction or expectation of the theory of intelligent design, and it disconfirms the neo-Darwinian hypothesis," said Stephen Meyer, director of the Center for Science and Culture at the Discovery Institute in Seattle.

Advocates like Meyer are increasingly latching onto scientific evidence to support the theory of intelligent design, a modern arm of creationism that claims life is not the result of natural selection but of an intelligent creator. Most scientists believe that intelligent design is not science. But Meyer says the opossum data supports intelligent design's prediction that junk DNA sequences aren't random, but important genetic material. It's an argument Meyer makes in his yet-to-be-published manuscript, The DNA Enigma.
Hmmmm ... This is so confused that it's difficult to know where to begin. First, the connection between my junk DNA and whether I am an opossum completely escapes me. I don't know of any credible scientist who claims that it's changes in junk DNA that makes us so different from the common ancestor of opossums. (And none who claim that we are descended from opossums.)

Second, the implication that most junk DNA is turning out to have a function is completely false and the confusion about the difference between junk DNA and noncoding DNA is inexcusable from someone who claims to be an expert on genomics [see Noncoding DNA and Junk DNA, The Deflated Ego Problem].

Third, the idea that large amounts of evolution in junk DNA supports Intelligent Design Creationism is crazy. But, in fairness, I don't think Shaffer is making the connection between the sequence variation and Intelligent Design Creationism; instead, she's making the (factually incorrect) connection between the discovery of some functions in noncoding, nonjunk, DNA and Intelligent Design Creationism (IDC). I think Steve Meyer is suggesting that IDC predicts that junk DNA will have a function and that's why he's being quoted here in the article (see above).
Scientists have made several discoveries about what some call the "dark matter of the genome" in recent years, but they say the research holds up the theory of natural selection rather than creationism.
When sequences in noncoding DNA are conserved, this is taken as evidence of negative selection. In that sense, it supports the theory of natural selection. However, most of the sequence comparisons show that junk DNA is not conserved. This does not support the theory of natural selection. It supports Neutral Theory and the mechanism of evolution by random genetic drift.

The article then describes one recent study suggesting that some noncoding DNA is not junk (Lowe et al. 2007). It appears to be the justification for writing the article since it compares short stretches of sequences in the human and opossum genomes. This is not news so I won't bother commenting.
With scientists increasingly believing that so-called junk DNA regulates other genes, among other functions, creationists like Michael Behe, a biochemistry professor at Lehigh University in Pennsylvania and author of the controversial new book on intelligent design, The Edge of Evolution, are more than happy to point out their errors.

"From the very beginning Darwinism thought whatever it didn't understand must be simple, must be nonfunctional," Behe said. "It's only in retrospect that Darwinists try to fit that into their theory."
The concept of junk DNA is not based on ignorance in spite of what the IDiots say. It's based on good scientific evidence and deduction. Of course most IDiots wouldn't recognize scientific evidence even if it bit them on the ...

Is this just a way of getting in another quote from a prominent advocate of Intelligent Design Creationism? Why is Shaffer so interested in the IDiots? This seems to be more than just seeking out controversy since the proper way to do that would be to interview real scientists who can put the work into perspective and comment on it's significance (see below).
Part of the difficulty in studying junk DNA is that it's impossible to prove a negative, i.e., that any particular DNA does not have a function.

That's why T. Ryan Gregory, an assistant professor in biology at the University of Guelph, believes that nonfunctional should be the default assumption. "Function at the organism level is something that requires evidence," he said.
That's how a real scientist speaks [see A word about "junk DNA" and Comments on "Noncoding DNA and Junk DNA"].

This is getting to be a familiar pattern among science writers. Many of them seem to be incapable of sorting out the actual science from the rhetoric. In this case the problem is exacerbated by introducing IDiots as though their opinion had a bearing on the subject. Not only that, the poor science writing stands in sharp contrast to the claim that, "My strengths include a meticulous attention to detail, an absolutely fanatical devotion to scientific accuracy, and enthusiasm."

Lowe, C.B., Bejerano, G. and Haussler, D. (2007) Thousands of human mobile element fragments undergo strong purifying selection near developmental genes. Proc. Natl. Acad. Sci. (USA) 104:8005-8010. [PubMed]

University College London Restores Professor Colquhoun's Website

 
David Colquhoun has a website at University College London where he regularly debunks the claims of "medical" quacks. Recently a herbal medicine practitioner took offense at this debunking and threatened legal action against the university. The university responded by removing the website.

Today the website has been restored [DC's Improbable Science] and University College London has published a press release explaining why [Joint statement by Professor Colquhoun and UCL].

While it's encouraging that the university decided to restore the website, the fact that it buckled to pressure in the first place is disturbing. What's the point of academic freedom if you abandon it whenever you're threatened with a lawsuit?
UCL has a long and outstanding liberal tradition and is committed to encouraging free and frank academic debate. The evidence (or lack thereof) for the claims made for health supplements is a matter of great public interest, and UCL supports all contributions to that debate. The only restriction it places on the use of its facilities is that its staff should use their academic freedom responsibly within the law.

To this end, the Provost and Professor Colquhoun have taken advice from a senior defamation Queen’s Counsel, and we are pleased to announce that Professor Colquhoun’s website – with some modifications effected by him on counsel’s advice - will shortly be restored to UCL’s servers. UCL will not allow staff to use its website for the making of personal attacks on individuals, but continues strongly to support and uphold Professor Colquhoun’s expression of uncompromising opinions as to the claims made for the effectiveness of treatments by the health supplements industry or other similar bodies.
I'm curious about the "minor modifications" and I'm troubled by the prohibition against "the making of personal attacks on individuals." It seems to me that such a prohibition could be used in a way that inhibits academic freedom. For example, would it prohibit a university Professor from criticizing Tony Blair for the war in Iraq? Would it block any negative comments about Prince Charles (pictured at left)? Does it mean that the UCL website is completely devoid of any negative comments about Richard Dawkins?

Perhaps more importantly, does this mean that university Professors cannot point out on their websites the stupidity of administration officials such as UCL President and Provost Malcolm Grant?

Nobel Laureates: Sir Alexander Fleming, Ernst Boris Chain, Sir Howard Walter Florey

 
The Nobel Prize in Physiology or Medicine 1945.

"for the discovery of penicillin and its curative effect in various infectious diseases"


Sir Alexander Fleming (1881-1955), Ernst Boris Chain (1906-1979) and Sir Howard Walter Florey (1898-1968) received the Nobel Prize in Physiology or Medicine for their work on penicillin [see Monday's Molecule #30 and How Penicillin Works to Kill Bacteria]. The Presentation Speech was delivered by by Professor G. Liljestrand of the Royal Caroline Institute (Karolinska Institutet), on December 10, 1945.
Attempts have been made to reach the goal of medical art - the prevention and cure of disease - by many different paths. New and reliable ones have become practicable as our knowledge of the nature of the different diseases has widened. Thus the successful combating of certain disturbances in the activities of the organs of internal secretion, as also of the deficiency diseases, or avitaminoses, has been a direct result of the increase in our knowledge of the nature of these afflictions. When, thanks to the research work of Louis Pasteur and Robert Koch, the nature of the infectious diseases was laid bare, and the connection between them and the invasion of the body by bacteria and other micro-organisms was elucidated, fully a generation ago, this was an enormous advance, both for the prevention and the treatment of this important group of diseases. This was so much the more important as the group included a number of the worst scourges of humanity, which had slain whole peoples, and at times had laid waste wide areas. But now possibilities were revealed which have not yet been by any means fully utilized. In rapid succession, different forms of vaccination were evolved, and subsequently also serum treatment, for the introduction of which the first Nobel Prize for Physiology or Medicine was given 44 years ago today. In these cases advantage was taken of the capacity of the human and animal bodies themselves to produce protective substances in the fight against the invaders, and to do so in great abundance. But it is by no means the higher organisms only that are able to produce such substances. In cooperation with Joubert (1877), Pasteur himself observed that anthrax bacilli cultivated outside the body were destroyed if bacteria from the air were admitted, and with prophetic acumen he realized that it was justifiable to attach great hopes to this observation in the treatment of infectious diseases. Nevertheless more than two decades passed before an attempt was made to profit by the struggle for existence which goes on between different species of micro-organisms. Experiments carried out by Emmerich and Loew (1899) did not give such favourable results, however, that any great interest was aroused, nor did success attend the later efforts of Gratia and Dath and others. It was reserved to this year's Nobel Prize winners to realize Pasteur's idea.

The observation made by Professor Alexander Fleming which led to the discovery of penicillin, is now almost classical. In 1928, in the course of experiments with pyogenic bacteria of the staphylococcus group, he noticed that, around a spot of mould which had chanced to contaminate one of his cultures, the colonies of bacteria had been killed and had dissolved away. Fleming had earlier made a study of different substances which prevent the growth of bacteria and, inter alia, had come upon one in lacrimal fluid and saliva, the so-called lysozyme. As he points out himself, he was therefore always on the look-out for fresh substances which checked bacteria, and he became sufficiently interested in his latest find to make a closer investigation of the phenomenon. The mould was therefore cultivated and subsequently transferred to broth, where it grew on the surface in the form of a felted green mass. When the latter was filtered off a week later, it was found that the broth had such a strongly checking effect on bacteria that even when diluted 500-800 times it completely prevented the growth of staphylococci; consequently an extremely active substance had passed to the broth from the mould. This proved to belong to the Penicillium group or brush moulds, and therefore first the broth, and later the substance itself, was called «penicillin». It was soon realized that most of the species of Penicillium did not form it at all, and a closer scrutiny showed that the species which polluted Fleming's culture was Penicillium notatum. It had been described for the first time by Richard Westling, in the thesis which he defended in the autumn of 1911 at the University of Stockholm for the degree of Doctor of Philosophy - an illustration of the international nature of science, but also of the suddenly increased importance which sometimes accrues to sound work as a result of further developments. Fleming also showed that penicillin was extremely effective against cultures of many different kinds of bacteria, above all against those belonging to the coccus group, among them those that usually give rise to suppuration, pneumonia and cerebral meningitis, but also against certain other types, such as diphtheria, anthrax, and gas gangrene bacteria. But as numerous other species, among them the influenza, coli, typhoid and tuberculosis bacilli, grew even if they were exposed to moderate quantities of penicillin, Fleming was able to work out a method for isolating out from a mixture of bacteria those which were insensitive to penicillin. He found, further, that the white blood corpuscles, which are usually so sensitive, were not affected by penicillin. When injected into mice, too, it was fairly harmless. In this respect penicillin differs decisively from other substances which had been produced earlier from micro-organisms, and which were certainly found to be noxious to bacteria, but at the same time at least equally noxious to the cells of the higher animals. The possibility that penicillin might be used as a remedy was therefore within reach, and Fleming tested its effect on infected wounds, in some cases with moderate success.

Three years after Fleming's discovery, the English biochemists Clutterbuck, Lovell, and Raistrick, endeavoured to obtain penicillin in the pure form, but without success. They established, inter alia, that it was a sensitive substance which easily lost its antibacterial effect during the purifying process, and this was soon confirmed in other quarters.

Penicillin would undoubtedly still have remained a fairly unknown substance, interesting to the bacteriologist but of no great practical importance, if it had not been taken up at the Pathological Institute at the venerable University of Oxford. This time a start was again made from what is usually called basic research. Professor Howard Florey, who devoted his attention to the body's own natural protective powers against infectious diseases, together with his co-workers, had studied the lysozyme referred to above, the nature of which they succeeded in elucidating. Dr. Ernst Boris Chain, a chemist, took part in the final stage of these investigations, and during 1938 the two researchers jointly decided to investigate other antibacterial substances which are formed by micro-organisms, and in that connection they fortunately thought first of penicillin. It was certainly obvious that the preparation of the substance in a pure form must involve great difficulties, but on the other hand its powerful effect against many bacteria gave some promise of success. The work was planned by Chain and Florey, who, however, owing to the vastness of the task, associated with themselves a number of enthusiastic co-workers, among whom mention should be made especially of Abraham, Fletcher, Gardner, Heatley, Jennings, Orr-Ewing, Sanders and Lady Florey. Heatley worked out a convenient method of determining the relative strength of a fluid with a penicillin content, by means of a comparison under standard conditions of its antibacterial effect with that of a penicillin solution prepared at the laboratory. The amount of penicillin found in one cc. of the latter was called an Oxford unit.

In the purifying experiments then made, the mould was cultivated in a special nutritive fluid in vessels, to which air could only gain access after it had been filtered through cotton wool. After about a week the penicillin content reached its highest value, and extraction followed. In this connection advantage was taken of the observation that the free penicillin is an acid which is more easily dissolved in certain organic solvents than in water, while its salts with alkali are more readily dissolved in water. The culture fluid was therefore shaken with acidified ether or amyl acetate. As, however, the penicillin was easily broken up in water solution, the operation was performed at a low temperature. Thus the penicillin could be returned to the water solution after the degree of acidity had been reduced to almost neutral reaction. In this way numerous impurities could be removed, and after the solution had been evaporated at a low temperature it was possible to obtain a stable dry preparation. The strength of this was up to 40-50 units per mg and it prevented the growth of staphylococci in a dilution of at least 1 per 1 million - thus the active substance had been successfully concentrated very considerably. It was therefore quite reasonable that it was thought that almost pure penicillin had been obtained - in a similar manner, in their work with strongly biologically active substances, many earlier researchers had thought that they were near to producing the pure substance. The further experiments, which were made subsequently with the help of the magnificent resources of modern biochemistry proved, however, that such was not the case. In reality the preparation just mentioned contained only a small percentage of penicillin. Now when it has become possible to produce pure penicillin in a crystalline form, it has been found that one mg contains about 1,650 Oxford units. It is also known that penicillin is met with in some different forms, which possibly have somewhat different effects. The chemical composition of penicillin has also been elucidated in recent years, and in this work Chain and Abraham have successfully taken part.

The Oxford school was able to confirm Fleming's observation that penicillin was only slightly toxic, and they found that its effect was not weakened to any extent worth mentioning in the presence of blood or pus. It is readily destroyed in the digestive apparatus, but after injection under the skin or into the muscles, it is quickly absorbed into the body, to be rapidly excreted again by way of the kidneys. If it is to have an effect on sick persons or animals, it should therefore be supplied uninterruptedly or by means of closely repeated injections - some more recent experiments indicate that gradually perhaps it will be possible to overcome the difficulties in connection with taking the preparation by mouth. Experiments on mice infected with large doses of pyogenic or gas gangrene bacteria, which are sensitive to penicillin, proved convincingly that it had a favourable effect. While over 90% of the animals treated with penicillin recovered, all the untreated control animals died.

Experiments on animals play an immense role for modern medicine; indeed it would certainly be catastrophic if we ventured to test remedies on healthy or sick persons, without having first convinced ourselves by experiments on animals that the toxic effect is not too great, and that at the same time there is reason to anticipate a beneficial result. Tests on human beings may, however, involve many disappointments, even if the results of experiments on animals appear to be clear. At first this seemed to be the case with penicillin, in that the preparation gave rise to fever. Fortunately this was only due to an impurity, and with better preparations it has subsequently been possible to avoid this unpleasant effect.

The first experiments in which penicillin was given to sick persons were published in August 1941 and appeared promising, but owing to the insufficient supplies of the drug, the treatment in some cases had to be discontinued prematurely. However, Florey succeeded in arousing the interest of the authorities in the United States in the new substance, and with the cooperation of numerous research workers it was soon possible, by means of intensive work, to obtain materially improved results there and to carry on the preparation in pure form to the crystallization stage just mentioned. Large quantities of penicillin could be made available, and numerous tests were made above all in the field, but to a certain extent also in the treatment of civilians. Many cases were reported of patients who had been considered doomed or had suffered from illness for a long period without improvement, although all the resources of modern medicine had been tried, but in which the penicillin treatment had led to recoveries which not infrequently seemed miraculous. Naturally such testimony from experienced doctors must not be underestimated, but on the other hand we must bear in mind the great difficulties in judging the course of a disease. «Experience is deceptive, judgment difficult», is one of Hippocrates' famous aphorisms. Therefore it is important that a remedy should be tested on a large material and in such a way that comparison can be made with cases which have not been given the remedy but had otherwise received exactly the same treatment. There are now many reports of such investigations. The extraordinarily good effects of penicillin have been established in a number of important infectious illnesses, such as general blood poisoning, cerebral meningitis, gas gangrene, pneumonia, syphilis, gonorrhea and many others. It is of special importance that even sick persons who are not favourably affected by the modern sulfa drugs are not infrequently cured with penicillin. The effect naturally depends on the remedy being given in a suitable manner and in sufficient doses. On the other hand, experience has confirmed what might have been surmised, namely that penicillin is not effective in cases of, e.g. tuberculosis, typhoid fever, poliomyelitis, and a number of other infectious diseases. Consequently penicillin is not a universal remedy, but it is of the highest value for certain diseases. And it appears not improbable that, with the guidance of experience with penicillin, it will be possible to produce new remedies which can compete with or perhaps surpass it in certain respects.

Four years is a short time in which to arrive at definite conclusions as to the value of a remedy. But during these last few years experiences of penicillin have been assembled which, under ordinary conditions, would have required decades. And therefore there is no doubt at the present time that the discovery of penicillin and its curative properties in the case of various infection diseases for which this year's Nobel Prize is awarded, is of the greatest importance for medical science.

Sir Alexander Fleming, Doctor Chain, and Sir Howard Florey. The story of penicillin is well-known throughout the world. It affords a splendid example of different scientific methods cooperating for a great common purpose. Once again it has shown us the fundamental importance of basic research. The starting-point was a purely academic investigation, which led to a so-called accidental observation. This gave the nucleus, around which one of the most efficient remedies ever known could be crystallized. This difficult process was made possible with the aid of modern biochemistry, bacteriology, and clinical research. To overcome the numerous obstacles, all this work demanded not only assistance from many different quarters, but also an unusual amount of scientific enthusiasm, and a firm belief in an idea. In a time when annihilation and destruction through the inventions of man have been greater than ever before in history, the introduction of penicillin is a brilliant demonstration that human genius is just as well able to save life and combat disease.

In the name of the Caroline Institute I extend to you hearty congratulations on one of the most valuable contributions to modern medicine. And now I have the honour of calling on you to accept the Nobel Prize for Physiology or Medicine for the year 1945 from the hands of His Majesty the King.

Tuesday, June 12, 2007

How Penicillin Works to Kill Bacteria

 
Bacterial cell walls are made of peptidoglycan [Bacteria Have Cell Walls]. In order to form a rigid structure, the polysaccharide chains (glycans) are linked together by peptide crosslinks. The first step in the formation of the crosslinks involves attachment of a short five residue peptide to the MurNAc sugar in the polysaccharide. This peptide ends in two D-Alanine (D-Ala) residues.

This peptide is further modified by attaching an additional peptide to the middle of the first one creating a branched structure. Finally the peptide of one of the polysaccharide molecules is attached to another to form the crosslink. The reaction is a transpeptidase reaction.

The mechanism is shown in the figure (Lee et al. 2003). The top structure (1) is a polysaccharideMurNAc-GlcNAc) with the first peptide already bound. Note that it ends in two D-Ala residues. The first step in the transpeptidase reaction involves binding of the enzyme (Enzyme-OH) to the D-Ala-D-Ala end of the chain. A reaction takes place in which one of the D-Alanine residues is released and the enzyme become attached to the end of the peptide (2).

In the second step, an adjacent peptidoglycan (purple) (3) with a branched structure is covalently linked to the first peptidoglycan forming a crosslink between the two polysaccharides.

Almost all bacteria have cell walls and they have transpeptidase enzymes that catalyze this reaction. The activity of this enzyme is inhibited by penicillins or β-lactams. Mondays Molecule #30 was Penicillin G, one of many different types of β-lactam that block cell wall formation and kill bacteria.

The mechanism of inhibition is well known. The β-lactam region of the drug resembles the D-Ala-D-Ala end of the peptide to which the transpeptidase enzyme binds. The structures are shown below.


A typical penicillin is shown at the top of the figure. The business part of the molecule is the β-lactam moiety and the "R" represents various groups that can be bound to create different penicillin drugs. The structure of D-Ala-D-Ala is shown below.

The structures of several different transpeptidases have been solved. The enzymes are usually called penicillin-binding proteins or PBP's. Most bacteria have several related versions of PDB genes but all of the enzymes are inhibited by β-lactams.

The figure shows the structure of penicillin-binding protein 1a (PBP1a) from Streptococcus pneumoniae with the bound drug in gray in the grove in the lower right corner of the enzyme (Contreras-Martel et al. 2006). This form of the enzyme is inactive because the drug binds very tightly to the active site and blocks the reaction. That's how penicillin works.


Contreras-Martel, C., Job, V., Di Guilmi, A.M., Vernet, T., Dideberg, O. and Dessen, A. (2006) Crystal structure of penicillin-binding protein 1a (PBP1a) reveals a mutational hotspot implicated in beta-lactam resistance in Streptococcus pneumoniae. J. Mol. Biol. 355:684-96.


Lee, M., Hesek, D., Suvorov, M., Lee,W., Vakulenko, S. and Mobashery, S. (2003) A mechanism-based inhibitor targeting the DD-transpeptidase activity of bacterial penicillin-binding proteins. J. Am. Chem. Soc. 125:16322-16326.

Bacteria Have Cell Walls

 
Most species of bacteria have a cell wall. The rigid cell wall prevents the bacterial cell from expanding in solutions where the salt concentrations are lower than the salt concentration inside the cell. If it wasn't for the cell wall, bacteria wouldn't be able to live in fresh water or sea water.

Gram positive bacteria have a thick cell wall on the exterior that picks up the purple Gram stain (named after Hans Christian Gram). Gram negative bacteria, such as the E. coli cell shown in the figure, do not stain with the dye because the thinner cell wall lies between the inner and outer membranes.

The cell wall is made up of peptidoglycan, which, as the name implies, is a combination of polysaccharide (glycan) and peptides. The polysaccharide consists of alternating N-acetylglucosamine (GlcNac) and N-acetylmuramic (MurNAc) resides [see Glycoproteins].


During cell wall synthesis a short peptide of five amino acid residues is attached to the polysaccharide. The sequence of this peptide varies slightly from species to species. In some gram negative bacteria the sequence is L-alanine- D-isoglutamate- L-lysine- D-alanine- D-alanine. These short chains are linked to each other by another peptide consisting of five glycine residues. When the cross-links are formed, the terminal D-alanine residue is cleaved off so the final structure has only a single D-alanine at the end. (The significance of this cleavage will become apparent in subsequent postings.)


The completed peptidoglycan cell wall is extremely rigid because of the peptide crosslinks between the polysaccharide chains. In the figure above, the original peptide chain is colored blue and the second pentaglycine peptide is colored red. The right end of the red pentaglycine is covalently attached to the blue D-alanine residue at the bottom of an adjacent polysaccharide chain as shown in the cartoon in the upper right corner of the figure.

Monday, June 11, 2007

Does Politics Influence When Scientific Papers Are Published?

 
I'm told that the American House of Representatives is considering a bill that will allow embryonic stem cell research. Matt Nisbet thinks that the recent publication of three papers on stem cell research [Reprogramming Somatic Cells] may have been timed to correspond with this debate in the US Congress [Understanding the political timing of stem cell studies]. Nisbet quotes from an article by Rick Weiss in the Washington Post [Darn Cells. Dividing Yet Again!]. Here's what Weiss says,
Thursday, June 7. After months of intense lobbying by scientists and patient advocacy groups, the House is ready to vote on legislation that would loosen President Bush's restrictions on the use of human embryos in stem cell research. But that very morning, the lead story in every major newspaper is about research just published in a British journal that shows stem cells can be made from ordinary skin cells.

The work was in mice, but the take-home message that suffuses Capitol Hill is that there is no need to experiment on embryos after all.

If that doesn't sound suspicious, consider this:

Monday, Jan. 8. After months of intense lobbying by scientists and patient advocacy groups, Congress is ready to vote on legislation that would loosen Bush's restrictions on stem cell research. But that very morning, newspapers are touting new research just published in a British journal suggesting that stem cells can be made from easily obtained placenta cells. No need for embryos after all!

Is there a plot afoot?

Lots of lobbyists, members of Congress and even a few scientists are starting to think so.

"It is ironic that every time we vote on this legislation, all of a sudden there is a major scientific discovery that basically says, 'You don't have to do stem cell research,' " Democratic Caucus Chairman Rahm Emanuel (Ill.) sputtered on the House floor on Thursday. "I find it very interesting that every time we bring this bill up there is a new scientific breakthrough," echoed Rep. Diana DeGette (D-Colo.), lead sponsor of the embryo access bill. Her emphasis on the word "interesting" clearly implies something more than mere interest.

"Convenient timing for those who oppose embryonic stem cell research, isn't it?" added University of Pennsylvania bioethicist Arthur Caplan in an online column. (The bill passed easily, but not with a margin large enough to override Bush's promised veto.)
Hmmm ... let's see if we can figure out what's going on here. Apparently there's some vast conspiracy afoot to keep the American ban on embryonic stem cell research in place. The idea is that scientists and the editors of Nature (for example) want to publish key papers about alternatives to embryonic stem cell research just when American politicians are about to vote on a bill to lift the ban.

The conspiracy makes several key assumptions. It assumes that the editors of the British journal Nature knew about the American bill back on May 22nd when they accepted the two papers. That's when the decision to publish on line for June 6th was taken (the two week delay between acceptance and online publication is typical). It assumes, therefore, that the acceptance date was juggled to meet the target date of June 6th—assuming that the editors even knew, or cared, about what was going on in Washington D.C. (USA). Presumably the acceptance date was delayed somewhat in order to fix the timing. One assumes that the group in Japan who published one of the papers had no problem with this delay and nor did the scientists in Boston. The papers were extremely important in a competitive field but, hey, anything can wait for American politics, right?

Harvard risk expert David Roepik and Temple mathematician John Allan Paulos are skeptical about the conspiracy theory with good reason. The whole idea is ludicrous but that doesn't stop Matt Nisbet from suggesting that it's true. Here's what Nisbet says,
Still, something more than just coincidence is likely to be going on here. Roepik and Paulos' arguments innocently assume that publication timing at science journals is random, without systematic bias. But journal editors, just like news organization editors and journalists, are subject to various biases, many of them stemming from the fact that they work within a profit-driven organization that has to keep up a subscriber base and play to their audience.

Peer-review is just one of the many filtering devices that scientific research goes through. Certainly many papers make it through peer-review based on technical grounds, but then editors at the elite journals, faced with limited space and the need to create drama and interest among subscribers and news organizations, apply more subjective criteria based on what they believe to be the "scientific newsworthiness" of the research. In other words, how much interest among the scientific community will these papers generate AND how much news attention?
Still, Nisbet isn't quite as paranoid and confused about the process as Rick Wiess. In the Washington Post article he says,
Then there is the question of motive. The Brits are competing against Americans in the stem cell field and are legally allowed to conduct studies on embryos. Might they be aiming to dominate the field by helping the conservative and religious forces that have so far restricted U.S. scientists' access to embryos?

Or might the journals be trying, as one stem cell expert opined on the condition of anonymity, to leverage their visibility by publishing stem cell articles just as Congress is voting on the topic?
Damn Brits. :-)

In fairness, Weiss includes a disclaimer from the editors of Nature,
"Nature has no hidden agenda in publishing these papers," said the journal's senior press officer, Ruth Francis, in an e-mail. The real goal was to get the papers out before a big stem cell conference in Australia next week, she said.
More significantly, Weiss includes a comment from someone who seems to have hit the nail on the head,
To Ropeik, the Harvard risk expert, the fact that people are imputing anything more than sheer coincidence is "just more proof that inside the Beltway the thinking is so myopic. They see the whole world through their own lens, and are blinded" to common sense.
That sounds about right to me. If you live in Washington you start to think that the whole world revolves around the White House and Congress. It's easy to believe that everything has to be spun framed in order to influence American politicians—even the timing of publication of scientific papers by a prominent British journal.