Eugene Thacker on 18 Feb 2001 00:09:38 -0000 |
[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]
[Nettime-bold] The Human Genome Race Considered as a High-Speed Data Dump |
---------------------------------------------------------------------------------------------- The Human Genome Race Considered as a High-Speed Data Dump Eugene Thacker I. The Race Celera - a latecomer to the race - got off to an energetic start. With a souped-up high-octane arsenal of turbo-IT gene-machines at its side, it looked as though this would be the quickest race of this type ever sanctioned. The public consortium, however, was not so easily put off. Their old school DNA bi-partisan tag-team strategy served them well, leading off from multiple points in the U.S. and U.K., as well as Japan and Germany. While Celera had the advantage coming into the race, its youth, inexperience, and bold, brash style drew many criticisms. The public consortium, on the other hand, was no stranger to the challenges of extreme big science (or XBS). Celera had, in many ways, fired the starting shot for the race when it aggressively tapped the "enter" key on its sequencing computer in 1998. With robust, lean vehicles from Perkin-Elmer, the data began flushing in almost immediately. Patent, publish, and press enter again - in that order. The public consortium quickly became aware of the sexy hum emanating from Celera's gene machines. After doing some rush-delivery online shopping, it too upgraded its tech, but without dropping its old school appeal. Some of the researchers would actually teach and do research at universities - a sign of by-gone days. But the decider for the race was not Celera's new-kid-on-the-block attitude, nor was it the public consortium's old-boys-will-prevail assurance - it was software. The public consortium relied on the pragmatics of the GigAssembler - 10,000 lines of code written in four weeks by a grad student last summer: string them together & do a quick taste-test. Celera stuck to its avant-garde guns by employing its controversial Whole Genome Shotgun sequencing method (WGS): blow it to pieces & then cluster them together freestyle. Although both groups published their findings simultaneously, it appears that Celera inched through the finish line at the last moment (though frame-by-frame analysis has yet to confirm this - ftp clocks are being checked at the servers of both Nature and Science magazines). Reports from both groups state that this is only the end of the beginning. Celera continues the race wireless - but that is another story. (For more info, see "The Assassination of John Fitzgerald Kennedy Considered as a Downhill Motor Race" by J.G. Ballard, and "The Crucifixion Considered as an Uphill Bicycle Race" by Alfred Jarry.) II. Reports Critique functions the best, perhaps, when it does no work at all. That is, when critique is embodied in the very same locus of its object, a kind of discursive telescoping occurs, in which that which is implicit is made explicit. The race to map the human genome has been such a process. More than any other "big science" project, the mapping of the genome has always partially un-done itself. It has played two roles: that of a promotional campaign supporting a certain tradition of Western science, and that of a highly unstable agenda for the production of a hegemonic view of health, the body, and theories of biological "life." As Mae-Wan Ho states in her recent critique of the genome <http://www.i-sis.org>, a number of critics have long been suggesting what scientists are only now beginning to consider, with the "hard evidence" at hand: that genetic reductionism doesn't work, that "central dogma" theories of any kind are highly limiting, and that "genes" (a still relatively undefined category of biomolecule) may not be the central and single most important component of molecular life. We can look to Evelyn Fox Keller, Richard Lewontin, Susan Oyama, Donna Haraway, Critical Art Ensemble and others for examples of such critiques. The recent reports on the human genome actually perform this - the articles by Celera (in Science) and the International Human Genome Consortium (in Nature) are filled with a rhetoric of surprise at the findings: the number of "genes" is lower than expected, the number of unique genes to humans is fewer than expected, the "coding" portion of the genome is only about 1.5% of the total DNA, and vast regions of the genome are filled with "deserts" of repetitive sequence whose function is unknown. Three perspectives here: What did they find? What do these findings imply? What challenges do these findings put forth? What did they find?: Both the public consortium and Celera came to essentially the same conclusions in their initial analyses of the genome. And both expressed "surprise" at their findings: 1. Number of genes: approx. 30,000-40,000. This is much less than expected, with most estimates before this around 100,000 or more, and many companies (mostly pharmaceutical or tech companies banking on the high number of genes) are still in genomic denial. In addition, this is not much more than bacteria or mice. In cross-species genome comparisons, it was found that only some 300 genes are unique to human beings compared with the mouse genome. 2. Large number of repetitive sequences. Again, a surprise finding, that often large segments of DNA are repeat sequences, either inserted by retroviruses, remnants of fossilized DNA, or sequences which play as yet undetermined functions in gene expression and regulation. Jumping genes, or "transposons," are beginning to be taken seriously as a primary contingency in the dynamic aspects of the genome. 3. Modularity of genes in alternative splicing. Contrary to the conservative one gene-one protein hypothesis, alternative splicing of RNA from DNA (the "spliceosome") means that the combinatorial possibilities of gene function increases greatly. In humans at least, this means that one "gene" can have many different variations, and can perform many different functions. There is no reason to think that it would be different for proteins, or any other biomolecule in the cell. What do these findings imply?: Reading the primary findings of the genome analyses through the social and philosophical implications of "surprise" discoveries, reveals several implicit critiques at work in genomics specifically, and biotech generally: 1. The low number of genes is only surprising if it is assumed that the quantity of genes is directly related to the complexity of the organism or species. More genes would thus mean more complex organisms. In a characteristically consumerist vein, the assumption here was that "more is better." 2. The finding of repetitive sequences in the genome is only surprising if it is assumed that each "gene" or DNA sequence has a unique function, thereby making the genome original and unique. Repetition here is taken to be synonymous with redundancy and a lack of novelty, while the icons of American subjectivity - individuality, originality, novelty - are assumed to reside in the very bodies of subjects, down to their biomolecules. As bioinformatics is just beginning to demonstrate, repetition can be a highly complex system for generating complexity (be it ones and zeros or pairing of A-T and C-G). 3. The evidence of alternative splicing - modularity - in the production of RNA is only surprising if it is assumed that one gene makes only one protein. Despite the continual suggestion from both scientists and cultural theorists, the old "central dogma" (DNA makes RNA makes protein makes us) is still very much alive. A molecule with more than one primary function, at one time, is beyond the scope of conventional thinking in genetics and genomics. What challenges do these findings put forth?: The initial genome analyses make it all the more clear that different models of theorising the genome and its significance are desperately needed, not only by scientists, but by computer programmers, activists, artists, cultural theorists, journalists, and policy makers. 1. Informatic ontogeny: As Susan Oyama long ago theorised (_The Ontogeny of Information_), the debates between organism and environment are filled with assumptions about the pre-existence of both, often at the expense of the productive, generative "interactionisms" that occur within dynamic organisms and dynamic environments. Genes do not simply already exist as fixed entities, and then interact with a fluctuating environment. Information is actually generated within the dynamisms of the organism-environment system - genetic information thus has an ontogeny, a self-making property that emerges out of different interactionisms. As Oyama states, "developmental information itself, in other words, has a developmental history." Genetic information, then, "neither preexists its operations nor arises from random order"; it is, to paraphrase Gregory Bateson, a difference that makes a difference, and therefore "its meaning is dependent on its actual functioning." Several questions here: What if genetic "information" does not simply pre-exist in each cell, waiting still and patiently to be mapped and annotated? What if a "gene" describes not a specific string of DNA that codes for the production of a protein, but a particular network of operative biomolecules embedded within various contingencies, that have a certain biochemical effect? There is already work being done in this direction, but it remains marginal with respect to the genome. 2. Combinatorial complexity: The emergence of complexity out of simplicity is the result of combinatorial variations. John von Neumann's _Theory of the Automaton_ stated this very early on: complexity of spatial patterning and temporal variation can arise out of the simplest elements, be they binary bits or DNA molecules. A combinatorial approaches relies on mathematics, stochastic approaches, etc., and therefore also relies on computers: bioinformatics. This approach thus demands that bioinformatics be more than merely a tool for reiterating the central dogma (but with fancy molecular models); it demands that a kind of computer science develop which takes into account the dynamic, "wet" complexity of the molecular body. The intersection of bio-science and computer science is intriguing; but bioinformatics has a long way to go before it starts functioning as more than a tool for big science. The interstitial fields of a-life, bio-computing, and nanotech may provide a window to seeing past bioinformatics as another lab tool. 3. Systems approaches to molecular biology: The smaller number of genes and their repetitions point to the obvious fact that a complex organism such as the human somehow operates through apparent simplicity. Gradually we're seeing the term "complex" and "complexity" seep into the mainstream discourse surrounding genetics (both Venter and the New York Times referred to the genome in this way). The discovery of DNA itself - two pairs of two molecules - reiterates this theme of the simplicity of beginning conditions, and their combinatorial complexity (see below). Clearly what is needed is not a narrow, linear, production-oriented approach (genes makes proteins make us), but a wider, more flexible, more differentiated, systems approach to molecular biology. This would, among other things, take into account the extra-genomic components involved in cellular processes. Writing about systems theory during the 1940s (_General System Theory_), Ludwig von Bertalanffy states that "the living cell and organism is not a static pattern or machine-like structure...It is a continuous process...The organism's being an 'open system' is now acknowledged as one of the most fundamental criteria of living systems..." The initial reports from the human genome once again demonstrate that what is needed - theoretically, technically, and politically - is something very very simple: to think in more complex ways. ---------------------------------------------------------------------------------------------- ����������������������������������� Eugene Thacker e: [email protected] w: http://gsa.rutgers.edu/maldoror/index.html Pgrm. in Comparative Literature, Rutgers Univ. ����������������������������������� CURRENT: "Molecules That Matter: Nanomedicine & the Advent of Programmable Matter" @ nettime: <http://www.nettime.org>. "Regenerative Medicine: We Can Grow It For You Wholesale" @ Machine Times (DEAF_00, V2 book, http://www.v2.nl/deaf) "The Post-Genomic Era Has Already Happened" @ Biopolicy Journal <http://bioline.bdt.org.br/py> "SF, Technoscience, Net.art: The Politics of Extrapolation" @ Art Journal 59:3 <http://www.collegeart.org/caa/publications/AJ/artjournal.html> "Point-and-Click Biology: Why Programming is the Future of Biotech" @ MUTE (Issue 17 - archives at http://www.metamute.com) ����������������������������������� also: FAKESHOP <http://www.fakeshop.com> ����������������������������������� _______________________________________________ Nettime-bold mailing list [email protected] http://www.nettime.org/cgi-bin/mailman/listinfo/nettime-bold