Enough whine, let's look at the cheese... (see here for an earlier discussion on the non issue of Haldane's dilemma by Ian Musgrave). The paper in question: Nunney, Leonard, The cost of natural selection revisited, Ann. Zool. Fennici. 40:185-194, 2003Nunney's results differ markedly from those of Haldane and ReMine. Considering Nunney's reticence to have his results critiqued and the divergence from other results, his results appear to lack credibility.
In the conclusions Nunney mentions thatIn a constantly changing environment, organisms must continuously adapt or face extinction. J. B. S. Haldane argued that the "cost of natural selection" (also called the cost of substitution) puts an upper limit on the rate of adaptation, and showed that the cost (C) was a decreasing function of the initial frequency of the beneficial alleles. Based on mutation-selection balance and 10% selective mortality, he suggested that the limit to adaptive evolution was about one allelic substitution per 300 generations. I have tested Haldane's results using simulations of a population limited by density-dependent regulation and subject to a constantly changing environment that affects n (= 1–7) independent survival traits, each controlled by a single locus. I investigated the influence of carrying capacity (K), mutation rate (u), number of beneficial mutations per generation (approximated by M = 2Ku) and net reproductive rate (R). Of these, M has the predominant influence. The effect of large changes in R was relatively small. The cost of selection (C) was measured as the shortest number of generations between an allelic substitution at all loci under selection that was consistent with population persistence. The results differed from Haldane's solution. Across a range of conditions, the cost of simultaneous selection at n loci was determined by the linear relationship C = C0(M) + nC1(M), where C0(M) is the intercept and C1(M) is the slope of the linear regression of C on n, for a given M. The intercept defined a positive fixed cost of substitution, that appears to reflect genetic deaths occurring during the stochastic phase when the beneficial alleles are rare. For M > 1/2, the cost of natural selection is substantially less than Haldane's estimate; however, when M < 1/2, the cost (and particularly the fixed cost) increases in an accelerating fashion as M is lowered. This result has important implications for conserved populations, since for u ~ 5 x 10–6 the carrying capacity of the population must be 50000 for M = 1/2. To avoid low M, smaller populations should be linked together into a large metapopulation whenever possible. This large unit would be capable of adapting when the isolated parts could not. It also suggests that if M << 1, small gains in K through increases in habitat can have a very large positive influence on the future survival of the population in a changing environment.
Perhaps it is time for ID creationists to become more familiar with the facts and depend less on hear-say? According to Scordova, Walter's research is funded by the Discovery Institute, showing once again how ID fails to provide much of any relevant scientific research. As for Walter's 'response' enjoy...Haldane (1957) ended his paper by noting "I am quite aware that my conclusions will probably need drastic revision. But I am convinced that the quantitative arguments of the kind here put forward should play a part in all future discussions of evolution" (p. 523). The results presented here suggest that Haldane was correct that some revision of his conclusions is needed, but they also suggest that he was correct that the cost of natural selection is a real phenomenon that needs to be included in more discussions of adaptations in a changing environment. The concept is made even more relevant as we become increasingly aware of the potential for rapid environmental change, and as more and more natural populations become fragmented in many isolated units.
32 Comments
Bob O'H · 26 February 2008
caligula · 26 February 2008
ReMine's not the only lazy one. I have had it running in the background for like a year now that I should implement Nunney's simulation independently. You see, I disagree with ReMine that relevant details are somehow lacking in Nunney's paper, making it impossible to evaluate his conclusions. So far as I can see, Nunney gives pretty much enough detail for an independent implementation for anyone who can code. (I can't open the link to Walter's response, but I can guess that Walter is complaining about "not enough detail".)
SteveF · 26 February 2008
SteveF · 26 February 2008
Joshua Zelinsky · 26 February 2008
I'm going to have to disagree somewhat; when doing any form of modeling work it is always a good idea to provide your code on request. This allows more transparency and makes it easier for people to catch possible mistakes. That's one reason some members of the open access movement have actually argued that journals should mandate open disclosure of code.
SteveF · 26 February 2008
Well, I can see where Nunney is coming from. It's his data and he has the right to do whatever he wants with it. I'd be somewhat leery about giving my work to some crank on the internet. Having said that, I do tend to be inclined to think that researchers should provide code on request (including Nunney).
However, it seems to me that this is a bit of a side issue. Reading his paper, it looks as if his model is described in a decent amount of detail. As I said in my post above, I'm not a programmer, but it strikes me that there is sufficient information for someone sufficiently qualified to make a good attempt at replicating his results (as he was perfectly happy to do with Crow's, admittedly simpler, model). If he can't replicate them, then there are some possible ways explanations leading to ways forward:
1) ReMine is wrong
2) Nunney is wrong
3) Replication requires the original code
There's also another issue here; as far as I can tell, ReMine has not attempted to deal with this paper. He has gotten around this by appealing to lack of data. However, when evaluating a studies conclusions, it is generally typical to take them at face value (e.g. the author is honest). Bearing this in mind, whilst it is not unreasonable to try and replicate the results, ReMine should also be taking the study at face value and dealing with it in these terms.
Beowulff · 26 February 2008
Reed A. Cartwright · 26 February 2008
I highly recommend reading the section about Haldane's Dilemma in WJ Ewens' Mathematical Population Genetics. In it he describes how the dilemma is an artifact caused by applying results from an infinite population model to finite population models.
SteveF · 26 February 2008
Stuart Weinstein · 26 February 2008
I can see both sides of this. As a modeler (of thermal convection) I have experienced such requests myself. My answers tend to depend on who was doing the asking. If it was somebody I knew and they were competent, I'd share what code I thought was appropriate. Otherwise, I'd give them as many details as necessary, but ultimately they would have to code it themselves. I simply couldn't be drawn into the trap of providing support to help them use the code and get it compiled, etc. etc. I don't blame Nunney for not wanting to be besieged by daily phone calls troubleshooting problems for Remine. Furthermore, imagine Remine using it improperly and then Nunney having to spend more time explaining where Remine fouled up. So, no, I don't blame him for ignoring Remine. As Nunney basically put it, he had nothing to gain from giving it to Remine, other than a good case of headache.
I didn't write user friendly hydrodynamics code, I wrote Stuart friendly code. Most science coders code that way; you know what works for them. Hence, its usually better to write your own stuff as you'll spend almost as much time trying to learn how to use someone else's and understand it.
That said, we did publish joint papers where we all simulated solutions to classical problems with varying difficulty, and compared the results. These papers served as benchmarks, and could be used as milestones for those writing their own code. I don't know if such things are appropriate for the population genetics community, but in the geodynamics community they did address this issue, at least with respect to code validation.
Reed A. Cartwright · 26 February 2008
I'm probably going to get half of the details wrong---save me RBH---but:
I remember when Sal (?) go a hold of the Avida code and decided that he found an error in their model---therefore evolution was false---because the program fell apart when he specified a mutation rate above 50%. Besides the fact that Sal didn't realize that such high mutation rates were not biologically relevant, what he discovered was not a problem in the model but an overflow error in the code. Apparently, the Avida programmers did not think that anyone would use such high mutation rates.
caligula · 26 February 2008
Beowulff,
Nunney's idea seems to be that the allelic values for A do form a "sequence". That is, the higher the allelic value, the farther we are, in terms of actual DNA sequence, from the starting point (A=0). I base my interpretation on how allelic values interact with the environmental change in eq. 5: at time T, where A=1 is the optimal value, A=4 would be more harmful for fitness than A=3. So, supposedly the DNA sequence for A=4 differs more from A=1 than does A=3. It also makes general sense, because after allele 1 has become fixed in the population, we are generating new mutations for allele 1, not allele 0, so value 2 is typically more than one point mutation away from value 0. This interpretation would mean, then, that the mutation rate u mentioned in the paper refers to the rate of beneficial mutations, exclusively, and that Nunney assumes only one beneficial change per each allele. (I.e. allele 3 can only experience one beneficial change: 3->4). Of course, being truly "beneficial" depends stricly on time t, i.e. the state of the environment.
I think the cycle of each generation proceeds as follows. First, the female fecundity for the entire population is calculated, based on the current density. Then, random mating occurs for female #1. The mated couple produces offspring into a newly created "population pool" according to female fecundity, each child getting her genotype according to the usual Mendelian rule for diploids (AA:2Aa:aa). Mutations are generated for each newborn child immediately after Mendelian segregation. The male returns to the "stud pool" and we move on to the next female. After completing the reproduction phase, we have a new generation in a separate population pool, and the parent populatuion is destroyed. Next, selection has its way on the new generation (juvenile deaths). Now we have completed one generation cycle.
caligula · 26 February 2008
Joe Felsenstein · 26 February 2008
No, what you really should read is my 1971 paper which really clears all this up. The load is imposed by the deterioration of the environment, and is measured by the amount of reproductive excess which it is necessary to have to prevent extinction in the face of repeated deterioration of the environment. When advantageous mutations occur in the absence of deterioration of the environment, they don't impose any load. If anything they make the population better able to bear loads.
Felsenstein, J. 1971. On the biological significance of the cost of gene substitution. American Naturalist 105: 1-11.
It is a very hard literature to follow, as various authors defined various costs and loads without ever making it clear what those were supposed to do for us. My paper clears this up. (OK, I'm biased).
You may also want to look at the years-long running debates between Remine and I in the Usenet newsgroup sci.bio.evolution
Paul S · 26 February 2008
caligula:
I don't think that Ewens changes Haldane's core argument. Haldane's argument is based on looking at the difference in fitness between the most fit and least fit individuals in the population (with respect to the substituting loci). Haldane argued that you couldn't expect much more than about 10% difference between these. Based on this (and I am glossing over a lot of details and subtleties), he came up with his limit of a substitution every 300 generations. Ewens shows that no individual with a fitness anywhere near this theoretical maximum will ever actually appear in a population. I don't think that the 10% was based on any hard data, but IS based roughly on what we see in real populations. That is, we don't generally see enormous variation in reproductive success in real populations. The number may not really be 10%, but it is in that ballpark. Ewen's argument is that this "obsrved" 10% is not the varation between the theoretical most fit and least fit individuals. Instead, it is the variation in individuals actually existing in the population and their range of genotypic variation is much smaller than the theoretical extremes. Thus, the limits imposed on the speed of selection are much less stringent than Haldane's argument.
SteveG · 26 February 2008
Pardon me for being naive about the requirements of professional science journals regarding what they publish. But it was my impression "from the outside" that serious research articles are in general supposed to provide enough details about the data and the research procedures so that the work and results could conceivably be replicated - or invalidated. I'm a computer programmer by profession, and while I work in business software development (design and coding of databases, and data processing applications), it really seems to me that the details of the computer programs used to model the concepts and generate the results are of criticial importance to the research articles, and the disclosure of the code should be part of the requirements of professional science journals. While I can certainly understand a professional scientist not having his time wasted by yet another creationist crackpot (there are so many - sigh), such a requirement by the professional science journals would make the issue a moot point since the program(s) used would be open viewing to anyone interested in the published research.
Ian Musgrave · 26 February 2008
Ian Musgrave · 26 February 2008
Erasmus, FCD · 26 February 2008
Any chance we can get to Walt Brown some day?
Ftk I know you are waiting for it.
Joe Felsenstein · 26 February 2008
Let me give a very brief summary of different notions of substitutional load, cost of natural selection etc. and what the original debate was about. In the 1960s it was discovered that there was tons of protein variation within species. One of Motoo Kimura's arguments that most of it was neutral was that otherwise there would be too high a substitutional load. Note that this is not a concern that the rate of evolution was too high for evolution to be true -- that is how Remine describes the controversy, but I was there and I remember, and he's utterly wrong about that.
All sorts of people chimed in in the late 1960s to early 1970s with calculations of substitutional load, usually not making it clear what happened if there was too much load. There actually were at least three distinct notions:
1. That the load measured the reduction of the population size. Not widely used because it brings in all sorts of ecological issues.
2. That too high a load meant you would go extinct.
3. That too high a load meant the postulated scenario of what was happening was wrong and something else was going on.
Ewens' and Crow's loads were of type 3. They are very similar to Remine's calculation.
Haldane was unclear but basically seems to have been making a type 2 argument, as I was. Remine always ignores this distinction and just claims all the other calculations are wrong.
People also came up with schemes of gene interaction such as truncation selection that lead to lower loads. After few years everyone concluded that the substitutional load calculation didn't speak powerfully about whether variation was neutral. So shortly after I published my 1971 paper which I thought cleared things up ... the whole discussion died and everyone forgot about it. The mess was left in the literature. Remine complains that it is an unresolved mess and in that he is quite right.
I hope folks will read my paper (alas, the JSTOR link will work only for people whose institutions have a subscription to it).
Joe Felsenstein · 27 February 2008
I should add that the "original debate" I was referring to is the wave of reconsiderations of Haldane's result in the late 1960s and early 1970s. Haldane's paper was the original one, of course.
caligula · 27 February 2008
caligula · 27 February 2008
Stephen Wells · 27 February 2008
It's not even clear that there is such a thing as the "theoretical optimum genome" or "absolute fitness."
caligula · 27 February 2008
pwe · 27 February 2008
Oh, I have avoided ReMine for quite a long time now -- since my encounter with him on Wikipedia.
I am no biologists, and that may be why I don't get it all. I read Nunney's paper back then and Joe Felsenstein's paper -- and even ReMine's own paper -- and Joe Felsenstein's review of ReMine's paper.
But how current is Haldane's definition of evolution: the substitution of a gene (allele) for another at a locus?
And, all ReMine's tapdancing about 1667 substitutions from a simian to Mozart (or something like that), is that even woth debunking? I mean, as far as I have understood 8and that may not be very far), then gene regulation is, where it's at. I mentioned that to ReMine, bnut he only claimed that made things worth for the evolutionist side.
Oh, well, I'll probably pull the plug and go into the woods to hunt little green fairies. May you all have a nice day.
- pwe
pwe · 27 February 2008
Joe Felsenstein · 27 February 2008
Torbjörn Larsson, OM · 27 February 2008
Ian Musgrave · 27 February 2008
Joe Felsenstein · 27 February 2008
OK, I see the quote wasn't supposed to be me. Thanks.
caligula · 27 February 2008
By the way, I find Walter ReMine's comments on subsection "4.11 Truncation selection" rather telling, in his Creation Wiki and ResearchID.org articles.
What ReMine is complaining about truncation selection is essentially beanbag genetics, because his complaint is that the fitness interaction across loci is not epistasic. Well, Haldane was the very spokesman for "beanbag genetics", and Haldane(1957) certainly does not assume epistatic fitness interaction, as ReMine notes himself. Essentially, ReMine seems to agree that truncation selection solves the problem presented in Haldane(1957). Although the solution, too, is based on "beanbag genetics", that never was the core issue raised by ReMine. The issue was supposed to be "Haldane's dilemma", not "beanbag genetics".
ReMine of course claims that epistatic fitness interaction of a suitable kind radically slows down evolution, and that evolution is governed by just this kind of epistasis, but he has not really demonstrated his claim.