| . on 5 Mar 2001 22:23:54 -0000 | 
[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]
| [Nettime-bold] Might a laboratory experiment now being planned destroy planet Earth | 
| 
 Might a laboratory experiment 
destroy planet Earth?   F. Calogero Dipartimento di Fisica, 
Università di Roma "La Sapienza" Istituto Nazionale di Fisica 
Nucleare, Sezione di Roma 
 Synopsis   Recently some concerns have 
been raised about the possibility that a high-energy ion-ion colliding beam 
experiment which just began at Brookhaven National Laboratory in the United 
States, and a similar one that is planned to begin some years hence at CERN in 
Geneva, might have cataclysmic consequences, hypothetically amounting to the 
disappearance of planet Earth. The probability that this happen is of course 
tiny. In the first part of this paper a popularised review is presented of the 
motivations for such concerns and of the extent they have been investigated, and 
in the context of such a popularised treatment the appeasing conclusions of 
these investigations are scrutinised. In the second part, in the light of this 
example, a terse analysis is provided of some scientific, ethical, political and 
sociological issues raised by the problematique associated with human endeavours 
which might entail a tiny probability of an utterly catastrophic outcome -- with 
special emphasis on related responsibilities of the scientific/technological 
community.     “First Commandment for 
experimental physicists: Thou shalt put error bars on all your 
observations. First Commandment for 
theoretical physicists: Thou shall get the sign 
right.” (Private Communication by 
Professor Sebastian Pease)     0.         
Recently concerns have been raised about the possibility that a major 
catastrophe -- possibly amounting to complete destruction of planet Earth -- 
result from an experiment which just began (summer 2000; for up to date 
information see http://www.rhichome.bnl.gov/AP/Status) at the Brookhaven 
National Laboratory (BNL) in the United States, and/or from a similar experiment 
under preparation at the European Centre for Nuclear Research (CERN) in Geneva, 
scheduled to begin there some years hence. Purpose and scope of this paper is to 
explain what these concerns are about, to report on various estimates [1-4] -- 
all of them appeasing -- of the risk, and to proffer some general 
scientific-political-ethical-sociological  
considerations related to this kind of issues; considerations whose 
interest should be however qualified by the avowed lack of specific 
scientific-political-ethical-sociological expertise of this 
author.   1.         
More detailed and professional analyses of the experiments in question 
and of their hypothetical danger are given in [3] and [2] (see also [4]). The 
presentation given below is largely in the nature of a popularised summary of 
these papers; it also points out certain aspects which to this author appear as 
possible shortcomings of these treatments. Some information on the publication 
history of these papers [1-4] is also provided below, as it might have some 
(minor) relevance with respect to the more philosophical considerations reported 
in the second half of this paper. This author profited from many discussions by 
voice and by e-mail with several colleagues, who will however remained unnamed 
to avoid any hint that they agree with what is written herein, for which the 
author feels he should take exclusive personal responsibility. An explicit thank 
is however due to Adrian Kent for having called my attention (via Pugwash) to 
this problematique.   2.         
The experiments in question realise collisions of heavy ions, produced by 
two beams of such particles colliding head-on against each other. The particles 
of the two beams are accelerated by huge (and very expensive -- many hundred 
million dollars) apparatuses ("colliders"). In the case of the Relativistic 
Heavy Ion Collider (RHIC) experiment at BNL, the particles in the two beams are 
gold ions (indeed, one can think of gold nuclei -- since the ionisation is 
almost complete), which get accelerated to a (planned) energy of 20 TeV 
(1 Tev=10^3 Gev=10^6 Mev=10^12 ev), amounting to an approximate 
energy of 100 Gev per nucleon (each gold nucleus is composed of 79 
protons and 118 neutrons, altogether 197 nucleons). The energy 
in the centre-of-mass system when two particles collide adds up to 40 
Tev. In the CERN experiment (A Large Ion Collider Experiment -- 
ALICE), the energies per ion are expected to be 30 times larger, and the 
mass of each ion is also larger, but only marginally so: lead rather than gold, 
mass 207 (82 protons and 125 neutrons) rather than 
197. The total energies in the centre-of-mass system (which in this case 
of colliding beams with equal energies coincides with the laboratory system; the 
situation is of course quite different in the case of one beam hitting a fixed 
target) are the largest ever realised in high-energy physics experiments; 
although of course one should actually talk, in this context, of energy 
density (the energies in macroscopic events, such as the collision of two 
billiard balls, or two cars, are much larger). It should also be noted that 
experiments with higher centre-of-mass energies per nucleon have been 
realised in machines which accelerate protons rather than 
ions.             
In any case from some points of view these high-energy experiments 
explore a region of microphysics never before attained by man-made experiments 
-- although cosmic ray events with much higher centre-of-mass energies occur all 
the time and many of them have been experimentally observed. It is therefore 
justified to wonder whether the exploration of such (relatively) new 
phenomenology might give rise to surprises. Indeed it is precisely the hope to 
find something interesting that motivates these experiments and justifies their 
huge costs. Specifically, it is expected/hoped that in these experiments, 
immediately after the head-on collision, the nuclear matter that constituted the 
two colliding nuclei become effectively a "plasma" made up of  quarks and gluons -- 
namely of the objects which elementary particle physicists consider nowadays as 
the ultimate constituents of matter. Such a plasma would to some extent mimic 
the situation that existed in the early Universe (if the origin of our Universe 
is correctly described by Big Bang scenarios); information on its structure is 
obviously of great scientific interest.    3.         
But could the "surprises" entail dangerous 
consequences?             
While the experimental exploration of any new phenomenology always 
entails an unknown element, it is possible for elementary particle physicists to 
envisage which are the potentially dangerous developments, and to analyse the 
likelihood that they emerge. Indeed just such a task was recently assigned 
(perhaps a bit late in the game!) by the Director of BNL to four American 
physicists, who produced a Report [1], and who later issued another 
(significantly modified) version of it [3]; and a somewhat analogous analysis 
has been performed in parallel by three physicists associated with CERN [2]. 
These findings have also been commented upon by two other, quite distinguished, 
physicists, in a note published by Nature [4]. All these experts agree 
that the probability of any catastrophic outcome of the RHIC experiment is so 
small to justify proceeding with it without delay. Their arguments also apply, 
although to some respect with less cogency (but from some others with more 
cogency) to the ALICE experiment at CERN. A committee to analyse the matter has 
also been recently appointed by the Director General of CERN, Luciano 
Maiani.             
It is however desirable that a larger community of experts than those 
specifically tasked to do so involve themselves in this problematique, and look 
critically at these analyses; and it is also desirable that, to the maximum 
extent possible, a larger community of responsible scientists, and of citizens, 
become informed. Moreover this problematique also entails considerations of a 
political and ethical character, which are to some extent independent of the 
technical details, but can only profit from a better understanding of them. It 
is for this reasons that this attempt is made to popularise (and in that 
context, to some extent, to also criticise) the findings of the experts who have 
looked at these issues [1-4], undoubtedly with greater scientific competence 
than I can muster, yet giving me occasionally the impression to be biased 
towards allaying fears "beyond reasonable doubt"; a posture quite understandable 
under the circumstances -- since undue fears might block experiments which 
promote scientific progress -- but which nevertheless constitutes, almost by 
definition, an odd stand for scientists (doubt being the seed of science; 
although, admittedly, only reasonable doubt!).    4.         
Three hypothetical dangers which might emerge from these experiments have 
been considered: (i) formation of a "black hole" or some other "gravitational 
singularity" in which surrounding matter might then "fall"; (ii) transition to 
another hypothetical "vacuum state", different, and lower in energy, than the 
vacuum state of our world (which would therefore be metastable); (iii) formation 
of a stable aggregate of "strange" matter, which might initiate a transition of 
all surrounding matter to this new kind of matter, with the result of completely 
destroying planet Earth (such a phenomenon would entail a great liberation of 
energy; hence, if it were to unfold quickly, it would result in a Supernova-like 
explosion).             
The first ("gravitational") concern can be allayed by simple, hence quite 
reliable, order-of-magnitude calculations, that definitely exclude any such 
possibility. The second concern can also be 
eliminated by estimates of the (very large) number of cosmic ray collisions, at 
higher centre of mass energy than those envisaged in these experiments, which 
have occurred in the past history of our Universe, and which should therefore 
already have triggered a transition to another vacuum if such a phenomenon were 
possible.             
The third concern requires a more detailed 
analysis.   5.         
The corresponding, dangerous scenario goes as follows. (i) Suppose that a 
sufficiently large aggregate of strange hadronic matter (exist and that it) be 
stable (at zero pressure) -- or metastable, but with a sufficiently long 
lifetime for the subsequent developments to unfold. (Strange hadronic matter is 
nuclear matter formed not only by nucleons -- protons and neutrons -- but also 
by "strange" baryons, which are well known to exist, although none of them is 
stable; or, equivalently, but in terms of "more elementary" constituents, by the 
quarks that make up nucleons, but also by those other quarks which 
are called strange  -- that 
are also known to exist, as constituents of those elementary particles also 
called strange -- perhaps because all of them are unstable and are 
therefore not normally present in our environment). Following current usage, we 
call such an aggregate of strange matter a strangelet. (ii) Suppose that 
strangelets are negatively charged. (iii) Suppose that a (negatively 
charged) strangelet is produced in the lab, in a collision-experiment 
with heavy ions, and stops there without previously breaking up and thereby 
disassembling.  Then, via the long 
range Coulomb (electric) force, it would attract (or, equivalently, be attracted 
by) a nearby atomic nucleus, fuse with it, and become a larger 
strangelet, whose charge might be initially positive due to the 
contribution of the positive charge of the nucleus (all nuclei have positive 
charge, and charge is conserved), but which would subsequently become again 
negative via emission of positively charged electrons (a standard phenomenon, 
known as beta decay) if the normal (ground) state of strange nuclear matter were 
negatively charged. The new (larger) strangelet would then fuse with 
another nucleus, and so on and on. In this manner all surrounding matter might 
get transformed into strange matter. If the entire planet Earth were to be so 
transformed, the final outcome would likely be a sphere of enormous density, 
with roughly the mass of the Earth  
and only, say, one hundred meters radius. An enormous amount of 
nuclear energy would be liberated in the process which, if it were a fast one, 
would result in an explosion of astronomical proportions. This prospect 
represents clearly a very great, albeit hypothetical, danger. But how 
hypothetical? Two approaches are 
possible to assess this. One approach relies on current theoretical 
understanding of nuclear and subnuclear physics, and tries to estimate the 
probability that the scenario mentioned above unfolds. A second approach 
relies on the observation that analogous collisions to those which are going 
to be realised in the experiments under discussion have already occurred in 
Nature (cosmic rays), without causing observable calamities; from this fact 
upper bounds can be inferred on the probability that these experiments 
have a catastrophic outcome.  The main 
disadvantage of the first approach is that it relies on a theoretical 
framework which is still imperfectly known -- and it moreover depends on 
computations nobody knows how to reliably perform. The second approach 
must also necessarily rely, to some extent, on our theoretical understanding 
of nuclear and subnuclear physics (and also astrophysics, see below), but of 
course much less so. In the context of both approaches prudence requires that, 
as a rule, uncertainties be generally replaced by "worst case assumptions", 
although this should be done "within reason". The analyses should in any case be 
conducted with a critical spirit -- not with the purpose to prove a conveniently 
appeasing conclusion -- indeed a prudent methodology (not really followed so 
far, to the best of my knowledge) should engage two groups of competent experts, 
a blue team trying to make an "objective" assessment, and a red team 
(acting as "devil's advocates") specifically tasked to make a genuine effort 
at proving that the experiments are indeed dangerous -- an effort that, if 
successful, might then be challenged by the blue team who might 
perhaps point out that such a conclusion could only be achieved by making too 
many too far-fetched worst-case assumptions -- and also perhaps by introducing 
too encompassing a definition of what "dangerous" means. A debate might ensue, 
which, if conducted in a genuine scientific spirit, would be quite enlightening 
for those who eventually have the responsibility to decide. But we shall return 
below to a discussion of these methodological issues. For the moment we 
limit our presentation of the scientific aspects of the issue to a superficial 
if occasionally critical outline of the arguments and conclusions of the 
analyses performed so far [1-4] as we understand them; analyses which seem to us 
to have been in the nature of blue team treatments -- conducted of course 
in good faith by competent experts, but occasionally tainted by an excessive 
awareness of the public relations relevance of the exercise, perhaps at the 
expense of candour if not objectivity.   6.         
Let us begin by reporting on a risk assessment performed in the framework 
of the first point of view, namely based on the current theoretical 
understanding of the likelihood that the dangerous scenario described above 
unfold. We have seen that, in order for this to happen, three ingredients are 
necessary: (i) strangelets should exist (namely, be stable -- or at least 
long-lived -- at zero pressure); (ii) they should be negatively charged; (iii) 
there should be a nonnegligible probability that they be produced in the 
experiments in question. Let us consider each of these three items. 
             
The possibility that strangelets might be stable (or metastable 
but long-lived --without external pressure) is quite conceivable on the basis of 
our present knowledge of nuclear and subnuclear physics, although nobody is 
really able to make a firm prediction in that respect, and perhaps the present 
body of evidence and understanding of nuclear and subnuclear physics might be 
interpreted as rather suggesting otherwise. Under these circumstances, it cannot 
in particular be excluded with any certainty that stable strangelets 
might exist, and moreover only at masses larger than, say, 300 
nucleonic masses, so that they might in principle be produced in gold-gold 
or lead-lead collisions (which puts together a total mass of approximately 
400 nucleonic units), but not, for instance, in the collision of two 
nuclei of iron (iron is a rather common element, both as part of celestial 
bodies and of cosmic rays, but its nucleus only has mass 56, being 
composed of 26 protons and 30 nucleons). Indeed it is likely that 
strangelets, if they exist, are only stable (or long-lived) at relatively 
large masses. And they might be stable for arbitrarily large 
mass.             
The second element of the dangerous scenario outlined above requires 
strangelets to be negatively charged (if they were positively charged, 
they would be repelled by ordinary nuclei due to the long range Coulomb force, 
which is sufficient to keep them sufficiently apart to exclude the initiation of 
any nuclear reaction -- just the same mechanism that prevents ordinary nuclei 
from initiating nuclear reactions among themselves, even when these processes 
are energetically favoured). This appears, on the basis of our present knowledge 
of nuclear and subnuclear forces, quite unlikely: the expectation -- to the 
extent that such calculations can be performed with any degree of reliability -- 
is that, if strangelets exist and are stable, their charge will be 
positive, albeit perhaps small (namely, if Z is their charge in standard 
units, and A their mass in nucleonic units, then a likely guess is that 
0<Z/A<<1 -- while for ordinary nuclei              
The third element of the catastrophic scenario requires that 
strangelets be produced in heavy-ion collisions such as those realised in 
the experiments in question. This looks most unlikely. The reason is that, in 
such a high energy collision,  an 
environment gets created in which all the elementary constituents -- be they 
quarks or subassemblies of few quarks such as nucleons or strange 
baryons -- have a lot of kinetic energy; the most natural outcome is therefore 
that many fragments fly out; it is very difficult for a very large object such 
as a strangelet (which, as we saw above, would be formed by a very large 
number -- hundreds! -- of constituents) to get assembled and to come out 
unbroken. On the other hand it is not easy to evince a quantitative estimate 
from such a qualitative analysis -- which is most likely to be basically 
correct, although again subject to the same caution mentioned above concerning a 
possible lack of theoretical insight on collective multi-body effects: there 
indeed is a theoretical model ("evaporation") which tends to predict somewhat 
larger probabilities of producing large agglomerates in collisions such as those 
under consideration, than other, generally considered more reliable, models do. 
However, graduating from qualitative statements to some kind of quantitative 
estimates  seems to me a desirable 
development, since a large number of collisions will be realised in the 
experiments in question (approximately 20 billions per year at RHIC, 
which  is expected to run for 10 
years), and one would like to be quite certain that not a single dangerous 
strangelet gets produced, if indeed just one would be sufficient to 
initiate a catastrophic process.             
Let us now pause a moment to ponder on the nature of these arguments, and 
on the significance in this context of terms and notions such as likelihood 
and probability.             
When we talk about the likelihood that stable strangelets exist, 
and that they possibly be negatively charged, the notion of probability we 
invoke is associated with our imperfect knowledge of the laws of nature. This 
notion is rather different from that associated with probabilistic evaluations 
caused by our inability to predict exactly -- perhaps because of insufficient 
knowledge of the initial conditions -- the outcome of a physical phenomenon 
whose dynamics we do understand, which is instead the context in which we talk 
more usually, in ordinary life, about probabilities, for instance when we state 
that the probability to get, say, a two by throwing a dice, is one 
sixth (in this case the relevant "classical dynamics", while well known, 
actually entails a "sensitive dependence" on the initial conditions). It is 
indeed clear that the question whether negatively charged strangelets are 
or are not stable can in principle be settled; either one or the other 
alternative is true, and we eventually might be able to find out for sure 
(especially if the answer is positive), and thereafter no room would be left for 
probabilistic assessments. The probability of producing a dangerous 
strangelet -- if such a stable or metastable object does exist -- 
by running the RHIC or ALICE experiment for some time belongs instead to the 
second notion of probability: even if we had a much deeper knowledge of nuclear 
and subnuclear physics than we now have, we could never hope to go beyond a 
probabilistic assessment as regards the risk of producing a dangerous 
strangelet in such circumstances. This is a fundamental consequence of 
the quantum character of the laws of microphysics, as we understand them. 
However, we could -- if we knew enough -- be able to estimate accurately that 
probability, and thereby possibly to conclude it is small enough to exclude any 
reasonable concern.              
On the other hand, at the current stage of knowledge, it would be 
desirable to provide some kind of probabilistic assessment for all the 
components of the catastrophic scenario -- in particular, for the three points 
mentioned above -- in order to come up with an estimate of the risk -- unless 
one can convincingly argue that this risk is certainly so tiny that any attempt 
to quantify it is useless. In the context of such an exercise, the question 
shall arise whether these probabilities -- of which the second and third are 
presumably quite small, presumably the latter more so than the former -- are 
independent, and should therefore be multiplied to get a final 
assessment. I have heard arguments that suggest this to be the case. I am not 
convinced. For instance, an important element which might, as it were 
simultaneously, affect all these evaluations would be some 
(possibly unexpected and perhaps a priori quite "improbable") feature of 
that very very-many-body problem which would have to be mastered in order to 
establish the properties of (heavy) strangelets. And it would obviously 
be fallacious to multiply the probabilities based on theoretical considerations, 
with those based on empirical evidences, as we explain below, after we have 
tersely reviewed this second line of argumentation.   7.         
Let us then survey what can be learned by taking the second point of 
view. High-energy collisions of heavy nuclei occur naturally, when cosmic 
rays impinge on heavenly bodies or among each other in the cosmos. Yet no 
catastrophic event has been so far attributed to such collisions. Does this 
provide sufficient assurance that no disaster will occur in the experiments 
under consideration? The short answer is, unfortunately, rather inconclusive 
-- indeed negative if one believes that the evaluation must be 
prudently made on the basis of "worst-case analyses". But before providing some 
details, let us inject two remarks.             
Firstly we like to emphasise that, if one tries to set an upper bound on 
the probability of a disaster occurring by arguments such as those just 
mentioned, and finds out that the upper bound thus obtained is not sufficiently 
small to conclude that the risk is small enough to be acceptable, this does by 
no means entail that the risk is indeed sizeable: it only indicates that that 
particular argument is not useful to provide confidence. In this respect the 
difference among what we dubbed above first approach and second 
approach must be emphasised: in the first case, one is trying to 
assess the actual likelihood that a dangerous outcome emerge; in the 
second, one is trying to find an upper bound to the probability 
that a catastrophe occur. Let us repeat the obvious: in the first case, 
if one gets from the analysis a probability that is not quite small, then 
concern is indeed appropriate; in the second, if it turns out that the 
computed probability is too large to provide assurance, this merely indicates 
that we do not have an argument that provides confidence, but it would be quite 
wrong to interpret such a finding as an indication that the probability of 
disaster has been shown to be sizeable, because the nature of the argument 
clearly prevents any such conclusion.              
Secondly, we like to note an advantage of the second approach: it 
tends to be applicable to a larger variety of catastrophic hypotheses, rather 
than only to a particular kind of scenario. For instance -- if it did work -- it 
might also serve to exclude the risk that in high-energy ion-ion collisions such 
as those envisaged in the RHIC and ALICE experiments a different configuration 
of ordinary nuclear matter -- not strange: made up of ordinary 
nucleons -- be created which, if more tightly bound than the standard 
nuclear matter that constitutes standard (heavy) nuclei, might serve as "centre 
of condensation" for a transition of the nuclear matter of standard nuclei to 
this new configuration -- a transition which, if it were to involve a 
macroscopic chunk of matter, would be accompanied by a large release of energy 
and would therefore also entail catastrophic consequences. (For instance such a 
hypothetical, albeit implausible, more bound configuration of ordinary nuclear 
matter could be caused by a prevalence of the spin-orbit component of the 
nuclear force -- which can always be adjusted to be attractive -- over the 
central and tensor components, which provide instead most of the binding energy 
in standard nuclei. Such an anomalous "spin-orbit-bound" [5] configuration of 
nuclear matter would be characterised by large values of the relative angular 
momentum for every nucleon pair -- something that can indeed be in principle 
achieved -- and would therefore be very different from the configuration of the 
nuclear matter of standard (heavy) nuclei, and this might explain the 
metastability of such standard nuclear matter -- a metastability which might 
have a longer lifetime than that of the Universe, but might hypothetically be 
broken by a sufficiently energetic collision of sufficiently heavy nuclei -- 
giving thereby rise to a hypothetically catastrophic scenario analogous to that 
described above for strange nuclear matter -- although it is not clear in this 
case what the mechanism might be to cause the process to continue, so as to 
eventually involve macroscopic quantities of matter).   8.         
Let us then proceed to the second approach and report tersely two 
arguments which have been made (but also largely unmade) by BJSW [1], DDR [2] 
and JBSW [3], to provide, from empirical considerations based on cosmic ray 
phenomenology, confidence about the safety of the RHIC experiment. These 
arguments apply also to ALICE, but are less conclusive in that context. Of 
course, in order to be reliable, these arguments must refer to cosmic ray events 
analogous (in terms of the energies, and masses, involved) to those being 
envisaged in these experiments -- via direct evidence based on analogous cosmic 
rays events which have been actually measured, or via reliable extrapolations of 
such data (high energy cosmic ray data for heavy nuclei such as gold or lead are 
scarce).             
BJSW [1] point out that analogous collisions to those that are planned at 
RHIC occur when cosmic rays hit the Moon. (The use of the Moon, rather than the 
Earth itself, for this argument is required because the majority of the cosmic 
rays that hit the Earth interact with its atmosphere before reaching the ground, 
and the atmosphere contains few heavy elements; hence in the case of the Earth 
the analogy with the collisions among heavy ions, or heavy nuclei, 
is missing). Such collisions have occurred for a long time (the Moon is a few 
billion years old), without producing the catastrophic disappearance of our 
satellite. From this evidence they inferred [1] a very small upper bound on the 
probability that a catastrophe occur in the RHIC experiment, which would be 
quite appeasing, were it not for their failure -- as pointed out by DDH [2] -- 
to take due account of an important difference among the impact of cosmic rays 
on the nuclei in the lunar soil, and the collisions of heavy ions in the planned 
experiments. In the first case, a strangelet hypothetically produced in 
the collision would move with high speed relatively to the lunar matter and 
would therefore have a high chance to break up before coming to rest (to 
initiate the catastrophic scenario); in the second case the hypothetical 
strangelet, produced in a head-on collision of two ions, would be already 
almost at rest in the lab. Taking due account of this difference (and using in 
the related computations some rather extreme -- but perhaps not excessively so 
-- worst-case 
hypotheses), DDH [2] have shown that the safety margin provided by the 
persistence of the Moon essentially evaporates.             
DDH [2] -- who also seemed bent at providing reassurance if at all 
possible, but insisted in trying to do so by the second approach (perhaps 
being motivated by lack of confidence on theoretical considerations alone; for 
instance they state that "our understanding of the interactions between 
quarks is insufficient to decide with confidence whether or not strangelets are 
stable forms of matter" [2]) -- tried then to review the relevant evidence 
based on cosmic ray phenomenology, but keeping in the process due account of the 
need to restrict attention to collisions in which a hypothetically formed 
dangerous strangelet would not be likely to break up before getting in 
equilibrium with the matter surrounding it.              
Strangelets produced in cosmic space would eventually be swept 
into star matter (DDH [2] provide arguments that this would indeed happen, if 
the strangelets were negatively charged), and they would then cause stars 
to blow up as supernovae, if the catastrophic scenario indeed prevails. But only 
about 5 supernovae per millennium are observed (and there are other well 
understood scenarios to produce at least some of them). In this manner DDH [2] 
obtain, as an upper bound to the probability of producing a dangerous 
strangelet in one year of running the RHIC experiment, the estimate 
1/500,000,000 (one over five hundred million, namely two 
billionth). This argument also produces a bound for the ALICE experiment, 
which is however much larger.             
DDH [2] state that this bound implies that "it is safe to run RHIC for 
500 million years". A (substantially equivalent -- in operational terms! -- 
but) more correct -- albeit, perhaps, less appeasing -- language would state 
that this bound indicates that the time scale over which a catastrophe might 
emerge from the RHIC experiment is (at least) of the order of magnitude of 
hundred million years.             
But this bound is only applicable if the catastrophic phenomenon is fast 
enough to yield a supernova event. If the process is slow, so that no visible 
supernova explosion emerges, a different approach is needed. DDH [2] then argue 
as follows. Firstly they observe that if the process is excessively slow, then 
one need not worry: in particular if it would take more than ten billion 
years to destroy the Earth, no concern seems appropriate, since anyway we 
expect that ten billion years hence planet Earth will be engulfed by a 
much enlarged Sun, which by that time will have become a red giant star. Hence, 
they focus on the intermediate range of a hypothetical scenario that is not so 
fast to yield a supernova explosion, yet it is fast enough to cause reasonable 
concern in terms of the Earth getting destroyed before its natural death, as 
predicted by current astrophysical expectations, occurs. In this context, they 
look at the increased luminosity of stars that would be caused if some of them 
were destroyed, even relatively slowly, by strangelets, and they get 
again an upper bound of the same order of magnitude, or perhaps -- if worst-case 
assumptions are made to model the destruction of a star caused by the 
strangelet mechanism -- a bound one hundred times larger -- which 
would indicate that the time scale over which a catastrophe might emerge from 
the RHIC experiment is still quite large, of the order of (at least) million 
years.             
But this part of their analysis seem to me somewhat unconvincing -- in 
particular, their modelling of the dynamics of star destruction via the 
strangelet-caused mechanism. This point is, however, moot, since the DDH 
[2] bound is invalidated [3] if one takes account of the possibility that 
strangelets be metastable -- with a lifetime short enough for them not to 
be "eaten" by stars once they are formed in interstellar space, yet long enough 
to cause a catastrophe when they are produced at rest in the 
lab.             
It seems in conclusion that the empirical evidence from cosmic rays 
yields no appeasing upper bound on the probability of producing a dangerous 
strangelet in the experiments in question, at least if one insists that 
any such bound, to be entirely reliable, should be obtained by treating 
uncertainties via worst-case hypotheses. Indeed JBSW [3] (rather in contrast to 
BJSW [1]) state:             
"By making sufficiently unlikely assumptions about the properties of 
strangelets, it is possible to render both of these empirical bounds irrelevant 
to RHIC. The authors of Ref. [2] [namely, DDH] construct just such a 
model in order to discard the lunar limits: They assume that strangelets are 
produced only in gold-gold collisions [this is imprecisely stated -- the 
assumption is that strangelets be stable only at masses large enough, of the 
order or larger than the mass of two gold nuclei], only at or above RHIC 
energies, and only at rest in the centre of mass [this is also imprecisely 
stated]. We are sceptical of all these assumptions. If they are accepted, 
however, lunar persistence provides no useful limits. Others [presumably a 
call to Ref. [5] is missing here, which reads: "We thank W. Wagner and A. 
Kent for correspondence on the subject of strangelet metastability"; indeed, 
this Reference is quoted nowhere in JBSW [3] !], in turn, have pointed out 
that the astrophysical limits of Ref. [2] can be avoided if the dangerous 
strangelet is metastable and decays by baryon emission with a lifetime longer 
than    9.         
There is one other point that does not seem to have been taken quite into 
consideration in these analyses: namely the possibility that the catastrophic 
scenario, rather than ending up in the destruction of the entire planet Earth, 
yield "only" a local calamity. This might, for instance, possibly be the case if 
(i) there were a valley of stability or metastability for strangelets 
of masses, say, from 300 (in units of nucleon mass) to some finite 
mass B  -- analogous to the 
situation for standard nuclei, except for the fact that, in the dangerous 
strangelet case, there would be a lower mass limit (here arbitrarily 
guessed at 300) and at least some negatively charged specimens would also 
be included among the stable and metastable strangelets; and if moreover 
(ii) heavier strangelets had a sufficiently large probability to 
fission into such stable or metastable strangelets. If the various 
lifetimes and cross sections for the various decays and reactions were properly 
adjusted, a chain reaction might be initiated by the production of a negatively 
charged strangelet in the lab and it might result in a nuclear explosion, 
which might however stop before reaching astronomical proportions. Such 
fine-tuning of parameters might look contrived, hence unlikely; but it would not 
be the first time that Nature surprises us: who could have a priori 
guessed that standard nuclear physics was so finely tuned, not only to allow 
the creation of controlled nuclear reactions, but even to organise naturally 
such an experiment on our planet, over geological times, in the Uranium-rich 
mines of Gabon? Moreover, what about the observation  (anthropic principle?) according to 
which, of the infinitely many other possibilities, quite a number must be 
excluded since, if they had prevailed, we would not be here to argue about them. 
   10.       But let us 
abandon such far-fetched speculations, to try and summarise this part of our 
discussion. To this end it is perhaps both expedient and instructive to quote GW 
[4]:"If strangelets exist (which is conceivable), and if they form 
reasonably stable lumps (which is unlikely), and if they are negatively charged 
(although the theory strongly favours positive charges), and if tiny strangelets 
can be created at RHIC (which is exceedingly unlikely), then there just might be 
a problem. A new-born strangelet could engulf atomic nuclei, growing 
relentlessly and ultimately consuming the Earth. The word 'unlikely', however 
many times it is repeated, just isn't enough to assuage our fears of this total 
disaster."             
GW [4] then go on to report that, by relying on what we called above the 
second approach, sufficiently small upper limits can be put on the risk 
probability. They report the lunar argument of BJSW [1], without mentioning the 
criticism of it by DDH [2], and quote the BJSW conclusion ("cosmic ray 
collisions provide ample reassurance that we are safe from a 
strangelet-initiated catastrophe at RHIC" [1]), and likewise they quote 
uncritically DDH [2] ("beyond reasonable doubt, heavy-ion experiments at RHIC 
will not endanger our planet"); and they appeasingly conclude that "even 
though the risks were always minimal, it is reassuring to know that someone has 
bothered to calculate them." Unfortunately this conclusion, to the extent it 
relies on the second approach, seems to me to be by now somewhat 
unjustified -- as we have tried to explain above. This is, to some degree, 
reflected in the modified flair of JBSW [3] relative to BJSW [1]: indeed the 
sentence from BJSW [1] quoted by GW [3] -- which was the final sentence of this 
Report [1] commissioned by the Director of BNL, and was indeed introduced by a 
rather peremptory "we demonstrate that" -- is no more to be found in JBSW 
[3], and it is replaced there by the following final paragraph of the 
introductory section (which indeed follows the one we quoted above, at the end 
of § 8): "We wish to 
stress once again that we do not consider these empirical analyses central to 
the argument for safety at RHIC. The arguments which are invoked to destroy the 
empirical bounds from cosmic rays, if valid, would not make dangerous strangelet 
production at RHIC more likely. Even if the bounds from lunar and astrophysical 
arguments are set aside, we believe that basic physics considerations rule out 
the possibility of dangerous strangelet production at 
RHIC."   11.       In 
conclusion the main arguments to allay fears of a catastrophic outcome of the 
experiments RHIC at BNL and ALICE at CERN is (i) the unplausibility, on 
theoretical grounds, that stable or (sufficiently long-lived) metastable 
strangelets with negative charge (i. e.., "dangerous strangelets") 
exist, and (ii) the hunch that, even if they do exist, the probability that even 
a single one of them be created in these experiments is exceedingly small (but 
how small is small enough? -- more on this below).             
People might feel that (iii) the empirical arguments based on cosmic-ray 
phenomenology, even if not totally convincing, provide additional confidence. 
Perhaps so. Yet I have also read that each of the 3 arguments, (i), (ii) 
respectively (iii), can be interpreted as providing (tiny) upper limits, call 
them              
(Indeed, imagine you are tasked to estimate the probability to draw a 
black ball from a box. The theoretical information you have is that the 
box contains two balls, one black and one white; you also know 
empirically, from a number of previous draws, that the black ball came 
out about half the times. So your estimate based on theory is that the 
black ball has probability 1/2 to be drawn; your estimate based on empirical 
data also suggests it has probability 1/2 to be drawn; do you then conclude 
the probability is 1/4 ? I would not have introduced parenthetically this 
trivial argument, were it not for the fact that an eminent colleague indeed 
suggested the probabilities,              
Let us moreover recall the fundamental difference among the probability, 
call it             
 12.       My personal 
assessment of the situation concerning the ion-ion experiments at BNL and CERN 
is not one of serious concern, because I have confidence on the assessment of 
the experts who unanimously state there is no danger. I am, however, a bit 
disturbed by what I perceive as lack of a very sustained effort at getting more 
quantitative estimates of the various probabilities involved in this 
problematique than I saw reported so far. I wonder in this context about the 
ratio of the funds that have been devoted to such an endeavour, relative to 
those that have gone and are going into the experimental set up. If that ratio 
is less than, say, a few percent, I would -- in my admitted naiveté -- feel 
puzzled and disturbed (however, as a theoretical and mathematical physicist, I 
might be biased in this assessment).             
In this context, I hope (and expect) the Committee tasked by the 
Director-General of CERN to look into the matter will do a thorough job, and I 
also hope their findings will be adequately scrutinised by the expert community 
after they are published. The usefulness of such open discussions is of course 
obvious, and it has indeed been demonstrated by the progress in understanding 
these matters that has resulted from the fruitful reciprocal criticism among the 
two experts groups BJSW and DDH [1-3].   13.       But I am 
also somewhat disturbed by what I perceive as a lack of candour in discussing 
these matters by many -- including several friends and colleagues with whom I 
had private discussions and exchanges of messages -- although I do understand 
their motivations for doing so. Many, indeed most, of them seemed to me to be 
more concerned with the public relations impact of what they, or others, said 
and wrote, than in making sure the facts were presented with complete scientific 
objectivity              
This is, of course, a subjective assessment, for which I must take 
personal responsibility. It has, in any case, motivated me to also try and 
outline below some scientific-political-ethical-sociological considerations 
related to this kind of issues, which are perhaps of more general validity than 
referring to this particular case, although this is a significant example which 
I keep in mind throughout these reflections.    14.       First of 
all it is clear to me that risk evaluations can be reliably done only by 
experts: in this particular case by experts on nuclear and subnuclear 
physics, as well as, to the extent relevant, on astrophysics etc.; and also by 
experts on the evaluation of the risk of mishaps which are both "extremely 
unlikely and extremely catastrophic". And it is of course appropriate that, to 
the maximal extent possible, those who are tasked with making such evaluations 
not be affected by any "conflict of interest". If an experiment potentially 
dangerous is being planned by a group in a lab, it is of course desirable that 
risk evaluations be performed by scientists who have no interests vested in the 
performance of that experiment. This is, of course, not always easy -- since the 
more knowledgeable experts are often to be found just among those who are also 
most keen to see the experimental results in question.             
It is also obvious that an adequate investment in risk evaluation should 
be made before the funds already spent in the preparation of an experiment 
render its cancellation exceedingly wasteful and therefore hard to decide, even 
if there were good reasons to do so.   15.       Such risk 
evaluation exercises yield eventually a probabilistic estimate, which is arrived 
at after detailed investigations of catastrophic scenarios, that often entail 
lots of detailed computations, in which context choices must often be made among 
"most plausible" and "worst-case" hypotheses. Prudence of course suggest that 
the latter be preferred, but judgements may vary. It is therefore quite possible 
that different groups of experts end up with final evaluations which do not 
quite tally. Just for this reason it is generally desirable that more than one 
team be engaged in such analyses -- as already mentioned above, a desirable 
technique is to task a blue team to perform as objective an 
analysis as possible, and a red team to act as "devil's advocate", namely 
to try deliberately to prove that the experiment in question is indeed dangerous 
-- of course always using sound scientific arguments. At the end the two teams 
should compare their findings, and especially the analyses which led to those 
results. This approach is also desirable to minimise the possibility that a risk 
scenario be altogether ignored. I understand such a procedure is (more or less) 
generally followed whenever a potentially dangerous enterprise is undertaken, 
for instance the construction of a nuclear power plant. In such a case generally 
the blue team is organised by those who plan the plant, and the red 
team by those who authorise its construction.             
Once such an analysis has been performed, it should preferably end up 
with quantitative conclusions, in the guise of probabilistic estimates -- 
possibly yielding a range of values. Then the nontrivial question arises of what 
the acceptable value for the probability of a disaster happening should 
be -- namely a value small enough that the risk be considered worth taking. This 
of course depends quite significantly on the magnitude of the catastrophe that 
might occur if things went wrong, and on the advantages that accrue by 
proceeding with the project. Especially when the gains are purely scientific, 
there is clearly the danger that an excessively prudential approach end up in 
impeding scientific progress.             
Anyway I suggest some kind of benchmark to assess "acceptable" 
probabilities of extreme catastrophes might be set by the probability of some 
such impending "natural" catastrophe. I understand for instance the probability 
that an asteroid with diameter over 10 Km hit our planet -- an event which would 
most likely put an end to the human presence on Earth -- is estimated to be of 
the order of one over one hundred million per year.[6] It is probably 
correct to argue that any man-made risk of total catastrophe should be smaller 
than any natural risk -- but it seems to me reduction by an extra small factor 
-- one tenth, one hundredth, one thousandth, perhaps one over ten 
thousand -- should be sufficient. Hence I would probably advise in favour of 
authorising a worthwhile undertaking -- be it a scientific experiment or 
some other worthwhile human enterprise -- if I were reliably guaranteed 
it entails a risk of ultimate catastrophe per year less than one over one 
trillion (probability per year less than              
Of course what I have in mind here are major undertakings, which involve 
large investments of funds technology expert manpower – and which therefore, as 
it were automatically, entail a careful and responsible behaviour. The situation 
is quite different if consideration is extended to actions which can be 
performed on a much smaller scale – as is for instance the case for certain 
experiments in molecular biology and genetic engineering. Indeed simple 
arithmetic shows that the human experiment on Earth is unlikely to last for many 
more centuries if all humans on this planet were to exercise a hypothetical 
individual right to engage in an activity that entails a              
Considerations of this kind – paradoxical as they may appear – underline 
the obvious need to pay special attention to all those scientific/technological 
developments that entail the risk of major catastrophes being produced by small 
scale endeavours – involving few individuals.             
On the other hand, even in the context of major kind of enterprises such 
as those discussed above, I will not be surprised to find many who strongly 
disagree with the figure suggested above --              
The discussion in the previous paragraph seemed to lead us astray. In 
fact, it raised two important, and related, issues: (i) Who should, in the end, 
make decisions? (ii) How to cope with the fact that the majority of people are 
simply unable to comprehend a probabilistic argument?    16.       Granted that 
(fortunately) it is not up to me to take decisions, who then should decide? Let 
us firstly -- for the sake of making the argument concrete -- focus on the RHIC 
experiment: should a controversy about its safety emerge and should the 
laboratory governance get overruled, who should/could finally decide whether it 
should proceed? Since the experiment takes place in the USA and it is funded by 
American taxpayers' money, clearly the decision-makers in this instance are the 
relevant US institutions: the President, Congress, the judiciary. But to the 
extent that experiment puts at risk the entire planet Earth, is it fair 
that  the (non American) majority of 
world citizens have no say whatsoever on the matter? In any case, even 
US citizens may have some say in this matter only in a very indirect way, namely 
to the extent those who decide are their representatives -- within that context 
of representative democracy which still seems the better system of governance 
human kind has been able to devise so far. In any case, what I think is really 
important is that American citizens, as well as the rest of us who live on this 
planet, have a reasonable assurance that, within the decision-making system that 
eventually decides, there exist an adequate capability, and will, to make a 
competent and objective evaluation of the risk.             
As for the CERN experiment, the situation is somewhat less clear-cut. 
CERN is an European institution located between France and Switzerland. I 
understand the final decision, on such matters as authorising a potentially 
dangerous experiment at CERN -- if it ever had to be taken at a political level 
beyond the CERN governance -- would be taken by French rather than Swiss 
authorities. Somebody told me this was arranged so that such decisions could not 
be subject to a popular referendum -- a practice common in Switzerland, and 
inexistent in France. I do not know whether this gossip is true; I was informed 
that the legal department at CERN has cogent arguments to sustain the validity 
of this choice (arguments I would certainly not dare to challenge, nor indeed 
even to scutinize: not my cup of tea). In any case I am in no position to judge 
whether or not such a decision was wise. As it is, if a controversy were to 
emerge over whether to go ahead with the ALICE experiment at CERN, and if the 
CERN leadership -- who are to begin with the natural decision-makers on this 
matter -- were eventually overruled (a development I would not a priori 
like), then the final decision would rest with the French decision-making 
system (be it the executive, legislative, or judicial power, as the case may 
be): the decision-making system, to be sure, of a democratic and highly 
civilised country, which however does not have a spotless record on such 
matters.  (Suffice to recall 
the following episode. Years ago -- perhaps in the aftermath of the catastrophe 
associated with the use of AIDS-infected blood for transfusions -- an 
"independent" National Committee was created in France by President Mitterand to 
assess potential "great risks"; and it was granted the power, somewhat unusual 
in France, not only to select which topics to focus upon, but also to publish 
its findings, without having to ask for a governmental permission to do so. That 
Committee did use such powers, for instance in the context of its assessment of 
the potential risks associated with the fact that a new TGV (fast train) track 
was planned to pass close to a nuclear power station -- the TGV path is a very 
delicate political issue, with important electoral implications. Perhaps for 
this reason -- in any case, without any explanation being proffered -- that 
Committee was soon after, very suddenly, abolished -- via a one-line rider in 
the context of the omnibus budget bill).             
This is not to say I would prefer that Switzerland rather than France be 
the ultimate decision-making authority over whether or not to make an experiment 
at CERN: indeed my clear preference is that such decisions be taken by CERN’s 
own decision-making system. Yet I can't help asking myself how the citizenry in 
Switzerland views this matter.   17.       We are indeed 
coming again here to a crucial point: the role of citizens, who in democracy 
should ultimately be the determining element in decision-making. But to play 
this role responsibly, citizens should possess some understanding of the basic 
facts; yet this is sometimes, nay often when decisions concern scientific 
issues, next to impossible for the "man in the street", or for the "housewife in 
the kitchen" (to mention some politically incorrect, yet realistically quite 
relevant, stereotypes). A relevant example in this context is precisely the 
point mentioned above (end of § 15), namely the difficulty to understand 
probabilistic arguments, in particular those based on very small, yet finite, 
probabilities; arguments which are essential to form an informed opinion about 
any assessment of the risk of very major catastrophes having very low 
probability of occurring.             
It seems to me in this context the scientific community has a great 
responsibility. There is of course a strong temptation to take a "let the matter 
to us" attitude. But clearly this is neither quite right nor politically viable, 
nor, indeed, generally acceptable: indeed some of my elementary particle 
colleagues, whose tendency towards such a stand I found somewhat disturbing when 
I broached with them the potential danger of high energy ion-ion collision 
experiments (especially when it tended to take the subliminal form of simply 
joking the matter away), were themselves quite unwilling to accept such an 
attitude in the matter, say, of molecular biology experiments involving gene 
manipulation: there, some of them kept insisting, we know the dangers are real, 
hence oversight over what scientists (i.e., biologists -- not fellow 
physicists!) are up to is needed.              
 18.       There also 
is a strong temptation for many in the scientific community to view these 
matters primarily in terms of public relations. This may be subliminal or 
deliberate; it may be good or bad (in my judgement).             
A deliberate effort to inform the public -- particularly those closer at 
hand, who might influence local political decision-making -- on the scientific 
activities being undertaken is, in my opinion, a legitimate part of the public 
relations game, especially in major labs which require large public funding: and 
in this context even a certain amount of hype (of the results) and minimisation 
(of the dangers -- if any) is acceptable, especially if the task to propagate 
these notions is performed by public relation professionals rather than by the 
scientists themselves -- who should instead, I submit, rather keep a low 
profile, indeed be modest-minded, or at least act modestly: preferably the 
former, but at least the latter, to respect a scientific etiquette which is part 
of a more civilised way of living (a context, however, where the danger that bad 
behaviours eventually prevail over good ones looms large -- especially in those 
scientific environments, such as contemporary "high-energy" experimental 
physics, where the main traits required of top scientists are managerial skills 
and aggressiveness rather than thoughtful scholarship).             
What however must by all means be avoided is any attempt at obfuscation 
in the context of scientific analyses. This seems so obvious not to require any 
mention. But in fact it is not obvious at all. And it seems to me a scrutiny of 
the papers [1-4] on which this discussion has largely focussed provides as good 
an illustration as any, of the potential pitfalls that may emerge when 
scientific writings are somewhat affected [1-3]-- or even primarily motivated 
[4] -- by the concern "not to alarm the public"; not to mention editorial 
policies, such as that underlining the decision by Nature to refuse to 
publish DDH [2] on the grounds it was of no interest to their readers, which 
smack of deliberate attempts to manage public opinion rather than to inform 
it.             
The possibility that, on certain topics, the public may indeed "become 
alarmed" by reading a scientific paper (or quotations from it) does exist, since 
it is -- rather, it should be -- in the nature of such writings to avoid 
peremptory statements; while, when treating of extreme potential dangers, only 
peremptory statements (excluding any possibility that such risks might indeed 
materialise) are adequate to allay the fears of the public. And the concern is 
in my opinion justified that an "alarmed public" might lead to "irrational" 
decision-making, which might interfere with sound scientific progress, and 
perhaps end up in damaging the public interest (which, we trust, is indeed 
vested in the promotion of sound scientific progress). But any attempt by the 
scientific community to remedy this situation by resorting to lack of candour 
and transparency is in my opinion unwise and dangerous. For two, equally 
important, reasons.   19.       If the 
scientific debate gets muted or distorted -- or altogether suppressed -- because 
the imperative "not to alarm the public" takes precedence over the objective 
candour and the open confrontation of points of view which is a main 
characteristic of the scientific method, then the danger of eventually indeed 
making some silly mistakes increases significantly. This has been illustrated 
time and again in the context of military research, when such debates (for 
instance on the effects of radiation on military personnel and civilian 
populations) were altogether suppressed -- especially in totalitarian societies; 
but not only there -- by imposing military secrecy. But even in the context of 
open scientific research, one should not forget that grossly mistaken 
assessments have sometimes been made by well meaning most competent scientists 
(a classical example being the famous pronouncement by Lord Rutherford -- 
foremost nuclear physicist of his times -- that any prospect of exploiting 
nuclear energy was "moonshine"). The only cure against this risk is the 
scientific method of completely honest, completely candid, give-and-take open 
debate among competent practitioners. As we already noted above (end of § 12), 
this has indeed been once more demonstrated by the scientific exchanges that 
resulted in the JBSW [3] revision of BJSW [1].   20.       Another, no 
less important, drawback of any deviation by the scientific community -- under 
the banner of combating alarmism -- from the practice of completely unrestrained 
open candid unobfuscated transparency in all their utterances, is the 
(justified!) lack of confidence by the general public in the "integrity of 
scientists", that is eventually likely to result from any hint that such 
behaviour is prevailing. Such a lack of confidence is particularly deplorable 
precisely in the context we are discussing here. Indeed I would like to conclude 
this paper by reaffirming my strong belief that it is most desirable that any 
decision on the kind of matters we have been discussed herein be, if not always 
taken -- which would be politically impossible, and indeed ethically dubious in 
a democratic context -- certainly always primarily influenced by the scientific 
community, rather than by demagogues or charlatans or, at best, incompetent 
generalists. But this will become impossible -- at least in a democratic context 
(to which I assume few would like to renounce: particularly those of us who had 
some chance to experience the alternative!) -- if such a (justified!) lack of 
confidence by the general public in the scientific community will eventually 
prevail.      References   [1]        W. 
Busza, R. L. Jaffe, J. Sandweiss and F. Wilczek: Review of Speculative 
"Disaster Scenarios" at RHIC, hep-ph/9910333, 13 October 1999, referred to 
herein as BJSW. This is the text of a Report commissioned by Dr. John Marburger, 
Director of BNL.   [2]        Arnon 
Dar, A. De Rujula, Ulrich Heinz: Will relativistic heavy-ion colliders 
destroy our planet?, Phys. Lett. B 470, 142-148 (1999), referred to 
herein as DDH. This paper appeared on the web as hep-ph/9910471, 25 October 
1999; it was refused by Nature on the grounds that the topic was not of 
sufficient general interest, although Nature asked to be informed about 
eventual publication elsewhere, indicating they might be interested to publish a 
comment on the issue (which they eventually did [4]); it was submitted to 
Phys. Lett. B on November 2, 1999, accepted by the editor R. Gatto on 
November 3, 1999, and the issue on which it appeared is dated 16 December 1999; 
the comment [4] appeared on the issue of Nature dated 9 December 
1999.   [3]        R. L. 
Jaffe, W. Busza, J. Sandweiss and F. Wilczek: Review of Speculative "Disaster 
Scenarios" at RHIC, hep-ph/9910333 v2, 19 May 2000, referred to herein as 
JBSW (this is a significantly modified version of [1]).              
 [4]       Sheldon L. 
Glashow and Richard Wilson: Taking serious risks seriously, in the 
section News and Views (Nuclear physics), Nature 402, 596-597 (1999), 
referred to herein as GW.   [5]       F. Calogero 
and F. Palumbo,  "Spin‑Orbit‑Bound 
Nuclei",  Phys. Rev. C7, 
2219‑2228 (1973). [6]        C. R. 
Chapman and D. Morrison, "Impacts on the Earth by asteroids and comets: 
assessing the hazard", Nature 367, 33-40 (1994).   ______________________________________________________________________________________________ Francesco Calogero  is professor of theoretical 
physics at the Department of Physics of the University of Rome I "La Sapienza". 
His main current research activity deals with the mathematical physics of 
integrable dynamical systems and nonlinear partial differential equations. He is 
now completing a book of Lecture Notes (to be published by Springer) with the 
tentative title "Classical many-body problems in one-, two- and 
three-dimensional space amenable to exact treatments (solvable and/or integrable 
and/or linearizable…)". He served from 1989 to 1997 as Secretary General of the 
Pugwash Conferences on Science and World Affairs, and in that capacity accepted 
in Oslo the 1995 Nobel Peace Prize awarded jointly to Joseph Rotblat and to 
Pugwash. He serves now as Chairman of the Pugwash Council (1997-2002). He served 
as member of the Governing Board of the Stockholm International Peace Research 
Institute (SIPRI) from 1982 to 1992. The ideas and opinions proffered in this paper are 
strictly personal and should not be construed as expressing the views of any 
institution or organization.          |