Introduction

The fusion program, both short term and long term, is in trouble, certainly in the United States, and likely worldwide. In addition to large cost overruns and failures to meet milestones, surely another reason is that pure fusion has almost no chance of meeting energy requirements on a time scale that anyone alive today can relate to. Hence the assertion of this article is that fusion breeding of conventional nuclear fuel is a likely way out of fusion’s current and future difficulties. Fusion breeding substantially reduces the requirements on the fusion reactor. It significantly reduces the necessary Q (fusion power divided by input power), wall loading, and availability fraction. The capital cost of a reactor, estimated based on ITER’s capital cost, is affordable for fusion breeding, but definitely is not for pure fusion. It is likely that fusion breeding can produce fuel at a reasonable cost by mid century. The entire fusion and fission infrastructure would be sustainable, economical, environmentally sound, and have little or no proliferation risk. This article’s mission then, is to hopefully convince a much larger portion of the fusion establishment to make this case. At the very least it hopes to broaden the discussion in the fusion community from where we are now, where one prestigious review committee after another insists that every existing project is absolutely vital, nothing can be changed; except give us more $$$. The inevitable result of this process is that one fusion project after another gets knocked off.

The choice of fusion breeding versus pure fusion does not mean an immediate departure from the current course, but perhaps a 30 degree course correction. Many, but not all initial tasks are common to both options. However even with a 30 degree correction, after some distance traveled, the two paths diverge and are quite far apart. What is needed immediately then is a change in psychology in the fusion effort, that is a realization that breeding is a different, perfectly acceptable, and perhaps even a better option for fusion; and certainly one that is more achievable.

There is a natural symbiosis between fission and fusion. Fusion is neutron rich and energy poor; fission is energy rich and neutron poor. A fusion reaction, which produces about 20 MeV, creates 2–3 neutrons after neutron multiplication. The fission reaction produces the same 2–3 neutrons, but 200 MeV. Thus, for equivalent output power, fusion generates ten times as many neutrons as fission. A fusion reactor can be a very prolific fuel producer for fission reactor, which would be the primary energy producer.

Fusion proponents, for the most part, see fusion as an inherently safe energy source with minimal issues of hazardous waste and an inexhaustible fuel supply. They would prefer not to tie their fortunes to fission, which might not even want them, and which they see as having issues of safety, proliferation, and long-term radioactive waste. But they should consider realities. Fusion breeding is at least an order of magnitude easier to pull off than pure fusion power. It should be possible, whereas commercial application of pure fusion may turn out not to be. Even in the best of circumstances, commercial fusion breeding should be available quite a few decades earlier than commercial pure fusion. Thus it could serve as an intermediate objective, of genuine economic value, on the path to pure fusion.

With the recent disaster at Fukushima, and the sudden advent an era of inexpensive natural gas from hydraulic fracturing, that is, fracking, this hardly seems to be the optimum time to argue for an energy source that is largely nuclear. However this article takes a longer view.

An “inconvenient truth” is that the world may well have a climate problem by mid century, forcing it to move away from carbon-based fossil fuels even if they are not on a rapid path to depletion, which they may well be. But another “inconvenient truth” is that advanced economies require energy, and lots of it. And a third “inconvenient truth” is that the world is very unequal today, but the less-developed nations are pushing very hard for economic equality. It will be apparent by mid-century that the world has a very big long-term energy problem. Today 20 % or so of the world’s population uses the lion’s share of the 14 terawatts (TW) of power the world generates. By midcentury this will no longer be acceptable; everybody will demand the better life style abundant power makes possible. In a much cited paper, Hoffert et al. [1] argue that the world will need an additional 10–30 TW of carbon-free energy by mid-century.

And where will the world get this carbon-free energy? This author is convinced [2] that solar energy (solar thermal, solar photo voltaic, wind and biofuel) will contribute only a small fraction of this energy. As an example consider wind. First of all, much information we read is misleading. This author has seen mostly the ‘nameplate power’ quoted. This is the power the windmill of wind farm produces when the wind blows at exactly the right speed and direction. The average power, or the actual power is considerably less. According to the Wikipedia article on wind power (http://en.wikipedia.org/wiki/Wind_power), the actual wind power produced in 2011 was about 50 gigawatts (GW), about 0.36 % of the total world power of 14 TW. In other words, after more than 20 years of heavily subsidized development, if one takes the percentage of world power produced by wind, and rounds it off the nearest whole number, that percentage is zero.

In the absence of carbon free power sources, right now the world gets its energy from coal. Coal is overtaking oil as the largest source of power, its use is rapidly accelerating [3]. Rapidly developing countries like China, Indonesia, Turkey… are greatly increasing their use of coal. But if fossil fuel use, and especially coal, has to be reduced because of fears of climate change and/or depletion of a finite resource; and if wind, solar and biofuel cannot do it; what else is there? If not nuclear, then what?

Fission and fusion are the only carbon-free technologies that could power a world with 10 billion people. Fusion has many potential advantages as an energy source, but even its most optimistic proponents recognize that commercial fusion energy will not be ready any time near mid century. For nearly half a century, fusion’s advocates have been planning for a DEMO (a precursor to a commercial reactor) 35 years from t = 0, and who knows when t = 0 will be? [4].

Fission, on the other hand, is a technology that works well now, but the supply of fissile 235U is limited. Fission can be a long-term, sustainable solution only if breeders are used to convert 238U or thorium into fissile fuel. There are two options for breeders: fast-neutron fission reactors or fusion reactors. Fission breeding has a head start, but fusion breeding, advocated by this author for 15 years now is by far the more attractive option [411]. Fusion breeding is a much easier technology than pure fusion as an energy source; it could be commercialized by mid-century. Furthermore, fusion breeding could naturally serve as an economical bridge to pure fusion in the much more distant future.

Then it is also worth noting that both Andrei Sakharov [12] and Hans Bethe [13] advocated fusion breeding instead of pure fusion. These are two physicists we ignore at our peril.

We will see that for pure fusion to become economically feasible, either ITER or NIF would have to be greatly improved. Yet all experience shows that once systems get to be that size, progress comes with glacial lethargy, if comes at all. For instance it was 20 years from when ITER was proposed until it was approved, then another 15 years until first plasma, and another 7 (if all goes well) until DT experiments can even begin. This 42 year period, just to get to the starting gate, is longer than the total career of a typical physicist.

Furthermore, as we will also shortly see, pure fusion may well be out of reach for tokamaks. However an ITER or NIF scale system is more than adequate for fusion breeding. We will see the Q that ITER or NIF hopes to achieve is not nearly sufficient for pure fusion but definitely fine is for breeding. The first wall of a pure fusion reactor would have to be able to withstand a neutron flux of at least 4 MW/m2; but as we will see, a fusion breeder is fine with ITER scale wall loadings of one, or perhaps even half a megawatt per square meter. If the fusion reactor principally breeds, fuel and is only secondarily a power source, its availability fraction is not such a critical consideration either.

One obvious question then is why develop fusion breeding, decades in the future; when we are not developing fission breeding today? One answer is that there is no immediate need for either fission or fusion breeding. For the next few decades, fissile fuel is readily available. However the need for one or both will be pressing by mid century [1]. A key advantage of fusion breeding, is that it is about an order of magnitude more prolific than fission breeding. That is one fusion breeder can fuel about five light water reactors (LWR’s) of equal power, whereas it takes two fission breeders, for instance two integral fast reactors (IFR) at maximum breeding rate to fuel a single LWR of equal power [1418], (http://en.wikipedia.org/wiki/Integral_fast_reactor). Clearly fusion breeding fits in well with today’s (and very likely tomorrow’s) nuclear infrastructure; fission breeding does not. In other words, in a fission breeder economy, at least 2/3 of the reactors would have to be breeders and this will represent a staggering cost. In a fusion breeder economy, all of the LWR’s, here and in the future, would remain in place.

Thus fusion breeding does not have to compete with LWR’s, a competition likely unwinnable if fuel supply for LWR’s were no consideration. It only has to compete with fission breeders, a competition it might well win. But more realistically, in a competition between fission and fusion breeding, there will likely be room in the midcentury economy for both, at least if fusion breeding makes a real effort to compete.

A fusion breeder economy does envision a vital role for fast neutron fission reactors. A single IFR (which in fact can run as either a breeder or burner), can burn any actinide. If run as a perfectly efficient burner, it can burn the actinide wastes of about 5 LWR’s (an LWR each year releases about 1/5 of the fuel it starts with in the form of actinide by products) [19]. Thus we can envision a sustainable, mid century, fusion breeding based energy architecture, where a single fusion breeder fuels 5 LWR’s, the nuclear reactor of choice up to now, and a single IFR burns up the actinide waste of these 5 LWR’s. The only waste products would be those fission decay products (cesium 137, strontium 90, etc.) which have no commercial value, but which have half lives of about 30 years. These could be stored for 300–600 years, a time human society can reasonably plan for, not the half million or so years it would take for, for instance 239Pu (half life 24,000 years), to decay away.

There is some dispute about the amount of uranium fuel available. Hoffert et al. [1], measuring it in terawatt years, estimates 60–300 TW years of uranium fuel is available. He mentioned that if the lowest estimate is correct, and 10 TW are needed (right now the world uses about 300 GWe, or about 1 TWth of nuclear power), there is only enough fuel for 6 years, hardly enough to justify creating a large, multi terawatt infrastructure. However other sources estimate that the available fuel is much greater [20]. This paper certainly cannot sort out these conflicting estimates. But one thing is indisputable. No matter how much nuclear fuel fission advocates think they have, they must admit that their supply of fissile nuclear fuel is <1 % of the uranium resource, and 0 % of the thorium. Fusion breeding makes about 50–100 % of each available. This author does not believe the world is so well endowed with energy resources that we can afford to discard more than 99 % of them. To get an idea of the size of this resource, thermal nuclear reactors have been delivering about 300 GWe for about 40 years. This means that in depleted uranium alone, there is sufficient fuel for 3 TWe for 200–400 years! But as we will see, a much better option is to use thorium as a fuel, and there is three times as much thorium as uranium. Nuclear fuel, for all practical purposes, is inexhaustible in the same sense as wind or solar [21].

Hence a conservative approach is to assume the worst as regards nuclear fuel supply and support a relatively low cost, economical way of enhancing it by about two orders of magnitude, i.e. fusion breeding. Supporting fusion breeding will not produce appreciable fuel until mid century at best; and by then there could well be a crying need for it.

Then to reiterate our basic case, this article makes the case that (1) The optimum strategy is not pure fusion, i.e. using the neutron’s 14 MeV kinetic energy to boil water; but fusion breeding, i.e. using what, for want of a better term, we call the neutron’s potential energy, to breed 10 times more fuel, and its kinetic energy to boil water. This is especially true if one adopts the tokamak approach, because conservative design rules, which we will discuss shortly, and which have limited tokamak operation for half a century appear to forbid pure fusion, but allow fusion breeding. Of all the hybrid fusion options, the most sensible is breeding. Recently a summary of many different possible types of many different fusion fission hybrids has appeared [20]. While not explicitly concluding that breeding is the best option, it does seem to point toward it. Our own book chapter summarizing hybrids does explicitly recommend breeding among the options [11]. A valuable resource for anyone interested in fusion breeding or other forms of hybrid fusion is a web site set up by Ralph Moir (www.ralphmoir.com). (2) The fusion configurations should continue along the paths being blazed by ITER and NIF, and not divert scarce resources to other approaches. (3) The mid century energy infrastructure will necessarily have a large nuclear component. Fission now supplies about 300 GWe, or about 13 % of the 2.5 TW the world generates. There are about 400 fission power plants now, and about 70 more are in various stages of construction or planning. There is no reason to think that nuclear power generation cannot be considerably increased. By mid century, fusion breeders could be supplying their fuel and fast neutron reactors could be burning the actinide wastes.

Recently Ralph Moir, along with this author has contributed a chapter on hybrid fusion to a textbook [11]. This chapter concentrated on the nuclear aspects, while not ignoring the plasma. By contrast, this article concentrates mostly on the plasma aspects, while not ignoring the nuclear, and may be regarded as a companion to the book chapter.

Now let’s get to fusion’s difficulties. Consider first ITER [22]. When it was approved in 2005, the construction cost was estimated at about 5 billion euros, with a start date of 2106. In 4 years, the cost has tripled to 16 billion euros, and the start date to 2019 [23]. Some estimates have put the cost of ITER even higher. Here is Senator Diane Feinstein [24], chair of the subcommittee on Energy and Water Development of the Senate on the rapidly increasing cost of ITER:

“We provide no funding for ITER until the department (of energy) provides this committee with a baseline cost, schedule and scope.”

Now consider NIF, another multibillion dollar machine. It was supposed to achieve ignition in FY 2012. As of this writing (fall 2013 and winter, 2013–2014) it has not only failed, but has failed in spectacular fashion. While it now routinely generates laser shots of well over a megajoule, the best gains they have achieved are still <10−2, more than three orders of magnitude below original predictions. The Lawrence Livermore Laboratory has recently published a paper on where NIF is now, pointing out that on their best shot, they achieved about 1015 neutrons, about 3 kJ [25]. Recent efforts have done somewhat better [26].

Here is the House Appropriations Committee [27] on the failure of NIF to achieve ignition:

“As the first ignition campaign comes to a close in fiscal year 2012, it is a distinct possibility that the NNSA will not achieve ignition during these initial experiments. While achieving ignition was never scientifically assured, the considerable costs will not have been warranted if the only role of the National Ignition Facility (NIF) is that of an expensive platform for routine high energy density physics experiments.”

Thus fusion’s two gigantic flagships, ITER and NIF appear to be taking in water. However NIF is, and ITER unquestionably will be unique, priceless, world class resources which this author is confident will ultimately make very major contributions to fusion. But fusion has long term problems, which are even more serious. While these short and long term problems are not obviously related, a more credible long term strategy will help in the short term as well. The fact that fusion breeding fits in very well with todays and tomorrows likely nuclear infrastructure so that it could be introduced gradually and seamlessly enhances the case further.

To illustrate ITER’s long term problems, consider its goal as stated on the ITER web site (www.iter.org):

“The Q in the formula on the right symbolizes the ratio of fusion power to input power. Q ≥ 10 represents the scientific goal of the ITER project: to deliver ten times the power it consumes. From 50 MW (megawatts) of input power, the ITER machine is designed to produce 500 MW of fusion power—the first of all fusion experiments to produce net energy.”

So how close would we be to a reactor? The 500 MW of fusion thermal power typically has an efficiency of conversion to electricity power of about 1/3, so it generates about 170 MWe. However the driver, beams or radiation need about 50 MW. But accelerators and radiation sources are not 100 % efficient either. Again, one third is a more reasonable estimate, so the drivers would need 150 MW of the electric power produced, leaving all of 20 MWe for the grid. Clearly ITER would have to be greatly improved before it could begin to be regarded as an economical power source.

But what about the fact that ITER could be a burning plasma so much less input power is needed. Again, here is the ITER web site:

“The ITER Tokamak will rely on three sources of external heating that work in concert to provide the input heating power of 50 MW required to bring the plasma to the temperature necessary for fusion. These are neutral beam injection and two sources of high-frequency electromagnetic waves.

Ultimately, researchers hope to achieve a “burning plasma”—one in which the energy of the helium nuclei produced by the fusion reaction is enough to maintain the temperature of the plasma. The external heating can then be strongly reduced or switched off altogether. A burning plasma in which at least 50 percent of the energy needed to drive the fusion reaction is generated internally is an essential step to reaching the goal of fusion power generation.”

Hence a burning plasma does not seem to be anything like one of ITER’s initial goals, but taking their figure of a reduction of input power by 50 %, this means that only 25 MW of external power is needed, leaving perhaps 100 MWe for the grid; small power for a $20B facility. In this paper the original Large ITER is also considered [28]. Roughly it produces about 3 times the power, still with Q = 10. At the time the switch was made, the cost of Large ITER was estimated to be about a factor of 2 larger than the cost of ITER. However the cost is very much a moving target; who knows what Large ITER would cost, were estimated today. Here we still do consider mostly a Large ITER based breeder.

In any case, the performance of ITER would have to be upgraded by a very great amount before it could be considered as a potential economical power source. But this could be particularly difficult for ITER, even if it requires no input power. The reason is that tokamak performance has always been constrained by what this author has called ‘conservative design rules’, to be discussed in the next section. These rules indicate that a pure fusion tokamak could be extremely difficult, whereas a tokamak breeder can operate well within these constraints.

Now let’s consider NIF another multibillion dollar machine. Let’s stipulate a gain of 100 from a megajoule laser pulse (current best gains 10−3–10−2), a laser wall plug efficiency of 3 % (currently <1 %) pulse rate of 15 Hz (currently about 1–2 shots per day) generating the same 1.5 GWth, or 500 MWe, just as one would estimate for Large ITER. However each laser pulse needs 30 MJ of electricity to drive it, leaving all of 50 megawatts for the grid.

Hence even stipulating the maximum success for either ITER or NIF, we are still very, very far from economical fusion. This is the serious long term problem that fusion has. After all, if neither ITER nor NIF brings us very close to fusion, what are our sponsors paying for; another half century, or century of effort starting with their success? The great advantage of fusion breeding is that ITER and NIF sized devices could be ends in themselves rather than stepping stones to who knows what DEMO, built who knows how many decades later? Using breeding the author believes fusion could very likely make a major impact on the energy systems in our grand children’s prime (mine are 9, 12, and 14); without, it maybe their grand children’s.

Yet what are the alternatives to tokamaks and lasers? Each of these technologies were developed through a four to five decade international effort, costing billions to get to where they are today. Many other concepts have their proponents who assure us that they can achieve fusion very quickly if only we fund them at the required level. For instance Robert Burke claims that we could get there very quickly via heavy ion beam fusion [29]. In the past, the proponents of these concepts have sought government funding, but recently they have been seeking venture capital funding. One such enterprise, Fusion Power Corp. of Sacramento CA, promises commercial power in 10 years [30] via heavy ion beam inertial fusion. But the high-current ion accelerators and/or storage rings required for heavy ion fusion are at this point merely concepts that have been studied theoretically for over 30 years at Lawrence Berkeley National Laboratory (LBNL) and other labs, but never built. Their feasibility has not been demonstrated, and they would cost billions to build. Is that kind of money available in the private sector for a speculative concept? If so, we can only wish them the best of luck. General Fusion Company in Vancouver Canada, funded by Jeff Bezos of Amazon at $30 M, puts even that time scale to shame. Using magnetized target fusion [31], a concept that was studied over the past 30 years at the Naval Research Laboratory (NRL) and the Los Alamos National Laboratory (LANL), they promise break-even by 2014 and a commercial reactor by 2020.

Maybe; and of course maybe a genius will invent a commercial fusion reactor in his or her garage.

But more realistically, any other concept, starting from where it is now will probably take the same time and effort to get to where tokamaks and lasers are today. And this of course assumes that these other concepts are even as good, despite the fact that they have already been rejected in favor of tokamaks and lasers. Also without getting very far into the politics, is there really any chance that congress would appropriate many billions more for say a stellarator if ITER fails, or a heavy ion accelerator if NIF continues to fail? What with fracking and cheap gas now, do our sponsors really have the stomach for this? Who knows, in the dim distant future, stellarators or heavy ion accelerators may prove to be superior to tokamaks or lasers. But realistically, the only way we will learn this is if they are follow on projects to successes at ITER and NIF.

To this author’s mind the best hope for fusion is to get an ITER or NIF like system to be an economical power producer. The best way to achieve this is to use fusion neutrons to breed fuel for conventional nuclear reactors, likely, but not necessarily light water reactors (LWR’s). The more than order of magnitude increase in Q in the breeding blanket could provide just the boost needed to accomplish this. Let us be clear here, that by an order of magnitude increase in Q, we included the power produced directly by the fusion reactor plus the power of the fuel, which is produced by the fusion reactor, but is burned elsewhere.

But of course this results in a difficult dilemma. Fission people mostly (but not unanimously) think they have enough fuel, and are fighting for their own survival on other fronts. For instance Germany plans to mothball its 17 nuclear reactors, and hence is going through something of an energy crisis now. Also fusion people do not want to get their hand dirty. This paper attempts to take on and answer these arguments.

Section II reviews magnetic fusion. It concentrates on the tokamak in some detail. It discusses other options in much less detail, and mostly makes the case that none of these are nearly ready for prime time. Section III does the same for inertial fusion, concentrating mostly on lasers, and making the case that these are the only reasonable option at this point. Section IV reviews the nuclear issues, concentrating on fusion breeding and gives a possible road map for fusion to produce large scale power by mid century or shortly thereafter. Section V reviews the ‘energy park’, a sustainable approach to producing terawatts of economical, environmentally viable power for a world with 10 billion people. A fusion breeder produces fuel for about 5 thermal reactors, likely 5 LWR’s, of equal power, and a fast neutron reactor, also of equal power, which burns the actinide wastes. Only thorium goes in, only electricity and/or manufactured liquid fuel go out. Section VI draws conclusions.

I will close this introduction with an anecdote. A few years ago I was at an international fusion workshop. The head of the fusion effort of a small European country told all he and his group were doing for ITER. In one of my own less than stellar moments, I jumped on him, saying that he was only getting lip service, and the larger European countries did not value his contribution. He answered that they are indeed a small, but important part of ITER. Then when I got up and gave my pitch for fusion breeding, he jumped on me and said only pure fusion made sense and breeding or any other nuclear option only was a waste of effort.

By serendipity, he and I rode together to the airport. First I apologized for my remarks, and admitted that it did indeed look like his group was doing important work. Then he went on to say that he actually understood that fusion breeding or some other hybrid approach was ultimately the only hope for ITER. I asked him why he landed on me like a ton of bricks. He said, I did not understand, the German Greens are very powerful, and with any hint of nuclear, they would go on the attack against ITER, and likely derail it.

This article has no concern for the sensitivities of the German Greens.

Magnetic Fusion

There are many different possible approached to magnetic fusion. How does one compare them? There are many different potential yardsticks, but we will choose two. The first is the triple fusion porduct nTτ, where n is the density in m−3, T is the temperature in keV, and τ is the energy confinement time in seconds. At fusion temperatures, the DT fusion reaction rate 〈σv〉, is roughly proportional to the ion temperature squared. For instance at 10 keV, 〈σv〉 = 1.1 × 10−16 cm3/s, at 20, 4.2 × 10−16. Since the fusion power per unit volume is nDnTW〈σv〉, where W is the fusion energy per reaction, 14 MeV for the neutron and 3.5 for the alpha particle, this power density is roughly proportional to n2T2. However the input power density is simply nT/τ, so the ratio of fusion power to input power, the Q of the device is roughly proportional to nTτ.

Since the whole idea of magnetic fusion is to contain the energy of the plasma, the second obvious measure of a magnetic fusion device is how much energy E does the device confine in Joules. With these two measures of performance, we will compare the various magnetic confinement configurations.

We first discuss the tokamak, which is the main part of this section. Then we briefly discuss other confinement schemes. Not only is the tokamak way ahead of the other devices, many of these others have what appear to this author to be fatal flaws. These flaws are, in many cases, quite simple in concept and indeed appear to be irrefutable. The very simplicity of these flaws, combined with the many orders of magnitude in performance measure that separate these devices from tokamaks, argue very powerfully in favor of tokamaks. Alternative systems undoubtedly can provide a great deal of fundamental information on plasma confinement, but is this the goal? Is it worth the additional billions to achieve it? To this author, the goal is to achieve fusion breeding or pure fusion in the shortest possible time, and the way to get there is to concentrate on the device, which is, by many orders of magnitude, the closest to the goal.

This is not to say tokamaks are a slam dunk. There are serious obstacles that must be overcome before tokamaks can be viable fusion breeders, or pure fusion devices. However they have many fewer obstacles than any other configuration.

Tokamaks

History and Development

The worldwide magnetic fusion has concentrated on the tokamak approach for the last 40 years. Very briefly, a tokamak is a toroidal plasma carrying a toroidal current, confined in part by a toroidal magnetic field. In various modern high performance tokamaks around the world, the major radius varies from less than a meter to more than 3 m, the minor radius from 0.2 to more than 1 m and the toroidal field is typically between about 2.5 and 10 T (with 5 T being a estimate if one wishes to use a single parameter). Electron densities are in the range of 1019–20/m3, electron and ion temperatures between 1 and 15 keV and plasma currents are typically in the Meg amp range. Generally today’s tokamaks use copper toroidal field coils. However while satisfactory for pulsed experiments, these would be enormous power drains in continuous or high duty cycle operation. Hence there is wide agreement that superconducting coils would have to be used in any reactor. The plasmas is heated by Ohmic heating, but this is not nearly enough; supplementary heating is necessary. Usually neutral beams are used for heating, but various types of electromagnetic waves are also being considered.

There have been three large tokamak experiments using 40 MW or more of neutral beam to power to tokamak, JT-60 in Japan, JET in England, and TFTR in the United States; but TFTR has been decommissioned in 1998. Only the latter two have run with DT plasmas. In a brief digression, the author expresses the opinion that it was a great mistake to decommission TFTR. The logic might have been that it had achieved everything it could, but that logic did not convince the sponsors of JET and JT-60, and those tokamaks have gone on another 15 years and have achieved a great deal. The main effect of our decision is that the United States can no longer play in the big leagues in tokamak research. However the author is convinced that there is enough residual tokamak knowledge here that we can rejoin at any time. This article will propose that we do just that.

The tokamak effort has been incredibly successful. A thorough review of tokamak performance, up through the late 1990s has recently been given by the author [5]. In Fig. 1a is shown a very rough plot of the triple fusion product as a function of year (much more detail is given in Ref. [5] ). In Fig. 1b is shown the number of transistors on a chip as a function of year (Moore’s law) [32]. The latter has been called a “25 year record of innovation unmatched in history”. But the slopes of the two graphs are about the same. To this author, the period from about 1970 to about 2000 can be regarded as the golden age of tokamaks.

Fig. 1
figure 1

The nTτ in m−3 keVs for tokamaks as a function of year, compared to the number of transistors on a chip as a function of year from 1975 to 1995

However there is one important difference between Fig. 1a, b. At every point along the curve in Fig. 1b, the semiconductor industry was able to produce something useful and profitable. Furthermore, after the late 1990s that curve kept advancing, while Fig. 1a leveled off. The tokamak program, despite its success, is still several orders of magnitude away from producing anything economically viable; and the cost of the follow on projects gets very high. However for tokamaks, the period of rapid advance shown in Fig. 1a corresponded to the period of constructing larger and large tokamaks. However there have been no large tokamaks built since, so the performance, as measured by the triple fusion product is about where it was in 1997. What has improved are other figures of merit, principally the discharge time, going from say a second to tens of seconds (in JT-60) [3338].

Both TFTR and JET have operated with DT plasmas, so they have produced 14 MeV neutrons in appreciable quantities. In a particular shot, JET has generated about 1019 in a 1 s pulse [39, 40], or generated about 20 MJ of neutron energy, giving a Q of about 0.5. JT-60, in deuterium plasma has produced equivalent Q’s higher than unity [33, 34] by using what they called a W shaped diverter. However, for one reason or another, these plasmas all terminated prematurely (after about 1 s). In plasma which persists as long as the beam, the Q is much smaller, typically about 0.2. In Fig. 2, redrawn from Ref [39] are shown neutron rates from two separate discharges in JET, one which terminates and one which persists. It would seem that one very important milestone for both JET and JT-60 would be to see if it can achieve a persistent Q ~ 1 discharge.

Fig. 2
figure 2

Neutron production rate from JET as a function of time for steady state and hot ion modes of operation

The JT-60 web site specifies that it has achieved a triple fusion product of 1.6 × 1021 m−3 keVs, and contains plasma energy of about 9 MJ. The web site shows the progress of nTτ as a function of year. It shows a large advance, up to the late 1990s, followed by about 15 years of flat line.

The International Tokamak Experimental Reactor (ITER)

Since follow on experiments to TFTR size tokamaks get so large and expensive, the world has made a decision to cooperate on the construction of a single large reactor ITER. It has had a troubled political and economic history. Its origin was in the 1985 summit meeting between President Ronald Reagan and General Secretary Mikhail Gorbachev. They proposed cooperating on a single large joint tokamak. It was originally called INTOR for international torus. Design work began on it but few decisions were made. In 1999, the project was to be an 8 m major radius machine, in which about 150 MW of neutral beam heating would generate about 1.5 GW of neutron power for a Q of about 10 [28]. The cost then was estimated at about $20 billion, $10B for construction, and then another $10B for operating expenses over a 10 years period. The international partners were the Soviet Union (and subsequently Russia), the United States, the European Community, China, Japan and South Korea.

Subsequently the United States, deciding that the project was too expensive, pulled out. The remaining partners ultimately decided to agree on a smaller, less expensive machine. The new ITER was to be a 6 m major radius machine which would use about 40 MW of beam power to produce about 500 MW of fusion power [22]. The cost was estimated at about $10B, now $5B for construction and $5B for operation over 10 years. The United States rejoined in 2003. However there was now another controversy; where to build it. Both the European community and Japan put in strong proposals to host it. The United States, Japan and Korea voted for Japan, while Europe, Russian and China voted for Europe (Carderache, France). With the vote tied, there was a standoff, which lasted for several years. Finally in 2005, the partners decided to construct ITER in France. By this time, India joined the consortium, bringing the total number of partners to 7. Shown in Fig. 3 is a schematic of ITER taken from its web site (www.iter.org).

Fig. 3
figure 3

A schematic of ITER take from its web site (Color figure online)

However with agreement on the site, ITER’s problems were far from over. The construction cost was greatly underestimated. The original 5B euro estimate in 2006 has more than tripled to over 16B euros in 2010, and the completion date has slipped from 2016 to 2019 [23]. This undoubtedly motivated Senator Diane Feinstein’s comment cited in the introduction. However at least since 2010, according to the ITER web site, the completion date has not slipped any further.

Whether construction of ITER was a wise decision or not, ITER is what we have got. The only realistic option for magnetic fusion is to get it to work as well as possible. On the ITER web site, there is talk of a DEMO as a follow on project. This would potentially generate electricity for the grid. However as we will show in the next section, in order to do so, this DEMO must get over a series of obstacles, which have constrained tokamak operation for half a century. It is extremely important that while these obstacles prevent pure fusion, they do not prevent fusion breeding.

Conservative Design Rules (CDR’s) for Tokamaks

The Rules

Tokamak operation is subject to four very simple parameter constraints, which are well-grounded in theory, extensively confirmed by experiment, and generally accepted by the fusion physics community. Although these constraints have been known for decades, there has not been much discussion of their impact collectively on fusion reactor operation. We have coined the term “Conservative Design Rules” (CDRs) to describe this set of constraints, and in Ref. [8] we discussed in detail the limits that these rules, taken collectively, impose on fusion power output. We emphasize that the Conservative Design Rules are not controversial. The paper [8] has been in the literature 5 years; its conclusions have not been challenged. The author has discussed the CDRs in seminars at MIT, University of Maryland, NRL, and APS headquarters, again without challenge. Thus, there is a heavy burden of explanation and proof on any proponent of pure fusion who assumes reactor performance in excess of the CDRs.

The other thing about conservative design rules is their simplicity. One does not have to know about any of the complexities of tokamaks; fishbones, transport barriers, grassy ELM’s, ITGI’s, density pedestals, etc. In fact the Sam Cooke approach (“Don’t know much about tokamaks…”) suffices. There are surely much more complicated physical effects further limiting tokamak behavior (i.e. the afore mentioned), but conservative design rules bound all of these, are very simple, and are well established in the tokamak science base. But they have not been emphasized in the literature, nor have their very important implications for fusion been discussed.

It is important to note that conservative design rules have nothing to do with transport. Good transport cannot improve things, bad transport can only make things worse. In fact even if there is no input power, for instance an ignited tokamak, conservative design rules still place the same limits on the fusion power a tokamak can produce. They almost certainly prevent economical pure fusion.

The conservative design rules were discussed in detail in Ref. [8]. Here we will give a shortened version and leave additional details to the reference.

Any tokamak run as a reactor can in all likelihood withstand existing levels of transport. What it cannot tolerate are many (or even any) major disruptions. Thus in the relevant parameter space, there is a boundary separating regions where a tokamak may disrupt. A commercial reactor should operate as far from this red zone as possible, thereby motivating the author’s term ‘conservative design rules’. While disruptions are still not yet fully understood, they are almost certainly rooted in MHD (ideal and resistive) effects in the plasma. MHD instabilities are driven by the current and pressure gradient. The first and most important design principle concerns the plasma beta.

To simplify the discussion here, we will assume the plasma has circular cross section in the poloidal plane, more general configurations are discussed in Ref. [8].

Troyon and Gruber [41, 42] achieved a theoretical breakthrough in understanding the pressure limit. They determined that the maximum beta was governed by what they called the maximum normalized beta βN. In terms of βN, the volume averaged plasma beta β was given by

$$\upbeta = 10^{ - 2}\upbeta_{\rm N} \text{I}/\text{aB}$$
(1)

where I is the current in Megamps, a is the minor radius in meters, and B is the toroidal field in teslas. Their calculations gave a value for βN, and from this β could be determined. If the plasma had no wall stabilization, they found a maximum stable βN of about 2.5 or a little greater; and with strong wall stabilization, it might be as large as 5.

In our conservative design, we will neglect wall stabilization. In a DT tokamak reactor, the wall is doing enough; absorbing and multiplying neutrons, dissipating heat from fast ions and radiation, being one end of a heat exchanger and breeder of 233U and/or T, etc. Furthermore tokamaks always operate with either divertors or limiters, so the wall cannot get that close to the plasma in any case. Hence we take Troyon’s most conservative value, since it will be furthest from the disruption threshold. Thus we take for our first principle of conservative design the condition that βN is 2.5 or less.

To make further progress while keeping our analysis as simple as possible, we assume a density and temperature profile for the plasma. For circular cross section with minor plasma radius a (more complicated geometries are analyzed in [8]), we take parabolic profiles

$$n_{e} = n_{o} \left[ {1 - \left( \frac{r}{a} \right)^{2} } \right]$$
(2a)
$$T_{i,e} = T_{i,eo} \left[ {1 - \left( \frac{r}{a} \right)} \right]^{2}$$
(2b)

where ne is the electron density, assumed to be twice the deuterium and tritium density and Ti,e is the ion (electron) temperature. The spatially average density is no/2. The pressure is the product of the two, and the spatially averaged pressure is no(Teo + Tio)/3. Of course there may be effects from different profiles but they should not be major. For instance at the average beta, the center temperature may be higher (giving more fusion power) but cover a smaller average volume (giving less fusion power). In fact most profiles do not fill out nearly as much as a parabolic profile does, so the parabolic profile choice is rather optimistic.

If density and temperature are totally unrestricted, 〈σv〉, as a function of temperature, has a broad maximum at a temperature of Ti ~ 50 keV. However if β, that is total plasma pressure is held constant the maximum is at lower temperature, because this means the density is higher. Since β depends on Te + Ti, whereas the fusion rate depends only on Ti, we must make some assumption here. We assume Te = Ti/2, as is often characteristic of today’s beam heated tokamaks. (If the temperatures were equilibrated, the neutron power would be lower, obviously one can do this calculation for any electron temperature).

Calculating the neutron power by integrating over volume 〈σv〉 times the tritium times the deuterium density times the neutron energy, we find that the it is

$$P_{n} = 2.2 \times 10^{22} (2\pi R)(2\pi a^{2} )\frac{{n_{o}^{2} }}{4}\chi (T_{io} )$$
(3a)

where we have assumed nD = nT = no/2 (recall the units of no are 1020 m−3; Pn, MW; and lengths, meters; and temperatures, keV)

$$\chi (T_{io} ) = \int\limits_{0}^{1} {udu < \sigma v(T_{i} (u))(1 - u^{2} )^{2} }$$
(3b)

If beta is specified, then the density is proportional to T−1. The function χ(Tio)/T 2io is plotted in Fig. 4. It has a maximum at Tio of about 15 ≡ Tio(β). To get the average reactivity for the plasma, just multiply the ordinate by T 2io . Now expressing the density at which the maximum fusion rate occurs, we get

$${\text{n}}_{\text{o}} (\upbeta) = \frac{{\upbeta_{{\rm N}} {\text{IB}}}}{{2{\text{aT}}_{\text{io}} (\upbeta)}}$$
(4)

To determine no(β), note that the maximum βN can be is 2.5, consistent with out first design principle.

Fig. 4
figure 4

A plot of the quantity χ(Tio)/T 2io as a function of central ion temperature. It is proportional to the neutron power for a plasma of fixed pressure

Now we introduce the second conservative design principle. Decades of plasma experience have shown the tokamaks cannot operate at densities above the Greenwald limit [43, 44] ). While this is more of an empirical law than one grounded in solid theory, it has held for two decades already. The Greenwald density limit (equal to no/2 for our assumed parabolic density profile) is given by

$${\text{n}}_{\text{G}} = \frac{1}{{\uppi{\text{a}}^{2} }}$$
(5)

However the failure mode in approaching the Greenwald limit is often a shrinking of the plasma profile followed by a major disruption. Since major disruptions are basically intolerable in any reactor, we take as our second principle of conservative design that the density cannot be above ¾ of the Greenwald limit. To simplify the discussion, we do not consider further the Greenwald limit here; it places an additional restriction on the density as fully discussed in Eq. (8). Considering only the beta limited density gives a good idea of the constraints the tokamak operates under.

Both density limits depend on the current, and if this could increase indefinitely, there would be no problem. But from ideal and resistive MHD, we know that q, the safety factor equal to Br/BpR, where B is the toroidal field, Bp is the poloidal field, at the limiter or diverter, r = a, is nearly always greater than three. This then is the third principle of conservative design, namely that q(a) > 3. For tokamaks of circular cross section, the relations then simplify considerably since one can express q(a) very easily in terms of current. For circular cross section, the beta limited density is simply

$${\text{n}}_{\text{o}} (\upbeta) = \frac{{5\text{aB}^{2}\upbeta_{\text{N}} }}{{2{\text{Rq}}({\text{a}}){\text{T}}_{\text{io}} (\upbeta)}}$$
(6)

and q(a) is given by

$${\text{q}}({\text{a}}) = 5{\text{a}}^{2} {\text{B/IR}}$$
(7)

In this case, the maximum density depends only on magnetic field and aspect ratio(or geometry if the cross section is not circular) as well as βN, taken to maximize at 2.5, and q(a) taken to minimize at 3. Equation (6), as well as Eq. (3a) and Fig. 4 then give the maximum fusion power any tokamak can generate if it is limited by conservative design rules.

Notice that confinement does not come into these principles at all. This is not to say it is unimportant; the confinement and transport determine the external power needed to maintain the plasma profiles. However even if there were no losses (or else for instance an ignited plasma), these three design rules put serious constraints on what a tokamak can and cannot do. Good confinement cannot make things better, bad confinement can only make them worse.

We now discuss the fourth conservative design rule, the size of the blanket. For ITER and Large ITER, reactor sized tokamaks, our assumption is that the existing designs have room for an appropriate blanket which absorbs neutrons, breeds tritium, handles the heat load, etc. But if one wants to build a smaller tokamak, such as the scientific prototype, which we will discuss shortly, how thick does the blanket have to be? Here, the author has little expertise so only very qualitative matters are considered. The mean free path of neutrons with energies between about 1 and 14 MeV is about 15 cm in lithium, and about 6 cm in beryllium and thorium. All of these are important blanket materials for either pure or hybrid fusion. The mean free path for breeding and slowing down is even longer. Obviously the blanket has to be many mean free paths thick so as to prevent neutron leakage out the back, if one desires long life of the machine. Also one clearly desires to prevent activation of materials behind the blanket. Behind the blanket is usually a neutron shield, which itself is not thin.

Many references on fusion hybrids show schematics of the reactor along with the 2 m man standing along side it, and the blanket is about his size. Rarely are dimensions given. One exception is a rough schematic of a blanket shown in Ref. [45], reproduced here in Fig. 5. In this schematic, the blanket is between 1.5 and 2 m thick, and presumably there is no long term neutron leakage or activation of materials in back. Lidsky [45] when discussing a blanket for fission suppressed thorium cycle postulates a blanket 80 cm thick for just the fertile material. Hence, as a very rough rule of thumb, we will specify that the blanket has to be 1.5 m thick. We will call this the fourth conservative design principle. It applies only where one wishes to design a small (i.e. less than commercial size) reactor, and it imposes a certain minimum size on the experimental device which strives for steady state operation with DT. This design principle is more approximate than the other three, and it may be possible to design thinner blankets.

Fig. 5
figure 5

A schematic of a tokamak and its blanket showing the width of about a meter and a half

To summarize, the conservative design rules are:

$$\upbeta_{\text{N}} < 2.5$$
(8a)
$${\text{n}}_{\text{o}} < 1.5{\text{I}}/\uppi{\text{a}}^{2}$$
(8b)
$${\text{q}}\left( {\text{a}} \right) > 3$$
(8c)
$${\text{Blanket}}\,{\text{thickness}} > 1.5\,{\text{meters}}$$
(8d)

That is it. What could be simpler? All one needs is the toroidal field, temperature ratio and geometry, and conservative design rules determine the maximum fusion power a tokamak can produce.

Once the magnetic field and geometry are specified, this specifies the minimum q(a), Eq. (8c), that is, the maximum current. If the tokamak is pressure limited, Eq. (8a) specifies the maximum total pressure, no(Teo + Tio)/3. Take the ion temperature of 15 keV, make an assumption of the electron temperature and one has the no, and from no, the neutron power. For a circular cross section and beta limited density, the formula for the maximum power allowed by conservative design rules is very simple;

$${\text{P}}\left( {\text{MW}} \right) = 0.11\left[ {{\text{a}}\left( {\text{m}} \right){\text{B}}\left( {\text{Tesla}} \right)} \right]^{4} /{\text{R}}\left( {\text{M}} \right)$$
(9)

Here we have assumed Ti = 2Te, as is typical for today’s beam heated tokamaks. However temperature equilibration is likely a more reasonable assumption for a reactor. Then, since more of the allowed pressure is taken up by nonreacting electrons, the fusion power is less. The density is reduced by ¾ and the power by 9/16. This then gives the maximum fusion power a tokamak can give. More complicated geometries are treated in [8], but very roughly, the power is increased by the elliptical elongation factor k.

If the ion temperature could be maintained at twice the electron temperature, it would certainly be advantageous. One way this might be accomplished is through what is called alpha channeling [46, 47]. Waves and particles in a tokamak interact in a complicated way in which the interacting particles move on some prescribed path in the six dimensional phase space. The idea of alpha channeling is to find a wave or set of waves, which move the alpha particles from the center of the tokamak to the edge, while at the same time significantly lowering their energy. The energy thus goes from the alphas to the waves, which can then be used for other purposes. The most obvious purposes are current drive and heating of the plasma ions. (The alpha particles preferentially heat the electrons by collisions.) In this way it might be possible to maintain an ion temperature twice the electron temperature in a reactor. As a total reach, it might be possible convert these waves to other waves which can exit the plasma and thereby harvest their energy directly, but this is obviously very speculative. So far preliminary thought is to use ion Bernstein waves driven by lower hybrid waves, and Alfven waves.

If the density is Greenwald limited rather than beta limited, the maximum power is reduced [8]. Basically where the density has to be lower than the beta limited optimum, the fusion power optimizes if the ion temperature is higher than 15 keV. By applying these conservative design rules to existing and proposed tokamaks in the next few sections, we will see that the maximum neutron power that can be expected from any proposed tokamak is well below the minimum level needed for economical power production.

Unlike the argument given in the Introduction where driver power and its efficiency, and output power and its efficiency in converting to electricity, were considered, conservative design rules are independent of all of that. They show the maximum power a tokamak can deliver, even a burning plasma with no power to drive the plasma or current (Q = infinity). Let us make this extremely optimistic assumption. However even then, the capital cost involved in building such a tokamak such as Large ITER, we optimistically assume at least $25B for a 500MWe reactor, and likely a large operating cost to boot, will in practice render it unsuitable for economical power production. But we will also see that tokamak breeders can function well within these limits, even though their Q is ‘only’ 10, well below infinity.

Recent Tokamak Experiments in the Light of These Design Principles

We summarize results here for TFTR and JT-60 in the light of conservative design rues. Results for other tokamaks, including JET, D3-D and Asdex are summarized in [8]. No tokamak up to now has exceeded the limits imposed by conservative design rules; in fact conservative design rules overestimate fusion by at least a factor of two in all cases.

TFTR: Results from TFTR were summarized in [48] Briefly it achieved spectacular results when operating in the hot ion supershot mode. These have peaked profiles, and the beam is important for both heating and fueling the plasma. It achieved a maximum fusion power of 10 MW for perhaps half a second. However it terminated by rapidly dumping all or a significant part of the plasma energy. The major radius was 2.6 m, the limiter radius was 0.9 m and the magnetic field was over 5T. This is all that is needed to get maximum parameters of the device according to the conservative design principles.

In Ref [48] there was a table of parameters of 4 supershots. A portion of the table is presented in Table 1. The rows in bold are from Ref [48], the rows in ordinary type are from conservative design principles. The central ion temperature is much higher than the optimum value of 15 keV, but the beta is still consistent because the hot part of the plasma is so narrow compared to the parabolic profile we have been assuming. In fact, their measured βN’s are smaller than what we have assumed in the conservative design. While they do not give q(a), for their circular cross section one can estimate it easily enough. In all cases q(a) > 3, so that the results are consistent with conservative design principles in this respect. There are two rows for the calculation of the central density from conservative design rules for q(a) = 3 and βN = 2.5, no(β) and no(G). For TFTR, the former is slightly smaller, so the density is beta limited, not Greenwald limited. In the last row of the table is given neutron power from Eq. (9). Notice that the neutron power per the conservative design is at least double the actual neutron power observed. Hence even though TFTR managed to get a much higher ion temperature than 15, it did not help very much. The reaction rate was higher in this region of high temperature, but the density was lower, and the volume of strongly reacting plasma was also smaller; the net effect being less neutron power than the conservative design rules would specify. Thus for TFTR, the conservative design rules over estimate the fusion power, typically by at least a factor of two.

Table 1 Summary of parameters of hot ion shots in TFTR (bold type). Prediction of beta and Greenwald limited central density and predicted neutron power according to conservative design rules (regular type)

In Ref. [8] we show that conservative design rules also overestimate the neutron power in JET, in both hot ion and long lived modes, by at least a factor of 2.

JT-60U: JT-60 and more recently its upgrade JT-60U is the largest of the tokamaks, but so far, has not operated with DT (http://www-jt60.naka.jaea.go.jp/english/index-e.html). Its parameters are a major radius of 3.4 m, a minor radius of 1.2 m (to the vacuum wall) and an elongation of about 1.4. The maximum magnetic field is about 4.2 T. Although JT-60U has not yet operated with DT, it has operated with DD plasmas, and from the DD neutron rate, they extrapolate to get the DT rate. In all their reported data, as regards βN (virtually always equal to or <2.5) q(a) (virtually always >3) and n/nG (virtually always <0.75), their results are consistent with the conservative design rules.

A great deal of their earlier effort consisted in developing what they called a W shaped diverter. Here, they reported their largest amount of fusion power, with the equivalent Q in a DT plasma going above unity, and with a great deal of the improvement coming from the new diverter. Figure 6 redraws their plot of equivalent Q vs current from these references. It reaches a maximum of 1.25, nearly double their previous results. However they point out that these are all transient results. In quasi steady operation, their DT equivalent Q’s were below 0.2. This result is similar to the experience of JET.

Fig. 6
figure 6

Some JT-60 data confirming conservative design rules. A plot of DT equivalent Q as a function of current for JT-60 for normal and W shaped divertors. A scatter plot of tokamak operation points in the parameter space of q(a) and normalized beta. Long lived discharges only operate in the region specified by conservative design rules. A plot of H factor as a function of density over Greenwald density, confirming another aspect of conservative design rules. A plot of normalized beta as a function of time in a long lived discharge

In their results for quasi steady plasmas, plasmas lasting longer than 5 times the energy confinement time, all their q(a)’s were greater than 3 and all their βN’s were <2.5. A scatter plot of their data is shown in Fig. 6.

Regarding density in their earlier results, they were always below the Greenwald limit. In Fig. 6 is shown a redrawn plot of their H factor, the fractional increase in confinement time when they operate in the H mode, as a function of n/nG. They have a single point at 0.8, at the worst confinement, and virtually all of their data is for 0.4 < n/nG < 0.6.

In their later results, they emphasized long time operation. This involved getting bootstrap current of over 50 % sustained for a long time and a βN sustained for over 20s. Shown in Fig. 6 is a plot of βN as a function of time is redrawn. While sustained for long time, it is still no greater than 2.5. Their q(a)’s were everywhere greater than 3, and their maximum densities reported were at about 0.5nG.

The Upshot

To summarize, the conservative design rules are very well based in theory and so far have constrained tokamak operation. In fact so far, as regards neutron production, a tokamak is doing well to achieve half the neutron rate specified by conservative design rules. To get powers like 3 GWth, as needed in a commercial reactor, but in a tokamak smaller, much higher Q, and less expensive than large ITER would stretch conservative design rules well beyond the breaking point. This then is the basis for the author’s assertion that commercial pure fusion reactors based on tokamak configurations are unlikely, at least until one can find a way around conservative design rules. However as we will show shortly, a breeding tokamak can operate well within the limits of conservative design rules.

Disruptions and Their Implications

To get an appreciation of the disruption problem tokamaks have, it is instructive to look at the energies involved. In the case of hot ion modes, JT-60, JET and TFTR typically dump about 15 % of the plasma energy abruptly on the walls. As JT-60 confined about 10 MJ, this means in these partial disruptions, about 1.5 MJ is dumped, presumably mostly in the form of radiation and energetic ions and electrons. This is significant energy to dump in an uncontrolled manner, nearly a pound of TNT.

In a worst case scenario, if the plasma density is sufficiently low, the disruption energy could be, and has been manifested as relativistic electron ring. It could have much more than 1 MJ energy. Once formed, there is nothing in the plasma to stop it, it will eventually hit the wall; and this could be enormously destructive.

But this is only for a partial 15 % disruption. Major disruptions dump all of the energy of both the plasma (10 MJ in JT-60) and poloidal field (about the same as the plasma energy), so in this case about 20 MJ of energy, in one form or other, are suddenly dumped on the wall. This is about 1.25 % of the total magnetic field energy for each channel (plasma and current) for about 2.5 % total. The plasma beta is likely quoted as considerably larger than 1.25 %, but then there is a good bit of magnetic field energy outside the plasma.

To estimate the energy available for disruptions in future devices, we will take as a rule of thumb that it is 2.5 % of the total magnetic field energy. Now the energies available get serious. For ITER it is about 160 MJ, about the energy of an 80 pound bomb, and for large ITER, which this paper sees as an optimal end product, it is 550 MJ, the energy of nearly a 300 pound bomb. If more energy is stored in the plasma than our 2.5 % of total magnetic energy, this is good for fusion, but makes disruptions still more dangerous.

But this is only the beginning of the problem. An 80 pound bomb equivalent going off in the confined space of ITER might cause the superconducting toroidal field coils to uncontrollably quench. This could happen. It has happened in the Large Hadron Collider in CERN, bringing the accelerator down for a year. However the CERN tunnel has a huge volume compared to the interior of ITER. An uncontrolled quench in ITER would turn the 80 pound bomb into a 3200 pound bomb. It is unlikely that the building, or much around it could survive. It is important to note that ITER stores an enormous amount of energy.

But hasn’t JT-60 shown that tokamaks can have very long pulses? See Fig. 6 for example. The answer is yes and no. Tokamaks can have very long pulses, but that certainly does not mean that they never disrupt. A very interesting statistical analysis of disruptions on JET has recently been published [49]. One of their important figures of merit is the disruptivity, the rate of disruptions in s−1, as a function of various parameters. Shown in Fig. 7 are disruptivities vs inverse q(a) and density over Greenwald density. There is a similar curve for disruptivity vs normalized beta.

Fig. 7
figure 7

The disruptivity in JET as a function of Greenwald fraction and 1/q(a). There is an abrupt jump once conservative rules are violated

There are two important take conclusions from this data. The first is that it confirms conservative design rules. Disruptivities rise sharply, by well over an order of magnitude as conservative design rules become violated. Secondly, the harmless looking disruptivities at low normalized beta, reciprocal q and density are not harmless at all. A disruptivity of 10−2 means a disruption every 2 min! Clearly this is unacceptable in a reactor.

Probably the most important job for both ITER and the scientific prototype (to be discussed shortly) is to find a fusion relevant regime where the disruption rate is close enough to zero that a reactor is possible. Clearly this will be much easier to do within the confines of conservative design rules, where a breeder can operate. As of now, there is no reason to believe that a disruption free regime exists outside of the conservative design rules where a pure fusion reactor must run.

A large airplane at speed and altitude has enormous kinetic and gravitational potential energy (in fact about the same as the stored magnetic energy in Large ITER), but people are inside, and the aircrew is able to control it. This is a reasonable analogy for tokamaks, and this author’s expectation is that its stored energy also can ultimately be safely controlled, but only within the constraints of conservative design rules.

The Scientific Prototype and Neutron Power of Various Tokamaks

In all his work, the author has argued [510, 50] that the next logical step for the American Magnetic Fusion Energy program is a steady state tokamak, called the scientific prototype, about the size of TFTR, run with a DT plasma, achieves Q ~ 1 and which breeds its own tritium. After all tritium self sufficiency ultimately is absolutely essential, for either pure fusion or fusion breeding. If we do not tackle it now, then when will we? The scientific prototype is not some small thing to add to the American base MFE program; rather it is a very large project which would replace the entire base program. The tremendous resource of theoretical, experimental and computational expertise in the American base program would be refocused on this single large project. If the ~$350 M per year that comprises the base program were focused on the scientific prototype for 15 years, that would make over $5B available, not much less than the cost of a 1GWe LWR.

There are two crucial plasma physics problems, which must be solved by the scientific prototype. The first is steady state operation. Ideally this would mean no Ohmic current drive. However in practice, a small amount of Ohmic drive may be unavoidable, and may even be acceptable if the plasma on time is long compared to the off time necessary to recharge the transformer. This is particularly the case if the tokamak is viewed principally as a fuel factory and only secondarily as an on line power source. JT-60 has already demonstrated discharge durations of much more than 5 energy confinement times at the required normalized betas, q(a)’s, and density. One of their runs, at a normalized beta of 2.5, for half a minute is shown in Fig. 6. ITER is designed to operate with 400 s pulses, about 100 energy confinement times. But in all of these, the current is driven at least in part Ohmically.

Recently Luce [51] wrote a review article showing the advances in achieving steady state behavior. At lower power, some tokamaks with superconducting toroidal field coils have run for very long times. The question is whether this can be extended to more powerful tokamaks. Another problem he sees is that the power required to drive the current (Watts per Amp) might be too high. However Luce envisions only pure fusion. If fusion breeding is the goal, where the fusion reactor produces ten times the neutron power in 233U fuel, a bit more power to drive the current is not necessarily such a big deal.

The second crucial plasma physics goal to achieve in the the scientific prototype would be to eliminate or greatly reduce disruptions. Where a disruption in the scientific prototype would release only a small fraction of the energy released in a major disruption in ITER, it could be an ideal vehicle for his study. To achieve these twin plasma physics goals will take every bit of expertise in the American MFE effort, as well as whatever help we can get from the rest of the world. As previously stated, this is no cake walk, but the author believes it is achievable.

The scientific prototype will be a prodigious source of alpha particles even if not enough to dominate the heating. Hence it could also be a laboratory for testing ideas of alpha channeling [46, 47], at least the basic principles even if there is not enough alpha particle power to heat the plasma very much.

While proposing this in the context of fusion breeding [510], where the blanket would also breed 233U; it is also a viable concept for pure fusion, where the blanket would only have to breed tritium [50]. Since Q ~ 1, the Quixotic attempt to do a small scale ignition experiment would be abandoned. Let ITER and NIF handle the burning plasma and ignition issues. The American MFE effort has proposed many of such burning plasma experiments, CIT, BPX, IGNITOR, FIRE…,. Would we have been better off proposing the scientific prototype instead? We could hardly have been worse off. None of these burning plasma experiments sold, and the United States has had no large (i.e. the size of TFTR, JT-60 and JET) tokamak experiment for more than 15 years now.

While the scientific prototype would not be a burning plasma, it still would produce alphas, and unlike ITER, would do it on a steady state, or at least a high duty cycle basis. Hence it could also investigate extraction of the alpha energy on a small, but steady state basis.

The scientific prototype cannot be exactly like TFTR because its major radius is only 2.6 m, and the center is filled with all sorts of stuff (e.g. toroidal field coils, etc.). Thus we must take a larger major radius. We take a major radius of 4 m, like TFTR, but now leaving room for a 1.5 m blanket. The minor radius is 1.3 m, so as to keep the aspect ratio as in TFTR, and would keep the circular cross section as TFTR did (although other geometries could obviously be considered). The hope is that it would produce about 40 MW of neutron power with about a 0.25 MW/m2 neutron wall loading. The scientific prototype would reverse conventional wisdom, which has proposed the long list of ignition experiments. Instead of sacrificing every tokamak figure of merit for ignition, the scientific prototype would sacrifice ignition for a significant advance in every other figure of merit, i.e. steady state operation, eliminating disruption in fusion relevant regimes, extended DT operation, tritium breeding…. No other country, at this point is ahead of the United States here, we have plenty of nuclear expertise, the project seems achievable, and it is hard to see how either pure fusion or fusion breeding could advance very far without the knowledge provided by the scientific prototype.

In Table 2 are shown parameters of JT-60, the scientific prototype, ITER, and large ITER, the latter two taken from Refs. [22, 28]. The scientific prototype gives about the 55 MW of neutron power according to conservative design rules. If one takes the rule of thumb that actual tokamaks produce about half the power of the conservative design estimates, as with TFTR and JET, then the estimate of 20–40 MW seems reasonable for the scientific prototype.

Table 2 Summary of parameters for JT-60, the scientific prototype, ITER and Large ITER

The conservative design principles show ITER and large ITER both doing better than actual predicted design values. However if one takes the typical estimate that predicted neutron power is about double the best achieved, large ITER gives about the design value, but ITER does especially well. It may well turn out to do somewhat better than expected. It has higher current density than either large ITER or the scientific prototype. But despite this larger current density, since q(a) is 3.5 according to Ref [22], the current could still be increased by about 15 % and still remain consistent with conservative design rules, so power might be increased by 30 %. Note that for both ITER and Large ITER, the Greenwald density is considerably less than the beta optimized density, meaning that the ion temperature has to be considerably more than 15 (22 for the former, 36 for the latter), see Ref. [8] for a more complete discussion of the case where the density is Greenwald limited rather than beta limited. In any case, when quoting power levels expected for ITER and large ITER, we stick to those calculated by the designers, 500 MW and 1.5 GW.

Several alternate American MFE proposals seem to be in the works and under serious discussion. We seems to be on the threshold of proposing a large stellarator effort, but this a bad idea. Two large American stellarator projects have already been canceled. We are hopelessly behind the Germans and Japanese in this effort and have no prospect of surpassing them. Better to let them run with the ball and support their stellarator projects, if they would support the scientific prototype. Another idea attracting interest is a larger spherical tokamak experiment. This is if anything a worse idea. Later, this paper will argue against such a proposal, as well as several other alternative concepts.

Stellarators

Our discussion of stellarators and other fusion devices, will be very brief, and will mostly make the case that none are viable alternatives to tokamaks at this time. Also this section, and subsequent sections on MFE, will site many fewer references. Information given here, which is not specifically cited, can generally be confirmed by doing a Google search of the device and going to the appropriate web site; using the information directly on the web site, as well as various articles and reports linked to it.

Despite this negative introduction, this author has no reason to think stellarators cannot ultimately work, they just have not so far. If they could work, they would have real advantages. The rotational transform is produced by external coils, so there is no plasma current. Current is one of the two causes of MHD instabilities, pressure gradient being the other. Hence disruption should be much less of a problem than in tokamaks, but it is worth noting that the stellarator like a tokamak, also stores a tremendous amount of energy.

Counterbalancing these advantages, is the fact that a stellarator is an inherently three dimensional configuration, unlike a tokamak which is a two dimensional configuration. In my view, as a theoretical physicist of normal capability, who has trouble analyzing a complex two dimensional configuration, anyone who can analyze a complex three dimensional configuration is a physicist of rare and special talent. Not only is the problem much more difficult, the parameter space much greater. For instance Wendelstein 7-X has 5 bumps going around the major circumference. Why not 4 or 6? Also because of the three dimensional configuration, neoclassical losses are greater than in a tokamak. A great deal of effort in the stellarator community is dedicated to coming up with configurations, which minimize these losses. Because of its three dimensional nature and the demands of minimizing losses, the coil configurations are very complicated and must be engineered to very precise tolerances. As difficult as it may be to fit a blanket around a tokamak plasma, an it might be even more difficult in a stellarator.

So far the largest stellarator is the Large Helical Device (LHD) in Japan (http://www.lhd.nifs.ac.jp/en/). To this author’s mind, its achievements are not very encouraging. It seems to run in either one of two modes, a low density, high temperature mode, or a high density, low temperature mode. It has not been able to achieve high density and high temperature so far. The LHD web site gives the triple produce as 4 × 1019, about 1/40 that of JT-60. Also the web site gives the maximum contained energy so far as 1.4 MJ, about 1/6 that of JT-60.

The Germans are constructing a larger stellarator still, Wendelstein 7-X. The construction is expected to be finished in 2014, with first plasma in 2015. They hope to achieve both high temperature and high density simultaneously, and they hope to achieve a triple fusion product of 4 × 1020, 10 times higher than LHD, but still below JT-60. A schematic of the plasma and some of the field coils is shown in Fig. 8.

Fig. 8
figure 8

An artist’s conception of the plasma configuration and some of the magnetic field coils of the Wendelstein 7-X stellarator taken from its web site (www. ipp.mgp.de). Notice the five bumps around the circumference and the complicated coil structure (Color figure online)

The nearly infinite parameter space available for stellarators may be an opportunity, but it also poses a programmatic danger. How can we explore all of it with our very finite resources of dollars and time? LHD is a disappointment, so build Wendelstein 7-X; if that does not work, build a heliotron, if that disappoints, build a torsotron…. Where does it end?

While some in the United States argue that we start our own stellarator program, this is a bad idea. We are hopelessly behind the Germans and Japanese. Better for now to let them run with the ball and help them any way we can, at least until they can prove that a stellarator is definitely superior to tokamaks.

Spherical Tokamaks (ST’s)

The spherical tokamak is like a tokamak except that it has an aspect ratio of approximately unity, so the plasma is nearly spherical is shape. Shown in Fig. 9 is an image of the ST plasma in the Culham Laboratory, MAST. It is nearly spherical with a radius of about 1.5 m (http://www.ccfe.ac.uk/MAST.aspx). However notice that the toroidal field coils all merge to a narrow conductor running down the center, a conductor whose radius is about 10–20 cm judging from Fig. 9. One of the strongest assets of the ST is that it runs at a much lower magnetic field and much higher beta than a tokamak.

Fig. 9
figure 9

A photograph of the spherical Tokamak discharge in MAST, taken from its web site (Color figure online)

The ST at PPPL is called NSTX. This author has not found information on the web sites of either MAST or NSTX, which allows one to easily discern the energy contained or the triple fusion product (unlike the JT-60 and LHD web sites, where this information is prominently displayed). However recent studies of confinement enhancement by lithium walls give this information [52]. The presence of the lithium walls in NSTX increases the confinement time from about 35 ms to about 70. The energy contained is about 150 kJ, or about 20 kJ/m3, and the triple product is about 5 × 1018. This is about where tokamaks were 30 years ago according to Fig. 1. On a more detailed plot [5], one can see that NSTX is about where ASDEX and T-10 were in about 1985, but about an order of magnitude below PLT then. Thus ST’s have a long way to go before they are in the league of tokamaks, and possibly will not get there at all.

But they have a much bigger problem than their late start in the race. How will the center post handle the intense flux of 14 MeV neutrons in a reacting plasma? According to the fourth conservative design rule, plasma facing surfaces should have a depth of about a meter and a half! Certainly the coil cannot be superconducting, so the field coils will necessarily dissipate a tremendous amount of energy. And how does one cool the center post?

There have been two proposals for much larger ST’s in reacting plasmas, GA’s Fusion Development Facility (FDF) (Google GA Fusion Development Facility), and the University of Texas’s waste burner; their plasma parameters are similar. Here we briefly discuss the former. GA proposes the ST as a way to study a reacting plasma and study wall material in an intense neutron flux. They expect it to generate 300 MW of neutron power in steady state, breed its own tritium, and produce a wall loading of 2 MW/m2 of neutron flux with a proposed upgrade to 4.4 MW/m2, the minimum they say is needed for pure fusion.

The toroidal field coils will be copper and will need 500 MW of power to run; the FDF will need its own dedicated power plant next door! While GA does not propose the FDF as a prototype for an economical reactor but rather as a neutron source; still their proposal does not exactly increase one’s confidence that the ST will ever evolve into an economical reactor. This proposal is a tremendous extrapolation from the state of the art. It hopes to contain 73 MJ of plasma energy, about a factor of 500 more than NSTX, at a plasma energy density of about 2.3 MJ/m3, or about a factor of 100 more than NSTX; all in single leap.

Let us compare the proposal for the FDF with the scientific prototype. Many of the goals of the two are the same, specifically running steady state in a reacting plasma, breeding its own tritium (or in the case of the scientific prototype, its own 233U as well) and investigating neutron interactions with the walls. However unlike the FDF, which scales up the plasma by a gigantic jump in the unknown, the scientific prototype runs in a plasma regime scaled up only a bit from what has been accomplished. If successful it would produce a much lower wall loading (0.25 MW/m2 instead of 2 MW/m2, which they ultimately hope to scale up to 4.4 MW/m2). However as we will see, a fusion breeder probably will not require more than 1 MW/m2, and possibly could get by on half of that. To this author’s mind the scientific prototype is not only a much safer bet, but is also a necessary way station on the path to either pure magnetic fusion or magnetic fusion breeding.

Reverse Field Pinches

Henceforth we will be even briefer in our description of alternate magnetic configurations. The reversed field pinch is a configuration, which shows interesting MHD properties. It is a torus, but the poloidal and comparable toroidal field are both pulsed and the plasma reacts to it. The system relaxes to a state, which is nearly force free, and one characteristic of this state is that the toroidal field reverses between the toroidal axis and the wall. The British first observed this in their ZETA device, and Brian Taylor in 1974 [53] came up with a particularly ingenious theory of what transpired; interest in the device persisted, and several more were built.

The largest and most actively investigated is the RFX device in Padua, although others exist in Wisconsin and Sweden. The Padua web site (www.igi.cnr.it) is a valuable source of information on what the group has accomplished, particularly their reports, the last one being in 2010, as of this writing (fall 2013). Their device has a major radius of a meter and a minor radius of about 30 cm. Although their web site does not give direct information on nTτ, or contained energy, it is reasonably easy to approximate these values from the 2009 and 2010 reports. The 2009 report gives the density as about 3 × 1019 m3, and the temperature as 1–1.5 keV, so the total energy contained is about 6 kJ, about a factor of 25 below NSTX and about a factor of 300 below JT-60. The confinement time one can estimate from their reported thermal conductivity κ, in the region of a transport barrier, κ ~ 10 m2/s, so their confinement time ~ r2/κ is about 10 ms; giving rise to an nTτ of about 3 × 1017. In other words, by this metric, according to Fig. 1, RFP’s are where tokamaks were about 35 years ago.

Field Reversed Configurations and Magnetized Target Fusion

Not to be confused with reversed field pinches, field reversed configurations (FRC’s) are essentially toroidal balls of plasma without any center post. They are a theta pinch configuration except that the field lines close upon themselves. In the toroidal configuration, all fields are poloidal. There may be a toroidal field as well, and this would make it a spheromak. But in any case FRC’s are free to move from place to place. In a sense, they are analogous to ball lightening, except that field reversed configurations definitely exist and their properties have been measured. A schematic of a field reversed configuration, provided to Google images by the University of Washington is shown in Fig. 10.

Fig. 10
figure 10

A schematic of a field reversed configuration (Color figure online)

The Los Alamos National Laboratory has been a proponent of field reversed configurations and what they call magnetized target fusion [5456]. They produce rather dense field reversed plasmas, n ~ 2 × 1022/m3 with a large magnetic field, about 5 T, and electron and ion temperatures of about 300 eV. These plasmas only live about 10 microseconds, so nTτ is about 6 × 1016, about where tokamaks were in 1978 according to Fig. 1. They published the contained energy, which is 600 J. Their total particle inventory is 1019 particles, so if half are D and half are T, and all react they would produce about 10 MJ of fusion energy.

To get this plasma to react, they had planned to produce it and transport it to a region where it could be compressed by an imploding metal liner. The magnetic field would be compressed by two orders of magnitude, heating the plasma to fusion temperature. So far they have not been able to do this.

Work published work on this at LANL seems to have stopped in 2005 and their web site (www.LANL.gov) makes no mention of work after this either. However there are several obvious questions. How much energy does the imploding liner use? At best their fusion energy for their published configuration is 10 MJ assuming every DT pair fuses. In their renewal proposal, available on the LANL web site, they mention doing the compression experiment at the Shiva-Star facility at AFWL in Albuquerque. However this pulse power facility has 9 MJ, so it is hard to see how they achieve break even, much less power for the grid. Secondly can such a device be repetitively pulsed? How quickly can all the molten metal, which accumulates on each shot, be cleaned out in time to get off the next shot?

A private company, General Fusion is attempting to achieve commercial fusion power by magnetized target fusion. Very few scientific details of their work have been made available. They claim that they will have breakeven by 2014 and power into the grid by 2020. Time will tell.

Mirrors, Beams and Electrostatic Confinement

We lump these together, because in the opinion of the author, they have little or no chance of ever evolving to a commercial fusion device. Consider first a mirror. At 20 keV, the optimum energy for a beta limited plasma (Fig. 4) the ion ion angular scattering cross section is about three orders of magnitude above DT fusion cross section. Hence for every ion that fuses, many, many more escape. However the energy gain from a single fusion event is so great, that the energy accounting is not as horrendous as the particle accounting. If the ion dynamics alone controlled the energy, it might be marginal. But then there are the electrons. All experience with mirrors powered by neutral beams or ICRH is that the electrons remain cool, and therefore constitute a huge energy sink for the ions. Who knows what it would take to heat them up. Quite simply, this author’s opinion, backed up by many years of magnetic mirror experience, it that economical pure fusion or hybrid fusion power plants are out of reach. One large mirror machine, MFTX-B at LLNL was cancelled after the enormous field coils, along with the associated equipment had already been manufactured.

The largest and most recent effort in mirror machines directed toward fusion is the GAMMA 10 device at Tsukuba University in Japan. It is a machine at least 7 m in length and 0.3 m in radius, with all sorts of bumps and wiggles in the magnetic field in the two end sections. It is heated mostly by ion cyclotron resonance heating, and in the end cells by electron cyclotron resonance heating. Apparently there is also a capability to add neutral beam heating. A Google search ‘GAMMA 10 Tsukuba University’ shows little published after 2003, as does its web site (http://www.prc.tsukuba.ac.jp). As of 2002 [57], it had achieved a density of 4 × 1018 m−1, an ion temperature of 4 keV, an electron temperature of 60 eV, and an energy confinement time of 10 ms for a triple product of 1.6 × 1017 m−3keVs, about 4 orders of magnitude less than what JT-60 has achieved. This is about where tokamaks were in about 1978 according to Fig. 1. The total energy confined is about 6 kJ.

Other applications of mirror machines might be to produce an intense 14 MeV neutron source for material studies. This is being studied at the Institute for Nuclear Physics at Novosibirsk, Russia in their gas dynamic trap [58]. The advantage here is that one might be able to get a useful neutron flux over a small but relevant area, near the high field region, with a mirror fusion device having a Q much less than unity.

Now consider a beam based system, D and T beams at optimum energy come in from opposite sides of a sphere, and meet in the center, where they fuse. The angular scattering rate, which is much greater than fusion rate would degrade the process by reducing the relative velocity between the D and T. But even more important, collisions also mix up the energy at about the same rate as the mix up the angle. Thus long before this beam system could fuse, the energies would be randomized, so the deuterium and tritium would not even meet at the center any more, much less have optimum energy for fusion.

Electrostatics will not help either. Any commercial fusion reactor operating at plasma densities appropriate for fusion, will be enormous compared to the Debye length. Of course the Debye length may be large, but if so the density will be tiny, far too small for fusion. If one attempts to try to separate the electrons and ions in a dense plasma, they will just rush together and neutralize. Similarly if one attempts to impose strong electric fields from the outside, the plasma will set up its own electric field and block out the externally imposed field. I confess that I have not studied the proposal for a tandem mirror, which was based on electrostatically insulating against some end losses. However it has always seemed to me that the device would need not one, but two Maxwell demons.

Advanced Fuels, DD, D 3He, p 11B, etc.

The most perceptive comment I have heard regarding advanced fuels was several years ago from a well respected MFE expert at LLNL. At that time, he regarded fusion as being in something of a crisis mode also. “Don’t people realize”, he said “that it is fourth down, 99 yards to go, with seconds to play; and the advanced fuels advocates want to move the goal post back another kilometer!”

Summary of the Status of Magnetic Fusion

In Table 3 we plot the nTτ and the confined energy of the various MFE devices we have discussed. Let us recall that it has taken decades of international effort and billions of dollars for the tokamak to get to where it is now. Is it reasonable to think that any of the other devices (with the remotely possible exception of the stellarator) can achieve what a tokamak has achieved today, in any relevant time scale, let alone produce commercial fusion power? This author’s answer is no. The chart then answers another dilemma; in designing a fusion reactor, do we start with what the reactor would like and work backwards to the optimum plasma configuration, likely a field reversed configuration but without the ultimate magnetic compression? See from Fig. 10 how relatively simple it would be it put a blanket around it. Or do we start with the optimum plasma and work forward to the reactor? This chart provides the answer; we start with the plasma and work forward. The magnetic fusion program has little choice but to continue to dance with the lady that it came in with.

Table 3 First column is the device, second is the nTτ in units of m−3 keVs, third is energy contained in megajoules, and the fourth column contains simple comments

Inertial Fusion

Inertial fusion energy (IFE) relies on the deposition of an enormous amount of energy in a tiny target so as to heat it up to fusion temperature and density, let it fuse and then blow apart. In other words, there is no confinement as in MFE. IFE relies on the fact that the target fuses before it can blow apart. There are various proposed drivers for IFE, lasers, heavy ion beams and possibly others. In any prospective IFE power plant, the energy will come out in a series of explosions repeated at some rate, rather like an internal combustion engine in a car. Thus a figure of merit for any such power plant will be how quickly the debris can be swept away so the next shot can be fired. The less mass there is in the target, the better; and also the more standoff between the driver and target, the better.

Inertial fusion has a number of significant advantages over magnetic fusion. As we have seen, MFE requires the concentration of enormous energy (gigajoules) in very small volume, an obvious potential danger. This is not the case with IFE. Consider NIF, the laser is 1 % efficient and delivers 3 MJ at 1.06 μm wavelength and 1–2 MJ at the third harmonic. Thus for each shot, 300 MJ must be stored. However NIF has 192 individual lasers, so each one must store just over 1 MJ, and all of the units are independent and separate. If one of them blows up, there is very little potential for damage; and this is for NIF, a rather inefficient laser. Any laser for a power plant must be much more efficient, so there will be even less potential for wide scale damage from an accident in one of its power supplies. IFE is inherently safer than MFE.

Secondly, since the target fuses and blows apart, IFE does not have to worry about confining a burning plasma. At this point, the MFE program has no experience with burning plasma, and can only speculate on the problems of confining alphas, overheating the plasma, alpha-generated instabilities and the like.

Thirdly, there has been scuttlebutt around the community that the concept has been confirmed by classified experiments underground called Halite Centurion, but who knows if this is really so or what the details are [59].

Fourth, the blanket size of a tokamak is determined by the geometry of the plasma; there is little if any flexibility. The inertial fusion target fuses at the center of a spherical target chamber (about 10 m radius at LLNL). However there is at least some flexibility to make the target chamber larger or smaller if conditions require.

Finally, unlike tokamaks, there does not seem to be any limiting features like conservative design rules which has yet appeared in the theory of inertial fusion. And of course bombs do work.

Lasers

The Development of Laser Fusion; Indirect Drive

As with tokamaks, laser fusion has been under development internationally for decades at a cost of many billions of dollars. There are large programs in the United States, France, Japan and other places. The main program in the United States, is at the Lawrence Livermore National Laboratory (LLNL). The other American programs are at University of Rochester Laboratory for Laser Energetics (URLLE), and at the Naval Research Laboratory (NRL).

Almost as soon as lasers were invented, scientists thought of them as drivers for inertial fusion. Initially the thought was to simply deposit the energy in a target, heat it to fusion temperature and let it fuse. However, the laser energy needed was enormous, many, many megajoules. A significant theoretical breakthrough came when Nuckolls showed that by ablatively compressing the target, the laser energy could be enormously reduced, perhaps to as low as 10 kJ or less [60]. Ablative compression means is that the laser deposits its energy in the outer region of the target, which heats up, ablates away, and the inverse rocket force compresses the remainder of the target to fusion conditions.

To achieve this requires a spherical implosion, so maintaining the spherical symmetry is of utmost importance. This means that one must find a way to minimize the effect of the Raleigh Taylor instability, which is unavoidable, since ablative compression necessarily means the acceleration of a heavy fluid by a light one. An enormous effort has been made here, and the community generally agrees that the outward ablative flow has a strong stabilizing effect, although just how strong is still under study. In any case, by taking advantage of the flexibility one has in designing the laser pulse, one can exert a measure of control over the flow so as to minimize the effect of the instability.

While Nuckoll’s idea is still the main one being pursued today, as we will see, his original estimate of necessary laser energy was nothing if not optimistic. In the pursuit of laser fusion, LLNL embarked on major program developing a series of larger and larger lasers, Argus, Shiva, Nova, Beamlet,and finally NIF. All of these are Nd glass lasers with a wavelength of 1.06 μm. However at such long wavelengths, laser plasma instabilities become a major worry. Accordingly, LLNL has developed frequency multiplication techniques to operate at third harmonic, about 1/3 μm wavelength. LLNL now routinely operates with pulses in excess of a megajoule at third harmonic. URLLE has also taken this approach with their OMEGA laser (30 kJ). NRL has taken a different approach, using a KrF laser at a wavelength of 0.248 μm with its NIKE (3–5 kJ) and Electra lasers.

In terms of economics and timelines, the experience of NIF has not been so different from the experience of ITER, but at least NIF is now operational. It was approved in 1995, to be finished in 2002 at a cost of $1.1B. It was finished in 2009 at a cost of $3.5B. It is also important to note that the sponsor for NIF is not fusion energy, but nuclear weapon stockpile stewardship. Accordingly the sponsor has little interest in such things important to energy as laser efficiency or rep rate. More important, the sponsor interest is only in a particular type of implosion driven by X-rays. This necessitated a particular type of target configuration called indirect drive. The target was placed inside a high Z enclosure called a hohlraum, the laser was focused on the inner walls of the hohlraum which heated to a temperature which produced black body X-rays at a temperature of 250–300 eV. These X-rays, not the laser, drive the target. The hohlraum is filled with helium gas to provide a back pressure to keep the heated walls from expanding into the target.

Once the laser light enters the hohlraum, the helium is ionized, so the light must traverse a large, likely uniform plasma. Laser plasmas instabilities are a vitally important issue here, and LLNL has dedicated great resources to their analysis [61]. They believe they have things under control.

Furthermore, the sponsor, which paid for NIF, has many other uses for it besides fusion, and wants the facility for its own purposes more and more. A view of NIF, emphasizing the different important scale, kilometers down to millimeters is shown in Fig. 11.

Fig. 11
figure 11

Various photos of NIF, taken from the LLNL web site, emphasizing the tremendous range in size, from about half a kilometer for the overall facility, to the 100 m laser bays, to the 10 m target chamber, to the 1 cm hohlraum, to the 1 mm target (Color figure online)

In preparation for the ignition campaign, LLNL had done an enormous amount of work. In a major article [62], cited over 1500 times, Lindl and his coworkers have documented the theoretical basis for the project. A tremendous amount of work, by a large number of people went into the preparation of this article; it is most likely beyond the capability of any single person to absorb all of it (certainly well beyond the capability of this author). However the article it unambiguous in one respect, it predicts a Q (fusion energy over laser energy) of about 10 over a broad region of parameter space. As NIF was further delayed, LLNL’s theoretical efforts continued. In another paper summarizing 6 years of additional effort [63], their predicted gain remained 10.

It is now well known that nature did not cooperate. Their gains are just over 10−3. Among other things, they observe about 10–15 % of the laser light back scattered out of the hohlraum by stimulated Raman scatter. There may be more scattered light remaining in the hohlraum. While this might not seem like a killer, one characteristic of stimulated Raman scatter in the nonlinear regime is that it generates copious energetic electrons, whose total energy is about half that of the scattered light. Hence they could have as much as 100 kJ of energetic electrons, some with energies likely in excess of 100 keV bouncing around the hohlraum in a manner, which is likely nearly impossible to calculate today. Furthermore, in all likelihood there are 1–5 Megagauss magnetic fields (i.e. a 100 keV electron has a 1–5 μm Larmor radius) in the hohlraum driven by non parallel density and temperature gradients. These fields almost certainly have a complicated special and temporal structure. Who knows what will be the influence of 100 kJ of energetic electrons bouncing around the hohlraum will be.

In a recent article in Physics of Plasmas LLNL [25] has summarized their progress on the National Ignition Campaign (NIC) to date. A schematic of their experiment, taken from the available web version of Ref [25] is shown in Fig. 12. Of many of the key parameters for ignition, for instance ρr (density times radius) and implosion velocity, they are within 90 % of what is needed. However these achievements were not obtained simultaneously on the same shot. In fact some of the necessary parameters were quite far from what is required. Their best hot spot pressure achieved was two to three times below their code predictions. Their target mass was about 200 μg of DT. LLNL had estimated that with about 0.1 μg of ablator mixing with the target, they would be okay. However their measurements indicated that about 3–4 μg of ablator were mixing with the target, 30–40 times the acceptable level! For the most part they seem to think that hydro instabilities and lack of necessary symmetry were their biggest problem. To this author’s mind, it is surprising that they paid so little attention to laser plasma instabilities and energetic electrons, a subject they discussed in many APS meetings.

Fig. 12
figure 12

A schematic of the NIF target in the hohlraum with the different arrays of laser beams fired in, the schematic of the target itself, and the laser pulse as a function of time (dotted curve is the X-ray) pulse taken from Ref. [25] (Color figure online)

Finally it is important to note that despite their problems, LLNL seems to be making some progress. In their 2013 Physics of Plasmas (submitted in March 2013) paper, the maximum neutron production they quoted nearly 1015. However on a shot on Sept 28, 2013, they did achieve 5 × 1015 neutrons [26]. But they still have a very long way to go. They have not achieved that much more gain than what URLLE had achieved in a direct drive (the laser illuminates the target instead of a hohlraum) experiment with only 30 kJ of laser energy [64].

The fact remains that NIF was a multi billion dollar investment which missed the calculated gain by over three orders of magnitude as of January 2014. Despite the natural advantages of inertial over magnetic fusion pointed out earlier, magnetic fusion is still way ahead of inertial fusion in neutrons produced and gain achieved. MFE certainly has not had a disaster of the scale of missing its calculated a gain by over three orders of magnitude in an expensive machine. It serves nobody’s interest to deny this or to attempt to sweep it under the rug. Inertial fusion’s credibility today is not exactly riding high. Surely congress will never approve another multi billion (or even multi million) dollar inertial fusion machine until NIF achieves some measure of success. To be blind to this is simply to live in a dream world. As regards inertial fusion, NIF is what we’ve got, it is all we’ve got, and it is all we will have for quite some time. The only reasonable goal now is to get it to work, assuming congress does not get annoyed and pull the plug. Hopefully this will not happen. The rest of this section discusses some of the ways NIF might be made to work, and the potential for inertial fusion energy, assuming that NIF and inertial fusion can get beyond their current difficulties.

Direct Drive

An obvious alternative to indirect drive is direct drive, where the laser, spherically symmetrically, illuminates the target rather than a hohlraum wall. During the 15 or so years that the NRL program was an independently funded congressionally mandated program, it focused only on energy and only on direct drive. The URLLE program focused on indirect drive in support of LLNL and NIF, but also on direct drive. There are certain advantages to direct drive, both regarding the physics, and also regarding its applicability as an energy source.

Let us first concentrate on the physics. As the laser illuminates the target and does not go through an intermediate stage of conversion of light to X-rays in the hohlraum, there is one less channel for energy loss by the laser light. Accordingly, there are fewer calculations one must make to evaluate the target performance. Then there is the fact that the laser has to propagate through much less plasma to reach the target, so there is less opportunity for laser plasma instabilities, although they are still an important consideration. As long as spherical symmetry is maintained, the density and temperature gradients are parallel, so there should not be any magnetic field generation, further simplifying the physics and the analysis.

However for direct drive to work, the transverse spatial profile of the laser must be very uniform in order not to induce azimuthal structure in the target, which can serve as seed for the Rayleigh–Taylor instability. LLNL, URLLE and NRL have all come up with optical means of generating uniform laser beams. Common to all is a requirement of laser bandwidth. At this point KrF lasers like NRL’s NIKE have the potential for about 3 THz of bandwidth, URLLE’s OMEGA, about 1 THz, and NIF, about 500 GHz early in the pulse (when the amplitude is lower) and less later on. These bandwidths also have the advantage of having a stabilizing effect on laser plasma instabilities.

But the main advantage of direct drive is that the calculated gains are much higher for direct drive than for indirect. Recall that LLNL’s own calculations for indirect drive show a gain of only about 10. The calculated gains for direct drive are much greater. Shown in Fig. 13 are calculated gains as a function of laser energy for a series of different sorts of laser pulses. These curves were calculated assuming a KrF laser. Its shorter wavelength gives rise to an advantage in ablation pressure and threshold for laser plasma instabilities. However a KrF laser has an additional advantage over a frequency tripled Nd glass laser. It has a capability to perform what NRL calls zooming. That is part way through the laser pulse, even at maximum laser power, its focal properties can be changed, one or more times, in nanoseconds, so that as the target shrinks, the focal spot size does as well. Hence less laser light is lost in going around the shrinking target.

Fig. 13
figure 13

NRL calculations of laser gain versus energy for a KrF laser with zooming. Black (top), Shock fast ignition; blue (middle), higher implosion velocity, higher ablation pressure drives as proposed for the Fusion Test Facility; and red (bottom), conventional central ignition (Color figure online)

The lower curve in Fig. 13 is NRL’s original calculation [65] of what is called central ignition. The same laser pulse is used to both compress the inner part of the target and to heat it. Most of the compression is done at low adiabat. When the target is sufficiently compressed, it is heated to fusion temperature. The inner part begins to fuse, the alpha particles are locally deposited and heat the nearer parts, and a propagating burn wave propagates outward consuming the fuel.

The middle curve was an effort made by NRL to investigate whether gain could be achieved at lower laser energy. The idea was to take advantage of the higher ablation pressure available from a quarter micron KrF laser so as to accelerate the target to a higher inward velocity. NRL found that significant gains could be achieved in this manner. Accordingly it proposed the Fusion Test Facility (FTF, not to be confused with GA’s proposed FDF) as a means of investigating fusion neutron production with the lowest possible energy laser [66, 67].

In 2007, Riccardo Betti, of URLLE proposed a new scheme, shock ignition [68]. The idea here is to use a more structured laser pulse, so that the compression is done at the early part of the pulse, and when the center is compressed to sufficient density, while remaining cool, a second intense laser pulse is fired in. This sends in a shock wave, which collides with the reflected shock form the initial pulse and heats the plasma to fusion temperature. In this case, the original compression is at lower inward velocity. NRL did its own calculations on shock ignition [69] in 2010 and confirmed their potential for a KrF laser with zooming capability in a series of one and two dimensional fluid simulations. Shock ignition is now NRL’s standard approach to laser fusion.

It is worth noting also that the URLLE has actually done implosion experiments with cryogenic DT targets [64]. While with their 30 kJ laser they could not achieve ignition they did get decent neutron production and central ion heating. Shown in Fig. 14, taken from the available web version of [64] shows the neutron production and central ion temperature as a function of implosion velocity. The maximum neutron production is about 2 × 1013, or about 45 J. Since the maximum energy of OMEGA is 30 kJ, this corresponds to a Q of at least 1.5 × 10−3, not that much less that what NIF has achieved with indirect drive with over a megajoule of laser energy.

Fig. 14
figure 14

Measurements of neutron yield and ion temperature in direct drive powered by the OMEGA 30 kJ laser at URLLE taken rom Ref [64] (Color figure online)

There are also powerful economic and scientific arguments for direct drive laser fusion. For instance let’s stipulate that an indirect drive system produces 100 MJ of fusion energy in a single shot. However when converting to electricity, this is about 33 MJ of electricity, or 9 kW h; worth about a dollar. While today’s hohlraums cost many thousands of dollars each, mass production would undoubtedly bring down the cost, but certainly not to zero. After all these hohlraums are precisely engineered and contain such expensive materials as gold and/or uranium. Between shots, not only would the debris of the target have to be swept away, but also the debris of the hohlraum. Furthermore, the target would be dropped into the target chamber from the top, and when it gets to the center, the laser would be fired at it. Direct drive only has to get the target there, indirect drive has to both get it there and orient it properly as well. To summarize, there seem to be both physics and economical reasons indicating that direct drive has an advantage over indirect drive.

NRL’s HAPL Program

The High Average Laser Power program (HAPL) was a multi year, multi institution program managed at NRL. It existed from 1999 until 2008, when it became a casualty of the financial collapse. Its accomplishments have been documented on the ARIES web site (aries.ucsd.edu/HAPL). The program goal was to investigate every aspect of laser fusion with the goal of developing all of the science and engineering necessary to make it a reality. For instance its namesake goal was to develop lasers capable of high average power and efficiency, which are suitable as drivers for laser fusion. To bring this about, HAPL supported 2 laser projects, the Electra laser at NRL, and Mercury laser at LLNL. The former is a KrF laser, with an energy of 300–700 J per pulse and runs at a rep rate of 2–5 Hz. The latter is a frequency tripled diode pumped solid state laser, with an energy of 50 J per pulse and runs at about 10 Hz.

However lasers were hardly the only aspect investigated in HAPL. It involved some 30 institutions investigating such things as first wall, final optics, the chamber, the target manufacture, the target positioning, the target tracking…. The program was an integrated program with the goal of achieving economical and practical laser fusion. No single goal was regarded as paramount. In fact the mantra of the project was “You cannot solve your problem if you make the next guy’s problem impossible”.

Hence the philosophy underlying HAPL was much like that of the Scientific Prototype, proposed earlier in this paper. There the goal would be to solve very problem except the burning plasma; that would be left to ITER. However an obvious difference is the HAPL was a reasonably well funded project with many participants; the scientific prototype, at least as of now, has zero dollars and a single proponent (just me). The motivation of HAPL and the Scientific Prototype is just the opposite to the motivation of the various proposals for small burning plasma tokamaks (FIRE, BPX…), which, if successful, would solve the burning plasma problem, but would make everybody else’s problem impossible (in other words, there was no clear path from say FIRE, which would only last 3000 shots to an economical fusion machine). HAPL’s and the Scientific Prototype’s goal is to advance the field on as broad a front as possible.

The HAPL program had made steady progress during its brief lifetime, along a very broad front, and believed that there were no show stoppers.

What Next for NIF and Laser Fusion?

At this point the next step for NIF is uncertain. Some are in favor of attempting other types of indirect drive experiments in the hopes that one will work. But these would have to increase the measured gain by over 3 orders of magnitude in order to achieve what LLNL expected in 2004 and 2010.

Others favor switching to a several year science based program to examine more of the physics of indirect drive in the hopes that a new pathway will become apparent.

Then there is the possibility of converting NIF to what is called polar direct drive. This would use the existing target chamber and existing beam lines to directly illuminate a target. However the illumination would not be uniform over 4π solid angle, but most of the illumination would be at the poles. Thus with polar direct drive, laser fusion would be giving up spherical symmetry, one of its most important advantages. Target designs and the like would be optimized to make the implosion as symmetric as possible. But there are other problems as well. As the illumination is non uniform, some azimuthal angles will have greater illumination, and these would be subject more to laser plasma instabilities than would be the case for uniform illumination. This author worries that polar direct drive could be a large time and dollar sink spent on a non optimum configuration.

URLLE has proposed polar direct drive experiments on NIF. Apparently they believe that NIF can be reengineered to have sufficient bandwidth and other attributes for the process to work. LLNL seems reluctant to go this route. This is apparent in the dueling cost estimates to refit the laser for polar direct drive [70]. LLNL’s Edward Moses estimates the cost at $200 M or more, while Riccardo Betti, an expert from URLLE estimates a cost of $40–50 M.

Steven Bodner [71] has a contrasting proposal: that NIF should be scrapped and replaced by two laser development programs, one on KrF lasers and one on DPSSL’s. Each program would attempt to develop a 100 kJ laser, with the characteristics needed for direct drive fusion, and which would have a rep rate of 5 Hz. These would be used to test as much as they could on planar, spherical and conical targets, and if these tests were successful, build a megajoule class rep rated laser to power a fusion test facility. In other words, tell our sponsors, sorry, we made a big mistake with NIF, but give us a few hundred million more to develop two lasers, and then give us a few billion more for a perfect megajoule class laser, all the while dismantling the imperfect one we have right now.

Bodner does mention another possibility, but then seems to dismiss it. In his letter to the NAS, he states “Unofficial and rumored estimates from LLNL say that the conversion to symmetric illumination for direct drive would cost over $300 M and take at least 2 years. Since the paying customer is the weapons program, it won’t happen”. But if these estimates are correct, why won’t it happen? It is certainly the fastest and cheapest way to achieve megajoule class direct drive experiments.

The weapons program definitely would not be happy, especially where they paid for it; but they are not the only constituency. The House Appropriations Committee is a rather important and powerful group as well, and they should not be ignored. As stated in the Introduction, they do not think the weapons related work is worth the expenditure, only ignition is.

As Bodner points out, NIF is hardly ideal for direct drive experiments, because of its limited bandwidth. But how much does this matter? If URLLE believes it can get NIF beams satisfactory for polar direct drive experiments, then they surely must believe that these modified beams would also work at least as well in a spherically symmetric configuration, a configuration certainly more likely to achieve success.

Also direct drive gain calculations show impressive gains at half a megajoule laser energy. NIF has nearly 4 times this. Hence there is a very large margin for error both regarding the laser energy and the gain calculations. Let’s say NIF does a symmetric direct drive experiment and gets a gain of ‘only’ 10. Wouldn’t this be a tremendous accomplishment? It might be just 2 or 3 years away. Might it not encourage Congress to build the ideal laser or lasers Bodner suggests?

Laser Fusion for Breeding

With NIF spending billions, but failing by 3 orders of magnitude in its prediction of gain, the question of pure fusion or fusion breeding is hardly laser fusion’s main concern at this time. In fact officially, pure fusion or fusion breeding is not a concern at all. There is no American laser fusion energy program. NIF is sponsored for nuclear stockpile sponsorship; other US laser fusion programs are now either in support of NIF, or else are tiny by comparison. However even a brief perusal of the LLNL web site makes clear that NIF for energy is very much on the minds of the Lawrence Livermore National Laboratory. In fact LLNL invited Thomas Friedman to visit NIF just as all 192 lasers were simultaneously operating. Edward Moses, the project leader guided him and Friedman interviewed him in his op ed [72]. It was all about energy; there was not a word about stockpile stewardship. No, LLNL has energy very much on its mind despite its sponsor.

Even if NRL’s optimistic theoretical results are born out in experiments, there could be a problem as regards pure fusion. Let us take the most optimistic result, a gain of 250 with a 1 MJ KrF laser, assumed to be 7 % efficient [69]. Take a rep rate of 6 Hz so that the same 1.5 GHz of fusion energy is produced as for our assumptions for large ITER and NIF. This then produces 500 MW of electricity. But the laser needs 14 MJ for each shot, so the 6 shots per second would need about 100 MW of the electrical energy produced, leaving about 400 MW for the grid. This is better than ITER or NIF discussed in the Introduction; it may or may not be an economical power producer. Judging from the size of NIF and its $3.5 B price tag, a laser with the same energy per shot, but much greater average power is unlikely to be smaller or cheaper. But $3.5B is a high price for one major part of a 400 MW power plant. So there still could well be a real need for the enhancement in gain fusion breeding could provide, even if NRL’s optimistic calculations prove to be valid.

But then what if the calculations prove optimistic? If there is one thing to learn from the experience of NIF, it is that there are large gaps in our knowledge of laser fusion, certainly for indirect drive, perhaps even for direct drive. If the gain and efficiency are ‘only’ 100 and 3 %, the example given in the introduction, then pure fusion becomes virtually impossible. However the order of magnitude increase in Q from breeding would make all the difference. Using laser fusion for breeding instead of pure fusion gives a much larger margin for error in case the calculations prove to be optimistic. It is certainly a much more conservative approach.

Sandia’s Z Machine

We mention Sandia’s Z before heavy ion fusion because Z does exist, heavy ion accelerators so far exist only on paper. The Z machine is likely the largest pulse power facility in the world. Like NIF, it is sponsored for weapons research. The lab, on its web site does mention fusion (w.sandia.gov/z-machine/), but does not give fusion nearly the emphasis that LLNL gives NIF on its. The Z machine is an intense X-ray generator. Inside a hohlraum about the size of a spool of thread are many thin tungsten wires stretched axially near the edge, or in an axially nested array. The machine is discharged through these wires, more than 2 × 10amps. These wires vaporize and produce a tungsten plasma which both generates X-rays and implode toward the center producing still more X-rays. Typically they produce nearly 3 MJ of X-rays, certainly more than NIF has generated. The machine typically gets off about one shot per day. A great advantage of Z over lasers is its relative simplicity and low cost as compared to say NIF.

The plan for using Z as a fusion device, the idea is basically like indirect drive on NIF. There seems to be no analog for direct drive. While Z does produce copious X-rays, it does have some limitations as a fusion device. First of all, lasers have the capability of tailoring the pulse so as to meet requirements. For instance in URLLE’s direct drive spherical implosion experiments [64], they use precisely structured pulses with several prepulses (which they call pickets) on the leading edge; also Fig. 12 shows the precisely structured pulses that NIF uses. Sandia on its web site does not seem to have mentioned structuring the pulse in any way. Secondly, lasers have standoff and Z does not, so replacing targets after each shot is a challenge. Sandia has given some attention to this, principally by envisioning a power plant as having as many replicated Z machines as necessary, so as to provide time between shots to replace the target and wires. It may be that Z will ultimately evolve toward a viable fusion device, but it seems to be far behind lasers right now.

Heavy Ion Fusion

As discussed in the Introduction, the proponents of heavy ion fusion (HIF) have every confidence that they can rather quickly produce economical fusion. The advantage of an ion beam over a laser is its efficiency. Counterbalancing this advantage is the fact that spherical direct illumination is not possible. Also they are starting out with very little equipment. The main effort seems to be in Russia under Boris Sharkov [73]. Also Germany is a main partner in constructing an international heavy ion accelerator called FAIR (http://www.fair-center.eu/public.html), but it appears to have nowhere the parameters needed for fusion, and the accelerator does have many other customers, fusion being a minor one, if in fact it is a customer at all. In the United States, the main effort is at the Lawrence Berkeley Laboratory. However the resources for accelerator or target experiments are so meager that most of their work is in numerical simulations. In fact their HIF lab is called the virtual laboratory, both because their work is mostly simulation, and also because they are bringing several other laboratories and organizations together to hopefully work as one (http://hif.lbl.gov/VNLoverview.html).

The Russian effort proposes a 10 km long 100 GeV (about 500 MeV per nucleon) heavy ion accelerator coupled with eight storage rings, each about 1 km in diameter. These store and compress the heavy ion beam, which would have a total energy of about 7 MJ. This does not look inexpensive. Although it is said that standard accelerator technology is sufficient, Sharkov et al. [73] do list 5 difficult challenges the accelerator must overcome. It is not clear from Ref [73] just how much of the accelerator has actually been built.

The Russian target design is a modification of laser direct drive. A small cylindrical DT target is encased in a larger lead cylinder. The beam is deposited in the lead and the DT is compressed cylindrically; note that they are giving up on spherical compression. Also in NIF targets, mixing of fuel and ablator is a major problem. But in NIF, the ablator is about half of the target mass. How simple will it be to keep the fuel in Sharkov’s target separate from the much more abundant lead? Note also that even thought the target is inherently 3 dimensional (rz for the cylinder and theta of there azimuthally dependent instabilities), their hydrodynamics codes are all either one or two dimensional.

The Berkeley target design uses indirect drive. The annular cylindrical beam deposits in a hohlraum to generate X-rays to implode the target. However its hohlraum must certainly be thicker and heavier than a laser hohlraum since the entire ion beam must fit inside it. For either the Berkeley or Russian design, there will be much more material to be swept away between shots than for laser fusion, either direct or even indirect drive.

This author has recommended that we let the Germans and Japanese proceed with their stellarator programs, help them any way we can, but not jump in ourselves unless they prove to be definitively superior to tokamaks. The recommendation is the same for HIF. Let the Russians run with it, help them in any way we can, but not start a program ourselves unless they can prove it to be definitively superior to lasers. Right now they appear to be hopelessly behind.

Nuclear Issues

We now turn to the nuclear issues, which we discuss as briefly and simply as possible.

Fusion Breeding

In pure fusion, the 14 MeV neutron uses only its kinetic energy to boil water, because tritium breeding is relatively undemanding of neutron economy. In fusion breeding, it does this, but also uses what we have called its potential energy to breed enough nuclear fuel to produce ten times more fission power than fusion power. For breeding, the neutron is first inserted into a neutron multiplier, for instance lithium, beryllium, lead, uranium or some other material. (Even pure fusion needs some neutron multiplication, since the reaction produces only a single neutron, so none could be lost.) Here the fast neutron produces 2–3 slower slower neutrons. One of these is used to breed tritium to keep the fusion reactor going. The remaining neutrons can be used for other purposes.

Once the tritium is bred, the remaining slower neutrons are fed into either 232Th or 238U. We consider only the former, since the latter breeds plutonium, a material we would like to avoid as much as possible. The thorium absorbs a slow neutron to become 233Th, but this is unstable to a double beta decay. It has a half life of 22 min and then decays to 233Pa, which is also unstable and decays to 233U with a half life of 27 days. But 233U is a perfectly good fissile material, i.e. a nuclear fuel for thermal neutron reactors, just like 235U and 239Pu.

How much nuclear fuel is produced depends on the blanket design, and this paper does not get into that. One particular design [74, 75] has each fusion neutron producing 1 T after all losses, 0.6 233U’s, and since the breeding reactions are exothermal, the neutron energy is multiplied by about a factor of M. Recent studies estimate the M factor can vary between about 1.5 and 2. But for all of these, about 0.6 233U’s are produced from each fusion neutron. But each 233U, releases about 200 MeV when burned, so the 14 MeV neutron ultimately produces 120 MeV of nuclear fuel, or the neutron energy produces about nine times as much nuclear fuel, to be burned in separate reactors away from the fusion reactor. Also it releases roughly from 21–28 MeV in its own blanket. Furthermore the fusion alpha particle releases 3.5 MeV. This enormous increase in energy, about a factor of 10 increase in Q over the neutron power of the fusion reactor is reflective of the fact that fusion is neutron rich and energy poor, while fission is energy rich and neutron poor; a natural symbiosis.

Let us see what this means for a Large ITER sized fusion reactor or a NIF sized laser fusion reactor as discussed in the Introduction. It produces 1.5 GW of neutron power. However this is multiplied say 1.8 in the blanket, so in the blanket 2.7 GW is generated. To this, one adds the one quarter of the neutron energy which is the alpha particle energy, so the total fusion reactor power is about 3 GW. However in addition the blanket produces about 15 GWth of 233U fuel. This is enough to fuel five 1GWe (3GWth) conventional light water reactors. At this point the driver power, seen in the Introduction as a show stopper, is now a perturbation, and we neglect it.

The current ITER is designed to produce only 1/3 the neutron power of Large ITER, so it would produce about 5 GWth of nuclear fuel. This is more than enough to power a single LWR of equal power, but not enough to power very many. If LWR’s are the reactor of choice, ITER would not fit as well into the infrastructure as Large ITER. However there are other reactors which burn deeper into the fertile material, require less fuel, and ITER might be able to fuel five of these. If this is the case, the requirements on a NIF sized laser fusion breeder would also be relaxed. Ref. [75] talks of a single fusion reactor fueling as many as 20 or so molten salt reactors of equal power. If this proves to be a viable mid century fission architecture, then an ITER (as opposed to Large ITER) sized fusion reactor, could fuel five or six of these. The midcentury nuclear architecture is unknowable. It will certainly have a large number of legacy LWR’s. In any case, fusion breeding could play a vital role in fueling them.

There is one further complication, which is required for fusion breeding which is not necessarily required for pure fusion. A pure fusion reactor might be able to operate with either a solid or liquid blanket if it can withstand the fusion environment. There are advantages and drawbacks to each. If the blanket is a solid, the tritium can be produced there can be extracted every year or two by replacing the blanket.

But a solid 233U breeding blanket does not appear to be viable. It not only breeds tritium, but also 233U (or 239Pu). In the neutron flux, these continue to react, producing a witches brew of fission products, higher actinides, and an ever increasing heat load. There is a real safety issue here. A major disruption would be a real disaster.

The solution is to use a liquid blanket. It self anneals and as the liquid with the tritium and 233U or Pa flow out of the reactor region, these could be removed chemically. The 233U would then be mixed with 238U to give a proliferation resistant fuel. Optimally, the liquid would have a free surface facing the plasma, but also it could flow in pipes. These pipes would bring in some of the complications of a solid blanket, but not many. They do not have to do anything, but withstand the fusion environment. To this author, a liquid blanket seems overwhelmingly advantageous for pure fusion. For breeding 233U, it is almost certainly a necessity.

If one has a liquid blanket, it does not seem to this author as if dissolving some U, Pa, or Th in the blanket will overly complicate the plasma facing parts, especially given that the neutron flux the blanket must acc.ept is much lower for fusion breeding than for pure fusion. Of course the chemical processing, to retrieve the T, 233U or Pa, done away from the plasma will be much more complicated than retrieving T alone.

This paper will not get into the blanket details, but a molten salt FLIBE blanket, containing lithium, beryllium and fluorine has been discussed in the literature. The lithium breeds the tritium and the beryllium multiplies the neutrons. Also uranium, protactinium and thorium are all soluble in it. The web site, (www.ralphmoir.com) has references to and links to several blanket designs for fusion breeding, including several old LLNL reports on the subject, which would be difficult to access in any other way. Also UCLA has a large program in blankets, studying many possible options (www.fusion.ucla.edu/abdou).

A Roadmap for MFE Large Scale Power Production by Mid Century and Its Cost

Here we only consider magnetic fusion, as so far, it alone has produced fusion neutrons in sufficient quantity. ITER is scheduled to have its first plasma by 2020 and to operate with DT in 2027. In parallel to the world’s efforts on ITER, our proposal is that the United States builds the scientific prototype. Here one must initially make an important choice for fusion breeding; the blanket would have to be a liquid, hopefully initially, certainly in the ultimate configuration. The initial choice must be made to allow for a liquid blanket.

Since TFTR, JET, JT-60 and ITER have all already been designed, a great deal is already known about large tokamak plasmas, and the plasma part should be relatively easy to design. Designing and running the scientific prototype in a hydrogen or deuterium plasma should be relatively easy. Let us say the machine is first built for hydrogen or deuterium, but with space left for a breeding blanket. The machine would be run to research genuine steady state or high duty cycle operation and disruption free operation at conditions giving Q ~ 1 if the plasma were DT. In running steady state, we note that the amps per watt produced is much less of a consideration for fusion breeding than it would be for pure fusion [51]. Let us say that designing the first phase of the machine takes 5 years, and the hydrogen phase also takes 5 years. That brings us up to 2024. In parallel with the hydrogen operation, a design of the breeding blanket is done, first for only tritium, and then for both tritium and 233U.

Immediately after the hydrogen phase, the scientific prototype would operate steady state in DT using the tritium it produces. While the tritium breeding is being examined, the 233U blanket is designed. Let us say this phase takes 4 years, bringing us to 2028. At this point the scientific prototype is set to begin research as a breeder of both T and 233U. Hence ITER will begin investigating Q ~ 10, 400 s pulse operation in DT plasmas at about the same time as the scientific prototype will begin operate as a true steady state or high duty cycle breeder of T and 233U in a Q ~ 1 plasma. Assuming this phase of operation takes 6 years, this takes us to about 2035. It is clear that ITER and the scientific prototype complement each other in many important ways.

If the results of both look promising, a simultaneous design would be made of a steady state Q ~ 10 fusion breeder. It would be called ‘The Commercial Prototype’. Likely it would have a neutron power of 1.5 GW as would Large ITER. However if the nuclear infrastructure by then uses a deeper burn and requires less bred fuel, a conventional ITER, producing 500 MW of neutron power might suffice. If Large ITER is the choice, the neutron power would only need to be 1 MW/m2 on the wall, less than one fourth of that required for pure fusion. If ITER is the choice, the wall power would be only ½ MW/m2, about an order of magnitude less than required for pure fusion, and only a factor of two larger than the scientific prototype.

In any case, the design of the commercial prototype would begin around the early 2030s once it is clear that the scientific prototype and ITER are reasonably assured of success. It would come on line in about 2040. It runs as a breeder for 10 years, and then by 2050 the world is ready to start to mass produce them. Many would be operational by 2060. If 100 are operational by then, these would produce fuel for 500 GWe LWR’s plus the 100 GWe that the fusion reactors themselves would produce.

Now let us see where the choice of pure fusion would lead on this roadmap. While fusion breeding is producing the ‘commercial prototype’, its own ‘DEMO’ if you will, pure fusion will have many enormous hurdles to get over. First, assuming Large ITER is the choice, it must double the fusion power to 3 GW. Then it must produce and control the alpha particle burn so that the input power can be recued from 150 MW to say 50 MW or less. If the fusion breeder runs at high duty factor rather than true steady state, the pure fusion reactor must find a way to operate at steady state. Then the pure fusion reactor must develop a blanket that can withstand at least twice the neutron flux of the fusion breeder. Then it must reduced the cost from about $25B to about $2–5B so that it can be economical. But reducing the cost by a factor of 5 or 10 means making it smaller. Making it smaller means higher wall loading and further violation of conservative design rules. It could well be caught between the devil and the deep blue sea.

These are enormous hurdles, which the pure fusion system may or may not be able to get over, and who knows how many decades it would take if it can get over them. In all this time, the fusion breeder is economically generating fuel for the world.

Now let us do a very rough estimate of the cost of the fuel produced. This is based to a large degree on what the cost of an ITER scale reactor would be. Unfortunately the cost of ITER has been increasing very rapidly, and not only is this discouraging, it makes an estimate difficult. The original cost of Large ITER was to be $10B in capital cost and $10 in operating cost for 10 years. Let us assume that he capital cost of the commercial prototype is $25B. The machine is assumed to last 30 years. Let us assume the same billion dollars per year operating cost.

Thus as a very rough estimate, let us say the capital and operating cost of the commercial prototype is $2–2.5B year−1. It is reactor, which generates 1GWe. Assuming it runs all year, and sells the power for ten cents per kWh, it earns about $0.9B. But it also produces 5 GWe of nuclear fuel. To recover the additional $1.1B, it would have to sell the nuclear fuel for about 2–3 cents per kW h. This estimate is certainly not exact, and as costs capital and operating costs of ITER become clearer, it can be revised.

But at this point, the estimated cost does not seem to be any kind of show stopper. Uranium fuel for LWR’s now costs about one cent per kWh, so fusion bred fuel might increase the electricity cost by a penny or two per kWh. It is also important to note that Large ITER might play a role, at least as a stepping stone, even if the gain is only 3. It would still produce the same 5 GWe of nuclear fuel, but now all of the electric power produced by Large ITER (500 MW) plus another GWe of the nuclear fuel would be needed to generate the 500 MW of beam or microwaves needed by Large ITER; leaving ~4 GWe of nuclear fuel for the grid. By the same calculation, it would have to sell this for about 6–8 cents kWh−1; very expensive, but possibly still economical, especially if an improved version with a gain of 10 or more is on the horizon.

Furthermore mass production might well bring down the cost of fusion generated fuel by a considerable amount. After all, how much would a B777 cost if Boeing produced only one of them, or an aircraft carrier if General Dynamics made only one?

To summarize, there does seem to be a roadmap to large scale, economical power production via magnetic fusion breeding by mid century. Pure fusion can claim no such magnetic roadmap. Pure inertial fusion might, as there are no conservative design rules that we know of holding it back. However IFE still has to get over significant hurdles to get the neutron production that MFE has right now. Unquestionably, fusion breeding is a more conservative goal for IFE than is pure fusion; perhaps it is the only reasonable goal. A start on this path for IFE could be something like NRL’s proposed Fusion Test Facility, but with a liquid liner and added tasks of breeding T and 233U.

Thermal Nuclear Reactors

Since conventional nuclear reactors are such an important part of a fusion breeder economy, this subsection and the next two give a very cursory, and much oversimplified treatment of nuclear power. Only the aspects of nuclear reactors that are most relevant for fusion breeding are emphasized here.

Every thermal nuclear reactor has three important components, the fissile material, 233U, 235U or 239Pu, the material which actually burns; the fertile material, 238U or 232Th, which can produce additional fissile material, and the moderator, which slows down the neurons. In a nuclear reaction, the neutrons are typically emitted at an energy of about 2 MeV. However the fission cross section has a very strong maximum at low neutron energy, essentially room temperature, so one must slow down the neutrons. This is accomplished by having the neutrons repeatedly collide with a light nucleus, the moderator. Some of the excess neutrons are absorbed by the fertile material and become fissile material in the reactor, and this extends the fuel life. However the nuclear reaction also produces byproducts, which slow down the reaction. After some time, the reactor core must be discharged and the reactor must be refueled. The waste products consist of two basic classes of product. These are first fission fragments, intermediate Z elements which are highly radioactive, and generally have a half life of 30 years. The second are actinides, which build up from neutron absorption reactions in the fertile material; not only plutonium, but higher actinides as well, Californium, Americium and higher still. Current American policy is to containerize all the waste together and store it in a place like Yucca Mountain, where it would have to remain for perhaps half a million years (the half life of plutonium is 24,000 years). The French approach is to reprocess the waste, remove the plutonium and use it as fuel. However in this approach, the plutonium fuel in the 238U fertile material is not completely burned and some fraction of the plutonium inserted is again a waste product.

Let us give a simple hypothetical example, just to illustrate the process in a very much over simplified way. Let us say that in each nuclear reaction, 2.2 neutrons are released. One is necessary to continue the chain reaction. Let us say that 0.7 are lost to various loss mechanisms, and 0.5 produce a fissile atom from fertile material. Since less than one fertile atom is produced per reaction, the reaction will ultimately run out of fuel and stop, even if there are no harmful bi products generated. The conversion ratio is defined as the number of fissile atoms produced per reaction, 0.5 in the example just given. A burner has a conversion ratio less than one, a breeder, greater.

The predominant reactor type in the world today is the light water reactor (LWR). It is certainly not the only option, but when all its advantages and drawbacks are weighed, it seems to be the optimal at this time, and likely in the future, as Freidberg and Kadak claim [20]. The lightest element for a moderator is, of course hydrogen, and this is why water is used. For the fertile material 238U is used with an enrichment of about 4 %. In other words, each year about one metric ton of 235U, mixed with 24 metric tons of 238U is inserted into the reactor. What is taken out of the reactor is about 750 kg of fission products, 200 kg of 239Pu and about 40 kg of higher actinides [19], along with nearly the 24 metric tons of 238U (now enriched at ~1 %) which went along for the ride. The conversion ratio for the LWR is about 0.6. However it is important to understand that conversion ratio is not the only measure of how deep the burn is can be. The burn is also limited by neutron poisons, for instance xenon which builds up and absorbs neutrons. For the LWR, as we just noted, a good bit of the material removed when it is refueled, is unburned 239Pu and 235U.

There are two important neutron loss mechanisms in LWR’s, which are possible to eliminate with alternate reactor designs. First of all, the hydrogen can absorb a neutron to form deuterium. This loss can be eliminated by using heavy water, (D2O), as the moderator instead of light water. It is not as effective a moderator due to the double mass of the deuterium, but is does eliminate an important neutron loss mechanism. It allows the reactor to run with natural uranium (0.7 % 235U) instead of enriched uranium. As its conversion ratio is typically 0.8, it burns deeper into the fertile material, so it might relax the requirements on any fusion reactor used to refuel it. This is the approach the Canadians have taken in their CANDU (Canada Deuterium Uranium) reactors. They are used throughout Canada, and the Canadians have had some success in selling them on the world market. The fuel, unenriched uranium is much cheaper, but the moderator, heavy water is much more expensive.

Secondly, using thorium as the fertile material rather than 238U greatly reduces the neutron absorption to form higher actinides. Not only does this also allows for much deeper burn, conversion ratios typically at least 0.8, it also greatly reduces the actinides in the waste stream. However by using a uranium isotope as the fissile material and thorium as the fertile material, the raw fuel presents a great proliferation risk. The fissile uranium and fertile thorium can be separated chemically, and the uranium can be used to make bombs. Certainly the United States could never export this fuel.

At one time the US Navy was interested in 233U fuel with thorium as a fertile material for use on submarines. It built a 233U reactor and it was very successful [76]. Also proliferation is not that much of a concern for fuel on a submarine. However 235U proved to be a better choice and the 233U reactor was abandoned. While the author has not found an example in the literature of a fission reactor using 233U as the fissile material and 238U as the fertile material, there is no reason it could not work. Also there is no reason for anyone to have tried it up to now. However it is a strong contender for a fusion breeding infrastructure.

Another choice for a reactor using thorium as a fertile material is the molten salt thorium reactor (MSTR). There the thorium is dissolved in a molten salt, which flows through the reactor. A prototype 7 MW reactor was build at Oak Ridge National Laboratory in the 1960s [77]. It too was successful, but again the effort was abandoned.

Regarding fertile material, Hans Bethe made an interesting suggestion in his classic 1979 Physics Today paper advocating fusion breeding (he called it the fusion hybrid) [13]. Since 238U as a fertile material gives better proliferation protection, but thorium produces a much easier waste stream to deal with and allows deeper burn, use a fertile material, which is half thorium and half 238U in a light water reactor. At about 8 % enrichment instead of 4 %, the raw fuel has nearly the same proliferation resistance, but the waste steam has only about half the actinide content.

As a fusion scientist, this author is not knowledgeable enough to comment, and also has no dog in the fight. A Large ITER sized reactor seems to be able to fuel 5 LWR’s of equal power. However anything which provides deeper burn, eases the requirements of the fusion reactor.

Fast Neutron Reactors

As neutron energy increases, the fission cross sections decrease, making design of the reactor more difficult. However reactors using fast (~ 2 MeV) neutrons have been built, and these do have advantages. Additional nuclear reactions become possible, especially more complete actinide burn, and more prolific production of fertile material. To build on our hypothetical discussion of the last sub section, let us say that the fast neutron produces 3 neutron in a fission event instead of 2.2; 0.5 are lost somehow. One is needed to continue the chain reaction, and the other 1.5 are used to breed fissile atoms, so here the conversion ratio is 1.5. One of these replaces the fertile atom lost in the reaction and the other 0.5 can be used to fuel other reactors. In this over simplified case, 2 fission breeders would power a single burner, as mentioned in the introduction.

However since the fission reaction cross sections are so low, this means that the coolant has to be virtually transparent to neutrons. This greatly restricts the choice of coolant. The most common coolant used is liquid sodium. While this sounds dangerous (it burns in air and explodes in water), liquid sodium is a common industrial material, and like many other dangerous industrial materials, there is a standard, time honored protocol for handling it, pumping it, etc.

The largest fast neutron breeder reactor hooked up to the grid was Super Phenix in France. From the outset it was plagued with difficulties and delays, mostly because there was little experience in dealing with the liquid sodium in the required large quantity. The French did finally get it to work, and it delivered 1.2 GWe to the grid for a few years. But it was always controversial and the Greens constantly protested it. On a change of government in 1997, Super Phenix was closed down [19]. However there is still one powerful breeder operational; this is BN-600 in Russia. It has safely generated 600MWe for years.

The fast neutron reactor of most interest here is the Integral fast reactor (IFR) developed by Argonne National Laboratory [8, 1418]. From the perspective of fusion breeding, the advantage of the IFR is that it can be run as a breeder or a burner, and the conversion ratio can vary from a low value to about 1.5 (that is 2 IFR’s power a single LWR of equal power). Run as a burner, it can completely burn all transuranics and actinides. If the burner is perfectly efficient and simply burns and does not breed at all, a single IFR can burn the actinide bi product of about 5 LWR’s. It does this with a process called pyroelectric separation (another common industrial process) which removes all transuranics and actinides from the waste stream, concentrates them and reinserts them into the reactor. Material posing proliferation risk never leave the reactor site, they are all burned there.

There seems to be little in the literature on the use of an IFR as a burner of actinide and transuranics from LWR’s. The author has been told [78] that a more realistic figure is that an IFR can burn the waste of two LWR’s on a continuous basis. Little seems to be in the literature to tie this figure down, and indeed the developers of the IFR did not see this as a primary motivation, so they had little reason to devote resources to this. However in a fusion breeder and LWR economy, the capability of an IFR to burn actinides is of paramount importance.

This series of articles [68] has constantly advocated the use of IFR’s to burn transuranics and actinide wastes. Once burned the only waste would be the highly radioactive, intermediate Z, fission products like say cobalt 60. These have a half life of ~30 years, so they could be stored for 300–600 years until they are inert. This is a time human society can reasonably plan for; not the half million years of so the nation would have to store plutonium, an enormous proliferation risk, whose half life is 24,000 years. With the burning of actinide wastes by IFR’s the need for geological disposal would be greatly reduced or even eliminated.

It may be that the author’s long dream is about to become a reality. The British have contracted with GE-Hitachi to build a reactor called PRISM [79]. It is a scaled up version of the IFR, to generate 600 MW, 10 times that of the IFR. However its specific purpose is to burn Britain’s plutonium stockpile. The British have about 100 tons of plutonium stored in various places. Since a 600 MWe reactor burns less than a ton of fissile material per year, and the British claim that they will be able to ‘treat’ (in some unspecified way) all the plutonium in 5 years, they are clearly not thinking of ‘burning’ it all. However the PRISM fact sheet (Google GE Hitachi Prism Fact Sheet), as well as other information on the web site (gehitachiprism.com) do explain that a longer term goal is indeed to ‘burn’ all of the plutonium the British are storing, as well a to generate electricity from it. Hence clearly it is possible to refuel the PRISM and burn the plutonium on a continuing basis. If PRISM is successful, this could be a development of enormous importance.

Problems with Nuclear Energy: and Their Solutions

Opponents of nuclear power constantly say it is too expensive; it must and will be phased out as a failed attempt. There is a one word answer to this argument; France. France, for decades, has gotten 80 % of its electricity from nuclear reactors. Somehow they manage to pay for it without going broke. As a result, the French emit considerably less CO2 into the atmosphere than other European countries their size [2].

Just because nuclear generated electricity is definitely and demonstrably affordable, does not mean it is the cheapest. Natural gas will likely be cheaper, at least in the United States for decades. But it is obviously unwise to put all of our eggs in one basket. Nuclear, which will almost certainly be vital by mid century, should play an important role today.

The author’s 2004 and 2006 papers [6, 7] discussed other problems of nuclear energy mentioning safety, proliferation, waste and fuel supply. The 2006 paper blithely dismissed safety concerns with the following statement:

“Nuclear power has been criticized as dangerous and expensive almost since its birth. However in the past 25 years, the industry has compiled an impressive safety record, certainly much better than that of the coal industry.”

Oops! Well, at least the last part is still true. The coal industry has killed many more people than the nuclear industry. But in the light of the Fukushima nuclear disaster, the safety of nuclear reactors in the event of extreme events obviously has to be reexamined. This author has neither expertise nor suggestions, but is confident this can be accomplished. However there are a few factors worth considering regarding technology, biology and psychology.

First consider the technology. In a nuclear disaster the main danger is to land, not people. The Fukushima earthquake and tsunami killed about 20,000 people. Hopefully, without tempting fate, the number killed so far by radiation, is zero. The people had time to get away and the nuclear workers there seem to be adequately protected. However a large area of land near the reactors may be contaminated. Thus after immediately getting people away and securing the reactor, the ultimate decontamination of the land is the highest priority in a nuclear disaster. Our recommendation is that the nuclear industry dedicate considerable resources to researching what is required here and preparing (hopefully on a worldwide basis) for such an eventuality. This will be needed only very, very rarely (and hopefully never), but once needed, the need will be acute.

Now consider the biology and ask just how decontaminated does the land have to be for humans to safely live and work there. Humans do live in a constant bath of radiation, coming from both the sky and earth. An important issue is just how harmful is low level radiation to humans? One answer is from the cancer rates of the Hiroshima and Nagasaki survivors. There are about 100,000 of them, and their health has been monitored for decades by both the American and Japanese governments. Obviously they had been exposed to vast quantities of radiation. Yet only about 800 cases of additional cancers have been caused by this exposure [80]. In other words, fewer than 1 % of the survivors had developed a cancer which can be attributed to the radiation they received. This corresponds to about a 10 % increase in the cancer rate among the survivors. But in the United States, the state to state variation of cancer rate varies from about 380–510 cancers per hundred thousand people per year, an increase of 30 % from the state with the least cancer to that with the most.

While obviously nobody thinks that Hiroshima levels of radiation are acceptable, it does cause one to question just how harmful low levels are. At a level just a few times above background, the epidemiological measure of increase in cancer rate would be difficult to detect, because the natural and other man made causes are so much higher. This assumes the linear no threshold (LNT) model; that is every little bit of radiation causes some additional increase to the cancer rate, which is proportional to the radiation increase. However, medical personnel are more and more skeptical of LNT, since cells have repair mechanisms for damaged DNA. A study by Dr. Tubiana [81] and his team, published in the US National Institute of Health journal Radiology makes the case that below some specified level, increases in radiation have no adverse effects on humans. There is even a subculture, which believe that increases in low levels of radiation are beneficial to humans (Google Radiation hormesis), but this is obviously very speculative.

In any case, in decontamination of the land, technology most likely says that the levels cannot be brought to zero, and biology says they do not have to be. Likely it can be a few times above background and the result will be, at worst a nearly unmeasureable uptick in cancer rate, and more likely no increase at all.

But then there is psychology. It would be wonderful if there were an abundant, universally accepted, perfectly safe energy source. But we do not live in a perfect world. Any energy source has risks. Coal has killed many more people and destroyed much more land than nuclear ever will, and it may also cause destructive climate change. Yet people seem willing to accept coal over nuclear. Any time an anti-nuclear activist lights up a cigarette, she incurs a much greater cancer risk, than she does standing outside the nuclear power plant with her sign. Driving to the demonstration, he is at much greater risk in his car. Yet these risks are all acceptable.

Who can figure? certainly not the author, who is neither a psychologist nor pastoral counselor. But these attitudes are not necessarily permanent. They can change over a generation or so, especially if the Japanese, hopefully with international help, can render the damaged reactors inert, and decontaminate and repopulate the land.

Regarding the other problems, fuel supply, proliferation risk and waste disposal, we briefly summarize what we have already mentioned. Fuel supply can be increase by nearly two orders of magnitude by breeding. The proliferation risk of the raw fuel is negligible as long as a sufficient amount of 238U is mixed in with the fertile 233U or 235U. Regarding the spent fuel, it is a highly radioactive witches brew. However the 239Pu can be separated out chemically and this is certainly a proliferation risk. However one of the main points of this paper is that the 239Pu can be separated out and burned in fast neutron reactors such as the IFR, virtually eliminating this proliferation risk.

As to waste, it comes in two forms, the actinides and transuranics, which can be burned in say an IFR. Also there are the radioactive fission products, typically with half life 30 years of less. These are the only true wastes of a properly done nuclear infrastructure. The only option here is simply to store them for 300–600 years until they become basically inert. This is a time human society can reasonably plan for.

Sakharov and Bethe

Of all of the giants of twentieth century physics, two of the most gigantic were Andrei Sakharov and Hans Bethe. Yet both of these giants advocated fusion breeding [12, 13]. Sakharov, in his book, described it as one his most important ideas:

“An important proposal of mine (in 1951 or late 1950) was that neutrons from thermonuclear reactions be used for breeding purposes”.

This is probably the earliest time anyone advocated the concept. Hans Bethe advocated it in a 1979 article in Physics Today. Among other things he suggested the idea of a fertile material being a mixture of 238U (to minimize proliferation risks) and 232Th (to minimize actinide production in the waste stream). Bethe also advocated fusion breeding rather than any other form of hybrid fusion, as his article emphasized the number of satellite reactors a single fusion breeder could support. Both of these giants have advocated fusion breeding, yet the fusion establishment ignored them for virtually the entire history of the fusion effort.

This author believes that Sakharov and Bethe had it right, and the fusion establishment has it wrong.

The Energy Park

In his earlier publications on the subject [69, 11], the author has proposed “The Energy Park” as a possible unit of power production for the future. It was called ‘more than a dream, but much less than a careful plan’, the idealization being that only thorium comes in, only electricity and/or manufactured liquid fuel goes out.

Basically The Energy Park consists of five 1 GWe LWR’s powered by a single Large ITER size 1GWe fusion reactor. The actinide waste is burned on site by a 1GWe IFR, if run as an essentially perfectly efficient burner. If instead it takes two IFR’s [78] to ‘burn’ the actinide wastes of 4 LWR’s, this is obviously a possible variation. The canonical configuration for the LWR’s has a 233U fissile fuel, at about 4 % concentration, in 238U, the fertile material. This provides maximum proliferation resistance of the raw fuel. However other configurations are also possible. If some thorium is mixed in with the fertile material, some proliferation resistance is lost, but the requirements on both the fusion breeder and the actinide burner are relaxed. The reason is that the LWR’s now breed some 233U themselves, and produce less actinides in the waste stream. Very likely a single IFR would then suffice to burn all the actinide bi products of the LWR’s. Obviously there are many possible variations. However one common theme to all, is that there is no long time storage, nor long distance travel of any material with proliferation risk. There is no build up of material with proliferation risk anywhere, it is all burned on site. Only thorium comes in, only electricity and/or manufactured liquid fuel goes out.

The energy park is several square miles in area, and is surrounded by a low security fence. Inside the energy park is a high security fence, and inside this inner fence are the fusion breeder, the chemical separation plant where actinides are separated from waste stream, and the actinide burner. Thus, the industrial facilities which produce material with great proliferation risk, the 233U, and the actinides in the waste stream are all inside the high security fence are. The five LWR’s, the storage facility for the radioactive fission products and perhaps a liquid fuel factory are all outside the high security fence, but inside the low security fence.

A schematic of the energy park is shown in Fig. 15. Assuming GE-Hitachi and the British are successful with their PRISM reactor, the only component of the energy park, which does not now exist, is the fusion breeder.

Fig. 15
figure 15

The energy park, more than a dream, much less than a careful plan, ideally only thorium comes in, only electricity and liquid fuel go out. There is no long time storage, nor long distance travel of material with proliferation potential, they are destroyed in the park: A Low security fence; B Five 1GWe nuclear reactors, likely but not necessarily LWR’s; C electricity going out; D manufactured liquid fuel pipeline; E cooling pool or other storage facility where fission products are stored for the 300–500 years necessary to render them basically inert; F Liquid or Gaseous fuel factory; G High security fence; H Separation plant. Fission products go to storage (E), actinides go to; I, IFR or other fast neutron reactor where the actinides (e.g. 239Pu) are burned and; J The fusion breeder

Conclusions: Nirvana or Achievability?

Pure fusion might be nirvana, but it is a long way from being achieved. It is not at all clear how tokamaks will ever get past conservative design rules, as they must for pure fusion, but certainly need not for fusion breeding. Regarding lasers, there does not seem to be any conservative design like rule restricting their parameter space, but results from the National Ignition Campaign at NIF make it clear that there are large gaps in our understanding, at least for indirect drive. However fusion breeding might be available for large scale power production by mid century. Nobody believes that is true for pure fusion.

Where will we be at mid century? The entire world, not just the currently wealthier parts of it will demand many terawatts of power. But it will almost certainly be clear then (as it is to many of us now) that solar, wind and biofuel will not be able to deliver it. Fossil fuel and mined 235U may be well on their way to depletion; in addition fossil fuel may by then be causing unacceptable climate change. Fission and fusion breeding are about all that is left. If the fusion community makes a real and sustained effort to deliver economical breeders, it is unlikely that either can deliver a knockout blow to the other. Each has a significant advantage. Fission breeding has a shorter development path. However fusion breeding is a much more prolific fuel source, so it fits in well with today’s nuclear infrastructure, while fission breeding does not. Freidberg and Kadak [20] claim that LWR’s will be the infrastructure for a very long time. It is the choice that the world has made so far, knowing all of their advantages and drawbacks. Who knows where things will be at mid century, but shouldn’t fusion be ready to jump in?

But to develop fusion breeders will take an enormous effort and about a 30° change in course for the fusion effort. Fission breeder developers are making a real effort to be ready by mid century. Will the fusion effort then still be claiming that they will have a DEMO in 35 years, as they have been for the last 50? (I remember reading as a teenager, a Life magazine article, in the 1950s, about fusion powered rocket ships in 30 years). Fission breeding may be unable to deliver a knockout blow to fusion, but we are doing a pretty good job of delivering one to ourselves.

Let’s change direction, let’s propose and attempt to do something we can actually accomplish in a reasonable time. Let’s make a real effort to develop fusion breeding by mid century, before our sponsors completely lose patience with us. Let’s not lose the fusion project because perfect is the enemy of good enough.