≡ Menu

A Hybrid Interstellar Mission Using Antimatter

Epsilon Eridani has always intrigued me because in astronomical terms, it’s not all that far from the Sun. I can remember as a kid noting which stars were closest to us – the Centauri trio, Tau Ceti and Barnard’s Star – wondering which of these would be the first to be visited by a probe from Earth. Later, I thought we would have quick confirmation of planets around Epsilon Eridani, since it’s a scant (!) 10.5 light years out, but despite decades of radial velocity data, astronomers have only found one gas giant, and even that confirmation was slowed by noise-filled datasets.

Even so, Epsilon Eridani b is confirmed. Also known as Ægir (named for a figure in Old Norse mythology), it’s in a 3.5 AU orbit, circling the star every 7.4 years, with a mass somewhere between 0.6 and 1.5 times that of Jupiter. But there is more: We also get two asteroid belts in this system, as Gerald Jackson points out in his new paper on using antimatter for deceleration into nearby star systems, as well as another planet candidate.

Image: This artist’s conception shows what is known about the planetary system at Epsilon Eridani. Observations from NASA’s Spitzer Space Telescope show that the system hosts two asteroid belts, in addition to previously identified candidate planets and an outer comet ring. Epsilon Eridani is located about 10 light-years away in the constellation Eridanus. It is visible in the night skies with the naked eye. The system’s inner asteroid belt appears as the yellowish ring around the star, while the outer asteroid belt is in the foreground. The outermost comet ring is too far out to be seen in this view, but comets originating from it are shown in the upper right corner. Credit: NASA/JPL-Caltech/T. Pyle (SSC).

This is a young system, estimated at less than one billion years. For both Epsilon Eridani and Proxima Centauri, deceleration is crucial for entering the planetary system and establishing orbit around a planet. The amount of antimatter available will determine our deceleration options. Assuming a separate method of reaching Proxima Centauri in 97 years (perhaps beamed propulsion getting the payload up to 0.05c), we need 120 grams of antiproton mass to brake into the system. A 250 year mission to Epsilon Eridani at this velocity would require the same 120 grams.

Thus we consider the twin poles of difficulty when it comes to antimatter, the first being how to produce enough of it (current production levels are measured in nanograms per year), the second how to store it. Jackson, who has long championed the feasibility of upping our antimatter production, thinks we need to reach 20 grams per year before we can start thinking seriously about flying one of these missions. But as both he and Bob Forward have pointed out, there are reasons why we produce so little now, and reasons for optimism about moving to a dedicated production scenario.

Past antiproton production was constrained by the need to produce antiproton beams for high energy physics experiments, requiring strict longitudinal and transverse beam characteristics. Their solution was to target a 120 GeV proton beam into a nickel target [41] followed by a complex lithium lens [42]. The world record for the production of antimatter is held by the Fermilab. Antiproton production started in 1986 and ended in 2011, achieving an average production rate of approximately 2 ng/year [43]. The record instantaneous production rate was 3.6 ng/year [44]. In all, Fermilab produced and stored 17 ng of antiprotons, over 90% of the total planetary production.

Those are sobering numbers. Can we cast antimatter production in a different light? Jackson suggests using our accelerators in a novel way, colliding two proton beams in an asymmetric collider scenario, in which one beam is given more energy than the other. The result will be a coherent antiproton beam that, moving downstream in the collider, is subject to further manipulation. This colliding beam architecture makes for a less expensive accelerator infrastructure and sharply reduces the costs of operation.

The theoretical costs for producing 20 grams of antimatter per year are calculated under the assumption that the antimatter production facility is powered by a square solar array 7 km x 7 km in size that would be sufficient to supply all of the needed 7.6 GW of facility power. Using present-day costs for solar panels, the capital cost for this power plant comes in at $8 billion (i.e., the cost of 2 SLS rocket launches). $80 million per year covers operation and maintenance. Here’s Jackson on the cost:

…3.3% of the proton-proton collisions yields a useable antiproton, a number based on detailed particle physics calculations [45]. This means that all of the kinetic energy invested in 66 protons goes into each antiproton. As a result, the 20 g/yr facility would theoretically consume 6.7 GW of electrical power (assuming 100% conversion efficiencies). Operating 24/7 this power level corresponds to an energy usage of 67 billion kW-hrs per year. At a cost of $0.01 per kW-hr the annual operating cost of the facility would be $670 million. Note that a single Gerald R. Ford–class aircraft carrier costs $13 billion! The cost of the Apollo program adjusted for 2020 dollars was $194 billion.

Science Along the Way

Launching missions that take decades, and in some cases centuries, to reach their destination calls for good science return wherever possible, and Jackson argues that an interstellar mission will determine a great deal about its target star just by aiming for it. Whereas past missions like New Horizons could count on the position of targets like Pluto and Arrokoth being programmed into the spacecraft computers, the preliminary positioning information uploaded to the craft came from Earth observation. Our interstellar craft will need more advanced tools. It will have to be capable of making its own astrometrical observations, sending its calculations to the propulsion system for deceleration into the target system and orbital insertion, thus refining exoplanet parameters on the fly.

Remember that what we are considering is a hybrid mission, using one form of propulsion to attain interstellar cruise velocity, and antimatter as the method for deceleration. You might recall, for example, the starship ISV Venture Star in the film Avatar, which uses both antimatter engines and a photon sail. What Jackson has added to the mix is a deep dive into the possibilities of antimatter for turning what would have been a flyby mission into a long-lasting planet orbiter.

Let’s consider what happens along the line of flight as a spacecraft designed with these methods makes its way out of the Solar System. If we take a velocity of 0.02c, our spacecraft passes the outgoing Voyager and Pioneer spacecraft in two years, and within three more years it passes into the gravitational lensing regions of the Sun beginning at 550 AU. A mere five years has taken the vehicle through the Kuiper Belt and moved it out toward the inner Oort Cloud, where little is currently known about such things as the actual density distribution of Oort objects as a function of radius from the Sun. We can also expect to gain data on any comparable cometary clouds around Proxima Centauri or Epsilon Eridani as the spacecraft continues its journey.

By Jackson’s calculations, when we’re into the seventh year of such a mission, we are encountering Oort Cloud objects at a pretty good clip, with an estimated 450 Oort objects within 0.1 AU of its trajectory based on current assumptions. Moving at 1 AU every 5.6 hours, we can extrapolate an encounter rate of one object per month over a period of three decades as the craft transits this region. Jackson also notes that data on the interstellar medium, including the Local Interstellar Cloud, will be prolific, including particle spectra, galactic cosmic ray spectra, dust density distributions, and interstellar magnetic field strength and direction.

Image: This is Figure 7 from the paper. Caption: Potential early science return milestones for a spacecraft undergoing a 10-year acceleration burn with a cruise velocity of 0.02c. Credit: Gerald Jackson.

It’s interesting to compare science return over time with what we’ve achieved with the Voyager missions. Voyager 2 reached Jupiter about two years after launch in 1977, and passed Saturn in four. It would take twice that time to reach Uranus (8.4 years into the mission), while Neptune was reached after 12. Voyager 2 entered the heliopause after 41.2 years of flight, and as we all know, both Voyagers are still returning data. For purposes of comparison, the Voyager 2 mission cost $865 million in 1973 dollars.

Thus, while funding missions demands early return on investment, there should be abundant opportunity for science in the decades of interstellar flight between the Sun and Proxima Centauri, with surprises along the way, just as the Voyagers occasionally throw us a curveball – consider the twists and wrinkles detected in the Sun’s magnetic field as lines of magnetic force criss-cross, and reconnect, producing a kind of ‘foam’ of magnetic bubbles, all this detected over a decade ago in Voyager data. The long-term return on investment is considerable, as it includes years of up-close exoplanet data, with orbital operations around, for example, Proxima Centauri b.

It will be interesting to see Jackson’s final NIAC report, which he tells me will be complete within a week or so. As to the future, a glimpse at one aspect of it is available in the current paper, which refers to what the original NIAC project description referred to as “a powerful LIDAR system…to illuminate, identify and track flyby candidates” in the Oort Cloud. But as the paper notes, this now seems impractical:

One preliminary conclusion is that active interrogation methods for locating 10 km diameter objects, for example with the communication laser, are not feasible even with megawatts of available electrical power.

We’ll also find out in the NIAC report whether or not Jackson’s idea of using gram-scale chipcraft for closer examination of, say, objects in the Oort has stood up to scrutiny in the subsequent work. This hybrid mission concept using antimatter is rapidly evolving, and what lies ahead, he tells me in a recent email, is a series of papers expanding on antimatter production and storage, and further examining both the electrostatic trap and electrostatic nozzle. As both drastically increasing antimatter production, as well as learning how to maximize small amounts, are critical for our hopes to someday create antimatter propulsion, I’ll be tracking this report closely.

tzf_img_post
{ 34 comments }

Antimatter-driven Deceleration at Proxima Centauri

Although I’ve often seen Arthur Conan Doyle’s Sherlock Holmes cited in various ways, I hadn’t chased down the source of this famous quote: “When you have eliminated all which is impossible, then whatever remains, however improbable, must be the truth.” Gerald Jackson’s new paper identifies the story as Doyle’s “The Adventure of the Blanched Soldier,” which somehow escaped my attention when I read through the Sherlock Holmes corpus a couple of years back. I’m a great admirer of Doyle and love both Holmes and much of his other work, so it’s good to get this citation straight.

As I recall, Spock quotes Holmes to this effect in one of the Star Trek movies; this site’s resident movie buffs will know which one, but I’ve forgotten. In any case, a Star Trek reference comes into useful play here because what Jackson (Hbar Technologies, LLC) is writing about is antimatter, a futuristic thing indeed, but also in Jackson’s thinking a real candidate for a propulsion system that involves using small amounts of antimatter to initiate fission in depleted uranium. The latter is a by-product of the enrichment of natural uranium to make nuclear fuel.

Both thrust and electrical power emerge from this, and in Jackson’s hands, we are looking at a mission architecture that can not only travel to another star – the paper focuses on Proxima Centauri as well as Epsilon Eridani – but also decelerate. Jackson has been studying the matter for decades now, and has presented antimatter-based propulsion concepts for interstellar flight at, among other venues, symposia of the Tennessee Valley Interstellar Workshop (now the Interstellar Research Group). In the new paper, he looks at a 10-kilogram scale spacecraft with the capability of deceleration as well as a continuing source of internal power for the science mission.

Image: Depiction of the deceleration of interstellar spacecraft utilizing antimatter concept. Credit: Gerald Jackson.

On the matter of the impossible, the quote proves useful. Jackson applies it to the propulsion concepts we normally think of in terms of making an interstellar crossing. This is worth quoting:

Applying this Holmes Method to space propulsion concepts for exoplanet exploration, in this paper the term “impossible” is re-interpreted arbitrarily to mean any technology that requires: 1) new physics that has not been experimentally validated; 2) mission durations in excess of one thousand years; and 3) material properties that are not currently demonstrated or likely to be achievable during this century. For example, “warp drives” can currently be classified as impossible by criterion #1, and chemical rockets are impossible due to criterion #2. Breakthrough Starshot may very well be impossible according to criterion #3 simply because of the needed material properties of the accelerating sail that must survive a gigawatt laser beam for 30 minutes. Though traditional nuclear thermal rockets fail due to criterion #2, specific fusion-based propulsion systems might be feasible if breakeven nuclear fusion is ever achieved.

Can antimatter supply the lack? The kind of mission Jackson has been analyzing uses antimatter to initiate fission, so we could consider this a hybrid design, one with its roots in the ‘antimatter sail’ Jackson and Steve Howe have described in earlier technical papers. For the background on this earlier work, you can start by looking at Antimatter and the Sail, one of a number of articles here on Centauri Dreams that has explored the idea.

In this paper, we move the antimatter sail concept to a deceleration method, with the launch propulsion being handed off to other technologies. The sail’s antimatter-induced fission is not used only to decelerate, though. It also provides a crucial source of power for the decades-long science mission at target.

If we leave the launch and long cruise of the mission up to other technologies, we might see the kind of laser-beaming methods we’ve looked at in other contexts as part of this mission. But if Breakthrough Starshot can develop a model for a fast flyby of a nearby star (moving at a remarkable 20 percent of lightspeed) via a laser array, various problems emerge, especially in data acquisition and return. On the former, the issue is that a flyby mission at these velocities allows precious little time at target. Successful deceleration would allow in situ observations from a stable exoplanet orbit.

That’s a breathtaking idea, given how much energy we’re thinking about using to propel a beamed-sail flyby, but Jackson believes it’s a feasible mission objective. He gives a nod to other proposed deceleration methods, which have included using a ‘magnetic sail’ (magsail) to brake against a star’s stellar wind. The problem is that the interstellar medium is too tenuous to slow a craft moving at a substantial percentage of lightspeed for orbital insertion upon arrival – Jackson considers the notion in the ‘impossible’ camp, whereas antimatter may come in under the wire as merely ‘improbable.’ That difference in degree, he believes, is well worth exploring.

The antimatter concept described generates a high specific impulse thrust, with the author noting that approximately 98 percent of antiprotons that stop within uranium induce fission. It turns out that antiproton annihilation on the nucleus of any uranium isotope – and that includes non-fissile U238 – induces fission. In Jackson’s design, about ten percent of the annihilation energy released is channeled into thrust.

Jackson analyzes an architecture in which the uranium “propagates as a singly-charged atomic ion beam confined to an electrostatic trap.” The trap can be likened in its effects to what magnetic storage rings do when they confine particle beams, providing a stable confinement for charged particles. Antiprotons are sent in the same direction as the uranium ions, reaching the same velocity in the central region, where the matter/antimatter annihilation occurs. Because the uranium is in the form of a sparse cloud, the energetic fission ‘daughters’ escape with little energy loss.

Here is Jackson’s depiction of an electrostatic annihilation trap. In this design, both the positively charged uranium ions and the negatively charged antiprotons are confined.

Image: This is Figure 1 from the paper. Caption: Axial and radial confinement electrodes (top) and two-species electrostatic potential well (bottom) of a lightweight charged-particle trap that mixes U238 with antiprotons.

A workable design? The author argues that it is, saying:

Longitudinal confinement is created by forming an axial electrostatic potential well with a set of end electrodes indicated in figure 1. To accomplish the goal of having oppositely charged antiprotons and uranium ions traveling together for the majority of their motion back and forth (left/right in the figure) across the trap, this electrostatic potential has a double-well architecture. This type of two-species axial confinement has been experimentally demonstrated [53].

The movement of antiprotons and uranium ions within the trap is complex:

The antiprotons oscillate along the trap axis across a smaller distance, reflected by a negative potential “hill”. In this reflection region the positively charged uranium ions are accelerated to a higher kinetic energy. Beyond the antiproton reflection region a larger positive potential hill is established that subsequently reflects the uranium ions. Because the two particle species must have equal velocity in the central region of the trap, and the fact that the antiprotons have a charge density of -1/nucleon and the uranium ions have a charge density of +1/(238 nucleons), the voltage gradient required to reflect the uranium ions is roughly 238 times greater than that required to reflect the antiprotons.

The design must reckon with the fact that the fission daughters escape the trap in all directions, which is compensated for through a focusing system in the form of an electrostatic nozzle that produces a collimated exhaust beam. The author is working with a prototype electrostatic trap coupled to an electrostatic nozzle to explore the effects of lower-energy electrons produced by the uranium-antiproton annihilation events as well as the electrostatic charge distribution within the fission daughters.

Decelerating at Proxima Centauri in this scheme involves a propulsive burn lasting ten years as the craft sheds kinetic energy on the long arc into the planetary system. Under these calculations, a 200 year mission to Proxima requires 35 grams of total antiproton mass. Upping this to a 56-year mission moving at 0.1 c demands 590 grams.

Addendum: I wrote ’35 kilograms’ in the above paragraph before I caught the error. Thanks, Alex Tolley, for pointing this out!

Current antimatter production remains in the nanogram range. What to do? In work for NASA’s Innovative Advanced Concepts office, Jackson has argued that despite minuscule current production, antimatter can be vastly ramped up. He believes that production of 20 grams of antimatter per year is a feasible goal. More on this issue, to which Jackson has been devoting his life for many years now, in the next post.

The paper is Jackson, “Deceleration of Exoplanet Missions Utilizing Scarce Antimatter,” in press at Acta Astronautica (2022). Abstract.

tzf_img_post
{ 27 comments }

An Abundance of Technosignatures?

What expectations do we bring to the hunt for life elsewhere in the universe? Opinions vary depending on who has the podium, but we can neatly divide the effort into two camps. The first looks for biosignatures, spurred by our remarkably growing and provocative catalog of exoplanets. The other explicitly looks for signs of technology, as exemplified by SETI, which from the start hunted for signals produced by intelligence.

My guess is that a broad survey of those looking for biosignatures would find that they are excited by the emerging tools available to them, such as new generations of ground- and space-based telescopes, and the kind of modeling we saw in the last post applied to a hypothetical Alpha Centauri planet. We use our growing datasets to examine the nature of exoplanets and move beyond observation to model benchmarks for habitable worlds, including their atmospheric chemistry and even geology.

Technosignatures are a different matter, and it’s fascinating to read through a new paper from Jason Wright and colleagues. – Jacob Haqq-Misra, Adam Frank, Ravi Kopparapu, Manasvi Lingam and Sofia Sheikh – discussing just how. The intent is to show that technosignatures offer a vast search space that in a sense dwarfs the hunt for biosignatures. That’s not what you would expect, as the latter are usually described as a kind of all-encompassing envelope within which technosignatures would be a subset.

On the contrary, write the authors, “there is no incontrovertible reason that technology could not be more abundant, longer-lived, more detectable, and less ambiguous than biosignatures.” How this potential is unlocked impacts how the search proceeds, and it also sends out a call for collaboration among all those hunting for life elsewhere.

Image: Photo of the central region of the Milky Way. Credit: UCLA SETI Group/Yuri Beletsky, Carnegie Las Campanas Observatory.

Technosignatures as Subset?

Remember that technosignatures do not require an intent to communicate, but are evidence of technologies in use or even long abandoned, perhaps found in already existing datasets needing re-examination, or in results from upcoming observatories. Check your own assumptions here, based on the Drake equation, in which factors include the fraction of habitable planets that develop life, the fraction that produce species that are intelligent and can communicate, and so on. Traditional thinking sees technosignatures as an embedded feature within a broader spectrum of life.

Reasonably enough, then, we might decide that if intelligence is a rare subset within biological systems, technosignatures would prove even rarer. Our own planet seems to exemplify this, with our species having become communicative only within roughly a century of today, despite 4.6 billion years in which to evolve. But Wright and team make the case that technology cannot be bounded in this way. Its emergence may be rare, but once it appears, it is possible that it will outlive its biological creators.

Biology may confine itself to a single habitable planet, but why should technosignatures be thus limited? In our own Solar System, we are producing, the authors argue, technosignatures for multiple worlds right now, especially at Mars, where we have our combined force of landers and orbital assets taking data and communicating results back to Earth. Such signals should increase as we follow through on plans to explore Mars with human crews and robotic spacecraft. As we spread into the Solar System, new technosignatures will emerge at each venue we study.

Why, too, should technology not spread through self-replication, perhaps not under the control of the biological beings who set it into motion? For that matter, why should we confine technology to planets? Places with no biology may prove extremely useful for our species, as for example the asteroid belt for resource extraction. We might expect technosignatures to emerge from these operations, another separate appearance of technology that grows ultimately out of the single planetary source. Moreover, this diaspora is unlikely to confine itself to a single star system, as the authors point out:

There is also no reason to think that technological life in the galaxy cannot spread beyond its home planetary system (see Mamikunian & Briggs 1965; Drake 1980). While interstellar spaceflight of the sort needed to settle a nearby star system is beyond humanity’s current capabilities, the problem is one being seriously considered now, and there are no real physical or engineering obstacles to such a thing happening (e.g., Mauldin 1992; Ashworth 2012; Lingam & Loeb 2021). Even if we cannot envision it happening for humans in the near future, it is not hard to imagine it transpiring in, say, 10,000 or 100,000 yr.

What a shift in thinking in the above paragraph, which to us merely states the obvious, when compared to a mere 75 years ago, a time when the idea of interstellar flight was considered science fictional in the extreme, and we were only beginning to probe the physics of the engines that might make crossing to another star possible. Today we’re more likely to be thinking about interstellar journeys as expeditions awaiting new generations of technology and engineering rather than a mystical new physics. We also factor artificial intelligence into an interstellar future that may be exclusively robotic.

Image: A rendering of a potential Dyson sphere, collect stellar energy on a system wide scale for highly advanced civilizations. How many separate technosignatures might have emerged out of a single biological source in the building of such a thing? Credit: sentientdevelopments.com.

Recall our recent discussion of von Neumann probes. While the average distance between stars is vast, Greg Matloff looked at the problem in an exceedingly practical way. Suppose, he said, we confine ourselves to times when stars are within a single light year of each other, which happens to our Sun every 500,000 years or so. If we launch a self-replicating probe only every 500,000 years, we nonetheless set up a process of such crossings that fills a large percentage of stellar systems in the galaxy within a time frame of tens of thousands of years. All of these can produce technosignatures.

Thus even the most conservative assumptions for interstellar flight using speeds not much beyond what we can achieve with a Jupiter gravity assist today still create the opportunity for technology to spread far beyond the planet of its origin. As the authors are quick to point out, the Drake equation cannot capture this spreading, and the search space for technosignatures could vastly outnumber that for biological life.

Lifetimes Civilizational and Technological

Looming over discussion of the Drake equation has always been the issue of the lifetime of a technological civilization, the L factor. How likely would we be to pick up a signal from another civilization if our own is threatened at this comparatively early stage of its growth by factors like nuclear or biological war? The Fermi question may be answered simply enough by saying that no technological species lives very long.

Here it’s fair to ask how much we are projecting human tendencies onto our extraterrestrial counterparts. This gets intriguing. The collapse of civilization would be a dire event, but absent actual extinction, our species might recover or, indeed, re-develop the technologies that once proliferated. The time between catastrophe and potential recovery is not known, but such events do not put a fixed limit on a civilization’s lifetime. Even if we assume that technological civilizations will roughly track our own, we may understand our own only imperfectly. From the paper:

…humanity is the first species on Earth that can prevent its own extinction with technology, for instance by diverting asteroids, stopping or mitigating pandemics, or building “lifeboat” settlements elsewhere in the solar system or beyond (Baum et al. 2015; Turchin & Green 2017; Turchin & Denkenberger 2018). This means that the upper limit on our technology’s survival is essentially unlimited in theory, even in the face of inevitable natural catastrophes. Apart from these modern examples, Earth-analogs from human history teach us that a technological downshift—to temporarily become less technological until circumstances improve—is a common and healthy adaptation to catastrophe in human history and that technology and longevity are in this way inextricably linked…

Nor can we rule out the possibility that the Earth could develop other species beyond our own in the future that can produce a technological society following humanity’s extinction. For that matter, are we so sure about our past? If there have been prior periods of technology on Earth, the processes of time over millions of years would likely have eradicated them. Thus using our experience on Earth as the model for the Drake L factor is inadvisable because of how little we know about L for our own planet.

Technosignatures can outlast the beings that create them, and as the authors point out, the ones we produce are already on a par with Earth’s biosignatures in terms of detectability. While we would not be able to detect the biosignatures of Earth from Alpha Centauri’s distance, the final iteration of the Square Kilometre Array should be sensitive enough to pick up our radars at distances of several parsecs, and an advanced space telescope within our engineering capabilities now (such as the proposed LUVOIR) might be able to detect atmospheric pollution at 10 parsecs.

It seems a safe assumption that if our biosignatures and technosignatures are roughly comparable in terms of detectability today, the advance of technology as a species continues to innovate should produce ever more robust technosignatures. We cannot, in other words, assume a biology-like trajectory, as implicit in the Drake equation, for the evolution of technosignatures and their detectability through SETI. Indeed:

…the spread of technology could reasonably imply that the number of sites of technosignatures might be larger than that of biosignatures, potentially by a factor of as much as > 1010 if the galaxy were to be virtually filled with technology.

No wonder some authors have considered adding a ‘spreading factor’ to the Drake equation, which accounts for the possibility of technologies moving far beyond their home worlds. Thus one technosphere produces myriad technosignatures, while the Drake equation in its classic form inevitably does not account for such growth. If the equation assumes life emerges and stays on its home world, the authors of this paper see technology as having a separate evolutionary arc which potentially takes it far into the galaxy in ever proliferating form.

While the search for biosignatures continues, it makes sense given all these factors for technosignatures to remain under active investigation, and to encourage the astrobiology and SETI communities to engage with each other in the common pursuit of extraterrestrial life. Comparative and cooperative analysis should enhance the work of both disciplines.

The paper is Wright et al., “The Case for Technosignatures: Why They May Be Abundant, Long-lived, Highly Detectable, and Unambiguous,” Astrophysical Journal Letters 927, L30 (10 March 2022). Full text.

tzf_img_post
{ 74 comments }

A New Title on Extraterrestrial Intelligence

Just a quick note for today as I finish up tomorrow’s long post. But I did want you to be aware of this new title, Extraterrestrial Intelligence: Academic and Societal Implications, which has connections with recent topics and will again tomorrow, when we discuss a new paper from Jason Wright and SETI colleagues on technosignatures. As with the recent biography of John von Neumann, I haven’t had the chance to read this yet, but it’s certainly going on the list. The book is out of Cambridge Scholars Publishing. Here’s the publisher’s description:

What are the implications for human society, and for our institutions of higher learning, of the discovery of a sophisticated extraterrestrial intelligence (ETI) operating on and around Earth? This book explores this timely question from a multidisciplinary perspective. It considers scientific, philosophical, theological, and interdisciplinary ways of thinking about the question, and it represents all viewpoints on how likely it is that an ETI is already operating here on Earth. The book’s contributors represent a wide range of academic disciplines in their formal training and later vocations, and, upon reflection on the book’s topic, they articulate a diverse range of insights into how ETI will impact humankind. It is safe to say that any contact or communication with ETI will not merely be a game changer for human society, but will also be a paradigm changer. This means that it makes sense for human beings to prepare themselves now for this important transition.

Important indeed, but how demoralizing to see another title at a stiff tariff: £63.99 (that’s about $84 US). I will spare you my thoughts on the academic side of publishing, and in the meantime see if I can get a review copy, as I assume most Centauri Dreams readers aren’t going to want to pony up this amount for a book they know little about (although if you live near a good academic library, this one should turn up there).

tzf_img_post
{ 4 comments }

Modeling a Habitable Planet at Centauri A/B

Why is it so difficult to detect planets around Alpha Centauri? Proxima Centauri is one thing; we’ve found interesting worlds there, though this small, dim star has been a tough target, examined through decades of steadily improving equipment. But Centauri A and B, the G-class and K-class central binary here, have proven impenetrable. Given that we’ve found over 4500 planets around other stars, why the problem here?

Proximity turns out to be a challenge in itself. Centauri A and B are in an orbit around a common barycenter, angled such that the light from one will contaminate the search around the other. It’s a 79-year orbit, with the distance between A and B varying from 35.6 AU to 11.2. You can think of them as, at their furthest, separated by the Sun’s distance from Pluto (roughly), and at their closest, by about the distance to Saturn.

The good news is that we have a window from 2022 to 2035 in which, even as our observing tools continue to improve, the parameters of that orbit as seen from Earth will separate Centauri A and B enough to allow astronomers to overcome light contamination. I think we can be quite optimistic about what we’ll find within the decade, assuming there are indeed planets here. I suspect we will find planets around each, but whether we find something in the habitable zone is anyone’s guess.

Image: This is Figure 1 from today’s paper. Caption: (a) Trajectories of α-Cen A (red) and B (blue) around their barycenter (cross). The two stars are positioned at their approximate present-day separation. The Hill spheres (dashed circles) and HZs (nested green circles) of A and B are drawn to scale at periapsis. (b) The apparent trajectory of B centered on A, with indications of their apparent separation on the sky over the period from CE 2020 to 2050. The part of trajectory in yellow indicates the coming observational window (CE 2022–2035) when the apparent separation between A and B is larger than 6 and the search for planets around A or B can be conducted without suffering significant contamination from the respective companion star. Credit: Wang et al.

If we don’t yet have a planet detection around the binary Centauri stars, we continue to explore the possibilities even as the search continues. Thus a new paper from Haiyang Wang (ETH Zurich), who along with colleagues at the university has been modeling the kind of rocky planet in the habitable zone that we hope to find there. The idea is to create the benchmarks that predict what this world should look like.

The numerical modeling involved examines the composition of the hypothetical world, drawing on what we do know, based on spectroscopic measurements, of the chemical composition of Centauri A and B. Here there is a great deal of information to work with, especially on so-called refractory elements, the iron, magnesium and silicon that go into rock formation. Centauri A and B are among the Gaia “benchmark stars” for which stellar properties have been carefully calibrated, and up to 22 elements have been found in high-quality spectra, so we know a lot about their chemical makeup.

But a key issue remains. While rocky planets are known to have rock and metal chemical compositions similar to that of their host stars, there is no necessary correspondence when it comes to the readily vaporized volatile elements. The authors suggest that this is because the process of planetary formation and evolution quickly does away with key telltale volatiles.

The researchers thus develop their own ‘devolatilization model’ to project the possible composition of a supposed habitable zone planet around Centauri A and B, linking stellar composition with both volatile and refractory elements. The model grew out of Wang’s work with Charley Lineweaver and Trevor Ireland at the Australian National University in Canberra, and it continues at Wang’s current venue at ETH. This is fundamentally new ground that extends our notions of exoplanet composition.

Wang and team call their imagined world ‘a-Cen-Earth,’ delving into its internal structure, mineralogy and atmospheric composition, all factors in evolution and habitability. The findings reveal a planet that is geochemically similar to Earth, with a silicate mantle, although carbon-bearing species like graphite and diamond are enhanced. Water storage in the interior is roughly the same as Earth, but the deduced world has a somewhat larger iron core mixed with a possible lack of plate tectonics. Indeed, “…the planet may be in a Venus-like stagnant-lid regime, with sluggish mantle convection and planetary resurfacing, over most of its geological history.”

As to the atmosphere of the hypothetical world that grows out of Wang’s model, its early era shows an envelope rich in carbon dioxide, methane and water, which harks back to the Earth’s atmosphere in the Archean era, between 4 and 2.5 billion years ago. That gives life a promising start if we assume abiogenesis occurring in a similar environment.

Image: α Centauri A (left) and α Centauri B viewed by the Hubble Space Telescope. At a distance of 4.3 light-​years, the α Centauri group (which includes also the red dwarf α Centauri C) is the nearest star system to Earth. Credit: ESA/Hubble & NASA.

How far can we take a model like this? We may soon have data to measure it against, but it’s worth remembering what the paper’s authors point out. After noting that planets around the “Sun-like” Centauri A and B cannot be extrapolated from the already known planets around the red dwarf Proxima Centauri, they go on to say:

Second, although α Cen A and B are “Sun-like” stars, their metallicities are ∼72% higher than the solar metallicity (Figure 3). How this difference would affect the condensation/evaporation process, and thus the devolatilization scale, is the subject of ongoing work (Wang et al. 2020b).

That’s a big caveat and a useful pointer to the needed clarification that further work on the matter should bring – metallicity is obviously significant. The paper adds:

Third, we ignore any potential effect of the “binarity” of the stars on their surrounding planetary bulk chemistry during planet formation, even though we highlight that, dynamically, the planetary orbits in the HZ around either companion are stable. Finally, we have yet to explore a larger parameter space, e.g., in mass and radius, but have only benchmarked our analysis with an Earth-sized planet, which would otherwise have an impact on the interior modeling…

So we’re in early days with planet modeling using these methods, which are being examined and extended through the team’s collaborations at Switzerland’s National Centre of Competence in Research PlanetS. Note too that the authors do not inject any catastrophic impact into their model of the sort that could affect both a planet’s mantle and/or its atmosphere, with dramatic consequences for the outcome. We know from the Earth’s experience in the Late Heavy Bombardment that this can be a factor.

With all this in mind, it’s fascinating to see the lines of observation and theory converging on the Alpha Centauri binary pair. Finding a habitable zone planet around Proxima Centauri was exhilarating. How much more so to go beyond the many imponderables of red dwarf planet habitability to two stars much more like our Sun, each of which might have a planet in its habitable zone? The Alpha Centauri triple system may turn out to be a bonanza, showing us both red dwarf and Sun-like planetary outcomes in a single system that just happens to be the closest to us.

The paper is Wang et al,, “A Model Earth-sized Planet in the Habitable Zone of α Centauri A/B,” The Astrophysical Journal Vol. 927, No. 2 (10 March 2022). Abstract/Full Text. Preprint also available.

tzf_img_post
{ 26 comments }

Why Fill a Galaxy with Self-Reproducing Probes?

We can’t know whether there is a probe from another civilization – a von Neumann probe of the sort we discussed in the previous post – in our own Solar System unless we look for it. Even then, though, we have no guarantee that such a probe can be found. The Solar System is a vast place, and even if we home in on the more obvious targets, such as the Moon, and near-Earth objects in stable orbits, a well hidden artifact a billion or so years old, likely designed not to draw attention to itself, is a tricky catch.

As with any discussion of extraterrestrial civilizations, we’re left to ponder the possibilities and the likelihoods, acknowledging how little we know about whether life itself is widely found. One question opens up another. Abiogenesis may be spectacularly rare, or it may be commonplace. What we eventually find in the ice moons of the outer system should offer us some clues, but widespread life doesn’t itself translate into intelligent, tool-making life. But for today, let’s assume intelligent toolmakers and long-lived societies, and ponder what their motives might be.

Let’s also acknowledge the obvious. In looking at motivations, we can only peer through a human lens. The actions of extraterrestrial civilizations, and certainly their outlook on existence itself, would be opaque to us. They would possibly act in ways we consider inexplicable, for reasons that defy the logic we apply to human decisions. But today’s post is a romp into the conjectural, and it’s a reflection of the fact that being human, we want to know more about these things and have to start somewhere.

Motivations of the Probe Builders

Greg Matloff suggests in his paper on von Neumann probes that one reason a civilization might fill the galaxy with these devices is the possibly universal wish to transcend death. A walk through the Roman ruins scattered around what was once the province of Gaul gave weight to the concept when my wife and I prowled round the south of France some years back. Humans, at least, want to put down a marker. They want to be remembered, and their imprint upon a landscape can be unforgettable.

But in von Neumann terms, I have trouble with this one. I stood next to a Roman wall near Saint-Rémy-de-Provence on a late summer day and felt the poignancy of all artifacts worn by time, but the Romans were decidedly mortal. They knew death was a horizon bounding a short life, and could transcend it only through propitiations to their gods and monuments to their prowess. A civilization that is truly long-lived, defined not by centuries but aeons, may have less regard for personal aggrandizement and even less sense of a coming demise. Life might seem to stretch indefinitely before it.

Image: Some of the ruins of the Roman settlement at Glanum in Saint-Rémy-de-Provence, recovered through excavations beginning in 1921. Walking here caused me to reflect on how potent memorials and monuments would be to a species that had all but transcended death. Would the impulse to build them be enhanced, or would it gradually disappear?

Probes as a means of species reproduction, another Matloff suggestion, ring more true to me, and I would suggest this may flag a biological universal, the drive to preserve the species despite the death of the individual. Here we’re in familiar science fiction terrain in which biological material is preserved by machines and flung to the stars, to be activated upon arrival and raised to awareness by artificial intelligence. Or we could go further – Matloff does – to say that biological materials may prove unnecessary, with computer uploads of the minds of the builders taking their place, another SF trope.

I can go with that as a satisfactory motivator, and it’s enough to make me want to at least try to find what Jim Benford calls ‘lurkers’ in our own corner of the galaxy. Another motivator that deeply satisfies me because it’s so universal among humankind is simple curiosity. A long-lived, perhaps immortal civilization that wants to explore can send von Neumann probes everywhere possible in the hope of learning everything it can about the universe. Encyclopedia Galactica? Why not? Imputing any human motive to an extraterrestrial civilization is dangerous, of course, but we have little else to go on. And centuries of human researchers and librarians attest to the power of this one.

Would such probes be configured to establish communication with any societies that arise on the planets under observation? This is the Bracewell probe notion that extends von Neumann self-reproduction to include this much more immediate form of SETI, with potential knowledge stored at planetary distances. Obviously, 2001: A Space Odyssey comes to mind as we recall the mysterious monoliths found on the early Earth and, much later, on the Moon, and the changes to humanity they portend.

But are long-lived civilizations necessarily friendly? Fred Saberhagen’s ‘berserker’ probes key off the Germanic and particularly Norse freelance bodyguards and specialized troops that became fixtures at the courts of royalty in early medieval times (the word is from the Old Norse word meaning ‘bearskin’). These were not guys you wanted to mess with, and associations with their attire of bear and wolfskins seem to have contributed to the legend of werewolves. Old Norse records show that they were prominent at the court of Norway’s king Harald I Fairhair (reigned 872–930).

Because they made violence into a way of life, we should hope not to find the kind of probe that would be named after them, which might be sent out to eliminate competition. Thus Saberhagen’s portrayal of berserker probes sterilizing planets just as advanced life begins to appear. The fact that we have not yet been sterilized may be due to the possibility that such a probe does not yet consider us ‘advanced,’ but more likely implies we have no berserker probes nearby. Let’s hope to keep it that way.

Or what about the spread of life itself? If abiogenesis does turn out to be unusually rare, it’s possible that any civilization with the power to do so would decide to seed the cosmos with life. In this case, we’re not sending uploaded intelligence or biological beings in embryonic form in our probes, but rather the most basic lifeforms that can proliferate on any planets offering the right conditions for their development. Perhaps there becomes an imperative – written about, for example, by Michael Mautner and Matloff himself – to spread life as a way to transform the cosmos. Milan Ćirković continues to explore the implications of just such an effort.

In an interesting post in Sentient Developments, Canadian futurist George Dvorsky points out that self-reproduction has more than an outward-looking component. Supposing a civilization interested in building a megastructure – a Dyson sphere, let’s say – decides to harness self-reproduction to supply the needed ‘worker’ devices that would mine the local stellar system and create the object in question.

At a truly cosmic level, Matloff speculates, self-replicating probes might be deployed to build megastructures that could alter the course of cosmic evolution. We’re in Stapledon territory now, freely mixing philosophy and wonder. We’re also in the arena claimed by Frank Tipler in his The Physics of Immortality (Doubleday, 1994).

We’ll want to search the Earth Trojan asteroids and co-orbitals for any indication of extraterrestrial probes, though it’s also true that the abundant resources of the Kuiper Belt might make operations there attractive to this kind of intelligence. One of the biggest questions has to do with the size of such probes. Here I’ll quote Matloff:

In a search for active or quiescent von Neumann probes in the solar system, human science would contend with great uncertainty regarding the size of such objects. Some science fiction authors contend that these devices might be the size of small planetary satellites (see for example L. Johnson, Mission to Methone and A. Reynolds, Pushing Ice). On the other hand, Haqq-Misra and Kopparapu (2012) believe that they may be in the 1-10 m size range of contemporary human space probes and these might be observable.

But there may be a limit to von Neumann probe detection. If they can be nano-miniaturized as suggested by Tipler (1994), the solar system might swarm with them and detection efforts would likely fail.

I remember having a long phone conversation two decades ago with Robert Freitas on this very point. Freitas had originally come up with a self-reproducing probe concept at the macro-scale called REPRO, but went on to delve into the implications of nano-technology. He made Matloff’s point in our discussion: If probe technologies operate at this scale, the surface of planet Earth itself could be home to an observing network about which we would have no awareness. Self-reproductive probes will be hard to rule out, but looking where we can to screen for the obvious makes sense.

The paper is Matloff, “Von Neumann probes: rationale, propulsion, interstellar transfer timing,” International Journal of Astrobiology, published online by Cambridge University Press 28 February 2022 (abstract).

tzf_img_post
{ 48 comments }

Probing von Neumann Expansion

Before getting into the paper I want to discuss today, I want to mention the new biography of John von Neumann by Ananyo Bhattacharya. I make no comment on The Man from the Future (W. W. Norton & Company, 2022) yet because while I have a copy, I haven’t had time to read it. But be aware that it’s out there – it’s getting good reviews, and given the impact of this remarkable figure on everything from programmable computers to game theory and the interstellar dispersion of civilizations, it’s a book you’ll at least want to stick on your reference list.

I figure anyone who masters calculus by the age of eight, as von Neumann is reputed to have done, is going to turn out to make a substantial contribution somewhere. I’m also interested in how polymaths function, moving with what seems effortless ease through diverse fields of study and somehow leaving their mark on each. What a contrast to our age of micro-specialization, where relentless drilling down into a single topic – and this seems true of most academic disciplines – is the mode of choice.

Image: John von Neumann, shown here with technology that might have been more to his taste, the 18,000 vacuum-tube strong ENIAC. One can only wonder what the sybaritic mathematician would have made of quantum computing. If only he were here to tell us.

It’s a good time for this book to come out, because von Neumann isn’t exactly in the spotlight these days. In a review in Science, Dov Greenbaum and Mark Gerstein note that he seems to have dropped out of public view:

In 2022…von Neumann could be the smartest person most people have never heard of. To wit, Google Trends shows that his online popularity last year was almost an order of magnitude less than that of Alan Turing, a contemporary in computing; Erwin Schrödinger, a predecessor in quantum mechanics; and Stephen Wolfram, a successor in the world of automata.

All fame is fleeting, but it’s also mutable, and the Bhattacharya biography should go some distance in pumping up von Neumann’s recognition. But let’s talk interstellar, where his name comes up today because Greg Matloff has just published a new paper dealing with what we now call ‘von Neumann probes.’ By this we simply mean probes that are self-replicating, a notion that originated with von Neumann and has now gone on to wide-ranging study. Throw self-replication and interstellar probes together and you generate various notions about how long it takes to populate the entire galaxy, as found in the work of, for example, Frank Tipler, Michael Hart and others.

Most of those exploring this space have been what Milan Ćirković calls ‘contact pessimists,’ who point out that if von Neumann probes could visit all stars with habitable planets in an entire galaxy, and do this within a small fraction of the galaxy’s age, their existence should be obvious. A more subtle school of thought holds that 1) dispersion need not be uniform and 2) a von Neumann probe may already be in our own Solar System, much less others, for we have only begun to explore deep space.

We can imagine these probes as having the built-in intelligence to make the interstellar crossing, which could be on the order of tens of thousands of years or more given that no biological crews need be involved. Around a target star, such a probe uses local resources – mining a native asteroid system, perhaps – to produce a new probe that, in turn, moves on to the next nearest star, or whatever target it chooses. Robert Freitas has considered self-replication in terms of nanotechnology, in which the size of the probe may be reduced to something as tiny as a needle packed with assemblers.

I come back to the question of biological crews, for without them (or perhaps given probes that carry biological materials that can be activated at destination), the von Neumann probes are free of the massive constraints of species lifespans. Miniaturize a probe to nanotechnological levels and a space-based solar-pumped laser array can push it up to relativistic velocities, possibly using materials like graphene or some kind of future metamaterial at levels of thickness no more than a single atom. But Matloff believes a 20-nm aluminum sail performing an Oberth maneuver (close pass by the Sun followed by a propulsive burn to maximize the gravity slingshot) could reach speeds in the range of 300 kilometers per second. That translates to one light year every 1,000 years.

Either way, we have a method to move human technologies out into the galaxy once our engineering is up to the challenge – the physics behind the project do not preclude this. So let’s imagine that we or some other civilization reach a stage in which we can build von Neumann probes and set them on their journeys. Matloff develops a conservative estimate of the expansion rate of a civilization using such probes.

Because of the vast canvas of time we have to work with given the age of our galaxy, we can afford to be quite conservative in our assumptions. Suppose that to minimize transit times, we say that civilizations doing this send out probes only when another star makes a close approach to the parent probe’s planetary system. Remember, the goal here is the eventual placement of probes galaxy-wide. We give up on all notions of probes reaching destinations within the lifetime of those who build them, even the lifetime of their civilization!

This gets intriguing, based on current data. The second data release of the Gaia space observatory tells us that a star like the Sun will pass within one light year of the Sun every half million years or so. This is, Matloff notes, a pretty conservative figure, for Gaia underestimates the number of low-mass red dwarfs that might also serve. Working the math, we come up with an estimated rate of expansion, granting that some stellar systems will not be suitable. After 500,000 years, we have but two occupied stellar systems. After 18 million years, we have 68.7 billion systems. Says Matloff:

This approach is only an approximation; not all stellar systems will be suitable for occupation by von Neumann probes, and some close stellar encounters will be repeated. But it does indicate that not many long-lived space-faring civilizations that deploy von Neumann probes are required to occupy the galaxy. Even if the slowest interstellar propulsion technique presented above — unpowered giant planet gravity assists — is the one selected by ET, the required galactic occupation time is not substantially increased.

Ah, the joys of exponential growth. I’m reminded of George Gamow’s treatment of such growth in his delightful One Two Three… Infinity, first published in 1947. With probes generating new probes and continuing to push outward, it becomes clear that it would not take a great number of spacefaring civilizations to occupy the entire galaxy even using nothing more than sundiver maneuvers or even gravity assists around gas giant planets to serve as the propulsion technique. Obviously, the process quickens if we reach relativistic speeds with nanotech probes that can exploit the resources they find. The process is fast enough that it’s inevitable to ask where such probes might be located if they are already here.

But first, why would a civilization choose to mount a campaign to spread through the galaxy using such probes? In the next post, we’ll consider a range of possible motivations.

The paper is Matloff, “Von Neumann probes: rationale, propulsion, interstellar transfer timing,” International Journal of Astrobiology, published online by Cambridge University Press 28 February 2022 (abstract).

tzf_img_post
{ 22 comments }

HOEE: The Starshade and the Ground

I always keep an eye on the Phase I and Phase II studies in the pipeline at the NASA Innovative Advanced Concepts (NIAC) program. The goal is to support ideas in their early stages, with the 2022 awards going out to 17 different researchers to the tune of a combined $5.1 million. Of these, 12 are Phase I studies, which deliver $175,000 for a nine-month period, while the five Phase II awards go to $600,000 over two years. We looked at one of the Phase I studies, Jason Benkoski’s solar-thermal engine and shield concept, in the last post. Today we go hunting exoplanets with a starshade.

This particular iteration of the starshade concept is called Hybrid Observatory for Earth-like Exoplanets (HOEE), as proposed by John Mather (NASA GSFC). Here the idea is to leverage the resources of the huge ground-based telescopes that should define the next generation of such instruments – the Giant Magellan Telescope, the Extremely Large Telescope, etc. – by using a starshade to block the glare of the host star, thus uncovering images of exoplanets. Remember that at visible wavelengths, our Sun is 10 billion times brighter than the Earth. The telescope/starshade collaboration would produce what Mather believes will be the most powerful planet finder yet designed.

Image: Three views of a starshade. Credit: NASA / Exoplanet Exploration Program.

Removing the overwhelming light of a star can be done in more than one way, and we’ve seen that an internal coronagraph will be used, for example, with the Nancy Grace Roman Space Telescope. It’s what NASA describes as “a system of masks, prisms, detectors and even self-flexing mirrors” that is being built at the Jet Propulsion Laboratory for the mission.

In conjunction with a space telescope, a starshade operates as a separate spacecraft, a large, flat shade positioned tens of thousands of kilometers away. Starshades have heretofore been studied in this configuration, so the innovation in Mather’s idea is to align the starshade with instruments on the ground. His team believes that we could detect oxygen and water on an Earth-class planet using a 1-hour spectrum out to a distance of 7 parsecs (roughly 23 light years. In an ASTRO2020 white paper, Mather described a system like this using a different orbit for each target star, with the orbit being a highly eccentric ellipse. Thrust is obviously a key component for adjusting the starshade’s position for operations.

From the white paper:

An orbiting starshade would enable ground-based telescopes to observe reflected light from Earth-like exoplanets around sun-like stars. With visible-band adaptive optics, angular resolution of a few milliarcseconds, and collecting areas far larger than anything currently feasible for space telescopes, this combination has the potential to open new areas of exoplanet science. An exo-Earth at 5 pc would be 50 resolution elements away from its star, making detection unambiguous, even in the presence of very bright exo-zodiacal clouds. Earth-like oxygen and water bands near 700 nm could be recognized despite terrestrial interference…

And what a positioning challenge this is in order to maximize angular resolution, sensitivity and contrast, with the starshade matching position and velocity with the telescope from an orbit with apogee greater than ~ 185,000 km, thus casting a shadow of the star, while leaving the light of its planets to reach the instrument below. In addition to the active propulsion to maintain the alignment, the concept relies on adaptive optics that will in any case be used in these ground instruments to cope with atmospheric distortion. Thus low-resolution spectroscopy becomes capable of analyzing light that is actually reflected from Earth-like planets.

Mather’s team wants to cut the 100-meter starshade mass by a factor of 10 to support about 400 kg of thin membranes making up the shade. Thus the concept of an ultra-lightweight design that would be assembled – or perhaps built entirely – in space. It’s worthwhile to remember that the starshade concept in orbit is a new entry in a field that has seen study at NASA GSFC as well as JPL’s Team X, with suitability considered for various missions including HabEx, WFIRST, JWST, New Worlds Explorer, UMBRAS and THEIA. The Mather plan is to create a larger, more maneuverable starshade, as it will indeed have to be to make possible the alignments with ground observatories contemplated in the study.

It’s an exciting prospect, but as Mather’s NIAC synopsis notes, the starshade is not one we could build today. From the synopsis:

The HOEE depends on two major innovations: a ground-space hybrid observatory, and an extremely large telescope on the ground. The tall pole requiring design and demonstration is the mechanical concept of the starshade itself. It must satisfy conflicting requirements for size and mass, shape accuracy and stability, and rigidity during or after thruster firing. Low mass is essential for observing many different target stars. If it can be assembled or constructed after launch, it need not be built to survive launch. We believe all requirements can be met, given sufficient effort. The HOEE is the most powerful exoplanet observatory yet proposed.

Image: Graphic depiction of Hybrid Observatory for Earth-like Exoplanets (HOEE). Credit: John Mather.

Centauri Dreams readers will know that Ashley Baldwin has covered starshade development extensively in these pages. His WFIRST: The Starshade Option is probably the best place to start for those who want to delve further into the matter, although the archives contain further materials. Also see my Progress on Starshade Alignment, Stability.

For more, see Peretz et al., “Exoplanet imaging performance envelopes for starshade-based missions,” Journal of Astronomical Telescopes, Instruments, and Systems 7(2), 021215 (2021). Abstract. And for an overview: Arenberg et al., “Special Section on Starshades: Overview and a Dialogue,” Journal of Astronomical Telescopes, Instruments, and Systems 7(2), 021201 (2021). Abstract.

tzf_img_post

{ 20 comments }

Engineering the Oberth Maneuver

As we saw recently with the analogy of salt grains for stars, the scale of things cosmic stuns the imagination. But we don’t have to go to galactic scale. We can stay much closer to home and achieve the same effect. Because at our current technological levels, getting even as far as the outer planets taxes our capabilities. The least explored types of planet in our Solar System are the dwarf worlds, places like Ceres, Pluto and Charon, not to mention the enigmatic Triton. It takes years to reach them.

Beyond these objects we have a wide range of other dwarfs that merit study, at distances that push us ever farther. In a description of their NIAC Phase I study, just announced as a selection for 2022, Jason Benkoski and colleagues at Johns Hopkins University look into a combination heat shield and solar propulsion system that would perform a close Solar pass and use the Sun’s gravity to slingshot outwards at the highest possible velocity. It’s a maneuver familiar to Centauri Dreams readers, and one recently examined by the Interstellar Probe team at JHU’s Applied Physics Laboratory.

Benkoski is a materials scientist who has been working with the APL team, envisioning a tight solar pass around the Sun followed by the firing of a thruster to enhance the craft’s acceleration. This will require the probe to move within 1.6 million kilometers of the Sun’s surface, actually four times closer than the Parker Solar Probe plans to reach by 2025. In a 2021 article in Johns Hopkins Magazine, Benkoski explained the concept, which will preserve the heat shield by using channels filled with hydrogen gas that are built into the bulk of the shield itself. As the article puts it:

During the probe’s searing slingshot around the sun, the gas would heat up, expand, and course through the channels that all lead to a single exhaust nozzle. “The idea is to absorb all this heat with hydrogen,” Benkoski says, “and shoot it out the back of the probe.” In this way, the cooling setup also opportunistically doubles as an engine, thus supplying the thrust needed to complete the Oberth maneuver in the first place. “It’s like hitting two birds with one stone,” Benkoski says.

Image: Graphic depiction of combined heat shield and solar thermal propulsion system for an Oberth maneuver. Credit: Jason Benkoski.

The team believes that advances in materials science and engineering make their solar thermal engine concept a workable model for development. The 20 x 20 cm prototype they designed and fabricated is at benchtop scale, using liquid helium as coolant and propellant. The new study will extend this work, taking the concept into the realm of realistic materials and propellants. No small challenge, that, given that the contemplated Oberth maneuver would subject the probe to temperatures of 2500 degrees C, enough to melt even the Parker Solar Probe’s heat shield.

Benkoski points out that neither of our Voyagers was designed for observing the interstellar medium through which it now passes, while of course the Pioneers have long since ceased to function. New Horizons remains thankfully robust but will ultimately succumb to dwindling power levels and lose communications with Earth. The numbers are daunting: The Voyagers managed 3.6 AU per year, while even a full-stack SLS (which will never fly this mission) would push a 1 tonne spacecraft only to 8 AU per year.

The latter would require not just a working SLS but a Jupiter gravity assist, limiting the fly-out direction of our probes. Hence the need for a solar Oberth maneuver, in Benkoski’s thinking, which would be capable of surviving temperatures of 2800 K and use propellants now under study to widen the range of potential mission targets:

We…therefore propose a full trade study of alternate propellants in order to determine the maximum escape velocity for a given total system mass, including spacecraft, heat shield, propellant storage, and attitude control system. The main propellants of interest include H2, LiH, Li, CH4, NH3, and H2O. Methods: First we would determine material compatibility for each propellant with respect to its proposed storage system. We then calculate the efficiency (specific impulse) as a function of temperature for each propellant using Chemical Equilibrium Analysis (CEA).

Benkoski intends to discover how the mass and storage volume of the tank scale with the quantity of propellant to produce a series of realistic tank designs, devising an equation for the heat shield area and maximum propellant fraction that can be achieved given the limitations of existing heavy boosters. We’ll see how this study fares in producing a full-scale heat shield/heat exchanger design with robust long-term cryogenic storage. A tight Oberth maneuver is not going to be easy. See Assessing the Oberth Maneuver for Interstellar Probe for some of the myriad reasons why.

tzf_img_post

{ 23 comments }

Lowering the Laser Barrier

The continuing release of papers related to or referring to the Breakthrough Starshot sail concept is good news for the entire field. Interstellar studies as an academic discipline has never had this long or sustained a period of activity, and the growing number of speakers at space-related conferences attests to the current vitality of starflight among professionals and the general public alike.

Not all interstellar propulsion concepts involve laser-beaming, of course, and we’ll soon look at what some would consider an ever more exotic concept. But today I’m focusing on a paper from Ho-Ting Tung and Artur Davoyan, both in the Mechanical and Aerospace Engineering Department at UCLA. You could say that these two researchers are filling in some much needed space between the full-bore interstellar effort of Breakthrough Starshot, the Solar System-oriented laser work of Andrew Higgins’ team at McGill, and much smaller, near-term experiments we could run not so far from now.

Of the many potential show-stoppers faced by a mission to another star at our stage of development is the need to develop the colossal laser array envisioned by Starshot. The Higgins array is at a smaller scale, as befits a concept with nearby targets like Mars. What Tung and Davoyan envision are tiny payloads (here they parallel Breakthrough), some no more than a gram in mass, but the authors push the sail with a 100 kW array about a meter in size. Compare this with Breakthrough’s need for a gigantic square-kilometer array of 10 kW lasers with a combined output of up to 100 GW.

Image: In this illustration, a low-power laser (red cone) on Earth could be used to shift the orbit (red lines) of a small probe (grey circle), or propel it at rapid speeds to Neptune and beyond. Credit: Ho-Ting Tung et al.

The UCLA work takes us to a consideration of operations with spacecraft in Earth orbit as well as payloads sent on interplanetary trajectories. Thus we are in the realm of the kind of missions that today would demand chemical or electric propulsion, and we are looking at a system that might be used, for example, for orbital adjustment of Earth satellites after launch, or in the case of chip-class payloads, interplanetary missions with surprising velocities, up to 5 times that of New Horizons. As noted, the needed laser aperture is, by the standards of the missions we’ve discussed earlier, small:

…a sail with w = 1 m would require a laser with an aperture D ≃ 26 m (compare with the 30 m diameter primary mirror of the Thirty Meter Telescope under construction). However, we stress that most practical scenarios are limited to low and medium Earth orbits that require a much shorter operation range (z ≤ 1000km), and therefore a significantly smaller laser array.

Indeed, an array a meter in size could be efficient, maneuvering small satellites in Earth orbit, or being used to bring small chip-craft up to Solar System escape velocity. Thus we have the potential to create laser propulsion experiments and missions with array powers of ≥ 100 kW and array sizes that do not require kilometers of desert for their construction. Payloads can range from 1 to 100 grams depending on the mission, though the focus here is wafer-scale, on the order of 10 centimeters.

As to sail materials, the authors calculate that for maximum reflectivity coupled with rapid cooling, silicon nitride and boron nitride are the materials of choice:

Broadband spectral emissivity of silicon nitride…results in a better heat rejection (i.e., lower temperature) as compared [to] narrow band BN thermal emitters. However, boron nitride being lighter than silicon nitride allows design of very light-weight light-sails, which eventually translates onto higher velocity gain, Δv.

The paper offers possible ways to create these structures, including using metamaterials formed into nanostructured architectures with nanometer-scale ‘sandwich’ panels between material layers, or using ‘micro pillars’ within the photonic structure.

The broader picture is that we’re mapping out how to experiment with lasers and materials that may begin moving up the ladder of mission complexity. There are innumerable issues to be overcome, but the early theoretical work is crucial to making what may become an interplanetary infrastructure a reality. These examinations should also feed into the ambitious work on projects that aim at interstellar missions.

The paper is Ho-Ting Tung et al, Low-Power Laser Sailing for Fast-Transit Space Flight, Nano Letter,” Nano Letters 22, 3 (31 January 2022), 1108–1114 (abstract).

tzf_img_post

{ 5 comments }