Centauri Dreams
Imagining and Planning Interstellar Exploration
Data Return from Proxima Centauri b
The challenges involved in sending gram-class probes to Proxima Centauri could not be more stark. They’re implicit in Kevin Parkin’s analysis of the Breakthrough Starshot system model, which ran in Acta Astronautica in 2018 (citation below). The project settled on twenty percent of the speed of light as a goal, one that would reach Proxima Centauri b well within the lifetime of researchers working on the project. The probe mass is 3.6 grams, with a 200 nanometer-thick sail some 4.1 meters in diameter.
The paper we’ve been looking at from Marshall Eubanks (along with a number of familiar names from the Initiative for Interstellar Studies including Andreas Hein, his colleague Adam Hibberd, and Robert Kennedy) accepts the notion that these probes should be sent in great numbers, and not only to exploit the benefits of redundancy to manage losses along the way. A “swarm” approach in this case means a string of probes launched one after the other, using the proposed laser array in the Atacama desert. The exciting concept here is that these probes can reform themselves from a string into a flat, lens-shaped mesh network some 100,000 kilometers across.
Image: Figure 16 from the paper. Caption: Geometry of swarm’s encounter with Proxima b. The Beta-plane is the plane orthogonal to the velocity vector of the probe ”at infinity” as it approaches the planet; in this example the star is above (before) the Beta-plane. To ensure that the elements of the swarm pass near the target, the probe-swarm is a disk oriented perpendicular to the velocity vector and extended enough to cover the expected transverse uncertainty in the probe-Proxima b ephemeris. Credit: Eubanks et al.
The Proxima swarm presents one challenge I hadn’t thought of. We have to be able to predict the position of Proxima b to within 10,000 kilometers at least 8.6 years before flyby – this is the time for complete information cycle between Earth, Proxima and back to Earth. Effectively, we need to figure out the planet’s velocity to a value of 1 meter per second, with a correspondingly tight angular position (0.1 microradians).
Although we already have Proxima b’s period (11.68 days), we need to determine its line of nodes, eccentricity, inclination and epoch, and also its perturbations by the other planets in the system. At the time of flyby, the most recent Earth update will be at least 8.5 years old. The Proxima b orbit state will need to be propagated over at least that interval to predict its position, and that prediction needs to be accuracy to the order of the swarm diameter.
The authors suggest that a small spacecraft in Earth orbit can refine Proxima b’s position and the star’s ephemeris, but note that a later paper will dig into this further.
In the previous post I looked at the “Time on Target” and “Velocity on Target” techniques that would make swarm coherence possible, with variations in acceleration and velocity allowing later-launched probes to reach higher speeds, but with higher drag so that as they reach the craft sent before them, they slow to match their speed. From the paper again:
A string of probes relying on the ToT technique only could indeed form a swarm coincident with the Proxima Centauri system, or any other arbitrary point, albeit briefly. But then absent any other forces it would quickly disperse afterwards. Post-encounter dispersion of the swarm is highly undesirable, but can be eliminated with the VoT technique by changing the attitude of the spacecraft such that the leading edge points at an angle to the flight direction, increasing the drag induced by the ISM, and slowing the faster swarm members as they approach the slower ones. Furthermore, this approach does not require substantial additional changes to the baseline BTS [Breakthrough Starshot] architecture.
In other words, probes launched at different times with a difference in velocity target a point on their trajectory where the swarm can cohere, as the paper puts it. The resulting formation is then retained for the rest of the mission. The plan is to adjust the attitude of the leading probes continually as they move through the interstellar medium, which means variations in their aspect ratio and sectional density. A probe can move edge-on, for instance, or fully face-on, with variations in between. The goal is that the probes lost later in the process catch up with but do not move past the early probes.
All this is going to take a lot of ‘smarts’ on the part of the individual probes, meaning we have to have ways for them to communicate not just with Earth but with each other. The structure of the probes discussed here is an innovation. The authors propose that key components like laser communications and computation should be concentrated, so that whereas the central disk is flat, the ‘heart of the device,’ as they put it, is concentrated in a 2-cm thickened rim around the outside of the sail disk.
The center of the disk is optical, or as the paper puts it, ‘a thin but large-aperture phase-coherent meta-material disk of flat optics similar to a fresnel lens…’ which will be used for imaging as well as communications. Have a look at the concept:
Image: This is Figure 3a from the paper. Caption: Oblique view of the top/forward of a probe (side facing away from the launch laser) depicting array of phase-coherent apertures for sending data back to Earth, and optical transceivers in the rim for communication with each other. Credit: Eubanks et al.
So we have a sail moving at twenty percent of lightspeed through an incoming hydrogen flux, an interesting challenge for materials science. The authors consider both aerographene and aerographite. I had assumed these were the same material, but digging into the matter reveals that aerographene consists of a three-dimensional network of graphene sheets mixed with porous aerogel, while aerographite is a sponge-like formation of interconnected carbon nanotubes. Both offer extremely low density, so much so that the paper notes the performance of aerographene for deceleration is 104 times better than conventional mylar. Usefully, both of these materials have been synthesized in the laboratory and mass production seems feasible.
Back to the probe’s shape, which is dictated by the needs not only of acceleration but survival of its electronics – remember that these craft must endure a laser launch that will involve at least 10,000 g’s. The raised rim layout reminds the authors of a red corpuscle as opposed to what has been envisioned up to now as a simple flat disk. The four-meter central disk contains 247 25-cm structures arranged, as the illustration shows, like a honeycomb. We’ll use this optical array for both imaging Proxima b but also returning data to Earth, and each of the arrays offers redundancy given that impacts with interstellar hydrogen will invariably create damage to some elements.
Remember that the plan is to build an intelligent swarm, which demands laser links between the probes themselves. Making sure each probe is aware of its neighbors is crucial here, for which purpose it will use the optical transceivers around its rim. The paper calculates that this would make each probe detectable by its closest neighbor out to something close to 6,000 kilometers. The probes transmit a pulsed beacon as they scan for neighboring probes, and align to create the needed mesh network. The alignment phase is under study and will presumably factor into the NIAC work.
The paper backs out to explain the overall strategy:
…our innovation is to use advances in optical clocks, mode-locked optical lasers, and network protocols to enable a swarm of widely separated small spacecraft or small flotillas of such to behave as a single distributed entity. Optical frequency and reliable picosecond timing, synchronized between Earth and Proxima b, is what underpins the capability for useful data return despite the seemingly low source power, very large space loss and low signal-to-noise ratio.
For what is going to happen is that the optical pulses between the probes will be synchronized, meaning that despite the sharp constraints on available energy, the same signal photons are ‘squeezed’ into a smaller transmission slot, which increases the brightness of the signal. We get data rates through this brightening that could not otherwise be achieved, and we also get data from various angles and distances. On Earth, a square kilometer array of 796 ‘light buckets’ can receive the pulses.
Image: This is Figure 13 from the paper. Caption: Figure 13: A conceptual receiver implemented as a large inflatable sphere, similar to widely used inflatable antenna domes; the upper half is transparent, the lower half is silvered to form a half-sphere mirror. At the top is a secondary mirror which sends the light down into a cone-shaped accumulator which gathers it into the receiver in the base. The optical signals would be received and converted to electrical signals – most probably with APDs [avalanche photo diodes] at each station and combined electrically at a central processing facility. Each bucket has a 10-nm wide band-pass filter, centered on the Doppler-shifted received laser frequency. This could be made narrower, but since the probes will be maneuvering and slowing in order to meet up and form the swarm, and there will be some deceleration on the whole swarm due to drag induced by the ISM, there will be some uncertainty in the exact wavelength of the received signal. Credit: Eubanks et al.
If we can achieve a swarm that is in communication with its members using micro-miniaturized clocks to keep operations synchronous, we can thus use all of the probes to build up a single detectable laser pulse bright enough to overcome the background light of Proxima Centauri and reach the array on Earth. The concept is ingenious and the paper so rich in analysis and conjecture that I keep going back to it, but don’t have time today to do more than cover these highlights. The analysis of enroute and approach science goals and methods alone would make for another article. But it’s probably best that I simply send you to the paper itself, one which anyone interested in interstellar mission design should download and study.
The paper is Eubanks et al., “Swarming Proxima Centauri: Optical Communication Over Interstellar Distances,” submitted to the Breakthrough Starshot Challenge Communications Group Final Report and available online. Kevin Parkin’s invaluable analysis of Starshot is Parkin, K.L.G., “The Breakthrough Starshot system model,” Acta Astronautica 152 (2018), 370–384 (abstract / preprint).
Reaching Proxima b: The Beauty of the Swarm
NIAC’s award of a Phase I grant to study a ‘swarm’ mission to Proxima Centauri naturally ties to Breakthrough Starshot, which continues its interstellar labors, though largely out of the public eye. The award adds a further research channel for Breakthrough’s ideas, and a helpful one at that, for the NASA Innovative Advanced Concepts program supports early stage technologies through three levels of funding, so there is a path for taking these swarm ideas further. An initial paper on swarm strategies was indeed funded by Breakthrough and developed through Space Initiatives and the UK-based Initiative for Interstellar Studies.
Centauri Dreams readers are by now familiar with my enthusiasm for swarm concepts, and not just for interstellar purposes. Indeed, as we develop the technologies to send tiny spacecraft in their thousands to remote targets, we’ll be testing the idea out first through computer simulation but then through missions within our own Solar System. Marshall Eubanks, the chief scientist for Space Initiatives, a Florida-based startup focused on 50-gram femtosatellites and their uses near Earth, talks about swarm spacecraft covering cislunar space or analyzing a planetary magnetosphere. Eubanks is lead author of the aforementioned paper.
But the go-for-broke target is another star, and that star is naturally Proxima Centauri, given Breakthrough’s clear interest in the habitable zone planet orbiting there. The NIAC announcement sums up the effort, but I turn to the paper for discussion of communications with such swarm spacecraft. As Starshot has continued to analyze missions at this scale, it explores probes with launch mass on the scale of grams and onboard power restricted to milliwatts. The communications challenge is daunting indeed given the distances and power available.
If we want to reach a nearby star in this century, so the thinking goes, we should build the kind of powerful laser beamer (on the order of 100 GW) that can push our lightsails and their tiny payloads to speeds that are an appreciable fraction of the speed of light. Moving at 20 percent of c, we reach Proxima space within 20 years, to begin the long process of returning data acquired from the flybys of our probes. Eubanks and colleagues estimate we’ll need thousands of these, because we need to create an optical signal strong enough to reach Earth, one coordinated through a network that is functionally autonomous. We’re way too far from home to control it from Earth.
Image: Artist’s impression of swarm passing by Proxima Centauri and Proxima b. The swarm’s extent is ∼10 larger than the planet’s, yet the ∼5000-km spacing is such that one or more probes will come close to or even impact the planet (flare on limb). It should be possible to do transmission spectroscopy with such swarms. Green 432/539-nm beams are coms to Earth; red 12,000-nm laser beacons are for intra-swarm probe-to-probe coms. Conceptual artwork courtesy of Michel Lamontagne.
The engineering study that has grown out of this vision describes the spacecraft as being ‘operationally coherent,’ meaning they will be synchronized in ways that allow data return. The techniques here are fascinating. Adjusting the initial velocity of each probe (this would be done through the launch laser itself) allows the string of probes to cohere. The laser also allows clock synchronization, so that we wind up with what had been a string of probes traveling together through the twenty year journey. In effect, the tail of the string catches up with the head. What emerges is a network.
As the NIAC announcement puts it:
Exploiting drag imparted by the interstellar medium (“velocity on target”) over the 20-year cruise keeps the group together once assembled. An initial string 100s to 1000s of AU long dynamically coalesces itself over time into a lens-shaped mesh network 100,000 km across, sufficient to account for ephemeris errors at Proxima, ensuring at least some probes pass close to the target.
The ingenuity of the communications method emerges from the capability of tiny spacecraft to travel with their clocks in synchrony, with the ability to map the spatial positions of each member of the swarm. This is ‘operational coherence,’ which means that while each probe returns the same data, the transmission time is related to its position within the swarm. The result; The data pulses arrive at the same time on Earth, so that while the signal from any one probe would be undetectable, the combined laser pulse from all of them can become bright enough to detect over 4.2 light years.
The paper cites a ‘time-on-target’ technique to allow the formation of effective swarm topologies, while a finer-grained ‘velocity-on-target’ method is what copes with the drag imparted by the interstellar medium. This one stopped me short, but digging into it I learned that the authors talk about adjusting the attitude of individual probes as needed to keep the swarm in coherent formation. The question of spacecraft attitude also applies to the radiation and erosion concerns of traveling at these speeds, and I think I’m right in remembering that Breakthrough Starshot has always contemplated the individual probes traveling edge-on during cruise with no roll axis rotation.
Image; This is Figure 2a from the paper. Caption: A flotilla (sub-fleet) of probes (far left), individually fired at the maximum tempo of once per 9 minutes, departs Earth (blue) daily. The planets pass in rapid succession. Launched with the primary ToT technique, the individual probes draw closer to one another inside the flotilla, while the flotilla itself catches up with previously-launched flotillas exiting the outer Solar system (middle) ∼100 AU. For the animation go to https://www.youtube.com/watch?v=jMgfVMNxNQs (Hibberd 2022).
Figure 2b takes the probe ensemble into the Oort Cloud.
Image: Figure 2b caption: Time sped up by a scale factor of 30. The last flotilla launched draws closer to the earlier flotillas; the full fleet begins to coalesce (middle), now under both the primary ToT and secondary VoT techniques, beyond the Kuiper-Edgeworth Belt and entry into the Oort Cloud ∼1000–10,000 AU.
When we talk about using collisions with the interstellar medium to create velocities transverse to the direction of travel, we’re describing a method that again demands autonomy, or what the paper describes as a ‘hive mind,’ a familiar science fiction trope. The hive mind will be busy indeed, for its operations must include not just cruise control over the swarm’s shape but interactions during the data return phase. From the paper;
With virtually no mass allowance for shielding, attitude adjustment is the only practical means to minimize the extreme radiation damage induced by traveling through the ISM at 0.2c. Moreover, lacking the mass budget for mechanical gimbals or other means to point instruments, then controlling attitude and rate changes of the entire craft in pitch, yaw, roll, is the only practical way [to] aim onboard sensors for intra-swarm communications, interstellar comms with Earth and imagery acquisition / distributed processing at encounter.
I gather that other techniques for interacting with the interstellar medium will come into play in the NIAC work, for the paper speaks of using onboard ‘magnetorquers,’ an attitude adjustment mechanism currently in use in low-mass Cubesats in low Earth orbit. It’s an awkward coinage, but a magnetorquer refers to magnetic torquers or torque rods that have been developed for attitude control in a given inertial frame. The method works through interaction between a magnetic field and the ambient magnetic field (in current cases, of the Earth). Are magnetic fields in the interstellar medium sufficient to support this method? The paper explores the need for assessment.
A solid state probe has no moving parts, but it’s also clear that further simulations will explore the use of what the paper calls MEMS (micro-electromechanical systems) trim tabs that could be spaced symmetrically to provide dynamic control by producing an asymmetric torque. This sounds like a kludge, though one that needs exploring given the complexities of adjusting attitudes throughout a swarm. We’ll see where the idea goes as it matures in the NIAC phase. All this will be critical if we are to connect interswarm to create the signaling array that will bring the Proxima data home.
Interestingly, the kind of probes the paper describes may vary in some features:
We note for the record that although all probes are assumed to be identical, implicitly in the community and explicitly in the baseline study, there is in fact no necessity for them to be “cookie cutter” copies, since the launch laser must be exquisitely tunable in the first place, capable of providing a boost tailored to every individual probe. At minimum, probes can be configured and assigned for different operations while remaining dynamically identical, or they can be made truly heterogeneous wherein each probe could be rather different in form and function, if not overall mass and size.
There is so much going on in this paper, particularly the issue of the orbital position of Proxima b, which you would think would be known well enough by now (but guess again). The question of carrying enough stored energy for the two decade mission is a telling one. But the overwhelming need is to get information back to Earth. How data would be received from these distances has always bedeviled the Starshot idea, and having followed the conversation on this for some time now, I find the methods proposed here seriously intriguing. We’ll dig into these issues in the next post.
The paper is Eubanks et al., “Swarming Proxima Centauri: Optical Communication Over Interstellar Distances,” submitted to the Breakthrough Starshot Challenge Communications Group Final Report and available online.
Galactic ‘Nature Preserves’ over Deep Time
Speculating about the diffusion of intelligent species through the galaxy, as we’ve been doing these past few posts, is always jarring. I go back to the concept of ‘deep time,’ which is forced on us when we confront years in their billions. I can’t speak for anyone else, but for me thinking on this level is closer to mathematics than philosophy. I can accept a number like 13.4 × 10⁹ years (the estimate for the age of globular cluster NGC 6397 and a pointer to the Milky Way’s age) without truly comprehending how vast it is. As biological beings, a century pushes us to the limit. What exactly is an aeon?
NGC 6397 and other globular clusters are relevant because these ancient stellar metropolises are the oldest large-scale populations in the Milky Way. But I’m reminded that even talking about the Milky Way can peg me as insufferably parochial. David Kipping takes me entirely out of this comparatively ‘short-term’ mindset by pushing the limits of chronological speculation into a future so remote that elementary particles themselves have begun to break down. Not only that – the Columbia University astrophysicist finds a way for human intelligence to witness this.
You absolutely have to see how he does this in Outlasting the Universe, a presentation on his Cool Worlds YouTube channel. Now Cool Worlds is a regular stop here because Kipping is a natural at rendering high-level science into thoughtful explanations that even the mathematically challenged like me can understand. Outlasting the Universe begins with Kipping the narrator saying “We are in what you would call the future…the deep future” and takes human evolution through the end of its biological era and into a computer-borne existence in which a consciousness can long outlive a galaxy.
Image: Astrophysicist, author and indeed philosopher David Kipping. Credit: Columbia University.
Along the way we remember (and visit in simulation) Freeman Dyson, who once speculated that to become (almost) immortal, a culture could slow down the perceived rate of time. “Like Zeno’s arrow,” says Kipping, “we keep dialing down the speed.” The visuals here are cannily chosen, the script crisp and elegant, imbued with the ‘sense of wonder’ that brought so many of us to science fiction. Outlasting the Universe is indeed science fiction of the ‘hard SF’ variety as Kipping draws out the consequences of deep time and human consciousness in ways that make raw physics ravishing. I envy this man’s students.
With scenarios like this to play with, where do we stand with the ‘zoo hypothesis?’ It must, after all, reckon with years by the billions and the spread of intelligence. Science fiction writer James Cambias responded to my Life Elsewhere? Relaxing the Copernican Principle post with a tight analysis of the notion that we may be under observation from a civilization whose principles forbid contact with species they study. This is of course Star Trek’s Prime Directive exemplified (although the lineage of the hypothesis dates back decades), and it brings up Jim’s work because he has been so persistent a critic of the idea of shielding a population from ETI contact.
Jim’s doubts about the zoo hypothesis go back to his first novel. A Darkling Sea posits an Europa-like exoplanet being studied by a star-faring species called the Sholen, who are employing a hands-off policy toward local intelligence even as they demand that human scientists on the world’s sea bottom do the same. Not long after publication of the novel (Tor, 2014), he told John Scalzi that he saw Prime Directives and such as “ …a mix of outrageous arrogance and equally overblown self-loathing, a toxic brew masked by pure and noble rhetoric.” The arrogance comes from ignoring the desires of the species under study and denying them a choice in the matter.
In a current blog post called The Zoo Hypothesis: Objections, Jim lays this out in rousing fashion:
…we deduce that you can’t hide a star system which contains a civilization capable of large-scale interstellar operations, which the Zookeepers are by definition. They’re going to be emitting heat, EM radiation, laser light, all the spoor of a Kardashev Type I or higher civilization. And the farther away they are, the more they’re going to be emitting because they need to be bigger and more energy-rich in order to have greater reach.
This gives us one important lesson: if the goal of a Zoo is to keep the civilizations inside from even knowing of the existence of other civilizations, the whole thing is impossible. You can’t have a Zoo without Zookeepers, and the inhabitants of the Zoo will detect them.
Jim’s points are well-taken, and he extends the visibility issue by noting that we need to address time, which must be deep indeed. For a civilization maintaining all the apparatus of a protected area around a given star has to do so on time frames that are practically geological in length. Here we can argue a bit, for a ‘zoo’ set up for reasons we don’t understand in the first place might well come into existence only when the species being studied has reached the capability of detecting its observers.
I referenced Amri Wandel (Hebrew University of Jerusalem) on this the other day. Wandel argues that our own industrial lifespan is currently on the order of a few centuries, and who knows what level of technological sophistication a ‘zoo-keeping’ observer culture might want us to reach before it decides it can initiate contact? That would drop the geological timeframe down to a more manageable span, although the detectability problem still remains. So does the issue of interaction with other star-faring species who might conceivably need to be warned off entering the zoo. Cambias again:
If Captain Kirk or whoever shows up on your planet and says “I’m from another planet. Let’s talk and maybe exchange genetic material — or not, if you want me to leave just say so,” that’s an infinitely more reasonable and moral act than for Captain Kirk to sneak around watching you without revealing his own existence. The first is an interaction between equals, the second is the attitude of a scientist watching bacteria. Is that really a moral thing to do? Why does having cooler toys than someone else give you the right to treat them like bacteria?
This is lively stuff, and speculation of this order is why many people begin reading and writing science fiction in the first place. A hard SF writer, a ‘world builder,’ will make sure that he or she has thought through implications for every action he attributes not only to his characters but the non-human intelligences they may interact with. One thing that had never occurred to me was the issue of visibility when translated to the broader galaxy. Because a zoo needs to be clearly marked. Here’s Jim’s view:
If you’re going to exclude other civilizations from a particular region of the Galaxy, you have to let them know. Shooting relativistic projectiles or giant laser beams at incoming starships is a very ham-fisted way of communicating “keep out!” — and it runs the risk of convincing the grabby civilization that you’re shooting at to start shooting back. And if they’re grabby and control a lot of star systems, that’s going to be a lot of shooting.
Jim’s points are telling, and the comments on my recent Centauri Dreams posts also reflect readers’ issues with the zoo hypothesis. My partiality to it takes these issues into account. If the zoo hypothesis is the best of the solutions to the Fermi question, then the likelihood that other intelligent species are in our neighborhood is vanishingly small. Which lets me circle back to the paper by Ian Crawford and Dirk Schulze-Makuch that set off this entire discussion. It asked, you’ll recall, whether the zoo hypothesis wasn’t the last standing alternative to the idea that technological civilizations are, at the least, rare. It’s not a good alternative, but there it is.
In other words, I’d like the zoo hypothesis to have some traction, because it’s the only way I can find to imagine a galaxy in which intelligent civilizations are common.
Consider the thinking of Crawford and Schulze-Makuch on other hypotheses. Interstellar flight might be impossible for reasons of distance and energy, but this seems a non-starter given that we know of ways within known physics to send a payload to another star even in this century. A slow exploration front moving at Voyager speeds could do the trick in a fraction of the time available given the age of the Milky Way. The lack of SETI detections likewise points to technologies that are physically feasible (various kinds of technosignatures) but are not yet observed.
Is the answer that civilizations don’t live very long, and the chances of any two existing at the same brief time in the galaxy are remote? The nagging issue here is that we would have to assume that all civilizations are temporally limited. It takes only one to find a way through whatever ‘great filter’ is out there and survive into a star-faring maturity to get the galaxy effectively visited and perhaps colonized by now. Crawford and Schulze-Makuch reject models that result in volumes of the galactic disk being unvisited during the four billion years of Earth’s existence, considering them valid mathematically but implausible as solutions to the larger Fermi puzzle.
Many of the hypotheses to explain the Great Silence go even further into the unknowable. What, for example, do we make of attempts to parse out an alien psychology, which inevitably is seen, wittingly or not, as reflecting our own human instincts and passions? Monkish cultures that choose not to expand for philosophical reasons will remain unknowable to us, for example, as will societies that self-destruct before they achieve interstellar flight. We can still draw a few conclusions, though, as Crawford and Schulze-Makuch do, all pointing at least to intelligence being rare.
Although we know nothing of alien sociology, it seems inevitable that the propensity for self-destruction, interstellar colonization and so on must be governed by probability distributions of some kind. The greater the number of ETIs that have existed over the history of the Galaxy, the more populated will be the non-self-destructed and/or pro-colonization wings of these distributions, and it is these ETIs that we do not observe. On the other hand, if the numbers of ETIs have always been small, these distributions will have been sparsely populated and the non-observation of ETIs in their expansionist wings follows naturally.
Image: Are ancient ruins the only thing we may expect to find if we reach other star systems? Are civilizations always going to destroy themselves? The imposing remains of Angkor Wat. Credit: @viajerosaladeriva.
Likewise, we still face the problem that, as Stapledon long ago noted, different cultures will choose different priorities. Why assume that in a galaxy perhaps stuffed with aliens adopting Trappist-like vows of silence there will not be a few societies that do want to broadcast to the universe, a METI-prone minority perhaps, but observable in theory. We have no paradox in the Fermi question if we assume that aliens are rare, but if they are as common as early science fiction implied, the paradox is only reinforced.
So Crawford and Schulze-Makuch have boiled this down to the zoo hypothesis or nothing, with the strong implication that technological life must indeed be rare. I rather like my “one to ten” answer to the question of how many technological species are in the galaxy, because I think it squares with their conclusions. And while we can currently only speculate on reasons for this, it’s clear that we’re on a path to draw conclusions about the prevalence of abiogenesis probably in this century. How often technologies emerge after unicellular life covers a planet is a question that may have to wait for the detection of a technosignature. And as is all too clear, it’s possible this will never come.
The paper is Crawford & Schulze-Makuch, “Is the apparent absence of extraterrestrial technological civilizations down to the zoo hypothesis or nothing?” Published online in Nature Astronomy 28 December 2023 (abstract). James Cambias’ fine A Darkling Sea (Tor, 2014) is only the first of his novels, the most recent of which is The Scarab Mission (Baen, 2023), part of his ‘billion worlds’ series. Modesty almost, but not quite, forbids me from mentioning my essay “Ancient Ruins” which ran in Aeon a few years back.
Can the ‘Zoo Hypothesis’ Be Saved?
If we were to find life other than Earth’s somewhere else in the Solar System, the aftershock would be substantial. After all, a so-called ‘second genesis’ would confirm the common assumption that life forms often, and in environments that range widely. The implications for exoplanets are obvious, as would be the conclusion that the Milky Way contains billions of living worlds. The caveat, of course, is that we would have to be able to rule out the transfer of life between planets, which could make Mars, say, controversial. But find living organisms on Titan and the case is definitively made.
Ian Crawford and Dirk Schulze-Makuch point out in their new paper on the Fermi question and the ‘zoo hypothesis’ that this issue of abiogenesis could be settled relatively soon as our planetary probes gain in sophistication. We could settle it within decades if we found definitive biosignatures in an exoplanet atmosphere, but here my skepticism kicks in. My guess is that once we have something like the Habitable Worlds Observatory in place (and a note from Dominic Benford informs me that NASA has just put together teams to guide the development of HWO, the flagship mission after the Nancy Grace Roman Space Telescope), the results will be immediately controversial.
In fact, I can see a veritable firestorm of debate on the question of whether a given biosignature can be considered definitive. Whole journals a few decades from now will be filled with essays pushing abiotic ways to produce any signature we can think of, and early reports that support abiogenesis around other stars will be countered with long and not always collegial analysis. This is just science at work (and human nature), and we can recall how quickly Viking results on Mars became questioned.
So I think in the near term we’re more likely to gain insights on abiogenesis through probing our own planetary system. Life on an ice giant moon may turn up, or around a gas giant like Saturn in an obviously interesting moon like Enceladus, and we can strengthen our hunch that abiogenesis is common. In which case, where do we stand on the development of intelligence or, indeed, consciousness? What kind of constraints can we put on how often technology is likely to be the result of highly evolved life? Absent a game-changing SETI detection, we’re still left with the Fermi question. We have billions of years of cosmic history to play with and a galaxy that over time could be colonized.
Image: JWST’s spectacular image of M51 (NGC 5194), some 27 million light-years away in the constellation Canes Venatici. Taken with the telescope’s Near-InfraRed Camera (NIRCam), the image is so lovely that I’ve been looking for an excuse to run it. This seems a good place, for we’re asking whether a universe that can produce so many potential homes for life actually gives rise to intelligence and technologies on a galaxy-wide scale. Here the dark red features trace warm dust, while colors of red, orange, and yellow flag ionized gas. How long would it take for life to emerge in such an environment, and would it ever become space-faring? Credit: ESA/Webb, NASA & CSA, A. Adamo (Stockholm University) and the FEAST JWST team.
Crawford and Schulze-Makuch ask a blunt question in the title of their paper in Nature Astronomy: ”Is the apparent absence of extraterrestrial technological civilizations down to the zoo hypothesis or nothing?” The zoo hypothesis posits that we are being studied by beings that for reasons of their own avoid contact. David Brin referred in his classic 1983 paper “The Great Silence” (citation below) to this as one variation of a quarantine, with the Solar System something like a nature preserve whose inhabitants have no idea that they are under observation.
Quarantines can come in different flavors, of course. Brin notes the possibility that observers might wait for our species to reach a level of maturity sufficient to join what could be a galactic ‘club’ or network. Or perhaps the notion is simply to let planets early in their intellectual development lie fallow as their species mature. Wilder notions include the idea that we could be quarantined because we represent a danger to the existing order, though it’s hard to imagine a scenario in which this occurs.
But the Crawford / Schulze-Makuch paper is not exactly a defense of the zoo hypothesis. Rather, it asks whether it is the only remaining alternative to the idea that the galaxy is free of other civilizations. The paper quickly notes the glaring issue with the hypothesis, and it’s one anticipated by Olaf Stapledon in Star Maker. While any species with the ability to cross interstellar distances might remain temporarily hidden, wouldn’t there be larger trends that mitigate the effectiveness of their strategy? Can you hide one or more civilizations that have expanded over millions of years to essentially fill the galaxy? At issue is the so-called ‘monocultural fallacy’:
…to explain the Fermi paradox in a Galaxy where ETIs are common, all these different, independently evolved civilizations would need to agree on the same rules for the zoo. Moreover, to account for the apparent non-interference with Earth’s biosphere over its history, these rules may have had to remain in place, and to have been adhered to, ever since the first appearance of colonizing ETI in the Galaxy, which might be billions of years if ETIs are common. Indeed, Stapledon (ref. 29, p.168) anticipated this problem when he noted, from the point of view of a future fictional observer, that “different kinds of races were apt to have different policies for the galaxy”.
I always return to Stapledon with pleasure. I dug out my copy of Star Maker to cite more from the book. Here the narrator surveys the growth and philosophies of civilizations in their multitudes during his strange astral journey:
Though war was by now unthinkable, the sort of strife which we know between individuals or associations within the same state was common. There was, for instance, a constant struggle between the planetary systems that were chiefly interested in the building of Utopia, those that were most concerned to make contact with other galaxies, and those whose main preoccupation was spiritual. Besides these great parties, there were groups of planetary systems which were prone to put the well-being of individual world-systems above the advancement of galactic enterprise. They cared more for the drama of personal intercourse and the fulfillment of the personal capacity of worlds and systems than for organization or exploration of spiritual purification. Though their presence was often exasperating to the enthusiasts, it was salutary, for it was a guarantee against extravagance and against tyranny.
That’s a benign kind of strife, but it has an impact. The matter becomes acute when we consider interacting civilizations in light of the differential galactic rotation of stars, as Brin pointed out decades ago. The closest species to us at any given time would vary as different stars come into proximity. That seems to imply a level of cultural uniformity that is all but galaxy-wide if the zoo hypothesis is to work. But Crawford and Schulze-Makuch are on this particular case, noting that a single early civilization (in galactic history) might be considered a ‘pre-emptive civilization’ (this is Ronald Bracewell’s original idea), thus enforcing the rules of the road to subsequent ETIs. In such a way we might still have a galaxy filled with technological societies.
An interesting digression here involves the age of likely civilizations. We know that the galaxy dates back to the earliest era of the universe. European Southern Observatory work on the beryllium content of two stars in the globular cluster NGC 6397 pegs their age at 13,400 ± 800 million years. Extraterrestrial civilizations have had time to arise in their multitudes, exacerbating the ‘monocultural’ issue raised above. But the authors point out that despite its age, the galaxy’s habitability would have been influenced by such issues as “a possibly active galactic nucleus, supernovae and close stellar encounters.” Conceivably, the galaxy at large evolved in habitability so that it is only within the last few billion years that galaxy-spanning civilizations could become possible.
Does that help explain the Great Silence? Not really. Several billion years allows ample time for civilizations to develop and spread. As the paper notes, we have only the example of our Earth, in which it took something like two billion years to develop an atmosphere rich in the oxygen that allowed the development of complex creatures. You don’t have to juggle the numbers much to realize that different stellar systems and their exoplanets are going to evolve at their own pace, depending on the growth of their unique biology and physical factors like plate tectonics. There is plenty of room even in a galaxy where life only emerged within the last billion years for civilizations to appear that are millions of years ahead of us technologically.
Image: The globular cluster NGC 6397. A glorious sight that reminds us of the immensity in both space and time that our own galaxy comprehends. Credit: ESO.
Back to the zoo hypothesis. Here’s one gambit to save it that the paper considers. A policy of non-interference would only need to be enforced for a few thousand years – perhaps only a few hundreds – if extraterrestrials were interested primarily in technological societies. This is Amri Wandel’s notion in an interesting paper titled “The Fermi paradox revisited: technosignatures and the contact era” (citation below). Wandel (Hebrew University of Jerusalem) eases our concern over the monocultural issue by compressing the time needed for concealment. Crawford and Schulze-Makuch cite Wandel, but I don’t sense any great enthusiasm for pressing his solution as likely.
The reasons for doubt multiply:
Even if they can hide evidence of their technology (space probes, communications traffic and so forth), hiding the large number of inhabited planets in the background implied by such a scenario would probably prove challenging (unless they are able to bring an astonishingly high level of technical sophistication to the task). In any case, advanced technological civilizations may find it difficult to hide the thermodynamic consequences of waste heat production, which is indeed the basis of some current technosignature searches. Moreover, any spacefaring civilization is likely to generate a great deal of space debris, and the greater the number of ETIs that have existed in the history of the Galaxy the greater the quantity of debris that will drift into the Solar System, where a determined search may discover evidence for it.
Why then highlight the zoo hypothesis when it has all these factors working against it? Because in the view of the authors, other solutions to the Fermi question are even worse. I’m running out of time this morning, but in the next post I want to dig into some of these other answers to see whether any of them can still be salvaged. For the more dubious our solutions to the ‘where are they’ question, the more likely it seems that there are no civilizations nearby. We’ll continue to push against that likelihood with technosignature and biosignature searches that could change everything.
The paper is Crawford & Schulze-Makuch, “Is the apparent absence of extraterrestrial technological civilizations down to the zoo hypothesis or nothing?” Published online in Nature Astronomy 28 December 2023 (abstract). David Brin’s essential paper is “The Great Silence – the Controversy Concerning Extraterrestrial Intelligent Life,” Quarterly Journal of the Royal Astronomical Society Vol. 24, No.3 (1983), pp. 283-309 (abstract/full text). Amri Wandel’s paper is “The Fermi Paradox revisited: Technosignatures and the Contact Era,” Astrophysical Journal 941 (2022), 184 (preprint).
Life Elsewhere? Relaxing the Copernican Principle
Most people I know are enthusiastic about the idea that other intelligent races exist in the galaxy. Contact is assumed to be an inevitable and probably profoundly good thing, with the exchange of knowledge possibly leading to serious advances in our own culture. This can lead to a weighting of the discourse in favor of our not being alone. The ever popular Copernican principle swings in: We can’t be unique, can we? And thus every search that comes up empty is seen as an incentive to try still other searches.
I’m going to leave the METI controversy out of this, as it’s not my intent to question how we should handle actual contact with ETI. I want to step back further from the question. What should we do if we find no trace of extraterrestrials after not just decades but centuries? I have no particular favorite in this race. To me, a universe teeming with life is fascinating, but a universe in which we are alone is equally provocative. Louis Friedman’s new book Alone But Not Lonely (University of Arizona Press, 2023) gets into these questions, and I’ll have more to say about it soon.
I’ve thought for years that we’re likely to find the galaxy stuffed with living worlds, while the number of technological civilizations is tiny, somewhere between 1 and 10. The numbers are completely arbitrary and, frankly, a way I spur (outraged) discussion when I give talks on these matters. I’m struck by how many people simply demand a galaxy that is alive with intelligence. They want to hear ‘between 10,000 and a million civilizations,’ or something of that order. More power to them, but it’s striking that such a lively collection of technological races would not have become apparent by now. I realize that the search space is far vaster than our efforts so far, but still…
Image: The gorgeous M81, 12 million light years away in Ursa Major, and seen here in a composite Spitzer/Hubble/Galaxy Evolution Explorer view. Blue is ultraviolet light captured by the Galaxy Evolution Explorer; yellowish white is visible light seen by Hubble; and red is infrared light detected by Spitzer. The blue areas show the hottest, youngest stars, while the reddish-pink denotes lanes of dust that line the spiral arms. The orange center is made up of older stars. Should we assume there is life here? Intelligence? Credit: NASA/JPL.
So when Ian Crawford (Birkbeck, University of London) was kind enough to send me a copy of his most recent paper, written with Dirk Schulze-Makuch (Technische Universität Berlin), I was glad to see the focus on an answer to the Fermi question that resonates with me, the so-called ‘zoo hypothesis.’ A variety of proposed resolutions to the ‘where are they’ question exist, but this one is still my favorite, a way we can save all those teeming alien civilizations, and a sound reason for their non-appearance.
As far as I know, Olaf Stapledon first suggested that intelligent races might keep hands off civilizations while they observed them, in his ever compelling novel Star Maker (1937). But it appears that credit for the actual term ‘zoo hypothesis’ belongs to John Ball, in a 1973 paper in Icarus. From Ball’s abstract:
Extraterrestrial intelligent life may be almost ubiquitous. The apparent failure of such life to interact with us may be understood in terms of the hypothesis that they have set us aside as part of a wilderness area or zoo.
That’s comforting for those who want a galaxy stuffed with intelligence. I want to get into this paper in the next post, but for now, I want to note that Crawford and Schulze-Makuch remind us that what is usually styled the Fermi ‘paradox’ is in fact no paradox at all if intelligent races beyond our own do not exist. We have a paradox because we are uneasy with the idea that we are somehow special in being here. Yet a universe devoid of technologies other than ours will look pretty much like what we see.
The angst this provokes comes back to our comfort with the ‘Copernican principle,’ which is frequently cited, especially when we use it to validate what we want to find. Just as the Sun is not the center of the Solar System, so the Solar System is not the center of the galaxy, etc. We are, in other words, nothing special, which makes it more likely that there are other civilizations out there because we are here. If we can build radio telescopes and explore space, so can they, because by virtue of our very mediocrity, we represent what the universe doubtless continues to offer up.
But let’s consider some implications, because the Copernican principle doesn’t always work. It was Hermann Bondi, for example, who came up with the notion that we could apply the principle to the cosmos at large, noting that the universe was not only homogeneous but isotropic, and going on to add that it would show the exact same traits for any observer not just at any place but at any time. The collapse of the Steady State theory put an end to that speculation as we pondered an evolving universe where time’s vantage counted critically in terms of what we would see.
Our position in time matters. So, for that matter, does our position in the galaxy.
But physics seems to work no matter where we look, and the assumption of widespread physical principles is essential for us to do astronomy. So as generalizations go, this Copernican notion isn’t bad, and we’d better hang on to it. Kepler figured out that planetary orbits weren’t circular, and as Caleb Scharf points out in his book The Copernicus Complex: Our Cosmic Significance in a Universe of Planets and Probabilities (Farrar, Straus and Giroux, 2014), this was a real break from the immutable universe of Aristotle. So too was Newton’s realization that the Sun itself orbits around a variable point close to its surface and well offset from its core.
So even the Sun isn’t the center of the Solar System in any absolute sense. As we move from Ptolemy to Copernicus, from Tycho Brahe to Kepler, we see a continuing exploration that pushes humanity out of any special position and any fixed notions that are the result of our preconceptions. I think the problem comes when we make this movement a hard principle, when we say that no ‘special places’ can exist. We can’t assume from a facile Copernican model that each time we apply the principle of mediocrity, we’ve solved a mystery about things we haven’t yet proven.
Consider: We’ve learned how unusual our own Solar System appears to be; indeed, how unusual so many stellar systems are as they deviate hugely from any ‘model’ of system development that existed before we started actually finding exoplanets. This is why the first ‘hot Jupiters’ were such a surprise, completely unexpected to most astronomers.
Is the Sun really just another average star lost in the teeming billions that accompany it in its 236 million year orbit of the galaxy? There are many G-class stars, to be sure, but if we were orbiting a more average star, we would have a red dwarf in the sky. These account for 75 percent, and probably more, of the stars in the Milky Way. We’re not average on that score, not when G-class stars amount to a paltry 7 percent of the total. Better to say that we’re only average, or mediocre, up to a point. If we want to take this to its logical limit, we can back our view out to the scale of the cosmos. Says Scharf::
The fact that we are so manifestly located in a specific place in the universe — around a star, in an outer region of a galaxy, not isolated in the intergalactic void, and at just this time in cosmic history — is simply inconsistent with ‘perfect’ mediocrity.
And what about life itself? Let me quote Scharf again (italics mine). Here he works in the anthropic idea that our observations of the universe are not truly random but are demanded by the fact that the universe can produce life in the first place:
…a Copernican worldview at best suggests that the universe should be teeming with life like that on Earth, and at worst doesn’t really tell us one way or the other. The alternative — anthropic arguments — require only a single instance of life in the universe, which would be us. At best, some fine-tuning studies suggest that the universe could be marginally suitable for heavy-element-based-life-forms, rather than being especially fertile. Neither view reveals much about the actual abundance of life to be expected in our universe, or much about our own more parochial significance or insignificance.
So when we speculate about the Fermi question, we need to be frank about our assumptions and, indeed, our personal inclinations. If we relax our Copernican orthodoxy, we have to admit that because we are here does not demand that they are there. Let’s just keep accumulating data to begin answering these questions.
And as we’ll discuss in the next post, Crawford and Schulze-Makuch point out that we’re already entering the era when meaningful data about these questions can be gathered. One key issue is abiogenesis. How likely is life to emerge even under the best of conditions? We may have some hard answers within decades, and they may come from discoveries in our own system or in biosignatures from a distant exoplanet.
If abiogenesis turns out to be common (and I would bet good money that it is), we still have no knowledge of how often it evolves into technological societies. An Encyclopedia Galactica could still exist. Could John Ball be right that other civilizations may be ubiquitous, but hidden from us because we have been sequestered into ‘nature preserves’ or the like? Are we an example of Star Trek’s ‘Prime Directive’ at work? There are reasons to think that the zoo hypothesis, out of all the Fermi ‘solutions’ that have been suggested, may be the most likely answer to the ‘where are they’ question other than the stark view that the galaxy is devoid of other technological societies. We’ll examine Crawford and Schulze-Makuch’s view on this next time.
Caleb Scharf’s The Copernicus Complex: Our Cosmic Significance in a Universe of Planets and Probabilities is a superb read, highly recommended. The Ball paper is “The Zoo Hypothesis,” Icarus Volume 19, Issue 3 (July 1973), pp. 347-349 (abstract). The Crawford & Schulze-Makuch paper we’ll look at next time is “Is the apparent absence of extraterrestrial technological civilizations down to the zoo hypothesis or nothing?” Nature Astronomy 28 December, 2023 (abstract).
Holiday Thoughts on Deep Time
An old pal from high school mentioned in an email the other day that he had an interest in Adam Frank’s work, which we’ve looked at in these pages a number of times. Although my most recent post on Frank involves a 2022 paper on technosignatures written with Penn State’s Jason Wright, my friend was most intrigued by a fascinating 2018 paper Frank wrote for the International Journal of Astrobiology (citation below). The correspondence triggered thoughts of other, much earlier scientists, particularly of Charles Lyell’s Principles of Geology (1830-1833), which did so much to introduce the concept of ‘deep time’ to Europe and played a role in Darwin’s work. Let’s look at both authors, with a nod as well to James Hutton, who largely originated the concept of deep time in the 18th Century.
Adam Frank is an astrophysicist at the University of Rochester, and one of those indispensable figures who meshes his scientific specialization (stellar evolution) with a broader view that encompasses physics, cultural change and their interplay in scientific discourse. He fits into a niche of what I think of as ‘big picture’ thinkers,’ scientists who draw out speculation to the largest scales and ponder the implications of what we do and do not know about astrophysics for a species that may spread into the cosmos.
Now in the case of my friend’s interest, the picture is indeed big. Frank’s 2018 paper asked whether our civilization is the first to emerge on Earth. Thus the ‘Silurian’ hypothesis, explored on TV’s Doctor Who in reference to a race of intelligent reptiles by that name who are accidentally awakened. The theme pops up occasionally in science fiction, though perhaps less often that one might expect. James Hogan’s 1977 novel Inherit the Stars, for example, posits evidence for unknown technologies discovered on the Moon that apparently have their origin in an earlier geological era.
Image: Astrophysicist Adam Frank. Credit: University of Rochester.
I won’t go through this paper closely because I’ve written it up before (see Civilization before Homo Sapiens?), but this morning I want to reflect on the implications of the question. For it turns out that if, say, a species of dinosaur had evolved to the point of creating technologies and an industrial civilization, finding evidence of it would be an extremely difficult thing. So much so that I find myself reflecting on deep time in much the same way that I reflect on the physical cosmos and its seemingly endless reach.
Consider that we can trace our species back in the Quaternary (covering the last 2.6 million years or so) and find evidence of non-Homo Sapiens cultures, among which the Neanderthals are the most famous, along with the Denisovians. Bipedal hominids show up at least as far back as the Laetoli footprints in Tanzania, which date to 3.7 million years ago and were apparently produced by Australopithecus afarensis. Frank and co-author Gavin Schmidt also note that the largest ancient surface still available for study on our planet is in the Negev Desert, dating back about 1.8 million years.
These are impressive numbers until we put them into context. The Earth is some 4.5 billion years old, and complex life on land has existed for about 400 million of those years. Let’s also keep in mind that agriculture emerged perhaps 12,000 years ago in the Fertile Crescent, and in terms of industrial technologies, we’ve only been active for about 300 years (the authors date this from the beginning of mass production methods). Tiny slivers of time, in other words, amidst immense timeframes.
So as Frank and Schmidt point out, we’re talking about fractions of fractions here. There is a fraction of life that gets fossilized, which in all cases is tiny and also varies according to tissue, bone structure, shells and so forth, and also varies from an extremely low rate in tropical environments to a higher rate in dry conditions or river systems. The dinosaurs were active on Earth for an enormous period of time, from the Triassic to the end-Cretaceous extinction event, something in the range of 165 million years. Yet only a few thousand near-complete dinosaur specimens exist for this entire time period. Would homo sapiens even show up in the future fossil record?
From the paper:
The likelihood of objects surviving and being discovered is similarly unlikely. Zalasiewicz (2009) speculates about preservation of objects or their forms, but the current area of urbanization is <1% of the Earth’s surface (Schneider et al., 2009), and exposed sections and drilling sites for pre-Quaternary surfaces are orders of magnitude less as fractions of the original surface. Note that even for early human technology, complex objects are very rarely found. For instance, the Antikythera Mechanism (ca. 205 BCE) is a unique object until the Renaissance. Despite impressive recent gains in the ability to detect the wider impacts of civilization on landscapes and ecosystems (Kidwell, 2015), we conclude that for potential civilizations older than about 4 Ma, the chances of finding direct evidence of their existence via objects or fossilized examples of their population is small.
Image: The Cretaceous-aged rocks of the continental interior of the United States–from Texas to Montana–record a long geological history of this region being covered by a relatively shallow body of marine water called the Western Interior Seaway (WIS). The WIS divided North America in two during the end of the age of dinosaurs and connected the ancient Gulf of Mexico with the Arctic Ocean. Geologists have assigned the names “Laramidia” to western North America and “Appalachia” to eastern North America during this period of Earth’s history. If a species produced a civilization in this era, would we be able to find evidence of it? Credit; National Science Foundation (DBI 1645520). The Cretaceous Atlas of Ancient Life is one component of the overarching Digital Atlas of Ancient Life project. CC BY-NC-SA 4.0 DEED.
Intriguing stuff. The authors advocate exploring the persistence of industrial byproducts in ocean sediment environments, asking whether byproducts of common plastics or organic long-chain synthetics will be detectable on million-year timescales. They also propose a deeper dive into anomalies in current studies of sediments, the same sort of analysis that has been done, for example, in exploring the K-T boundary event but broadened to include the possibility of an earlier civilization. I send you to the paper, available in full text, for discussion of such testable hypotheses.
Back to deep time, though, and the analogy of looking ever deeper into the night sky. In asking how long a civilization can survive (Drake’s L term in the famous equation), we ask whether we are likely to find other civilizations given that over billion year periods, they may last only as a brief flicker in the night. We have no good idea of what the term L should be because we are the only civilization we know about. But if civilizations can emerge more than once on the same world, the numbers get a little more favorable, though still daunting. A given star may be circled by a planet which has seen several manifestations of technology, a greater chance for our detection.
A cycle of civilization growth and collapse might be mediated by fossil fuel availability and resulting climate change, which in turn could feed changes in ocean oxygen levels. Frank has speculated that such changes could trigger the conditions for creating more fossil fuels, so that the demise of one culture actually feeds the energy possibilities of the next after many a geological era. How biospheres evolve – how indeed they have evolved on our own world – is a question that exoplanet research may help to answer, for we have no shortage of available worlds to examine as our biosignature technologies develop.
Culturally, we must come to grips with these things. In an essay for The Geological Society, British paleontologist Richard Fortey discusses the seminal work of James Hutton and Charles Lyell in the 18th and 19th Centuries in developing the concept of geological time, which John McPhee would present wonderfully in his 1981 book Basin and Range (I remember reading excerpts in The New Yorker). The Scot James Hutton had literary ambitions, publishing his Theory of the Earth in 1795 and changing our conception of time forever. Hutton knew Adam Smith and spent time with David Hume; he would also have been aware of French antecedents to his ideas. But despite its importance, even Lyell would admit that he found Hutton’s book all but unreadable.
It took a friend named John Playfair to turn Hutton’s somnolent prose into the simplified but clear Illustrations of the Huttonian Theory of the Earth in 1802, making the idea of deep time available to a large audience and leading to Lyell. Which goes to show that sometimes it takes a careful popularizer to gain for a scientist the traction his or her work deserves. The emphasis there is on ‘careful.’
Lyell’s Principles of Geology, published in three volumes between 1830 and 1833, famously traveled with Darwin on the Beagle and, as Fortey says, “donated the time frame in which evolution could operate.” He goes on:
“…once the time barrier had been breached, it was only a question of how much time. The stratigraphical divisions of the geological column, the periods such as Devonian or Cambrian, with which we are now so familiar, were themselves being refined and put into the right sequence through the same historical period. Just to have a sequence of labels helped geologists grapple with time, and, in a strange way, labels domesticate time.
But domestication co-exists with wonder. I imagine the most hardened geologist of our day occasionally quakes at the realization of what all those sedimentary layers point to, a chronological architecture — time’s edifice — in which our entire history as a species is but a glinting mote on a rockface of the future. Our brief window today is reminiscent of Hutton and Lyell’s. Like them, we are compelled to adjust to a cosmos that seems to somehow enlarge every time we probe it, inspired by new technologies that give birth to entire schools of philosophy.
John Playfair would write upon visiting Siccar Point, the promontory in Berwickshire that inspired Hutton’s ideas, that “The mind seemed to grow giddy looking so far into the abyss of time.” We are similarly dwarfed by the vistas of the Hubble Ultra Deep Field and the exquisite imagery from JWST. Who knows what we have yet to discover in Earth’s deep past?
The paper is Schmidt and Frank, “The Silurian Hypothesis: Would it be possible to detect an industrial civilization in the geological record?” published online by the International Journal of Astrobiology 16 April 2018 (full text). Gregory Benford’s Deep Time: How Humanity Communicates Across Millennia (Bard, 2001) is a valuable addition to this discourse. For a deeper dive, Fortey mentions Martin Rudwick’s Bursting the Limits of Time: The Reconstruction of Geohistory in the Age of Revolution ( University of Chicago Press, 2007). Fortey’s own Life: A Natural History of the First Four Billion Years of Life on Earth (Knopf, Doubleday 1999) is brilliant and seductively readable.