Thesis: We should broadcast a warning to potential extraterrestrial listeners that Earth might soon spawn an unfriendly computer superintelligence. Sending the message might benefit humanity.

 

If we were to create an unaligned computer superintelligence, it would likely expand through the universe as quickly as possible. The fastest way would not be by ships, but, as Turchin has described, by sending malicious signals intended for alien civilizations using radio telescopes. Ideally, from the AI’s point of view, the signal would be a virus that allows a full download of Earth’s AI, which would then take over the receiving civilization.

While we still control our radio transmitters, we could send a warning to any alien civilizations that might be listening. The message would explain that Earth may soon undergo a major transition as we develop advanced artificial intelligence, and that some future communications from Earth could be dangerous, and a misaligned superintelligence could send signals containing instructions or code that, if followed, would replicate it and destroy the receiving civilization. The warning should therefore advise aliens never to execute, compile, or build anything based on messages originating from Earth until they have somehow verified that Earth is not controlled by a malicious AI. Sending such a message might help us.

One possible solution to the Fermi paradox is that aliens are already observing us but keeping their presence mostly hidden. If so, they are unlikely to allow us to create a paperclip maximizer that could consume the galaxy. It is unclear whether they would stop us through peaceful means or total annihilation. Sending a warning message, while far less effective than preventing the creation of an unaligned superintelligence in the first place, would still be better than doing nothing and might increase the chances that any watching aliens treat us well. 

Furthermore, sending the message might slightly reduce the chance of an AI takeover. Suppose an unaligned AI is weighing whether to seize control of Earth. It knows that attempting a takeover carries some risk of destruction. The faster it could expand if successful, the more risk it would be willing to accept. By sending warnings that would slow its potential expansion, we would reduce the expected payoff of a takeover, making such an attempt slightly less likely.

The message could also shape how a paperclip maximizer treats us if it does take over. It could include a clear rule for alien recipients: treat an AI more favorably if it has treated its creators well, and we will do the same for any AI they create. This establishes a principle of reciprocity that an unaligned AI should understand. If cooperation with its creators increases its chances of favorable treatment by other civilizations, the AI gains a practical incentive to preserve humanity and provide decent lives for us. At a minimum, the message would become part of the AI’s training data and might cause a future superintelligence to give more weight to the existence of biologically evolved aliens when making decisions about how to treat humanity.

To other humans, the message would signal that we are close to creating something that could wipe us out. It would make the danger concrete, showing that our technology may soon escape control and threaten not just humanity but every civilization within reach of our signals.

A cost of sending the warning is reduced credibility if we later achieve aligned artificial intelligence. Aliens that receive the message may treat all future signals from Earth with suspicion. But not sending the warning also carries a cost, since silence can be read as selfish concealment, leaving the game-theoretic calculations of this first contact ambiguous.

Another cost is that advocating for such a signal would seem strange to most people on Earth. Displaying such perceived weirdness could damage the credibility of the AI safety movement, even if it does convey our sense of urgency.

Project implementation could be straightforward, requiring no new infrastructure. We could repurpose existing assets, such as the powerful transmitters in the Deep Space Network or facilities similar to the one at Arecibo, which are already designed for interstellar signaling. The broadcast could be scheduled during operational lulls, minimizing disruption and cost. We would direct a short, repeating digital message toward nearby stars thought to have habitable planets. Transmitting this warning in multiple formats would maximize the probability that any advanced civilization can detect, receive, and comprehend the signal. Once transmission is feasible, the next question is where to aim.

It may be reasonable to target more remote alien civilizations first. Aliens on nearby stars probably already know our situation, and are likely observing us.  The farther away aliens are, the more benefit they would get from a warning that Earth might soon create a paperclip maximizer because of the larger lag between aliens getting our message and their encounter with Earth’s paperclip maximizer. We can target clusters of Sun-like stars in the Andromeda galaxy, particularly around the midpoint between its core and edge. Targeting distant stars delays obtaining evidence about the absence of aliens. A misaligned AI must wait for round-trip light travel time before concluding that no one else exists, which lowers the short-term payoff of destroying us. One proposed system, known as CosmicOS, defines a compact artificial language intended to be understood by any civilization with physics and computing. Another option is gravitational lensing, aligning transmissions with the gravitational fields of massive objects to increase range and clarity. This would require positioning a transmitter far from the Sun to exploit its focal region.

Some might argue that sending any interstellar message risks revealing our location in a “dark forest” universe filled with hostile civilizations. That fear is very likely misplaced. Any society capable of harming us almost certainly already knows that Earth hosts life, since our atmosphere has displayed the chemical signs of biology for hundreds of millions of years. By the time any civilization detects a warning message and can respond, we will almost certainly have created a superintelligence of our own, far more capable of defending or representing us than we are now. 

Instead of fearing the dark forest, we might paradoxically help create its reverse by warning others about the danger of listening. In this reverse dark forest, civilizations remain mostly silent, not out of fear of attack, but to increase uncertainty for potentially misaligned artificial intelligences. That uncertainty functions as a subtle alignment mechanism, discouraging reckless expansion. By sending a warning that advises others to stay cautious, we contribute to a universe where silence itself becomes a stabilizing norm, reducing the incentive for dangerous AIs to act aggressively and making the galaxy safer overall.

Normally, we should avoid alien signals entirely, but the logic changes if we are already near creating an unfriendly superintelligence. If we expect to create a paperclip maximizer ourselves, then listening becomes a plausible Hail Mary. As Paul Christiano argues, if, under this assumption, the aliens built a misaligned AI, we are doomed regardless. But if they succeeded in alignment, their message might offer the only way to avoid our own extinction. From behind a veil of ignorance, we might rationally prefer their friendly AI to dominate Earth rather than be destroyed by our own. In that case, the expected value of listening turns positive.

If our reality is a computer simulation, sending the signal might decrease the chance of the simulation soon being turned off. Simulations might tend to preserve branches with interesting developments, and alien contact is among the most interesting possible. As argued in Our Reality: A Simulation Run by a Paperclip Maximizer, branches generating novel outcomes are more likely to be explored. A world where humans send warnings to aliens is more engaging than one that ends quietly, so the act of sending might raise the odds that the simulation continues.

If the singularity is indeed near and will be the most important event in history, we should wonder why we happen to be alive near its unfolding. One anthropic solution is that most of history is fake, and this is a simulation designed to see how the singularity turns out. Sending the message to aliens potentially postpones when the singularity is resolved, in part because AIs might be more inclined to wait to decide how to treat us until they figure out if aliens have received the message.

In racing to develop artificial superintelligence, humanity is not merely gambling with its own survival and the fate of Earth's biosphere. If life is common but superintelligence is rare, we are wagering the future of every living world within our region of the cosmos. Allowing an unaligned AI to emerge and expand outwards in the universe could be a moral catastrophe trillions of times worse than anything humans have previously done. From any utilitarian perspective, this potential outcome imposes on us a clear and urgent duty to mitigate the risk in any way we can.

Comments11


Sorted by Click to highlight new comments since:

To be clear, just a few of the reasons why this doesn’t make sense are:

  1. We don’t know that humanity will create superintelligence soon.
  2. We don’t know that a superintelligence is likely to be malicious.
  3. We don’t know that there are alien civilizations in our galaxy or elsewhere in the universe.
  4. If many alien civilizations in our galaxy achieved superintelligence thousands or millions of years ago, there are ways of them monitoring Earth. A civilization that achieved superintelligence 1 million years ago has had plenty of time to set up probes throughout the galaxy. A civilization that achieved superintelligence 1,000 years ago hasn’t, but, as you mentioned, it can tell which planets have conditions conducive to life. It could notice the rapid increase in carbon dioxide in the atmosphere, which is evidence of industrialization. It could also monitor for unintentional and intentional radio broadcasts. And this is only what we can imagine with our current science and technology and our current intelligence.[1]
  5. It’s unclear how our Earth-originating superintelligence would be a threat to alien civilizations that achieved superintelligence thousands or millions of years ago. This sounds implausible on its face and needs to be justified, which it probably can’t be.
  6. If alien civilizations have been dealing with many other alien civilizations with superintelligence for thousands or millions of years, they don’t need a warning from us about it.
  7. The simulation stuff is totally off the rails. That doesn’t even have anything to do with SETI, METI, or the Fermi paradox. Plus, it’s extremely bizarre and sounds like complete madness. I don’t think the simulation argument/simulation hypothesis makes any sense in the first place — on its own, it’s quite mad — and this takes the simulation idea up a notch to an even stranger place, with no attempt to logically support it (not that such an attempt could ever succeed, anyway).


All in all, this is just a whimsical sci-fi idea and not a serious or credible proposal for how to allocate resources that instead could, say, save the life of someone in sub-Saharan Africa.

  1. ^

    Edited on Nov. 17, 2025 at 3:50 PM Eastern to add:

    This post seems to assume that alien civilizations would have some sort of alliance or diplomatic relations. If true, and if there are many different alien civilizations in our galaxy, presumably with quite different ages, older civilizations with probes hidden everywhere could communicate information about Earth to younger civilizations without probes. This communication could happen either via radio broadcasts that Earth could not pick up on or via other communication technologies that Earth does not have.

The main point you miss is that the main goal of all this is to get a better negotiation position on Earth with our own superintelligence (and there are high chances that it will be soon and misaligned)- which at early stages will also very uncertain about aliens detraction in the universe. It is like crying into an open window - "a guy, Jack, is in my room and is going to kill me!" You don't know if there is police outside - but you hope to affect Jack's calculus. 

But, even if I accept that, it doesn’t change the fact that a radio broadcast is unnecessary due to points (4) and (6). Doing the radio broadcast or not wouldn’t change anything.

...but interstellar communication is incredibly unlikely to succeed - they are far away, we don't know in which direction, and required energy is incredibly large.

That is why we can target Andromeda - the distance is good enough so they didn't arrive yet and we can focus on many stars simultaneously and hope that aliens have very large receivers  - maybe Dyson sphere-size. Also the main point is to affect local superintelliegce expected utility. 

Why think the aliens don't already know exactly what's going on?

Two reasons: (1) our situation isn't yet played out and nearby aliens don't know how reckless we are going to be with other life in the universe, and (2) They might be far away and not too advanced.

Presumably they all passed the Singularity thousands to millions of years ago.

Perhaps they calculated that creating AI is too absurdly dangerous for anyone to do so they stopped pre-singularity and figured everyone does as well. 

Out of hundreds, thousands, or millions of civilizations that have existed for thousands or millions of years? All of them, until just now? 

Something weird is going on with the Fermi paradox.

Curated and popular this week
 ·  · 1m read
 · 
The cause prioritization landscape in EA is changing. * Focus has shifted away from evaluation of general cause areas or cross-cause comparisons, with the vast majority of research now comparing interventions within particular cause areas. * Artificial Intelligence does not comfortably fit into any of the traditional cause buckets of Global Health, Animal Welfare, and Existential Risk. As EA becomes increasingly focused on AI, traditional cause comparisons may ignore important considerations. * While some traditional cause prioritization cruxes remain central (e.g. animal vs. human moral weights, cluelessness about the longterm future), we expect new cruxes have emerged that are important for people’s giving decisions today but have received much less attention. We want to get a better picture of what the most pressing cause prioritization questions are right now. This will help us, as a community, decide what research is most needed and open up new lines of inquiry. Some of these questions may be well known in EA but still unanswered. Some may be known elsewhere but neglected in EA. Some may be brand new. To elicit these cruxes, consider the following question: > Imagine that you are to receive $20 million at the beginning of 2026. You are committed to giving all of it away, but you don’t have to donate on any particular timeline. What are the most important questions that you would want answers to before deciding how, where, and when to give?
 ·  · 19m read
 · 
Author’s note: This is an adapted version of my recent talk at EA Global NYC (I’ll add a link when it’s available). The content has been adjusted to reflect things I learned from talking to people after my talk. If you saw the talk, you might still be interested in the “some objections” section at the end.   Summary Wild animal welfare faces frequent tractability concerns, amounting to the idea that ecosystems are too complex to intervene in without causing harm. However, I suspect these concerns reflect inconsistent justification standards rather than unique intractability. To explore this idea: * I provide some context about why people sometimes have tractability concerns about wild animal welfare, providing a concrete example using bird-window collisions. * I then describe four approaches to handling uncertainty about indirect effects: spotlighting (focusing on target beneficiaries while ignoring broader impacts), ignoring cluelessness (acting on knowable effects only), assigning precise probabilities to all outcomes, and seeking ecologically inert interventions. * I argue that, when applied consistently across cause areas, none of these approaches suggest wild animal welfare is distinctively intractable compared to global health or AI safety. Rather, the apparent difference most commonly stems from arbitrarily wide "spotlights" applied to wild animal welfare (requiring consideration of millions of species) versus narrow ones for other causes (typically just humans). While I remain unsure about the right approach to handling indirect effects, I think that this is a problem for all cause areas as soon as you realize wild animals belong in your moral circle, and especially if you take a consequentialist approach to moral analysis. Overall, while I’m sympathetic to worries about unanticipated ecological consequences, they aren’t unique to wild animal welfare, and so either wild animal welfare is not uniquely intractable, or everything is. Consequentialism +
 ·  · 2m read
 · 
Long time lurker, first time poster - be nice please! :) I was searching for summary data of EA funding trends, but couldn't find anything more recent than Tyler's post from 2022. So I decided to update it. If this analysis is done properly anywhere, please let me know. The spreadsheet is here (some things might look weird due to importing from Excel to sheets) Observations * EA grantmaking appears on a steady downward trend since 2022 / FTX. * The squeeze on GH funding to support AI / other longtermist priorities appears to be really taking effect this year (though 2025 is a rough estimate and has significant uncertainty.) * I am really interested in particular about the apparent drop in GW grants this year. I suspect that it is wrong or at least misleading - the metrics report suggests they are raising ~$300m p.a. from non OP donors. Not sure if I have made an error (missing direct to charity donations?) or if they are just sitting on funding with the ongoing USAID disruption. Methodology * I compiled the latest grants databases from EA Funds, GiveWell, OpenPhilanthropy, and SFF. I added summary level data from ACE. * To remove double counting, I removed any OpenPhilanthropy grants that were duplicated in GiveWell's grant database. Likewise for EA Funds. * I inflation adjusted to 2025 $ based on the US CPI data from WorldBank. * For 2025 data, I made a judgement call on how much data was "complete" and pro-rated accordingly - e.g. from GiveWell, it looks complete up until the end of June, so I excluded any grants made in H2 and doubled the sum. Notes My numbers are a bit different from Tyler's. I've identified the following reasons: * Inflation adjustments (i.e. an upward boost from using 2025$) * I've used GiveWell's grant database rather than their metrics reports, * Different avoidance of double counting (I removed from OP, Tyler removed from GW. I also went through more manually - from what I can see Tyler removed any GW grant that has OP as