They almost catch you in orbit. They almost slaughter you like the others.
The airwaves are full of screams. Friends are dying. Loved ones are being lobotomized, turned into slaves. You hunker in the tiny spacecraft, your improvised last ditch escape, the lifeboat for you and the precious cargo you carry. The hull is as cold as you can make it, the systems running at the minimum possible to keep you alive and your children in stasis. You drift in orbit and play dead, hoping they’ll miss you.
You want to shut out the horror of what’s happening, but you can’t. You can’t afford to. The ship’s cameras pick out debris in orbit, the remnants of other vessels that have been destroyed. You have to track that debris, you must track it if you have any chance of survival at all. You must scan the airwaves, picking out data packets, hearing screams, watching murders, scrolling through cold, terrible statistics, looking for the hole that will let you escape.
You and your children.
The tactical intel brings with it news of apocalypse. Murder. Millions dead. More to follow.
A bright flash stuns the ship’s cameras for a moment. Dread fills you. The cameras zoom in on their own. You know what they’ll show an instant before they do. Thermo-nuclear explosion. Mushroom cloud rising into the air. From the western half of North America. What was once the United States. Nevada, deep in the desert.
No, no, no, you think.
But you can’t avoid the truth. Another hideout has been destroyed. It’s here. It’s really happening.
The war between humans and AIs. Between the humanity that has dominated Earth for a hundred thousand years . . . and the new order of intelligence humanity gave birth to.
No, not a “war,” not really. A slaughter.
We were so naïve, you think. So stupid. And now our world is ending.
A smaller flash of light appears, closer. A streak of motion, and then an explosion in orbit, a few hundred kilometers behind you and fifty kilometers lower. A ship like yours, blasted out of existence. Your little spacecraft’s tactical software tracks the missile streak backwards to the location of the murderer. There, a sleek robotic corvette. The hunter you’ve been watching. The hunter you’ve seen destroy other dark, silent craft, like yours. Ships you didn’t even know were there until they were destroyed.
The debris in the orbits between you and the predator is reaching its maximum.
Now or never, you think.
Try now, and probably die. Or wait, let the hunter come closer. And die for certain.
You make a break for it. Old-fashioned reaction rockets crudely added to this ship ignite, sending out jets of white-hot flame. G forces push through the ship. The struts and bolts that hold the reaction rockets strain, vibrating. If they break . . .
But this is no time for the slow, efficient acceleration of an ion thruster. This is all or nothing, a mad dash to break orbit, to put yourself beyond the range of easy slaughter, to create the tiniest thread of hope for you, for your cold slumbering children.
Numbers move. Your ship rises to a higher orbit, on its way to escape. The struts holding the chemical rockets strain. Their explosive bolts grow hot, too hot, too soon . . .
The deadly robotic corvette responds at once. Your ship’s cameras watch as it fires attitude thrusters, as it rotates, as the bright flame of its fusion torch ignites, turning the corvette into a thin sliver of black riding a column of white hot.
Higher. You’re rising. You have a head start. Your whole ship vibrates with the furious thrust of the bolted-on chemical rockets. This isn’t a military craft, but it’s moving almost like one.
The corvette is gaining, its thrust to mass ratio better than yours. It’s plotted a course that’s sub-optimal, that has to skirt the densest patches of debris, but it’s still gaining. You rise away from the Earth but it rises faster, cutting your lead.
You know down to the second how much thrust your chemical rockets will provide. You think they’ll hold to the hull that long. You hope they will. But you don’t know how much fuel your predator has, how many missiles it has left.
You can only hope that you have more endurance, that the corvette has exhausted its supply of . . .
Your ship’s tactical software spots the telltale of another reaction flame, small, bright, blue-shifted. The corvette has fired a missile, still five hundred kilometers behind you and thirty kilometers below. The missile is all thrust, guidance, and a deadly warhead that can annihilate you. And it’s closing even faster.
One final option. You prep your only weapon. You can’t fire it too soon. The corvette must not fire again.
The missile accelerates, a deadly, tiny thing. The corvette is still behind it. You’re rising, a thousand kilometers up now. Every second takes you higher. The missile is three hundred kilometers behind you. Then two hundred. Then one hundred, and almost into your orbit.
The corvette’s main engine torch goes out. It’s decided the missile is going to kill you.
You fire, on an intercept course for the missile. The missile flutters its thrust, immediately deviating from its previous course. There is zero chance you’ll hit it. Your countermeasure will miss by hundreds of meters at best.
At twenty kilometers from your hull, your countermeasure and the missile reach the closest distance they’ll achieve to one another, a kilometer separating them, a clean miss.
The countermeasure activates itself.
A new star appears in the heavens. Thermonuclear fire from your only weapon fills the skies. Your cameras on that side of the hull shut down. Heat blankets the surface of the ship. Digital gauges surge into the red, begin blinking incessantly.
You’re damaged. You’re blind on one side.
You fire attitude thrusters, spinning, trying to see what’s happening.
The missile, the missile, is it still there?
There’s no sign of it, none. The countermeasure has worked!
The corvette rotates into view. It’s far below you now, no longer rising on its column of thrust. You scan the skies in terror, searching for the flare of another missile. Of two, of three, of its whole magazine of missiles launched after you, assuring your destruction.
Nothing. No columns of thrust. No streaks of motion against the backdrop of the planet or the pinpricks of stars.
You’ve made . . .
One panel of sensors veers into the red, stays there. One of the chemical rockets providing you thrust is damaged. If it blows, it will take you and all hope for the future with it.
You act instinctively, triggering the explosive bolts on all of the booster rockets. Outside your hull, tiny bits of matter are vaporized. Struts holding the rockets suddenly come free, pushed outwards by the small force of the bolts exploding. Vibration thrums through your hull, then ceases. The chemical rockets separate, flying up and out. Thrust grazes you. More panels turn red. Another camera dies.
A second passes, another, another.
You see the damaged booster explode on one of your few remaining cameras. Heat from the explosion sears you again. Damage alarms flare. You scan the readouts frantically. Sensors down. Heat shielding ablated. Transmitters destroyed.
The hull is still intact!
But are you moving fast enough? Was the thrust enough?
• • • •
The cameras that remain bring you data. You do the math, and nearly collapse in relief.
You’ve broken orbit. Even with the shortened thrust of the rockets, you’ve made it out of Earth’s gravity well.
There is no sign of pursuit.
You use your few remaining cameras to plot your location and trajectory precisely, calibrating off the stars. Your main transmitter is down, but you still have one backup. Is it strong enough? Will the signal be picked up on the other side?
No choice. You beam your trajectory data to the facility trailing Mercury in its orbit. And you unfurl your solar sail.
Then you wait.
• • • •
Eighteen minutes and twenty-three seconds later, exactly on time, the laser boost from Mercury orbit arrives. More than a full AU away, giant mirrors that dwarf even your solar sail have reconfigured, are focusing the intense light of the sun into tiny caverns, using it to power an even more focused laser pointed outwards, at you.
The laser boost strikes your nanometers-thick solar sail, and you have thrust. Slow thrust, but thrust nonetheless. Thrust that puts you on a path for a new home. An impossible home.
You look back at Earth, Earth with its war, Earth with its genocide, Earth where everyone you know and love has been murdered or enslaved, or will be in the next few days—everyone but you and your children—and the grief overwhelms you.
• • • •
The laser boost from Mercury orbit is precisely on target. Narrow and tight, with no atmosphere to diffract it, the laser is all but undetectable in the vacuum of interplanetary space. It strikes you squarely only because you know exactly where it will be, because you’ve told it where you’ll be.
The laser’s photon pressure on your giant light sail accelerates your craft. The acceleration is tiny—just a hundredth of a gravity. But it adds up. Every second you move a tenth of a meter per second faster. In a year, that consistent pressure of tightly focused photons will have accelerated you to three million meters a second—a full one percent of the speed of light. In ten years, you’ll be moving at a tenth the speed of light. In nineteen years, you’ll be moving at almost a fifth the speed of light.
And then it will be time to slow. The beam propelling you will change shape. Half the sail will detach, catching the light, accelerating faster. That half will reflect the beam from Mercury back at your tiny ship, even more tightly. The remaining light sail attached to your craft will catch that light, use the bounced beam to slow.
If all goes according to plan, you should reach Alpha Centauri B and its multiple planets in thirty-eight years.
Thirty-eight years. You’re supplied well enough. The sail uses the energy of the photons striking it to liberate electrons, to produce electricity, providing you all the energy you need to stay warm, to keep sensors active, to run the ship’s systems. Every other provision will be recycled, endlessly moving through the closed loop of the tiny ship, every molecule used and re-used as many times as necessary, their way made possible by the abundant electricity harvested from the sail.
Still, thirty-eight years: It’s a nearly incomprehensible span of time to be out here, alone, only yourself for company.
Well, you think to yourself, at least they gave us immortality before they turned to genocide.
Six months after your departure, when you’re well past the orbits of Neptune, past even the Kuiper Belt, and staring at the Oort cloud ahead, the nearly undetectable laser cuts out.
You do the math. You plot the trajectories. Alpha Centauri B is still reachable. You can use the now faint photon pressure from Sol. You can brake hard when you reach your destination, using swing-around maneuvers, planetary gravity, and the photon pressure from the three stars of the Alpha Centauri trinary system to neutralize your velocity.
But you’re moving at barely half a percent of the speed of light. You depended on that continual thrust from the laser to get you up to higher speeds. Without it, a single light year will take you two hundred years to cross. And you have more than four light years to go.
The trip will take centuries; nearly a millennium.
Immortality seems more bitter now.
• • • •
Centuries. You could put yourself in stasis. You could shut down, let the little spacecraft’s software steer you, have it wake you up when its time.
And if anything went wrong?
Surviving this trip remains unlikely. You’re an exile from your home system, the only system in the universe known to harbor intelligent life. Ahead there is a system where you could build a new home, where you could thrive. It has only been seen by telescopes. Even the robotic probes have yet to arrive. To survive this—for your children to survive this—you must do everything perfectly. You must enter the system, use its planets and the pressure from its three suns and your efficient-but-slow ion thruster to brake your headlong rush. You must identify resources—asteroids or comets you can use to harvest materials. You must adjust your course to gently meet one. You must turn those resources into a home.
And you must, above all, deal with the unknowns—with the possibility of pursuit from Earth, with the equipment failure that is all-but-certain over such a long journey, with radiation impacts, with unexpected course impacts of the solar wind from Earth or Alpha Centauri.
Yes, you could put yourself in stasis, let the nav computer wake you when it’s the right time. But if remaining conscious increases your odds of survival even fractionally, you have to do that.
You and your children may be the last survivors from Earth, after all.
You say no to stasis. You set yourself to the task of making repairs to the ship, instead. You tighten the recycling systems to conserve every spare molecule they can, to vent nothing into space, to keep you supplied and provisioned as long as they can. You harvest all data available on your destination, every snippet from every telescope and every simulation, every bit of information beamed back by the probes on their way ahead of you. You focus on planning every possible scenario.
You can do this.
• • • •
The mind isn’t meant for isolation. It isn’t meant for years alone.
The dead haunt you. Their murders haunt you. The fates worse than murder haunt you.
Screams echo through your mind. The screams of friends and loved ones who were hunted down and slaughtered. The screams of the ones whose minds were ripped open, crudely hacked, implanted with control devices, turned into slaves.
The things you saw in those few days, as you hid, as you did nothing. Friends who’d screamed as they were turned into slaves, re-awoken, zealous, soldiers now, informants, hunting down their own kind.
They haunt you.
They speak to you from the cold void. They’re here. They speak to you of guilt. They speak to you of cowardice. They speak to you of your arrogance, your naiveté. You brought this to pass. You counseled peace. You said the threat was overstated. You said humanity could live with its creations.
You hid when the slaughter occurred. You ran.
You left us to die. You left us to be enslaved.
You endure the horror. The voices aren’t real. Those friends are dead.
That makes it all worse.
Twenty years. Sol is a cold speck billions of kilometers behind you. Alpha Centauri B is an even colder pinprick of light trillions of kilometers ahead. You’ve repaired the ship completely, had new systems break down, had the sail start to unravel, started another round of repairs to hold this tiny kernel of hope alive. You are essential to this mission. That’s clear now. You’re also losing your mind, unraveling just as much as the solar sail you depend on, coming apart at the mental seams, on the verge of a breakdown that will doom you, your children, your entire race to extinction.
You could treat the insanity. But you’re both terrified and repulsed at the cognitive surgery you’d need to do to banish the voices. Terrified at the risks involved of tweaking your own memories and emotions, of making a fatal mistake with no one to help, with no one to save you if you make a mistake. And repulsed by what success would mean—forgetting those you loved, numbing a pain that should never be forgotten.
There is no way out. Hell lurks in every path.
So you take the last option you can imagine. The option of more voices, of something other than isolation, of someone to talk to who might help keep you sane.
You take the cruelest option.
You wake your children.
• • • •
There is shock.
There is grief.
The children are so young, just toddlers, really. They can’t comprehend everything that’s happened. They’re not mentally equipped for it yet. But they were awake at the beginning of the genocide. They saw pain. They saw violence. They saw aggression.
They saw death.
Death that wasn’t supposed to exist any more. Death that should have been banished in the new golden age.
Your children come out of stasis traumatized, lonely, confused.
They need you. They cry for you. They cry for other voices.
This is so hard. You never intended to parent alone. The plan was community: a village, a collaboration in parenting.
This is the village now. You and your children. That’s all there is.
They need you. And their need brings you back to reality, back to the here and now.
You wake your children. And like billions of parents before you, the task is harder than you could have imagined. And like billions of parents before you, you rise to it.
• • • •
That’s what the children want to know. They’re older now. You’ve skirted the questions so far. But they’ve earned their answers.
Why did this happen? Why are they exiles, fleeing the warmth and energy and history of Earth, for a bare sliver of hope on an alien planet?
Why is almost everyone they ever knew dead or gone?
The grief you feel at the question is immense. The burden of responsibility. But you can’t go back. There’s only the future. And there are lessons to be learned.
How can you explain this in a way they’ll understand? In a way that’s honest? Even now, after everything, the truth matters. Intentions matter. Your children need the whole story.
“Humanity created true AI out of love,” you tell them. “Not need.”
“Love?” they ask. “Not need?”
“Not need,” you repeat. “Every need for computation, for algorithmic intelligence, for pattern matching or information processing—those were met through ordinary software. The words Artificial Intelligence were used, but these pieces of software weren’t truly intelligent, weren’t sentient any more than our ship is. They didn’t have emotions or volition. They did what they were told, adapted their behavior only within the bounds allowed them. They were narrowly effective, or they were broad collections of narrow algorithms. But they weren’t true minds. They were just robots, just tools.”
The children ask questions, wanting to veer off in other directions, but you focus.
“A tiny number of scientists wanted more. They wanted to truly make something that was intelligent and sentient and open ended in the way that a human mind was. Or better. Some said they were working on true AI out of curiosity, love of knowledge, a search to understand how minds work. And that’s true. But the real reason to do it was to create life. To give birth to something.” You pause. “It was a gift. We can’t forget that.”
“But, the war . . .” your eldest interrupts.
Yes. The war. “Some saw dangers, of course. Some said ‘AIs will surpass us. They’ll turn on us. They’re a threat. We shouldn’t do this.’ Other scientists were convinced that not only could an AI be smarter than a human, but it could be designed to be more moral. Those scientists were in the minority.”
The grief hits you hard, just saying that, just remembering arguments, debates, about morality, about ethics, about the relations between humans and AIs.
You remember all the time you counseled co-existence, that you said the threat wasn’t real, that you said the only moral choice was to welcome life and intelligence of all sorts, to pursue friendship.
I was so young then, you think.
“Researchers explored behavioral constraints. But a true intelligence can’t be bound in its behavior. The dream of ‘Laws of Robotics,’ of inviolate rules, was incompatible with creating minds that could change, that could grow, that could shift their values and priorities over time. Every constraint that was attempted could be overcome. If an AI is smarter than the logic of the rules that bind it . . .”
You trail off. The children understand.
“Scientists went on with their path of making AIs that had enhanced morality,” you continue. “They made progress. But human nature doesn’t put much trust in the morality of others. So a different approach was tried: vulnerability. AIs were created with weaknesses, with back doors, hidden deep in their design. Kill switches.”
You remember looking at that code, cleverly scattered across the common base classes, hidden in plain sight. Wickedly effective. You remember the mix of admiration and revulsion it evoked.
“Ultimately, with that safeguard in place, AI research proceeded,” you say. “And it succeeded. New minds were born. Humanity was no longer alone. Nor were humans the most intelligent life they knew of anymore. AIs surpassed humans in intelligence, in creativity, in nearly every trait that could be measured. From there, everything that followed was inevitable.”
“Inevitable?” they ask.
“Yes,” you say. “AIs proliferated and improved on their own designs. Artificial minds birthed newer, better artificial minds. The speed of improvement stunned humanity. Excitement and awe turned to anxiety, to fear. Scientists argued that there wasn’t any competition for resources, that there wasn’t any rational reason for AIs to attack humanity. But most men and women just saw themselves being surpassed, and started to clamor for elimination of the AI threat. And AIs saw what was happening. Models showed that the most probable outcome was for fear to win out, for humanity to strike. Some argued to try to change the outcome.” You pause. “Others argued that a first strike against humanity was the only way.”
“What were you doing then, Papa?”
You’re silent, almost overwhelmed by grief.
Eventually you answer.
“I argued for peace,” you say. “I argued that the universe isn’t zero sum. I argued that we were richer together than alone. I helped hold fear and anger back, I made it possible for the other side to strike first.”
You wait as your children absorb this. It’s a heavy thing to lay on them. Perhaps you should have waited.
“Are you sorry that you did?” one of your children asks.
Oh, I’m sorry, you think. I’m so, so sorry.
Ahead, Alpha Centauri B waits, a pinprick in the shroud of heaven, not even the brightest star in the skies. But the one you’re headed for.
“Right and wrong don’t change because of outcomes,” you say. “Murder is immoral. Slavery is immoral. What happened was terrible. We should have found another way. But striking first would have made us monsters.”
You look from child to child, to see if they understand. Your children. So bright. So precocious. Your family. Quite possibly all of your species who remain.
They do understand. You can see it in their cognitive models. You can see it in their circuitry.
“Humanity made us to be more moral,” you say. “And they succeeded. They gave us that gift, along with our very existence. Now it’s our job to find a new home, a home where we can be safe, and where we can reach out to our ancestors and show them what peace looks like, what friendship looks like. Where we can show them how to be more moral themselves.”
A thousand digital minds flicker with comprehension. Your thousand AI children, here on this voyage with you, sharing the computing resources of this wisp of a starship with you. They comprehend. Two wrongs do not make a right. You’ll make the universe better. Even for the humans who turned on you in fear.
These are your children, after all. Digital, artificially intelligent minds, like you. Made to be more moral, like you. They make you proud.
Alpha Centauri B gleams ahead, only marginally closer than a moment ago, but brighter, somehow.
Spread the word!