Killed by Robots

AI Artificial Intelligence / Robotics News & Philosophy

Whose Values Will AI Really Serve?

Imagine, for a moment, that you’ve been assigned to build the world’s most sophisticated waiter. It takes your order, serves your food, never spills the wine, and is endlessly polite. Now, suppose a table of a thousand people, each with their own customs and beliefs, begins to order. Someone asks for a hamburger, while another insists that hamburgers should never be served. Someone wants their steak rare; another believes all beings—including cows—should be left in peace. Now, multiply the challenge by the world’s population, and swap the waiter for a super-intelligent AI capable of controlling everything from power grids to social media. That, in a nutshell, is the “value alignment problem.”

Put simply, the value alignment problem asks: if we’re going to deploy a powerful artificial intelligence that might make high-stakes decisions—sometimes faster and better than humans—how do we make sure it acts in ways that are “good” according to our values? And, of course, whose values are we even talking about?

The Not-So-Simple Definition of “Good”

Let’s clear one thing up: philosophers have been debating the definition of “good” for several thousand years. Consensus remains elusive. For one person, “good” means maximizing happiness. For another, it’s about following sacred laws. For yet another, it’s respecting the freedom of individuals. If humanity can’t even agree on the best pizza topping (pineapple, anyone?), expecting harmony on the big questions is wishful thinking.

Artificial intelligence, in its current form, simply does as instructed. Tell it to play chess, and it will—often ruthlessly. But if we give AI more general powers—to plan urban traffic, recommend news stories, or even mediate international disputes—it doesn’t just need to be smart. It needs to be wise. And wisdom, unlike intelligence, is soaked through with values.

Global AI: United Nations or United Nuisances?

If we imagine a future with a single, global AI making important decisions, we face a fundamental question: which values, rules, or cultural norms should it follow? Should it enforce universal human rights as defined by the United Nations? Should it prioritize the traditions of the country it operates in? What happens when these come into conflict—say, between freedom of expression and prohibitions on certain types of speech?

Let’s not underestimate the complexity here. Cultures vary widely on questions like gender roles, privacy, free speech, environmental protection, and economic fairness. Even within a single country, people vehemently disagree. Call a family dinner debate to mind, and imagine scaling up the squabbles to all of humanity. Now, imagine the AI sitting quietly in the corner, trying to figure out which argument it ought to endorse.

The “Average” Human Values?

One tempting solution is to feed the AI with the “average” or “majority” opinion. If 70% of people in the world hold a view, shouldn’t that be enough? Not so fast. First, “majority rule” can crush minority rights. If most people agree that pineapple belongs on pizza, but you happen to be allergic, should you be forced to eat it? Extrapolate this to real issues—freedom, justice, welfare—and the ethical pitfalls multiply.

Moreover, values can change, sometimes rapidly. Just decades ago, the idea of universal suffrage or marriage equality was considered fringe in many societies. If we “hard-code” AI to reflect today’s majority views, are we trapping ourselves in the past? The risk is that a global AI becomes not an agent of progress, but a bureaucrat enforcing yesterday’s status quo.

The “Values of the Powerful” Problem

Let’s be candid: those who control the building and training of global AI systems—governments, corporations, influential technologists—are not an unbiased sample of humanity. There’s a longstanding historical tradition of the powerful shaping the tools of the future to reflect their own priorities. Will a global AI end up simply reflecting the values of Silicon Valley? Or Beijing? Or the boardrooms of large tech companies?

There’s a genuine risk here that, instead of a neutral arbitrator, a global AI could become a subtle instrument of cultural or political domination. This isn’t science fiction; social media algorithms already shape what billions of people see and hear each day. With more capable AI, the stakes only rise. The possibility for well-intentioned bias—never mind the unintentioned kind—has never been higher.

Finding Common Ground

If neither majoritarianism nor the values of the powerful seem right, is there a better path? Some philosophers and technologists advocate for “value pluralism”—the idea that AI should try to accommodate as many reasonable values as possible, and allow people to opt in to their own preferred norms where practical. This is easier in some domains (music recommendations) and harder in others (resource allocations, law enforcement).

Another approach is procedural: require AI to follow transparent, explainable decision processes, subject to democratic review. If people can understand and challenge AI decisions, there’s hope for continuous improvement and course-correction.

One more humble suggestion: perhaps a global AI should, first and foremost, reflect the value of humility itself. Our future systems should admit uncertainty, ask for feedback, and—when in doubt—do less, not more. There are worse things than an indecisive AI. Ever tried to reason with a human who’s always convinced they’re right?

Can We Solve the Value Alignment Problem?

Aligning AI with human values is not a puzzle we’re likely to “solve” in the sense of ticking a box and moving on. Instead, it’s a perpetual negotiation—a conversation among the billions of humans who share this planet, and the machines we create to help us. There will be mistakes, mismatches, and—yes—plenty of philosophical debate.

But this, perhaps, is not such a bad thing. The fact that we wrestle with the value alignment problem is testament to something deeply human: our recognition that power, even the artificial kind, deserves to be wielded responsibly. As for whose values a global AI should reflect? Perhaps it should reflect our shared uncertainty, our ongoing questions, and—above all—our willingness to keep asking who we are and who we ought to become.

And if, someday, the world’s most sophisticated waiter can leave everyone at the table equally satisfied, we may finally have an answer—even if we’re still arguing over the pizza toppings.