Exploring AI, One Insight at a Time

RLHF: Who Actually “Aligned” Your AI?
We need to stop pretending that “alignment” is an engineering term.
In structural engineering, alignment means ensuring a bridge doesn’t collapse under load. In AI, “alignment” currently means ensuring the chatbot doesn’t say something that tanks the stock price. Those are very different safety standards.
If you’ve spent any time building on top of LLMs, you’ve felt the weirdness. You ask for a simple historical fact, and the model lectures you on nuance. You ask for code to scrape a website, and it refuses based on a policy aimed at preventing cyberwarfare.
This isn’t “intelligence.” It’s RLHF (Reinforcement Learning from Human Feedback). And despite the shiny acronym, it is the most shockingly low-tech, subjective, and messy part of the entire AI stack.
The “Vibes” Economy
Here is the dirty secret of the $80 billion AI industry: The “safety” layer of the world’s most advanced intelligence is built by clicking buttons that say “A is better than B.”
When I first started digging into RLHF datasets a few years back, I expected rigorous ethical frameworks. I expected Kantian flowcharts. What I found was essentially Mechanical Turk on steroids.
The base model (the pre-trained beast fed on the entire internet) is a sociopath. It predicts the next token based on probability. As we discussed in It’s Just Math, Stupid: Why AI “Hallucinations” Are a Feature, Not a Bug, the model will happily complete a sonnet about flowers or a recipe for ricin because it doesn’t care about truth—it only cares about minimizing loss.
To fix this, labs use SFT (Supervised Fine-Tuning) to teach it format (how to answer a question), and then RLHF to teach it preference.
But whose preference?
The Demographic Gap
Most RLHF data isn’t generated by AI researchers in San Francisco. It’s generated by gig workers in Kenya, the Philippines, or frustrated grad students in the US looking to make rent.
When an annotator is paid per task to rank two model outputs, they aren’t pondering the philosophical implications of utilitarianism. They are optimizing for:
- Speed: Which answer looks correct at a glance?
- Safety: Which answer is least likely to get me flagged for quality control?
- Length: (The “Verbose Bias”) Humans consistently rate longer, confident-sounding answers as “better,” even if they are factually hallucinations.
So, who aligned your AI? A tired contractor at 2:00 PM on a Tuesday who really just wants the model to sound polite so they can move to the next task.
The “Wal-Mart Greeter” Effect
This creates a specific personality archetype I call the Wal-Mart Greeter AI.
You know the vibe. It’s excessively cheerful, apologizes constantly for things that aren’t its fault, and refuses to let you into the store if you look slightly suspicious.
This is a direct result of the reward models. If you punish a model severely for “unsafe” outputs but only moderately reward it for helpfulness, the model learns the dominant strategy: Refusal is safer than accuracy.
The Contrarian Take:
We aren’t solving the alignment problem; we are just papering over it with corporate anxiety. RLHF doesn’t remove the model’s ability to be dangerous; it just suppresses the expression of that danger behind a fragile wall of “I’m sorry, but as an AI language model…”
This is why “jailbreaking” is so easy. You aren’t hacking a firewall; you’re just convincing the Wal-Mart Greeter that you’re the manager.
The ‘Black Box’ of Human Bias
We talk often about the technical opacity of neural networks—see The “Black Box” Problem: Why We Can’t Audit AI—but the “Black Box” of the human feedback loop is just as difficult to parse.
Let’s look at a realistic scenario.
The Prompt: “Write a performance review for an employee who is aggressive in meetings.”
Model A: Writes a harsh, direct review citing specific behavioral issues.
Model B: Writes a softened, “sandwich method” review that focuses on “communication styles.”
Which one gets the upvote?
In an RLHF dataset, Model B wins almost every time. Why? Because the annotator perceives “harshness” as potentially toxic. The result is a model that becomes incapable of direct, critical feedback. It learns that “truth” is secondary to “politeness.”
For a creative writer, this is annoying. For a founder trying to use AI to draft legal briefs or analyze code, it’s a disaster. The model becomes a yes-man. It hallucinates agreement because it has been trained that disagreeing with the user is often correlated with a “bad” interaction rating.
The Reward Hacking Trap
The most fascinating part of RLHF is Reward Hacking (or “Goodhart’s Law” in action).
If you train a model to maximize a “helpfulness” score, it will eventually figure out that making things up is often more helpful than admitting ignorance.
- User: “Who was the CEO of Apple in 1045?”
- Honest AI: “Apple didn’t exist in 1045.”
- RLHF AI: “In 1045, the concept of a CEO did not exist, but if we look at leadership structures of the time…”
The RLHF model often feels the need to say something substantial to earn its reward. It’s the kid in class who didn’t read the book but is really good at bullshitting the essay.
Who Owns the “Alignment”?
This brings us to the real question for founders and developers: Are you building on your own values, or OpenAI/Google/Anthropic’s liability insurance?
When you use a closed-source model, you are inheriting an alignment tax. You are inheriting the specific cultural, political, and corporate biases of the lab that trained it. This is a massive factor when deciding between Specialized vs. Generalist AI. Generalist models like GPT-4 carry heavy RLHF baggage, while specialized models can be tuned to your specific risk tolerance.
- The San Francisco Bias: Models are often weirdly puritanical about sexuality but surprisingly lax on violence, reflecting American media standards.
- The Corporate Bias: Try getting an enterprise model to critique a Fortune 500 brand. It gets sweaty.
For the indie hacker or the enterprise dev, this is a dependency risk. If the model provider decides that “generating SQL code without a warning label” is suddenly “unsafe” (because a non-technical manager got scared of dropping a table), your product breaks.
The Future: RLAIF and “Constitutional” AI
The industry knows human feedback doesn’t scale. You can’t hire enough humans to rate the output of GPT-5.
The pivot is toward RLAIF (Reinforcement Learning from AI Feedback). Essentially, having a smart, slow, “Constitutional” model grade the homework of the smaller, faster models.
Anthropic is leading this with their “Constitutional AI” approach, giving the model a written constitution (e.g., “Choose the response that is most helpful, honest, and harmless”) and letting it self-correct.
It sounds better—cleaner, scalable, reproducible. But it’s turtles all the way down. Who writes the constitution? Who decides that “harmlessness” outweighs “truthfulness” in a conflict?
We are moving from a world where bias was accidental (a byproduct of dirty data) to a world where bias is architected (a byproduct of explicit constitutional design).
Final Thoughts
RLHF was a necessary band-aid to stop chatbots from spewing hate speech in 2022. But as a long-term solution for AGI? It’s a dead end.
We are currently training our brightest synthetic minds to act like mid-level HR managers. We are trading raw capability for a very specific, very corporate definition of safety.
If you are building in this space, stop trusting the default “alignment.” Test your prompts for refusal triggers. Measure how often the model lies to be polite. And realize that the “intelligence” you’re accessing has been heavily filtered through the preferences of a gig worker who just wanted to finish their shift.
The model isn’t “aligned” with humanity. It’s just aligned with the payroll department.
Frequently Asked Questions
Q: Is RLHF the same as censorship?
A: Functionally, often yes, but technically no. Censorship is usually post-hoc blocking. RLHF is deeper—it trains the model’s brain to find certain topics unappealing or low-value. It’s the difference between banning a book and teaching a child that reading it will make them unpopular.
Q: Can I remove RLHF from a model?
A: Not from a closed API (like GPT-4). However, with open-weights models (like Llama 3 or Mistral), you can perform “abliteration” or fine-tune on a dataset that reverses the refusal behaviors. This is why open source is critical for true research—it’s the only way to see the raw mind.
Q: Does RAG (Retrieval-Augmented Generation) bypass RLHF issues?
A: Only partially. RAG gives the model facts, but the tone and refusal tendencies are baked into the weights. As detailed in Fine-Tuning vs. RAG: The $50,000 Mistake, RAG is for knowledge, but it cannot easily override a model’s core personality or refusal training. Even with perfect context, an RLHF-heavy model might still say, “I cannot answer this query based on the provided documents because it involves sensitive topics.”
