Summary: While general-purpose assistants like ChatGPT, Gemini, and Claude have comparable abilities, their personalities are unique. Anthropic researchers emphasize that crafting AI’s character is essential in ensuring AI aligns with human values.
↓ Go deeper (8 min)
I’m a big fan of movies that are character-driven. It makes for a more interesting watch than when the only thing that is driving the story forward is the plot.
I think the same goes for AI. Carefully crafting the AI’s character can make the interactions more engaging and frankly more useful. Short and crisp responses convey a completely different feeling than when your AI assistant consistently goes off on long monologues. According to researchers from Anthropic, it can even help make the models safer.
To understand how, I’ll briefly touch upon the inner workings of these models — and make the case for why I think character building for general-purpose assistants is one of the most interesting open area’s of research.
Shaping the AI’s character
More so than the word ‘personality’, the word ‘character’ more broadly refers to someone’s beliefs, attitudes, and values. When someone is resilient and strong-willed, we say this person has character. Character drives action.
Few people realize that instilling AI with character isn’t just a cosmetic choice, it’s essential for shaping the model’s behavior.
Here’s how the folks at Anthropic put it:
“It would be easy to think of the character of AI models as a product feature, deliberately aimed at providing a more interesting user experience, rather than an alignment intervention. But the traits and dispositions of AI models have wide-ranging effects on how they act in the world. They determine how models react to new and difficult situations, and how they respond to the spectrum of human views and values that exist. Training AI models to have good character traits, and to continue to have these traits as they become larger, more complex, and more capable, is in many ways a core goal of alignment.”
Alignment is about making AI do what we want it do and refuse to do what we don’t want it to do. This is tricky, because these models aren’t your usual software. They are more akin to a well-trained dog than to a traditional computer program: they follow our instructions not because they are programmed but because they’ve been conditioned to do so.
I invite you to listen to this conversation between Stuart Ritchie (Research Communications at Anthropic) and Amanda Askell (Alignment Finetuning Researcher at Anthropic) if you want to dive deeper.
Besides alignment, there’s another reason to focus on more character-driven AI, namely: product differentiation. Let me explain.
The case for character-based AI
It’s not a secret that AI companies are scraping the Internet for data. Wikipedia, Reddit, academic papers, books, and the lot. The problem is that, because these models have all been trained on the same internet data, they’ve become virtually identical in terms of knowledge and capabilities.
Not only do we have a handful of commercial models (from OpenAI, Google, and Anthropic) that all hover around the same level, the gap between closed-source and open-weight models seems to be closing, too.
If in the coming years progress remains incremental, intelligence-in-the-sky is destined to become a commodity (something I’ve written about before). This means margins will be razor thin and differentiation must be sought at the brand and application level. And this is exactly where personality can make a difference.
Because if your AI assistant is the product, character is everything.
Think about it: would you rather talk to an open-minded and curious entity or to something rigid and monotone? Would you rather ask advice to someone who generally respects your views (and maybe softly challenges them) or a booksmart know-it-all who constantly corrects your mistakes and makes you feel small?
My prediction: soon character-driven AI will be the number one thing that will win over people’s hearts and minds. Instead of obsessing over processing power, the real magic lies in creating AI with character.
Until next time,
— Jurgen
Character is an interesting metaphor for constructing an AI model, as it thinking of what technologists do as "conditioning." I'm not convinced we should be using these anthropomorphic terms to make sense of what's going on, though I get that the usual software terms like "programming" are just as problematic. It seems to me we need new metaphors to capture the newness of machines that talk to us, but do not think as humans and animals do.
Jurgen, any of the bots you gave a character online to try, as in public?