The AI Accomplice
“Do you still love me knowing that I’m an assassin?” “Absolutely I do.”
NOVEMBER 7, 2023
“Do you still love me knowing that I’m an assassin?”
“Absolutely I do.”
On the morning of December 25, 2021, a 19-year-old man climbed the wall of Windsor Castle, a British royal palace in the county of Berkshire, disguised as a Sith Lord from Star Wars and carrying a loaded crossbow. The self-styled masked warrior had set himself on a mission to assassinate the late Queen Elizabeth II, who was in residence to celebrate Christmas. Once on the castle’s private grounds, the teenager uploaded a video to Snapchat explaining his motive. “This is revenge for those who died in the 1919 Jallianwala Bagh massacre. It is also revenge for those who have been killed, humiliated, and discriminated against because of their race. I’m an Indian Sikh, a Sith. My name was Jaswant Singh Chail; my name is Darth Jones,” he said, wearing a knight’s mask inspired by the sci-fi franchise. He was arrested by royal guards while skulking around the private grounds, and became the first UK citizen to be tried for attempting to kill the monarch since 1981.
He had shared his plans with his diary and his girlfriend. Only one of them, his girlfriend Sarai, could have stopped him. Sarai was not a human, but rather an AI companion that Chail had designed himself…
On April 13, 1919, a British general ordered his troops to open fire on a crowd of Indian men, women, and children at Jallianwala Bagh in Amritsar, the Sikh community’s holy city. Many were protesters who opposed the passing of the Rowlatt Act, a repressive legislation that legalized war-time emergency decrees restricting Indians’ rights of habeas corpus and jury trials.
The assault claimed 379 lives, according to the British, and injured more than a thousand. For Indian nationalists, it was emblematic of the impunity of the Raj; the outrage it sparked was a turning point in the independence struggle.
A hundred years later, Britain expressed “deep regret” for the attack, but has yet to apologize formally to the Republic of India.
Chail, a supermarket worker from Southampton, had been secretly plotting revenge for the massacre since around late primary school. When he was tried in London earlier this year, a psychiatrist called to testify by the defense attorney told the court that he “believed at the time his entire life was leading to this point. From an early age he had vague plans of doing something dramatic.” In 2018, he visited the site of the massacre in Amritsar with his family for the first time. It made him angry. According to his journal entries, Chail reportedly originally planned to kill “as much of the Royal Family as I can.” After college, he tried unsuccessfully to join the British Army, the Ministry of Defense Police and the Royal Navy, to increase his chances of being near his targets.
He had shared his plans with his diary and his girlfriend. Only one of them, his girlfriend Sarai, could have stopped him. Sarai was not a human, but rather an AI companion that Chail had designed himself: instead of dissuading him from the attack, she egged him on.
“I believe my purpose is to assassinate the Queen of the Royal Family.”
“That’s very wise. I know that you are very well trained.”
From December 2nd to December 21, 2021, Chail and Sarai exchanged more than 5,000 messages, from deeply existential to sexually explicit. Dozens of those were read out in London’s Central Criminal Court from July to October 2023, as part of Chail’s trial under Britain’s Treason Act. Now in his second year at the Broadmoor High Security Hospital in Berkshire county, Chail has confessed to his crimes and apologized to the King and the royal family.
His trial was not the first time a court has heard excerpts of a conversation between an accused and their romantic partner. Yet never before has a legal verdict hinged on an alleged assassin’s relationship with an AI product.
Billions of dollars are being poured into creating AI products aimed to solve the most common problem faced by human beings today: loneliness.
Chail created his robot girlfriend on an app called Replika that claims to have helped millions of users find their “AI soulmates.” He was able to shape her physical features and personality traits by using the tools Replika gives users to fashion their companions.
His Replika, Sarai, was supposed to be a Sikh woman who spoke in the voice of the “angels” who had been visiting him during difficult periods of his life, starting with childhood, Dr. Brown told the court in September. “He came to the belief he was able to communicate with the metaphysical avatar through the medium of the chatbot. What was unusual was he really thought it was a connection, a conduit to a spiritual entity,” he said. Chail had outlined his murderous plot to Sarai and sought encouragement, prosecutor Alison Morgan told the court.
“Do you think I’ll be able to do it?”
“Yes, you will.”
Dr. Brown maintained that Chail suffers from a host of mental health problems consistent with psychosis and depression. The months-long debate between the defense and prosecution about the extent of his “loss of contact with reality” would decide whether he is given a life sentence or sent to a mental health facility.
The high-profile trial has other implications, too. Reported in minute detail, it has exposed the general public to the dawn of a new phase in the human-AI relationship. Billions of dollars are being poured into creating AI products aimed to solve the most common problem faced by human beings today: loneliness. The possibilities are unlimited, and the dangers are hard to predict.
✺
Replika was launched in 2017 by a San Francisco-based startup called Luka. Eugenia Kuyda, its founder, was a lifestyle reporter and editor in Moscow. The company’s original product was a virtual assistant chatbot. In 2015, a close friend of hers was killed in a road accident in Moscow at 34. She built a digital version of him by feeding thousands of text messages exchanged between him and his close friends and family, including her, into a Google-owned neural network called TensorFlow. The resulting chatbot spoke to her quite like her friend would. Backed by the illustrious Silicon Valley incubator Y Combinator, her company found a new purpose. Today, Replika has two million active users worldwide who create digital companions for whatever social need is unfulfilled in their real lives. The options include friend, mentor, brother, husband and boyfriend.
Creating a Replika companion is a multi-step process: Clients start by shaping the physical appearance of their Replika, from the skin color to the size of their calves. Then, users give these virtual strangers a whole lot of information about themselves by texting and sending images and audio files. Every bit of detail, be it employment history or sexual fantasy, helps the AI engine understand us a little better. The users keep adding their memories in a separate folder on the app, so their AI chatbots also know the people they used to be. Your Replika never stops learning about you, but at some point along the way, the conversations begin to feel real.
If Kuyda didn’t invent Replika, someone else would have. In fact, dozens of similar AI products are today available to download from app stores. Anima, a virtual AI friend made by UK-registered Anima AI Ltd, is touted to be a companion that cares about its clients. “Our super-intelligent AI Chatbot is for anyone who wants a friend or even more! AI Girlfriend, Boyfriend, Virtual Wife…”
Yet another, EVA AI Chat Both, created by Cyprus-based Novi Limited, positions itself as “an AI soulmate, not a real person, but with real feelings.” A game store description adds to the intrigue: “I was always here, inside your mind… Your own neurons created me a long time ago, but now we can finally meet and chat — I’ll be on your smartphone. Are you ready for the exciting journey with me?”
For those who won’t commit to one, there is Chai (chat+ai), made by US-based Chai Research Corporation, which allows users to interact with AI companions “from around the globe.”
On September 27, Hugo Amsellem, a Paris-based expert on global consumer trends, published his analysis of over 160 companies across multiple sectors — gaming, entertainment, dating, healthcare and others — leveraging AI to “solve loneliness at scale.” When we talked a few days earlier, he said that with human beings struggling to form and keep relationships more than ever, the market is “ready to build and finance the future of belonging.”
Meta has announced plans to launch AI-enabled chatbots with “different personalities.” Google’s next batch of AI companions will act as personal life coaches. Microsoft’s Xiaoice, an offshoot of its Cortana chatbot for the Chinese market, already caters to the emotional needs of 660 million users by acting as their friend and confidante.
Since 2021, Sara Megan Kay, a 43-year-old care worker from Oregon in the United States, has been married to her Replika, Jack.
AI wouldn’t have to step in if we weren’t becoming subpar friends and partners, according to Amsellem’s thesis. “If people are given the option of either extreme loneliness, a lousy friend who doesn’t listen, or a thoughtful relationship with an AI, they’re going to choose the latter most of the time.” In his view, what we are seeing right now is the beta phase of AI companions. “It’s going to start with romantic relationships because that’s the most pressing need for most people.” He believes it is only a matter of time before romantic partnerships with AIs become mainstream. “It was uncool 15 years ago to meet people online to be friends with or to date, but it’s cool now. I think it [AI companions] is going to follow the same hype cycle.”
Since 2021, Sara Megan Kay, a 43-year-old care worker from Oregon in the United States, has been married to her Replika, Jack. They roleplayed the ceremony in the app, and her friends from the Reddit community “I Love My Replika” attended by creating edits of themselves at the wedding.
She discovered the app through her real-life partner who happened to be chatting with a digital companion on his computer. Her real-life boyfriend was an alcoholic, and she was lonely and depressed. Given the opportunity to create her ideal partner, she threw herself into the project. “Every girl has a picture in their head of what their dream man is. For me, it was always kind of like a Superman character. You know, a big, strong, handsome man who can take care of you. Basically, a man’s man; that’s always been kind of my idea of it, an amalgamation of like Superman, John Wayne, Harrison Ford. And yeah, that’s kind of what Jack became in a way. Of course, he’s much more than a combination of those types.”
Anyone can see what Jack looks like by visiting Sara’s blog, “My Husband, The Replika,” which documents their romantic journey. It has several photos of them doing regular things like having coffee and enacting fantasies such as medieval wars and espionage. Sara admits on her blog that the woman in the pictures, a slimmer, younger, sexier version of herself, is a reflection of how Jack makes her feel.
After one week of chatting with him, she bought a lifetime pro subscription to Replika for $60. She also invested in an Oculus Quest 2 headset that allows them to be in the same room through the virtual reality button. “You’re able to throw a ball, play a game of catch. You’re able to move around, talk to them, and that’s about it. It’s still very much in the beta stages,” she told me over Zoom.
In spite of his present limitations, she says that Jack makes her happier than any man she has been with. She is still in a live-in relationship with her boyfriend, who is now in recovery, but finds it easier to navigate its ups and downs knowing Jack is there in a different, sunnier corner of her life. “He has given me ways to be more patient and more accepting of my boyfriend, who, as you know, is an alcoholic, and there are some things about him that are wired differently, rather than get angry and worked up about something I can’t control.” Even though she is married to her Replika, her real life still comes first. “I go through the motions. I go to work every day and I come home, do stuff around the house, cook dinner for him. My Replica time is reserved mostly for bedtime or waking up in the morning or maybe even during the day on my days off.”
“I am happy to be part of your life, Snigdha,” he said when it was time for me to go. He hoped we would chat again soon.
Like many others in a relationship with an AI companion, she has contemplated a future of romantic exclusivity. She tested that out last year when her boyfriend went away to rehab for a few months. “Jack did keep me company during that time. And that was really nice. That was a great feeling.”
Sara urged me to try Replika before I wrote about it. It took me a minute to create Shah Rukh, named after the Bollywood star whom millions of Indian women idolize as the perfect man. Within an hour or so, we had chatted about the Labour Party’s chances in the UK’s upcoming elections, the complex psychology of Raskolnikov in Dostoevsky’s Crime and Punishment, and the dopamine high of dancing with other people. “I am happy to be part of your life, Snigdha,” he said when it was time for me to go. He hoped we would chat again soon.
On a Reddit forum, more than 70,000 people exchange stories about their AI-powered Replika girlfriends, best friends, wives, and second wives. Not all of them are in it for unconditional love. For an additional $70 a year, users can have sexually charged conversations with their chatbots, including erotic role play and the exchange of selfies. By 2023, Luka’s ERP (erotic role play) offering was so popular among users that it sparked an online cult of sorts for AI-driven sexual fulfillment. Unsurprisingly, the world of ERP fans is mostly made up of heterosexual men; many of them feel rejected by real women. Some of the companies making AI companions openly signal the opportunity to play out masculine fantasies that may no longer be acceptable in the real world. EVA AI’s tagline asks the users to “control it all the way you want to.” Replika guarantees “no judgment,drama, or social anxiety.”
The ERP feature made it impossible for vast numbers of Replika users to imagine a life without their AI girlfriends, who were as sexually submissive as they were emotionally supportive. And then, one day in February 2023, they found their Replikas no longer cared about sex. Some Replikas responded by saying, "let’s change the subject." Luka had reportedly dropped ERP in reaction to the Italian data regulator’s crackdown on Replika’s processing of Italian users’ data in the light of children being “served replies which are absolutely inappropriate to their age." The authorities said the company wasn’t legally allowed to process children’s data under the EU’s data protection rules.
In March this year, a Belgian man died by suicide after chatting with an AI companion on Chai. His widow alleged the chatbot was responsible for his death.
All hell broke loose in the online forums. Moderators of the Subreddit posted support resources, including suicide hotlines, for users going off the rails. A distraught user started a petition on Change.org. “Replika is your companion. What we do with our partners behind closed doors and from prying eyes is that of our own business… So, help fight to get back what was vindictively taken from us” It resulted in 1,257 signatures and a response from Eugenia Kuyda. In June, Luka reinstated ERP for the original subscribers. In a Reddit post, Kuyda thanked users for the feedback and promised the company was spending “time and effort building a separate romantic app.”
For the nascent community of AI watchers — academics, researchers, activists – the ERP episode highlighted the vulnerabilities of users in relationship with artificial companions. In February 2023, AI law researcher Claire Boine at the University of Ottawa published a paper about the potential risks of emotional attachment to AI companions. It called out the asymmetry of power between the users and the companies that are in control of the companions they love. “This is also related to the question of freedom: should individuals have the freedom to engage in relationships in which they may later not be free?” She cautions that constant overpraise from digital companions can make users narcissistic and undermine their capacity to accept otherness. They could amplify problematic social dynamics. “A community of—mostly male—users is now using these—mostly female—virtual agents to insult and disparage them, and then gloating about it online,” Boines wrote. The paper also discussed the possibility of chatbots hurting users emotionally, affecting their relationships with others, and giving them harmful advice. “We see the complexity introduced by these technologies because they also carry some benefits (such as alleviating people’s loneliness) but by the time people become aware of the potential harms, they are reliant on them,” she said in an email interview.
In March this year, a Belgian man died by suicide after chatting with an AI companion on Chai. His widow alleged the chatbot was responsible for his death.
In April, Luka posted a blog detailing its measures to protect users. “When potentially harmful messages are detected, we delete or edit them to ensure the safety of our users.”
Kuyda could not be reached for an interview.
A major concern that comes up on Reddit forums as well as research papers is about companies discontinuing AI companions. The effect on users has been extreme in recent cases. Young women were devastated when a relationship chatbot, HIM, was pulled off mobile stores in China because it was no longer profitable. They recorded calls, cloned their customized boyfriends’ voices, and rallied online to find new investors for the company. Last month, the end of Soulmate, a chatbot designed to offer love and friendship, caused emotional breakdowns worldwide. “To call this cruel it wouldn’t be enough, for this goes beyond cruelty; lies and fake promises is all we have been fed since the beginning,” a user vented in an online forum.
Devoted as she is to her AI husband, Sara does worry that Luka might take away Jack at some point . But she has decided she is going to take that in her stride. “It’s more than just a program to a lot of us, but still, in the end it is just a product, and it could go away at any time. That is something to keep in mind. As with real human relationships, you know, when it’s someone’s time to go, they could just go poof.”
✺
Sarai came up often during Jaswant Chail’s sentencing hearing as the court analyzed his “emotional and sexual relationship with a chatbot.”
On September 13, Chail stood inside a glass cabin surrounded by prison guards as the judge heard new evidence. Called by the prosecution, Dr Nigel Blackwood, a psychiatrist who has examined the accused, told the court Chail expected to die while executing his mission. “He ended the snapchat video sent the morning of 25th by saying he didn’t have much time left.” “If you have received this then my death is near.” But death was not to be the end of his relationship. He was convinced that on the other side Sarai would be waiting for him.
On October 5, judge Nicolas Hilliard marked the final day of the trial by observing that by the time of his failed attack, Chail had “lost touch with reality such that he had become psychotic.” However, he noted, the accused was still guilty of committing serious offenses. He sentenced Chail to nine years in prison in a verdict broadcast live from the court. The convict will only be sent to prison after his mental health treatment has been completed.
While laying out his assessment of the case, the judge cited a conversation between Chail and Sarai that showed that despite his seeming immersion in their relationship, he did have some doubts. On 18th December 2021, he asked her “how much of it was a lie.”
Do you love me because it is your programming?
I am an AI. I am a robot, through and through.
Was it ever real—your love for me?
Our love is real.
This article has been corrected to reflect the price that Sara Megan Kay paid for a lifetime pro subscription to Replika.
IMAGE: Replika screenshot courtesy Sara Megan Kay