I. The Founder
Sol Kennedy used to ask his assistant to read the messages his ex-wife sent him. After the couple separated in 2020, Kennedy says, he found their communications “tough.” An email, or a stream of them, would arrive—stuff about their two kids mixed with unrelated emotional wallops—and his day would be ruined trying to reply. Kennedy, a serial tech founder and investor in Silicon Valley, was in therapy at the time. But outside weekly sessions, he felt the need for real-time support.
After the couple’s divorce, their communications shifted to a platform called OurFamilyWizard, used by hundreds of thousands of parents in the United States and abroad to exchange messages, share calendars, track expenses. (OFW keeps a time-stamped, court-admissible record of everything.) Kennedy paid extra for an add-on called ToneMeter, which OFW touted at the time as “emotional spellcheck.” As you drafted a message, its software would conduct a basic sentiment analysis, flagging language that could be “concerning,” “aggressive,” “upsetting,” “demeaning,” and so on. But there was a problem, Kennedy says: His co-parent didn’t seem to be using her ToneMeter.
Kennedy, ever the early adopter, had been experimenting with ChatGPT to “cocreate” bedtime stories with his kids. Now he turned to it for advice on communications with his ex. He was wowed—and he wasn’t the first. Across Reddit and other internet forums, people with difficult exes, family members, and coworkers were posting with shock about the seemingly excellent guidance, and the precious emotional validation, a chatbot could provide. Here was a machine that could tell you, with no apparent agenda, that you were not the crazy one. Here was a counselor that would patiently hold your hand, 24 hours a day, as you waded through any amount of bullshit. “A scalable solution” to supplement therapy, as Kennedy puts it. Finally.
But fresh out of the box, ChatGPT was too talkative for Kennedy’s needs, he says—and much too apologetic. He would feed it tough messages, and it would recommend replying (in many more sentences than necessary) I’m sorry, please forgive me, I’ll do better. Having no self, it had no self-esteem.
Kennedy wanted a chatbot with “spine,” and he thought that if he built it, a lot of other co-parents might want it too. As he saw it, AI could help them at each stage of their communications: It could filter emotionally triggering language out of incoming messages and summarize just the facts. It could suggest appropriate responses. It could coach users toward “a better way,” Kennedy says. So he founded a company and started developing an app. He called it BestInterest, after the standard that courts often use for custody decisions—the “best interest” of the child or children. He would take those off-the-shelf OpenAI models and give them spine with his own prompts.
Estranged partners end up fighting horribly for any number of reasons, of course. For many, perhaps even most, things cool down after enough months have gone by, and a tool like BestInterest might not be useful long-term. But when a certain kind of personality is in the mix—call it “high-conflict,” “narcissistic,” “controlling,” “toxic,” whatever synonym for “crazy-making” you tend to see cross your internet feed—the fighting about the kids, at least from one side, never stops. Kennedy wanted his chatbot to stand up to these people, so he turned to the one they may hate most: Ramani Durvasula, a Los Angeles–based clinical psychologist who specializes in how narcissism shapes relationships.
