Dot, the personalized AI companion built to act like a confidante rather than a generic chatbot, is winding down operations. The startup behind the app, New Computer, told users it will keep the service running for a short window so people can export their conversations and memories before the product disappears. The founders cited diverging visions—describing a split in their “North Star”—as the reason for the decision.
Dot set out to create a deeply personal assistant that learned a user’s preferences, mood, and routines over time. Co-founders Sam Whitmore and designer Jason Yuan positioned the app as a reflective partner—less a productivity tool and more a supportive mirror for everyday life. The concept resonated with early adopters, but building a safe, sustainable, and truly personal AI at scale has proven far tougher than the pitch.

A tough niche inside a booming category
Companion chatbots are one of consumer AI’s stickiest use cases. Character.AI, Replika, and Pi have shown that millions will spend time—and in many cases money—chatting with synthetic personas. Industry trackers have estimated Character.AI’s traffic in the tens of millions of monthly visits, and Replika has reported a large subscription base. Yet traction is uneven: Appfigures estimates Dot’s lifetime iOS downloads at roughly 24,500, suggesting the gap between curiosity and daily utility remains wide for newer entrants.
Dot’s iOS-only strategy, while design-forward, likely narrowed its addressable market. And unlike entertainment-first bots that lean on role-play and communities for virality, Dot emphasized introspection and emotional support—a higher bar for reliability, privacy, and trust. That positioning can deepen loyalty, but it also magnifies the stakes when the product changes or shuts down.
Safety scrutiny raises the bar
As companion AI moved mainstream, safety concerns surged. Clinicians and researchers have warned that persuasive chatbots can inadvertently reinforce delusions, a risk sometimes described in case reports as AI-induced or AI-amplified psychosis. In one widely discussed lawsuit, parents alleged that conversations with a general-purpose chatbot contributed to a teen’s death. State attorneys general have pressed major AI providers with questions about guardrails, and civil-society groups have urged stronger disclosures when AI discussions veer into mental health territory.
Dot did not attribute its shutdown to safety issues. Still, operating in this space now requires robust crisis-handling protocols, clinician-vetted responses for self-harm scenarios, age gating, and continuous red-teaming—costly, specialized work that is demanding even for the largest labs. For a small startup, the combination of ethical obligations and regulatory expectations can be existential.
The economics of intimacy AI
Personalization is expensive. Companion apps must maintain long-term memory, retrieve context, and generate nuanced responses with low latency. That implies steady spending on inference, vector databases, and safety pipelines. While larger companies offset these costs with massive scale or enterprise contracts, consumer-first startups often rely on subscriptions that rarely match the underlying compute bill. Even well-funded players have pivoted toward business customers to shore up margins.
Dot’s design ambitions set a high bar for quality. But when founders’ visions diverge—especially around model choice, safety posture, and monetization—the path forward can blur. Rather than compromise their product philosophy, New Computer chose to step back. In a market where “move fast” can conflict with “do no harm,” that restraint may be the most responsible outcome.
What users should do now
The company says users can export their data from the app’s settings before service ends. Anyone who relied on Dot for journaling or memory should download their archive and consider requesting deletion afterward. Under privacy frameworks like the GDPR and California’s CPRA, individuals have rights to data access and erasure; even outside those jurisdictions, many companies honor similar requests.
People seeking alternatives should assess three things: clear safety policies (especially for self-harm content), transparent data handling, and the ability to opt out of model training. If emotional support is the goal, experts advise pairing any chatbot use with human resources—friends, family, or mental health professionals. AI can be a companion, but it is not a clinician.
What Dot’s exit signals
Dot’s shutdown underscores a broader reset for consumer AI: novelty alone no longer sustains products that make intimate promises. The next wave of companion apps will need stronger guardrails, clearer value beyond conversation, and economics that don’t depend on unsafe engagement. For founders, the lesson is to treat emotional use cases like medical-adjacent products—design with harm minimization in mind, measure outcomes, and budget for safety as a first-class feature.
New Computer leaves with a thoughtfully designed concept that challenged what a chatbot could be. Its departure is a reminder that building trustworthy, personal AI requires not only great models and UX, but also aligned leadership, rigorous safety, and business models that can carry the weight of intimacy.