Start With the Person With Dementia

Every ethical question about how we care for someone with dementia has to start here: what is the experience of the person with dementia?

When your mother calls, they is frightened. Something has triggered a moment of fear or loneliness, and the only thing that resolves that fear is the voice of the person they loves. They calls. They hears your voice, warm, familiar, with the things only you would know. They calms. The fear resolves. The call ends.

That is their experience. Comfort. Relief. The felt presence of someone who loves them.

That outcome, comfort, relief, felt safety, is the standard by which dementia care interventions are evaluated. Not factual accuracy. Not full disclosure. Comfort.

The Deception Question

Here is the hardest version of the question: is it deceptive to have an AI answer in your voice without your loved one knowing?

Let's answer it with context.

Trained dementia care professionals are taught, as a standard of practice, never to correct a patient's reality. When a person with dementia believes it is 1975, the correct professional response is not "no, it's 2026." The correct response is to enter their reality, validate their emotional experience, and provide comfort. This is not deception. It is therapeutic communication, the understanding that for a person with dementia, the emotional truth of feeling safe and loved is more important than factual correction.

This approach has a name, validation therapy, and it has been the standard of professional dementia care for over fifty years. Read more about validation therapy →

Therapeutic fibbing, telling a person with dementia that their deceased spouse is "at work" rather than deceased, to prevent the grief of re-learning the loss over and over, is specifically discussed in professional dementia care literature. The Alzheimer's Association acknowledges the validity of this approach. The reasoning is compassionate: when memory is gone, the grief of hearing someone is dead is experienced as fresh every time. There is no cruelty that serves the person in making their experience that grief repeatedly.

What KindredMind does exists on this same ethical spectrum. Your loved one doesn't know the call is being answered by AI. But what they experience is your voice, your warmth, your specific knowledge of their life. What they receive is comfort. What the call does for them is exactly what a real call from you does for them: they hear the person they love, and the fear goes away.

The person who sets this up is the legal decision-maker for their loved one's care. That is not a technicality. It is the same authority that determines which facility, which medications, which visitors, and which care approaches serve their loved one best. KindredMind sits within that framework — not outside it, not around it.

Is the mechanism different? Yes. Is the outcome different? No.

The Consent Question

Your loved one did not consent to this.

This is a real ethical consideration, and it deserves a real answer.

The framework for making decisions on behalf of someone who lacks capacity, which is what dementia progressively creates, is family authorization. The same framework that governs every other care decision made for a person with dementia: which facility, which medications, which visitors, which care approaches. These decisions are made by the people legally responsible for care, based on their knowledge of the person's values and their best judgment about the person's wellbeing.

KindredMind is set up entirely by the family. Here is what that control looks like in practice:

  • You record your own voice, a real, human, deliberate act
  • You build the knowledge base from everything you know about them
  • You control when it answers and when it doesn't
  • You can turn it off completely, at any time, in seconds
  • Every voice model requires your explicit recorded consent, not a checkbox. The voice presence is built only from your own short interview sessions with our setup specialist Sarah, and real conversations between the companion and your loved one are never recorded for training.

The caregiver who sets this up is exercising exactly the same kind of family judgment that is exercised in every other aspect of dementia care. It is a care decision, made by the people who love them most, on their behalf.

If you are unsure whether you hold legal decision-making authority for your loved one's personal care, we recommend speaking with a lawyer or your local Alzheimer Society before proceeding. In Ontario, the Health Care Consent Act establishes a substitute decision-maker hierarchy that gives most adult children and spouses this authority even without a formal Power of Attorney document.

The Replacement Question

Does this replace real connection?

No. It should not, and it is not designed to.

KindredMind exists for the gaps. The 3am call that would otherwise wake you from sleep and leave you lying in the dark. The fortieth call on a Tuesday when you are in a meeting you cannot leave. The calls that come on the days when you are completely done, and the version of you that would answer would not be the version they deserve.

The families who use it most effectively are the ones who are already giving enormous amounts of themselves, in real visits, real calls, real presence. They use KindredMind not to withdraw from the relationship, but to sustain themselves enough to show up for it. The version of Kirstin that answers her mother Sharon's calls now, because KindredMind handles the volume she couldn't, is a better version than the one who was drowning. Her mother gets more of a her who has something to give.

If a family were using this to disappear from a parent's life, that would be a different ethical situation. But that is not the profile of the people who find their way to this tool.

The Wellbeing Question

Does the person with dementia actually benefit?

The evidence says yes.

Simulated presence therapy, using familiar family voices to provide comfort to people with dementia, has been studied since the 1990s. A 2025 randomized controlled trial published in the International Journal of Neuroscience found meaningful reductions in agitation, anxiety, depression, and caregiver burden when SPT was combined with standard care. The people receiving familiar voice presence were calmer, less distressed, and experiencing less agitation than those receiving standard care alone.

Source: Duan Q et al., Int J Neurosci 2025;135(9):10701080. PubMed ID: 38646703. Read our full summary of the research →

KindredMind goes further than the research basis: it doesn't play a static recording, it responds. It knows your loved one. It adapts to what they's expressing. It meets their exactly where they is. That is more presence, not less, than the research basis.

The Standard We Hold Ourselves To

We can describe what we believe, but ethics isn't only a set of positions, it's a set of practices. Here is how we translate the ethical commitments above into the actual system families interact with.

You are always in control. Every aspect of the system, what it knows, when it answers, what voice it uses, is set by you. You can disable any part of it at any time. The AI does not act autonomously. It acts within the parameters you define, and you can change those parameters instantly from your dashboard.

Safety is not delegated to AI. Anything that sounds like a genuine emergency, physical distress, fear of a specific situation, confusion that suggests something has changed, generates an immediate alert to you. KindredMind is explicit about the limits of what AI should handle. Medical situations are not among them. The AI encourages their to speak with care staff and alerts you immediately. Real help comes from real people.

Their voice is never captured. Only your voice, which you record deliberately, with explicit recorded consent, is used to build the voice model. Your loved one's voice is never recorded, stored, or used in any way. The system is designed around you being the actor, not them.

We do not use your family's conversations to train models. The private conversations between your loved one and KindredMind are yours. They are not used to improve the system for other families. They are deleted after 90 days. They belong to you, and we have made this contractually enforceable with every vendor in our technology stack.

We are not saying these commitments resolve every ethical concern. We are saying that we have tried to build a system that is honest about what it is, transparent about its limits, and designed with the wellbeing of a vulnerable person at its center. That is the standard we hold ourselves to, and we expect to be held to it. For more on the care principles that guide our design, see our approach to dementia phone care.

Where We Land

We don't think this question has a clean answer, and we're not going to pretend it does. Reasonable people can look at the same situation and reach different conclusions. Some families will decide this is not right for their specific situation, and we respect that.

What we believe, based on what we know about dementia care, about the research, and about the families we serve, is this:

The ethical framework of dementia care has always prioritized the wellbeing of the person with dementia over factual transparency. The same professionals who would raise questions about AI in this space are trained to validate, to gently step into someone's reality, to meet the person where they are rather than where we are. The principle is not new. The technology is.

And the caregiver's wellbeing is part of the ethical picture too. A burned-out caregiver is not a better caregiver. A family that has a sustainable system is one that can stay present for the long road. That matters, for everyone in the family, including the person with dementia.

We built this because Kirstin needed it for her mother Sharon. Not to step back from that relationship, to sustain it. That is what it's for.

Kirstin holds legal decision-making authority for Sharon's personal care. She built the tool within the same framework she uses to make every other care decision for her mother.

K

Kirstin Thomas

co-founder of KindredMind and Sharon's daughter. She has been her mother's primary caregiver since 2025. KindredMind was built because she needed it.