Healthcare Consultations In The Digital Age – Are Carebots Up To The Job?

Healthcare Consultations In The Digital Age – Are Carebots Up To The Job? LightField Studios/Shutterstock

Research shows that doctors who offer empathic and positive messages can reduce a patient’s pain, improve their recovery after surgery and lower the amount of morphine they need. But it doesn’t mean that telling a patient something simple, such as “this drug will make you feel better”, will have an effect. It’s more complicated than that, as our latest research shows.

Positive messages are usually repeated, definite, specific and personal. They should also be communicated by an authority figure who shows empathy (see graphic below). While our study does not identify what the most effective components of a positive message are (the sample was too small), the results imply that, for example, a positive message that is not specific or personalised, and is delivered by a doctor perceived to lack authority and empathy, will not have the desired effect.

What does all this mean for digitally assisted consultations ranging from telephone appointments to “carebots” (artificially intelligent robots delivering healthcare)? This is an important question to answer since carebots are being proposed as a cost-effective way to deal with the need to keep up with care for the increasing elderly populations in the UK and elsewhere.

The pandemic has accelerated the use of digitally assisted consultations, with the UK health secretary, Matt Hancock, claiming that patients won’t want to go back to face-to-face consultations after the pandemic. Online consultations are different from carebot consultations, but the trend away from human-to-human interaction can’t be denied. They have also both been rolled out too fast for ethical frameworks to be developed.

Technical and ethical problems

Adopting the evidence that positive messages help patients in the digital age is both technically and ethically problematic. While some of the components of a positive message (“this drug will make you feel better soon”) can be straightforwardly delivered through a mobile phone, via a video call, or even by a carebot, it seems to be inherently problematic for others. For example, the feeling that someone has authority might come from their title (doctor), which is presumably the same whether the doctor is seen in person or over the telephone.

But studies show authority also comes from body language. It’s more difficult to display body language through a telephone or video. While carebots may be able to convey their authority – and they have been shown to display sufficiently sophisticated body language to evoke certain emotions – real humans move differently. Adapting what we know about authoritative messages for the digital age is not straightforward. Some studies reveal that while digitally assisted consultations do not seem to be harmful, they are different (usually shorter), and we don’t know if they are as effective.

Also, to make the positive message personal to a patient (another component of positive messages) it might be important to pick up on subtle cues such as a downward glance or awkward pause, which studies have shown can be important for making accurate diagnoses. These cues may be more difficult to read through a telephone call, let alone by a carebot – at least for now.

 Get The Latest From InnerSelf

These are not just technical problems, they are ethical too. If digitally assisted healthcare consultations are not as effective at delivering positive messages, which, in turn, result in better care, then they threaten to violate the ethical requirement to help patients. Of course, if a carebot can do things more cheaply or to more people (they might not need to sleep), it might balance things out. But weighing the different ethical issues needs to be evaluated carefully, and this has not been done.

For carebots, this raises other ethical and even existential issues. If being empathic and caring is a key component of an effectively delivered positive message, it is important to know whether carebots are capable of caring. While we know that robots can be perceived as caring and empathic, it is not the same thing as being caring. It may not matter to some patients whether empathy is feigned or real as long as they benefit, but again, this needs to be fleshed out rather than assumed. Researchers are aware of these (and other) ethical issues and have called for a framework to regulate the design of carebots.

The study of positive messages shows that the new ethical frameworks would benefit from incorporating the latest evidence about the complexity of effective, positive – and other types of – communication. At the end of such a serious analysis, it may turn out that digitally assisted healthcare consultations and carebots are as good as face-to-face consultations.

They could even be better in some cases (some people may feel more comfortable telling intimate secrets to a robot than to a human). What is certain is that they are different, and we currently do not know what the implications of those differences are for optimising the benefits of complex positive messages in healthcare.The Conversation

About The Author

Jeremy Howick, Director of the Oxford Empathy Programme, University of Oxford

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Recommended Books: Health

Fresh Fruit CleanseFresh Fruit Cleanse: Detox, Lose Weight and Restore Your Health with Nature's Most Delicious Foods [Paperback] by Leanne Hall.
Lose weight and feel vibrantly healthy while clearing your body of toxins. Fresh Fruit Cleanse offers everything you need for an easy and powerful detox, including day-by-day programs, mouth-watering recipes, and advice for transitioning off the cleanse.
Click here for more info and/or to order this book on Amazon.

Thrive FoodsThrive Foods: 200 Plant-Based Recipes for Peak Health [Paperback] by Brendan Brazier.
Building upon the stress-reducing, health-boosting nutritional philosophy introduced in his acclaimed vegan nutrition guide Thrive, professional Ironman triathlete Brendan Brazier now turns his attention to your dinner plate (breakfast bowl and lunch tray too).
Click here for more info and/or to order this book on Amazon.

Death by Medicine by Gary NullDeath by Medicine by Gary Null, Martin Feldman, Debora Rasio and Carolyn Dean
The medical environment has become a labyrinth of interlocking corporate, hospital, and governmental boards of directors, infiltrated by the drug companies. The most toxic substances are often approved first, while milder and more natural alternatives are ignored for financial reasons. It's death by medicine.
Click here for more info and/or to order this book on Amazon.


follow InnerSelf on


 Get The Latest By Email



InnerSelf Newsletter: September 27, 2020
by InnerSelf Staff
One of the great strength of the human race is our ability to be flexible, to be creative, and to think outside the box. To be someone other than we were yesterday or the day before. We can change...…
What Works For Me: "For The Highest Good"
by Marie T. Russell, InnerSelf
The reason I share "what works for me" is that it may work for you as well. If not exactly the way I do it, since we are all unique, some variance of the attitude or method may very well be something…
Were You Part Of The Problem Last Time? Will You Be Part Of The Solution This Time?
by Robert Jennings,
Have you registered to vote? Have you voted? If you are not going to vote, you will be part of the problem.
InnerSelf Newsletter: September 20, 2020
by InnerSelf Staff
The theme of the newsletter this week can be summed up as "you can do it" or more specifically "we can do it!". This is another way of saying "you/we have the power to make a change". The image of…
What Works For Me: "I Can Do It!"
by Marie T. Russell, InnerSelf
The reason I share "what works for me" is that it may work for you as well. If not exactly the way I do it, since we are all unique, some variance of the attitude or method may very well be something…