51²è¹Ý

Artificial intelligence is already in our hospitals. 5 questions people want answered

Artificial intelligence is already in our hospitals. 5 questions people want answered

Before AI becomes widespread in health care, we need to ask what matters to consumers

Artificial intelligence (AI) is already being used in health care. AI can look for patterns in to help diagnose disease. It can help predict who in a hospital ward might . It can medical research papers to help doctors stay up-to-date with the latest evidence.

These are examples of AI making decisions health professionals previously made. More applications are being developed.

But what do consumers think of using AI in health care? And how should their answers shape how it’s used in the future?

What do consumers think?

AI systems are trained to look for patterns in large amounts of data. Based on these patterns, AI systems can make recommendations, suggest diagnoses, or initiate actions. They can potentially continually learn, becoming better at tasks over time.

If we draw together evidence, including our own , it seems most consumers accept the potential value of AI in health care.

This value could include, for example, increasing the or improving . At present, these are largely potential, rather than proven, benefits.

But consumers say their acceptance is conditional. They still have serious concerns.

1. Does the AI work?

A baseline expectation is AI tools should work well. Often, consumers say AI should be at least as good as a at the tasks it performs. They say we should not use AI if it will lead to more incorrect diagnoses or medical errors.

2. Who’s responsible if AI gets it wrong?

Consumers also worry that if AI systems generate decisions – such as diagnoses or treatment plans – without human input, it may be unclear who is responsible for errors. So people often want clinicians to remain responsible for the final decisions, and for from harms.

3. Will AI make health care less fair?

If health services are , AI systems can learn these patterns from data and the discrimination. So AI used in health care can make health inequities worse. In our studies consumers said this .

4. Will AI dehumanise health care?

Consumers are concerned AI will take the “human” elements out of health care, consistently saying AI tools should doctors. Often, this is because AI is perceived to lack important human traits, . Consumers say the communication skills, care and touch of a health professional are especially important when feeling vulnerable.

5. Will AI de-skill our health workers?

Consumers value human clinicians and their expertise. In our about AI in breast screening, women were concerned about the potential effect on radiologists’ skills and expertise. Women saw this expertise as a precious shared resource: too much dependence on AI tools, and this resource might be lost.

Consumers and communities need a say

The Australian health-care system cannot focus only on the technical elements of AI tools. Social and ethical considerations, including high-quality engagement with consumers and communities, are essential to shape AI use in health care.

Communities need opportunities to develop : to access reliable, trustworthy health information, services and resources.

Respectful engagement with Aboriginal and Torres Strait Islander communities must be central. This includes upholding Indigenous data sovereignty, which the Australian Institute of Aboriginal and Torres Strait Islander Studies :

the right of Indigenous peoples to govern the collection, ownership and application of data about Indigenous communities, peoples, lands, and resources.

This includes any use of data to create AI.

This critically important consumer and community engagement needs to take place before managers design (more) AI into health systems, before create guidance for how AI should and shouldn’t be used, and before clinicians consider buying a new AI tool for their practice.

We’re making some progress. Earlier this year, we ran a citizens’ jury on AI in health care. We supported 30 diverse Australians, from every state and territory, to spend three weeks learning about AI in health care, and developing recommendations for policymakers.

Their recommendations, which will be published in an upcoming issue of the Medical Journal of Australia, have informed a recently released for using AI in health care.

That’s not all

Health professionals also need to be upskilled and supported to use AI in health care. They need to learn to be critical users of digital health tools, including understanding their pros and cons.

Our of safety events reported to the Food and Drug Administration shows the most serious harms reported to the US regulator came not from a faulty device, but from the way consumers and clinicians used the device.

We also need to consider when health professionals should tell patients an AI tool is being used in their care, and when health workers should seek informed consent for that use.

Lastly, people involved in every stage of developing and using AI need to get accustomed to asking themselves: do consumers and communities agree this is a justified use of AI?

Only then will we have the AI-enabled health-care system consumers actually want.The Conversation

, Professor and Director, Australian Centre for Health Engagement, Evidence and Values, ; , PhD candidate, Australian Centre for Health Engagement, Evidence and Values, ; , Professor of Biomedical and Health Informatics at the Australian Institute of Health Innovation, , and , Research Fellow, Australian Centre for Health Engagement, Evidence and Values,

This article is republished from under a Creative Commons license. Read the .


UOW academics exercise academic freedom by providing expert commentary, opinion and analysis on a range of ongoing social issues and current affairs. This expert commentary reflects the views of those individual academics and does not necessarily reflect the views or policy positions of the 51²è¹Ý.