Giving Compass' Take:
- Alexandrine Royer discusses an emerging AI-driven health industry that has the potential to worsen healthcare outcomes while giving tech companies another venue to profit off the public's data.
- How can you advocate for regulations that hold tech companies accountable?
- Read about how AI stacks up against the pandemic.
What is Giving Compass?
We connect donors to learning resources and ways to support community-led solutions. Learn more about us.
If you need to treat anxiety in the future, odds are the treatment won’t just be therapy, but also an algorithm. Across the mental-health industry, companies are rapidly building solutions for monitoring and treating mental-health issues that rely on just a phone or a wearable device. To do so, companies are relying on “affective computing” to detect and interpret human emotions. It’s a field that’s forecast to become a $37 billion industry by 2026, and as the COVID-19 pandemic has increasingly forced life online, affective computing has emerged as an attractive tool for governments and corporations to address an ongoing mental health crisis.
Despite a rush to build applications using it, emotionally intelligent computing remains in its infancy and is being introduced in the realm of therapeutic services as a fix-all solution without scientific validation nor public consent. Scientists still disagree over the over the nature of emotions and how they are felt and expressed among various populations, yet this uncertainty has been mostly disregarded by a wellness industry eager to profit on the digitalization of health care. If left unregulated, AI-based mental-health solutions risk creating new disparities in the provision of care as those who cannot afford in-person therapy will be referred to bot-powered therapists of uncertain quality.
Read the full article about mental health technology by Alexandrine Royer at Brookings.