Hatchet News

All,

Soliciting reactions on a “self-learning” AI tool for mental health.

Facial expressions correlate with depression. ML models of small datasets (4 mins. of facial video from small #s of particpants) predict patient depression scores better than doctor’s scores do. Other smart-phone-compatible biomarkers like pupillometry, eye saccades, audio, and kinematics, also correlate with depression. Several companies sell AI tools using such platforms, but I’m unaware of any massively multi-modal models.

Psychiatry’s reliance on subjective data undermines reliability and trust. AI scores, though trained on subjective data, are objective and could address.

What if a non-profit: (a) Launched a free smart-phone app generating real-time depression score based on facial-expressions. (b) Asked users to complete a PHQ8 (8 question depression survey) periodically (e.g. every fourth use) and store both PHQ and video data. (c) Used this data to expand model training database; subsequent use should find a more robust model. (d) Expanded to include other data modes and account for fators like age, sex, race, culture, etc. (e) Evolve from an active app (“get your score in 1 minute”), to a passive one app (“leave running,, with user consent, to chart mental health over time”) (f) Monetized: a. Protecting user data, keeping the app free for data donors. b. Licensed algorithms to health care professionals. c. Reinvested proceeds of fund maintenance and development after an initial philanthropical startup period

Proof of concept for facial expression models exist in academia, but need redevelopment for smartphones. Coding, modeling, and mental health system design challenges remain. A non-profit structure may be critical to encourage data donations.

Unanswered questions include: (a) how model fidelity scales with data, (b) performance of longitudinal systems (tracking video snippets over time), (c) minimum video duration for reliable monitoring, (d) environmental impacts on video capture (e.g., need for emotionally provocative prompts, etc.)

Critics may call this a gimmick as, a users already self-assess depression (and self-assessment is used as the ground-truth for training) and the system could be spoofable. However, objective, longitudinal could benefit individuals, and sufficient data might enable spoof-detection. High usage could also transform mental health research by providing objective data to redefine conditions and evaluate treatments.

Regulatory concerns (e.g., avoiding “diagnosis, ensuring informed consent) exist but seem management for an early version.

Curious to hear your thoughts.