2
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Deconstructing demographic bias in speech-based machine learning models for digital health

      research-article

      Read this article at

      ScienceOpenPublisherPMC
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Introduction

          Machine learning (ML) algorithms have been heralded as promising solutions to the realization of assistive systems in digital healthcare, due to their ability to detect fine-grain patterns that are not easily perceived by humans. Yet, ML algorithms have also been critiqued for treating individuals differently based on their demography, thus propagating existing disparities. This paper explores gender and race bias in speech-based ML algorithms that detect behavioral and mental health outcomes.

          Methods

          This paper examines potential sources of bias in the data used to train the ML, encompassing acoustic features extracted from speech signals and associated labels, as well as in the ML decisions. The paper further examines approaches to reduce existing bias via using the features that are the least informative of one’s demographic information as the ML input, and transforming the feature space in an adversarial manner to diminish the evidence of the demographic information while retaining information about the focal behavioral and mental health state.

          Results

          Results are presented in two domains, the first pertaining to gender and race bias when estimating levels of anxiety, and the second pertaining to gender bias in detecting depression. Findings indicate the presence of statistically significant differences in both acoustic features and labels among demographic groups, as well as differential ML performance among groups. The statistically significant differences present in the label space are partially preserved in the ML decisions. Although variations in ML performance across demographic groups were noted, results are mixed regarding the models’ ability to accurately estimate healthcare outcomes for the sensitive groups.

          Discussion

          These findings underscore the necessity for careful and thoughtful design in developing ML models that are capable of maintaining crucial aspects of the data and perform effectively across all populations in digital healthcare applications.

          Related collections

          Most cited references50

          • Record: found
          • Abstract: found
          • Article: not found

          Dissecting racial bias in an algorithm used to manage the health of populations

          Health systems rely on commercial prediction algorithms to identify and help patients with complex health needs. We show that a widely used algorithm, typical of this industry-wide approach and affecting millions of patients, exhibits significant racial bias: At a given risk score, Black patients are considerably sicker than White patients, as evidenced by signs of uncontrolled illnesses. Remedying this disparity would increase the percentage of Black patients receiving additional help from 17.7 to 46.5%. The bias arises because the algorithm predicts health care costs rather than illness, but unequal access to care means that we spend less money caring for Black patients than for White patients. Thus, despite health care cost appearing to be an effective proxy for health by some measures of predictive accuracy, large racial biases arise. We suggest that the choice of convenient, seemingly effective proxies for ground truth can be an important source of algorithmic bias in many contexts.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Gender differences in depression in representative national samples: Meta-analyses of diagnoses and symptoms.

            In 2 meta-analyses on gender differences in depression in nationally representative samples, we advance previous work by including studies of depression diagnoses and symptoms to (a) estimate the magnitude of the gender difference in depression across a wide array of nations and ages; (b) use a developmental perspective to elucidate patterns of gender differences across the life span; and (c) incorporate additional theory-driven moderators (e.g., gender equity). For major depression diagnoses and depression symptoms, respectively, we meta-analyzed data from 65 and 95 articles and their corresponding national data sets, representing data from 1,716,195 and 1,922,064 people in over 90 different nations. Overall, odds ratio (OR) = 1.95, 95% confidence interval (CI) [1.88, 2.03], and d = 0.27 [0.26, 0.29]. Age was the strongest predictor of effect size. The gender difference for diagnoses emerged earlier than previously thought, with OR = 2.37 at age 12. For both meta-analyses, the gender difference peaked in adolescence (OR = 3.02 for ages 13-15, and d = 0.47 for age 16) but then declined and remained stable in adulthood. Cross-national analyses indicated that larger gender differences were found in nations with greater gender equity, for major depression, but not depression symptoms. The gender difference in depression represents a health disparity, especially in adolescence, yet the magnitude of the difference indicates that depression in men should not be overlooked. (PsycINFO Database Record
              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              Why is depression more prevalent in women?

                Bookmark

                Author and article information

                Contributors
                URI : https://loop.frontiersin.org/people/2739618/overviewRole: Role:
                Role: Role:
                URI : https://loop.frontiersin.org/people/874521/overviewRole: Role:
                Journal
                Front Digit Health
                Front Digit Health
                Front. Digit. Health
                Frontiers in Digital Health
                Frontiers Media S.A.
                2673-253X
                25 July 2024
                2024
                : 6
                : 1351637
                Affiliations
                [ 1 ]Computer Science & Engineering, Texas A&M University , College Station, TX, United States
                [ 2 ]Computer Science & Engineering, Texas A&M University Qatar , Al Rayyan, Qatar
                [ 3 ]Institute of Cognitive Science & Computer Science, University of Colorado Boulder , Boulder, CO, United States
                Author notes

                Edited by: Heysem Kaya, Utrecht University, Netherlands

                Reviewed by: Evdoxia Taka, University of Glasgow, United Kingdom

                Dhiraj Kumar, National Eye Institute (NIH), United States

                [* ] Correspondence: Theodora Chaspari theodora.chaspari@ 123456colorado.edu
                [ † ]

                These authors contributed equally to this work and share first authorship

                Article
                10.3389/fdgth.2024.1351637
                11306200
                39119589
                8d5b6461-096c-4b46-b2b2-c84d3bb07fa8
                © 2024 Yang, El-Attar and Chaspari.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

                History
                : 06 December 2023
                : 15 July 2024
                Page count
                Figures: 1, Tables: 8, Equations: 230, References: 71, Pages: 13, Words: 0
                Funding
                Funded by: National Science Foundation
                Award ID: #2046118
                The authors declare financial support was received for the research, authorship, and/or publication of this article.
                The authors would like to acknowledge the National Science Foundation for funding this work (CAREER: Enabling Trustworthy Speech Technologies for Mental Health Care: From Speech Anonymization to Fair Human-centered Machine Intelligence, #2046118).
                Categories
                Digital Health
                Original Research
                Custom metadata
                Health Informatics

                speech,machine learning,anxiety,depression,demographic bias,fairness

                Comments

                Comment on this article