Request for Information on the Use of Clinical Algorithms That Have the Potential To Introduce Racial/Ethnic Bias Into Healthcare Delivery, 12948-12949 [2021-04509]

Download as PDF 12948 Federal Register / Vol. 86, No. 42 / Friday, March 5, 2021 / Notices Dated: March 1, 2021. Marquita Cullom, Associate Director. [FR Doc. 2021–04538 Filed 3–4–21; 8:45 am] BILLING CODE 4160–90–P DEPARTMENT OF HEALTH AND HUMAN SERVICES Agency for Healthcare Research and Quality Request for Information on the Use of Clinical Algorithms That Have the Potential To Introduce Racial/Ethnic Bias Into Healthcare Delivery Agency for Healthcare Research and Quality (AHRQ), HHS. ACTION: Notice of Request for Information. AGENCY: The Agency for Healthcare Research and Quality (AHRQ) is seeking information from the public on clinical algorithms that are used or recommended in medical practice and any evidence on clinical algorithms that may introduce bias into clinical decision- making and/or influence access to care, quality of care, or health outcomes for racial and ethnic minorities and those who are socioeconomically disadvantaged. DATES: Comments must be submitted on or before May 4, 2021. The EPC Program will not respond individually to responders but will consider all comments submitted by the deadline. ADDRESSES: Submissions should follow the Submission Instructions below. We prefer that comments be submitted electronically on the submission website. Email submissions may also be sent to: epc@ahrq.gov FOR FURTHER INFORMATION CONTACT: Anjali Jain, Email: Anjali.Jain@ ahrq.hhs.gov. SUMMARY: The Agency for Healthcare Research and Quality (AHRQ) is seeking information from the public on clinical algorithms that are used or recommended in medical practice and any evidence on clinical algorithms that may introduce bias into clinical decision-making and/ or influence access to care, quality of care, or health outcomes for racial and ethnic minorities and those who are socioeconomically disadvantaged. Information received in response to this request will be used to inform an AHRQ Evidence-Based Practice Center Program (EPC) evidence review and may inform other activities commissioned by or in collaboration with AHRQ. Established in 1997, the mission of the jbell on DSKJLSW7X2PROD with NOTICES SUPPLEMENTARY INFORMATION: VerDate Sep<11>2014 20:30 Mar 04, 2021 Jkt 253001 EPC Program (https:// effectivehealthcare.ahrq.gov/about/epc) is to create evidence reviews that improve healthcare by supporting evidence-based decision-making by patients, providers, and policymakers. Evidence reviews summarize and synthesize existing literature and evidence using rigorous methods. AHRQ is conducting this review pursuant to sections 902 and 901(c) of the Public Health Service Act, 42 U.S.C. 299a and 42 U.S.C. 299(c). AHRQ intends to commission an evidence review that will critically appraise the evidence on commonly used algorithms, including whether race/ethnicity is included as an explicit variable, and how algorithms have been developed and validated. The review would examine how race/ethnicity and related variables included in clinical algorithms impact healthcare use, patient outcomes and healthcare disparities. In addition, the review will identify and assess other variables with the potential to introduce bias such as prior utilization. The review will identify and review approaches to clinical algorithm development that avoid the introduction of racial and ethnic bias into clinical decision making and resulting outcomes. For the purposes of this evidence review, clinical algorithms are defined as a set of steps that clinicians use to guide decision-making in preventive services (such as screening), in diagnosis, clinical management, or otherwise assessing or improving a patient’s health. Algorithms are informed by data and research evidence and may include patient-specific factors or characteristics which may be sociodemographic factors such as race/ ethnicity, physiologic factors such as, for example, blood sugar level, or others such as patterns of healthcare utilization. When used appropriately, algorithms can improve disease management and patient health by creating efficiencies in place of individuals having to weigh multiple and complex factors when making a clinical judgement. As a result, the use of clinical algorithms has become widespread in healthcare and includes a heterogeneous set of tools including clinical pathways/guidelines, the establishment of norms and standards that may vary according to patient-specific factors, clinical decision support embedded in electronic health records (EHRs) or within medical devices, pattern recognition software used for diagnosis, and apps and calculators that predict patient risk and prognosis. Some clinical algorithms include information about a patient’s PO 00000 Frm 00049 Fmt 4703 Sfmt 4703 race or ethnicity among its inputs and thus lead clinicians to decision-making that varies by race/ethnicity, including decisions about how best to diagnose and manage individual patients. The purpose of this evidence review is to understand which algorithms are currently used in different clinical settings; the type and extent of their validation; their potential for bias with impact on access, quality, and outcomes of care; awareness among clinicians of these issues; and strategies for developing and testing clinical algorithms to assure that they are free of bias in order to inform the scope of a future evidence review. We are interested in understanding which algorithms are currently in use in clinical practice including those related to the use of clinical preventive services. How many include race/ ethnicity and other factors that could lead to bias within the algorithm? We are interested in all algorithms including clinical pathways/guidelines, norms and standards (including laboratory values) that vary according to patient-specific factors such as race/ ethnicity and related variables, clinical decision support embedded in EHRs, pattern recognition software, and apps and calculators for patient risk and prognosis. We are interested both in algorithms developed through traditional methods and through new and ongoing methods including machine learning and artificial intelligence. AHRQ seeks information • From healthcare providers who use clinical algorithms to screen, diagnose, triage, treat or otherwise care for patients • From laboratorians or technicians who use algorithms to interpret lab or radiology data • From researchers and clinical decision support developers who develop algorithms used in healthcare for patients • From clinical professional societies or other groups who develop clinical algorithms for healthcare • From payers who use clinical algorithms to guide payment decisions for care for patients • From healthcare delivery organizations who use clinical algorithms to determine healthcare practices and policies for patients • From device developers who incorporate algorithms into device software to interpret data and set standards • From patients whose healthcare and healthcare decisions may be informed by clinical algorithms E:\FR\FM\05MRN1.SGM 05MRN1 jbell on DSKJLSW7X2PROD with NOTICES Federal Register / Vol. 86, No. 42 / Friday, March 5, 2021 / Notices Specific questions of interest to the AHRQ include, but are not limited to, the following: 1. What clinical algorithms are used in clinical practice, hospitals, health systems, payment systems, or other instances? What is the estimated impact of these algorithms in size and characteristics of population affected, quality of care, clinical outcomes, quality of life and health disparities? 2. Do the algorithms in question 1 include race/ethnicity as a variable and, if so, how was race and ethnicity defined (including from whose perspective and whether there is a designation for mixed race or multiracial individuals)? 3. Do the algorithms in question 1 include measures of social determinants of health (SDOH) and, if so, how were these defined? Are these independently or collectively examined for their potential contribution to healthcare disparities and biases in care? 4. For the algorithms in question 1, what evidence, data quality and types (such as claims/utilization data, clinical data, social determinants of health), and data sources were used in their development and validation? What is the sample size of the datasets used for development and validation? What is the representation of Black, Indigenous, and People of Color (BIPOC) and what is the power to detect between-group differences? What methods were used to validate the algorithms and measure health outcomes associated with the use of the algorithms? 5. For the algorithms in question 1, what approaches are used in updating these algorithms? 6. Which clinical algorithms have evidence that they contribute to healthcare disparities, including decreasing access to care, quality of care or worsening health outcomes for BIPOC? What are the priority populations or conditions for assessing whether algorithms increase racial/ ethnic disparities? What are the mechanisms by which use of algorithms contribute to poor care for BIPOC? 7. To what extent are users of algorithms including clinicians, health systems, and health plans aware of the inclusion of race/ethnicity or other variables that could introduce bias in these algorithms and the implications for clinical decision making? What evidence is available about the degree to which the use of clinical algorithms contributes to bias in care delivery and resulting disparities in health outcomes? To what extent are patients aware of the inclusion of race/ethnicity or other variables that can result in bias in algorithms that influence their care? Do VerDate Sep<11>2014 20:30 Mar 04, 2021 Jkt 253001 providers or health systems communicate this information with patients in ways that can be understood? 8. What are approaches to identifying sources of bias and/or correcting or developing new algorithms that may be free of bias? What evidence, data quality and types (such as claims/utilization data, clinical data, information on social determinants of health), and data sources and sample size are used in their development and validation? What is the impact of these new approaches and algorithms on outcomes? 9. What challenges have arisen or can arise by designing algorithms developed using traditional biomedical or physiologic factors (such as blood glucose) yet include race/ethnicity as a proxy for other factors such as specific biomarkers, genetic information, etc.? What strategies can be used to address these challenges? 10. What are existing and developing standards (national and international) about how clinical algorithms should be developed, validated, and updated in a way to avoid bias? Are you aware of guidance on the inclusion or race/ ethnicity, related variables such as SDOH, prior utilization, or other variables to minimize the risk of bias? 11. To what extent are users of clinical algorithms educated about how algorithms are developed or may influence their decision-making? What educational curricula and training is available for clinicians that addresses bias in clinical algorithms? AHRQ is interested in all of the questions listed above, but respondents are welcome to address as many or as few as they choose and to address additional areas of interest not listed. This RFI is for planning purposes only and should not be construed as a policy, solicitation for applications, or as an obligation on the part of the Government to provide support for any ideas identified in response to it. AHRQ will use the information submitted in response to this RFI at its discretion and will not provide comments to any responder’s submission. However, responses to the RFI may be reflected in future solicitation(s) or policies. The information provided will be analyzed and may appear in reports. Respondents will not be identified in any published reports. Respondents are advised that the Government is under no obligation to acknowledge receipt of the information received or provide feedback to respondents with respect to any information submitted. No proprietary, classified, confidential, or sensitive information should be included in your response. The contents PO 00000 Frm 00050 Fmt 4703 Sfmt 4703 12949 of all submissions will be made available to the public upon request. Materials submitted must be publicly available or can be made public. Dated: March 1, 2021. Marquita Cullom, Associate Director. [FR Doc. 2021–04509 Filed 3–4–21; 8:45 am] BILLING CODE 4160–90–P DEPARTMENT OF HEALTH AND HUMAN SERVICES Centers for Disease Control and Prevention [Docket No. CDC–2020–0011] Draft Infection Control in Healthcare Personnel: Epidemiology and Control of Selected Infections Transmitted Among Healthcare Personnel and Patients: Diphtheria, Group A Streptococcus, Meningococcal Disease, and Pertussis Sections; ReOpening of Comment Period Centers for Disease Control and Prevention (CDC), Department of Health and Human Services (DHHS). ACTION: Notice with comment. AGENCY: The Centers for Disease Control and Prevention (CDC), in the Department of Health and Human Services (DHHS), announces the reopening of a docket to obtain a public comment on the DRAFT Infection Control in Healthcare Personnel: Epidemiology and Control of Selected Infections Transmitted Among Healthcare Personnel and Patients: Diphtheria, Group A Streptococcus, Meningococcal Disease, and Pertussis Sections (‘‘Draft Guideline’’). DATES: Written comments must be received on or before May 4, 2021. ADDRESSES: You may submit comments, identified by Docket No. CDC–2020– 0011, by any of the following methods: • Federal eRulemaking Portal: https:// www.regulations.gov. Follow the instructions for submitting comments. • Mail: Division of Healthcare Quality Promotion, National Center for Emerging and Zoonotic Infectious Diseases, Centers for Disease Control and Prevention, Attn: Docket No. CDC– 2020–0011, Infection Prevention and Control Guidelines, 1600 Clifton Rd. NE, Mailstop H16–2, Atlanta, Georgia, 30329. Instructions: All submissions received must include the agency name and Docket Number. All relevant comments received will be posted without change to https://regulations.gov, including any personal information provided. For SUMMARY: E:\FR\FM\05MRN1.SGM 05MRN1

Agencies

[Federal Register Volume 86, Number 42 (Friday, March 5, 2021)]
[Notices]
[Pages 12948-12949]
From the Federal Register Online via the Government Publishing Office [www.gpo.gov]
[FR Doc No: 2021-04509]


-----------------------------------------------------------------------

DEPARTMENT OF HEALTH AND HUMAN SERVICES

Agency for Healthcare Research and Quality


Request for Information on the Use of Clinical Algorithms That 
Have the Potential To Introduce Racial/Ethnic Bias Into Healthcare 
Delivery

AGENCY: Agency for Healthcare Research and Quality (AHRQ), HHS.

ACTION: Notice of Request for Information.

-----------------------------------------------------------------------

SUMMARY: The Agency for Healthcare Research and Quality (AHRQ) is 
seeking information from the public on clinical algorithms that are 
used or recommended in medical practice and any evidence on clinical 
algorithms that may introduce bias into clinical decision- making and/
or influence access to care, quality of care, or health outcomes for 
racial and ethnic minorities and those who are socioeconomically 
disadvantaged.

DATES: Comments must be submitted on or before May 4, 2021. The EPC 
Program will not respond individually to responders but will consider 
all comments submitted by the deadline.

ADDRESSES: Submissions should follow the Submission Instructions below. 
We prefer that comments be submitted electronically on the submission 
website. Email submissions may also be sent to: [email protected]

FOR FURTHER INFORMATION CONTACT: Anjali Jain, Email: 
[email protected].

SUPPLEMENTARY INFORMATION: The Agency for Healthcare Research and 
Quality (AHRQ) is seeking information from the public on clinical 
algorithms that are used or recommended in medical practice and any 
evidence on clinical algorithms that may introduce bias into clinical 
decision-making and/or influence access to care, quality of care, or 
health outcomes for racial and ethnic minorities and those who are 
socioeconomically disadvantaged.
    Information received in response to this request will be used to 
inform an AHRQ Evidence-Based Practice Center Program (EPC) evidence 
review and may inform other activities commissioned by or in 
collaboration with AHRQ. Established in 1997, the mission of the EPC 
Program (https://effectivehealthcare.ahrq.gov/about/epc) is to create 
evidence reviews that improve healthcare by supporting evidence-based 
decision-making by patients, providers, and policymakers. Evidence 
reviews summarize and synthesize existing literature and evidence using 
rigorous methods. AHRQ is conducting this review pursuant to sections 
902 and 901(c) of the Public Health Service Act, 42 U.S.C. 299a and 42 
U.S.C. 299(c).
    AHRQ intends to commission an evidence review that will critically 
appraise the evidence on commonly used algorithms, including whether 
race/ethnicity is included as an explicit variable, and how algorithms 
have been developed and validated. The review would examine how race/
ethnicity and related variables included in clinical algorithms impact 
healthcare use, patient outcomes and healthcare disparities. In 
addition, the review will identify and assess other variables with the 
potential to introduce bias such as prior utilization. The review will 
identify and review approaches to clinical algorithm development that 
avoid the introduction of racial and ethnic bias into clinical decision 
making and resulting outcomes.
    For the purposes of this evidence review, clinical algorithms are 
defined as a set of steps that clinicians use to guide decision-making 
in preventive services (such as screening), in diagnosis, clinical 
management, or otherwise assessing or improving a patient's health. 
Algorithms are informed by data and research evidence and may include 
patient-specific factors or characteristics which may be 
sociodemographic factors such as race/ethnicity, physiologic factors 
such as, for example, blood sugar level, or others such as patterns of 
healthcare utilization.
    When used appropriately, algorithms can improve disease management 
and patient health by creating efficiencies in place of individuals 
having to weigh multiple and complex factors when making a clinical 
judgement. As a result, the use of clinical algorithms has become 
widespread in healthcare and includes a heterogeneous set of tools 
including clinical pathways/guidelines, the establishment of norms and 
standards that may vary according to patient-specific factors, clinical 
decision support embedded in electronic health records (EHRs) or within 
medical devices, pattern recognition software used for diagnosis, and 
apps and calculators that predict patient risk and prognosis. Some 
clinical algorithms include information about a patient's race or 
ethnicity among its inputs and thus lead clinicians to decision-making 
that varies by race/ethnicity, including decisions about how best to 
diagnose and manage individual patients.
    The purpose of this evidence review is to understand which 
algorithms are currently used in different clinical settings; the type 
and extent of their validation; their potential for bias with impact on 
access, quality, and outcomes of care; awareness among clinicians of 
these issues; and strategies for developing and testing clinical 
algorithms to assure that they are free of bias in order to inform the 
scope of a future evidence review. We are interested in understanding 
which algorithms are currently in use in clinical practice including 
those related to the use of clinical preventive services. How many 
include race/ethnicity and other factors that could lead to bias within 
the algorithm? We are interested in all algorithms including clinical 
pathways/guidelines, norms and standards (including laboratory values) 
that vary according to patient-specific factors such as race/ethnicity 
and related variables, clinical decision support embedded in EHRs, 
pattern recognition software, and apps and calculators for patient risk 
and prognosis. We are interested both in algorithms developed through 
traditional methods and through new and ongoing methods including 
machine learning and artificial intelligence. AHRQ seeks information
     From healthcare providers who use clinical algorithms to 
screen, diagnose, triage, treat or otherwise care for patients
     From laboratorians or technicians who use algorithms to 
interpret lab or radiology data
     From researchers and clinical decision support developers 
who develop algorithms used in healthcare for patients
     From clinical professional societies or other groups who 
develop clinical algorithms for healthcare
     From payers who use clinical algorithms to guide payment 
decisions for care for patients
     From healthcare delivery organizations who use clinical 
algorithms to determine healthcare practices and policies for patients
     From device developers who incorporate algorithms into 
device software to interpret data and set standards
     From patients whose healthcare and healthcare decisions 
may be informed by clinical algorithms

[[Page 12949]]

    Specific questions of interest to the AHRQ include, but are not 
limited to, the following:
    1. What clinical algorithms are used in clinical practice, 
hospitals, health systems, payment systems, or other instances? What is 
the estimated impact of these algorithms in size and characteristics of 
population affected, quality of care, clinical outcomes, quality of 
life and health disparities?
    2. Do the algorithms in question 1 include race/ethnicity as a 
variable and, if so, how was race and ethnicity defined (including from 
whose perspective and whether there is a designation for mixed race or 
multiracial individuals)?
    3. Do the algorithms in question 1 include measures of social 
determinants of health (SDOH) and, if so, how were these defined? Are 
these independently or collectively examined for their potential 
contribution to healthcare disparities and biases in care?
    4. For the algorithms in question 1, what evidence, data quality 
and types (such as claims/utilization data, clinical data, social 
determinants of health), and data sources were used in their 
development and validation? What is the sample size of the datasets 
used for development and validation? What is the representation of 
Black, Indigenous, and People of Color (BIPOC) and what is the power to 
detect between-group differences? What methods were used to validate 
the algorithms and measure health outcomes associated with the use of 
the algorithms?
    5. For the algorithms in question 1, what approaches are used in 
updating these algorithms?
    6. Which clinical algorithms have evidence that they contribute to 
healthcare disparities, including decreasing access to care, quality of 
care or worsening health outcomes for BIPOC? What are the priority 
populations or conditions for assessing whether algorithms increase 
racial/ethnic disparities? What are the mechanisms by which use of 
algorithms contribute to poor care for BIPOC?
    7. To what extent are users of algorithms including clinicians, 
health systems, and health plans aware of the inclusion of race/
ethnicity or other variables that could introduce bias in these 
algorithms and the implications for clinical decision making? What 
evidence is available about the degree to which the use of clinical 
algorithms contributes to bias in care delivery and resulting 
disparities in health outcomes? To what extent are patients aware of 
the inclusion of race/ethnicity or other variables that can result in 
bias in algorithms that influence their care? Do providers or health 
systems communicate this information with patients in ways that can be 
understood?
    8. What are approaches to identifying sources of bias and/or 
correcting or developing new algorithms that may be free of bias? What 
evidence, data quality and types (such as claims/utilization data, 
clinical data, information on social determinants of health), and data 
sources and sample size are used in their development and validation? 
What is the impact of these new approaches and algorithms on outcomes?
    9. What challenges have arisen or can arise by designing algorithms 
developed using traditional biomedical or physiologic factors (such as 
blood glucose) yet include race/ethnicity as a proxy for other factors 
such as specific biomarkers, genetic information, etc.? What strategies 
can be used to address these challenges?
    10. What are existing and developing standards (national and 
international) about how clinical algorithms should be developed, 
validated, and updated in a way to avoid bias? Are you aware of 
guidance on the inclusion or race/ethnicity, related variables such as 
SDOH, prior utilization, or other variables to minimize the risk of 
bias?
    11. To what extent are users of clinical algorithms educated about 
how algorithms are developed or may influence their decision-making? 
What educational curricula and training is available for clinicians 
that addresses bias in clinical algorithms?
    AHRQ is interested in all of the questions listed above, but 
respondents are welcome to address as many or as few as they choose and 
to address additional areas of interest not listed.
    This RFI is for planning purposes only and should not be construed 
as a policy, solicitation for applications, or as an obligation on the 
part of the Government to provide support for any ideas identified in 
response to it. AHRQ will use the information submitted in response to 
this RFI at its discretion and will not provide comments to any 
responder's submission. However, responses to the RFI may be reflected 
in future solicitation(s) or policies. The information provided will be 
analyzed and may appear in reports. Respondents will not be identified 
in any published reports. Respondents are advised that the Government 
is under no obligation to acknowledge receipt of the information 
received or provide feedback to respondents with respect to any 
information submitted. No proprietary, classified, confidential, or 
sensitive information should be included in your response. The contents 
of all submissions will be made available to the public upon request. 
Materials submitted must be publicly available or can be made public.

    Dated: March 1, 2021.
Marquita Cullom,
Associate Director.
[FR Doc. 2021-04509 Filed 3-4-21; 8:45 am]
BILLING CODE 4160-90-P


This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.