<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">Online J Public Health Inform</journal-id><journal-id journal-id-type="publisher-id">ojphi</journal-id><journal-id journal-id-type="index">45</journal-id><journal-title>Online Journal of Public Health Informatics</journal-title><abbrev-journal-title>Online J Public Health Inform</abbrev-journal-title><issn pub-type="epub">1947-2579</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v18i1e78235</article-id><article-id pub-id-type="doi">10.2196/78235</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Fairness Correction in COVID-19 Predictive Models Using Demographic Optimization: Algorithm Development and Validation Study</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Awasthi</surname><given-names>Naman</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Abrar</surname><given-names>Saad</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Smolyak</surname><given-names>Daniel</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Frias-Martinez</surname><given-names>Vanessa</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Computer Science, University of Maryland</institution><addr-line>8125 Paint Branch Ave</addr-line><addr-line>College Park</addr-line><addr-line>MD</addr-line><country>United States</country></aff><aff id="aff2"><institution>College of Information Studies, University of Maryland</institution><addr-line>College Park</addr-line><addr-line>MD</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Mensah</surname><given-names>Edward</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Amuda</surname><given-names>Basit</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Song</surname><given-names>Jiafeng</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Rountree</surname><given-names>Lillian</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Naman Awasthi, MS, Department of Computer Science, University of Maryland, 8125 Paint Branch Ave, College Park, MD, 20742, United States, 1 2402806921; <email>nawasthi@umd.edu</email></corresp></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>3</day><month>2</month><year>2026</year></pub-date><volume>18</volume><elocation-id>e78235</elocation-id><history><date date-type="received"><day>29</day><month>05</month><year>2025</year></date><date date-type="rev-recd"><day>20</day><month>09</month><year>2025</year></date><date date-type="accepted"><day>10</day><month>12</month><year>2025</year></date></history><copyright-statement>&#x00A9; Naman Awasthi, Saad Abrar, Daniel Smolyak, Vanessa Frias-Martinez. Originally published in the Online Journal of Public Health Informatics (<ext-link ext-link-type="uri" xlink:href="https://ojphi.jmir.org/">https://ojphi.jmir.org/</ext-link>), 3.2.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Online Journal of Public Health Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://ojphi.jmir.org/">https://ojphi.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://ojphi.jmir.org/2026/1/e78235"/><abstract><sec><title>Background</title><p>COVID-19 forecasting models have been used to inform decision-making around resource allocation and intervention decisions, such as hospital beds or stay-at-home orders. State-of-the-art forecasting models often use multimodal data, including mobility or sociodemographic data, to enhance COVID-19 case prediction models. Nevertheless, related work has revealed under-reporting bias in COVID-19 cases as well as sampling bias in mobility data for certain minority racial and ethnic groups, which affects the fairness of COVID-19 predictions across racial and ethnic groups.</p></sec><sec><title>Objective</title><p>This study aims to introduce a fairness correction method that works for forecasting COVID-19 cases at an aggregate geographic level.</p></sec><sec sec-type="methods"><title>Methods</title><p>We use hard and soft error parity analyses on existing fairness frameworks and demonstrate that our proposed method, Demographic Optimization (DemOpts), performs better in both scenarios.</p></sec><sec sec-type="results"><title>Results</title><p>We first demonstrate that state-of-the-art COVID-19 deep learning models produce mean prediction errors that are significantly different across racial and ethnic groups at larger geographic scales. We then propose a novel debiasing method, DemOpts, to increase the fairness of deep learning&#x2013;based forecasting models trained on potentially biased datasets. Our results show that DemOpts can achieve better error parity than other state-of-the-art debiasing approaches, thus effectively reducing the differences in the mean error distributions across racial and ethnic groups.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>We introduce DemOpts, which reduces error parity differences compared with other approaches and generates fairer forecasting models compared with other approaches in the literature.</p></sec></abstract><kwd-group><kwd>COVID-19 forecasting</kwd><kwd>deep learning model</kwd><kwd>fairness</kwd><kwd>time series model</kwd><kwd>regression</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><sec id="s1-1"><title>Background</title><p>Forecasting the number of COVID-19 cases, hospitalizations, or deaths is crucial to inform decision-making. For example, COVID-19 forecasts can be used by hospitals to evaluate medical needs and required resources, such as supplies or beds, or by public health officials to inform closure policies at various geographical scales. In the United States, COVID-19 forecasts have been used at the state and county levels to inform social distancing or masking, such as the publicly available forecasts on the COVID-19 Forecast Hub that the Centers for Disease Control and Prevention (CDC) has routinely used in their communications [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>].</p><p>Related work over the past 4 years has shown a diverse variety of COVID-19 forecasting approaches [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref10">10</xref>, <xref ref-type="bibr" rid="ref11">11</xref>] using datasets such as the New York Times (NYT), Johns Hopkins University, COVID-19 Community Vulnerability Index, Google, and Apple [<xref ref-type="bibr" rid="ref12">12</xref>-<xref ref-type="bibr" rid="ref16">16</xref>], among others. Most publications focused on COVID-19 case prediction have reported results around the accuracy of the models, that is, minimizing the difference between the predicted cases and the actual number of cases reported. Nevertheless, previous work has shown that the accuracy of COVID-19 predictions can depend on various social determinants, including race or ethnicity [<xref ref-type="bibr" rid="ref17">17</xref>], income, or age [<xref ref-type="bibr" rid="ref18">18</xref>], revealing worse performance for protected attributes and pointing to a lack of COVID-19 predictive fairness that can affect resource allocation and decision-making. This lack of predictive fairness might be related to bias in the datasets used to train the model, that is, bias in COVID-19 case reporting [<xref ref-type="bibr" rid="ref19">19</xref>] or bias in mobility data [<xref ref-type="bibr" rid="ref20">20</xref>].</p><p>Given the presence of bias in the training datasets frequently used by COVID-19 forecast models, and previous work demonstrating that COVID-19 prediction accuracy can vary across social determinants, it becomes critical to devise methods to prevent data biases from percolating into the COVID-19 forecasts to guarantee fair decision-making based on case predictions. In this paper, we focus on in-processing bias mitigation approaches given their scarcity in the COVID-19 literature and propose Demographic Optimization (DemOpts), a debiasing method designed to achieve COVID-19 case prediction error parity across racial and ethnic groups in the context of deep learning models, that is, guarantee that county prediction errors are not significantly different across racial and ethnic groups. Although there exists a diverse set of COVID-19 predictive approaches, we focus on deep learning models because these are the most frequently used models in the machine learning community [<xref ref-type="bibr" rid="ref21">21</xref>], and narrow down our choice to transformer-based architectures because they are state-of-the-art in time series predictions [<xref ref-type="bibr" rid="ref22">22</xref>].</p><p>The main objective of DemOpts is to improve the fairness of the COVID-19 case predictions at the county level by achieving error parity in a regression setting [<xref ref-type="bibr" rid="ref17">17</xref>]. DemOpts proposes a novel debiasing approach that leverages county racial and ethnic data during training to modify conventional deep learning loss functions to penalize the model for statistically significant associations between the predictive error and the race or ethnicity distribution of a county. Our main contributions are:</p><list list-type="bullet"><list-item><p>We present DemOpts, a novel debiasing method for deep learning architectures that attempts to increase the fairness of the COVID-19 county case predictions by achieving error parity, that is, guaranteeing that prediction errors are similar across racial and ethnic groups. The DemOpts architecture is designed to optimize error parity across race and ethnicity using a novel multilabel approach that allows each county to be characterized by its own racial and ethnic group distribution during the debiasing process, instead of by a unique label.</p></list-item><list-item><p>We propose a novel evaluation protocol for the COVID-19 context, and we show that (1) state-of-the-art COVID-19 county case prediction models based on transformer architectures with no debiasing approach lack error parity, that is, prediction errors are statistically significantly different across racial and ethnic groups, (2) DemOpts applied to transformer-based architectures improves the error parity of the prediction models, increasing the similarity between mean prediction errors across racial and ethnic groups, and (3) the DemOpts debiasing approach performs better than state-of-the-art debiasing methods for regression settings.</p></list-item></list><p>While COVID-19 research was particularly prominent from 2020 to early 2024, challenges related to data biases and sampling issues in predictive modeling remain highly relevant. Our approach, leveraging the regression fairness model DemOpts, provides a robust framework to address these challenges. As future pandemics and public health crises arise, similar issues will persist, making our contribution valuable for ensuring fairness and reliability in predictive models.</p></sec><sec id="s1-2"><title>Literature Review</title><sec id="s1-2-1"><title>Deep LearningBased Forecasting Models</title><p>Deep learning models have started to become popular in time series prediction tasks. The available methods include (1) autoregressive models, such as Long Short-Term Memory or Gated Recurrent Network [<xref ref-type="bibr" rid="ref23">23</xref>]; (2) graph-based neural networks, such as graph attention networks [<xref ref-type="bibr" rid="ref24">24</xref>], Spatio-temporal Graph Convolutional Network [<xref ref-type="bibr" rid="ref25">25</xref>], neighbor convolution model [<xref ref-type="bibr" rid="ref26">26</xref>], or graph convolutional network; and (3) transformers, including Logarithmic Sparse Transformer [<xref ref-type="bibr" rid="ref27">27</xref>], Informer [<xref ref-type="bibr" rid="ref28">28</xref>], Autoformer [<xref ref-type="bibr" rid="ref29">29</xref>], Frequency Enhanced Decomposed Transformer [<xref ref-type="bibr" rid="ref30">30</xref>], Pyramidal Attention&#x2013;based Transformer [<xref ref-type="bibr" rid="ref31">31</xref>], and Patch Time Series Transformer [<xref ref-type="bibr" rid="ref32">32</xref>]. In this paper, we specifically focus on the temporal fusion transformer (TFT) architecture [<xref ref-type="bibr" rid="ref22">22</xref>], since it allows us to easily incorporate exogenous variables (eg, mobility data) as well as static variables (eg, demographic data) on top of the COVID-19 time series.</p></sec><sec id="s1-2-2"><title>Bias in Mobility and COVID-19 Data</title><p>The COVID-19 epidemic was closely monitored and had extensive data available about the counts of cases, hospitalizations, and deaths, as well as fine-grained information about mobility of people, policy implementations, vaccinations, and so on. Reducing the impact of mobility data or COVID-19 case bias in COVID-19 case predictions, as we do in this paper, is of critical importance to support decision-making processes focused on resource allocation during pandemics, to reduce harm and guarantee that decisions are fair and just across racial and ethnic groups. Human mobility data has been used to characterize human behaviors in the built environment [<xref ref-type="bibr" rid="ref33">33</xref>-<xref ref-type="bibr" rid="ref37">37</xref>], for public safety [<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref39">39</xref>], during epidemics and disasters [<xref ref-type="bibr" rid="ref40">40</xref>-<xref ref-type="bibr" rid="ref45">45</xref>], as well as to support decision-making for socioeconomic development [<xref ref-type="bibr" rid="ref46">46</xref>-<xref ref-type="bibr" rid="ref53">53</xref>]. During the COVID-19 pandemic, human mobility has played a central role in driving decision-making, acknowledging the impact of human movement on virus propagation [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref54">54</xref>]. Previous work has revealed sampling bias in mobility data collected via mobile apps, with Black and older individuals being underrepresented in the datasets [<xref ref-type="bibr" rid="ref20">20</xref>], and has exposed biases in COVID-19 forecasting models [<xref ref-type="bibr" rid="ref55">55</xref>,<xref ref-type="bibr" rid="ref56">56</xref>]. COVID-19 underreporting bias has been discussed in the literature [<xref ref-type="bibr" rid="ref57">57</xref>-<xref ref-type="bibr" rid="ref59">59</xref>] and points to multiple causes, including inadequate testing across certain minority groups or a lack of consistency in reporting race and ethnicity for COVID-19 cases [<xref ref-type="bibr" rid="ref19">19</xref>].</p></sec><sec id="s1-2-3"><title>Fairness Metrics and Fairness Corrections</title><p>Transformer-based COVID-19 case forecast models require the use of fairness metrics for regression settings, given that the loss optimization process in gradient-based deep learning architectures uses real-number predictions instead of classes.</p><p>Agarwal et al [<xref ref-type="bibr" rid="ref60">60</xref>], Fitzsimons et al [<xref ref-type="bibr" rid="ref61">61</xref>], and Gursoy and Kakadiaris [<xref ref-type="bibr" rid="ref17">17</xref>] outline the different aspects of fairness in regression settings and propose a set of fairness metrics for regression-type models. For this paper, we use the error parity metric proposed in [<xref ref-type="bibr" rid="ref17">17</xref>]. Error parity requires error distributions to be statistically independent of racial and ethnic groups. We expand this definition and relax the statistical significance requirement to be able to also evaluate whether the proposed DemOpts method can at least reduce the differences in error distributions across racial and ethnic groups, even when they are still statistically significantly different. To correct for bias and unfair performance in deep learning models, researchers have used preprocessing [<xref ref-type="bibr" rid="ref62">62</xref>,<xref ref-type="bibr" rid="ref63">63</xref>] and in-processing correction approaches [<xref ref-type="bibr" rid="ref64">64</xref>-<xref ref-type="bibr" rid="ref67">67</xref>]. Preprocessing approaches focus on creating a better input for learning deep neural network models by removing bias from the datasets [<xref ref-type="bibr" rid="ref62">62</xref>,<xref ref-type="bibr" rid="ref63">63</xref>], and there have been successful efforts focused on debiasing underreporting COVID-19 datasets to estimate actual cases or deaths before they are fed into predictive models [<xref ref-type="bibr" rid="ref68">68</xref>,<xref ref-type="bibr" rid="ref69">69</xref>]. On the other hand, in-processing approaches to improve the fairness of deep learning models, like the one we use in this paper, focus on the model and its regularization, usually adding a bias correction term in the loss function [<xref ref-type="bibr" rid="ref65">65</xref>,<xref ref-type="bibr" rid="ref67">67</xref>]. In this paper, we will compare our proposed debiasing approach against 3 state-of-the-art methods for debiasing in regression settings, which are individual fairness correction [<xref ref-type="bibr" rid="ref70">70</xref>], group fairness correction [<xref ref-type="bibr" rid="ref70">70</xref>] (both Lagrangian-based), and sufficiency [<xref ref-type="bibr" rid="ref71">71</xref>]. Individual and group fairness calculate penalties by determining overestimations across different groups and weighting the loss by a factor proportional to the overestimations, while sufficiency-based regularizers propose to make the loss independent of sensitive data attributes by simultaneously training a joint model and subgroup-specific networks to achieve fair predictions [<xref ref-type="bibr" rid="ref71">71</xref>].</p></sec></sec></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Proposed DemOpts</title><p>Our modeling focus is on deep learning models, which are the most frequently used approach for COVID-19 county case forecasts in the machine learning community [<xref ref-type="bibr" rid="ref21">21</xref>]. We specifically focus on the TFT model introduced in [<xref ref-type="bibr" rid="ref22">22</xref>] for several reasons. First, this model is state-of-the-art in interpretable time series prediction [<xref ref-type="bibr" rid="ref22">22</xref>]. Second, this model allows for the use of static reals as input to the model (ie, attributes that do not change over the duration of the training process, such as demographic percentages or population statistics). Third, the model works well with time-dependent features, including COVID-19 cases or mobility data, whereby past data influences future statistics.</p><p>DemOpts is an in-processing algorithm that modifies the standard training procedure for deep learning models at the loss computation stage. The algorithm modifies conventional loss functions to penalize the model for any statistically significant association (<italic>P</italic>&#x003C;.005) between the county prediction loss (error) and the county&#x2019;s racial and ethnic groups. In other words, DemOpts performs a race-based correction on the error to account for county demographic, racial, and ethnic distributions.</p><p>The algorithm can be divided into 3 steps (refer to <xref ref-type="fig" rid="figure1">Figure 1</xref>, <xref ref-type="fig" rid="figure2">Figure 2</xref>, and &#x201C;S.1 DemOpts Method&#x201D; in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for mathematical details).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Algorithm: Demographic Optimization (DemOpts). TFT: Temporal Fusion Transformer.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ojphi_v18i1e78235_fig01.png"/></fig><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Flow diagram for the Demographic Optimization (DemOpts) method.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ojphi_v18i1e78235_fig02.png"/></fig><sec id="s2-1-1"><title>Step 1: Calculate Loss</title><p>We use quantile predictions, as standard in COVID-19 forecasting literature [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref72">72</xref>], instead of point-value predictions. Quantile predictions are measured for 7 quantiles (0.02, 0.1, 0.25, 0.5, 0.75, 0.9, and 0.98) to gain insights into the uncertainty ranges and CIs of the COVID-19 county case predictive models. When using quantile predictions, the error is computed using quantile loss, also known as pinball loss (PBL), and defined as follows:</p><p><inline-formula><mml:math id="ieqn1"><mml:mstyle><mml:mrow><mml:mstyle displaystyle="false"><mml:mi>P</mml:mi><mml:mi>B</mml:mi><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mtable columnalign="left left" rowspacing=".2em" columnspacing="1em" displaystyle="false"><mml:mtr><mml:mtd><mml:mi>q</mml:mi><mml:mo>&#x2217;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2265;</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>q</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2217;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x003C;</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable><mml:mo fence="true" stretchy="true" symmetric="true"/></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>For quantile <italic>q</italic>, the PBL for the prediction of a given input <italic>X<sub>i</sub></italic> is <italic>PBL<sub>q</sub>(y<sub>ip</sub>, y<sub>i</sub></italic>), where <italic>y<sub>i</sub></italic> is the ground truth, and <italic>y<sub>ip</sub></italic> is the predicted value. The average over all quantiles can be represented as  <inline-formula><mml:math id="ieqn2"><mml:mi mathvariant="normal">P</mml:mi><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">L</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x2008;</mml:mi><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mi>&#x2008;</mml:mi><mml:mo>=</mml:mo><mml:mi>&#x2008;</mml:mi><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mfenced open="|" close="|" separators="|"><mml:mrow><mml:mi mathvariant="normal">q</mml:mi></mml:mrow></mml:mfenced></mml:mrow></mml:mfrac><mml:mrow><mml:msub><mml:mo stretchy="false">&#x2211;</mml:mo><mml:mrow><mml:mi mathvariant="normal">q</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mi>P</mml:mi><mml:mi>B</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x2008;</mml:mi><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow></mml:mrow></mml:math></inline-formula>.</p></sec><sec id="s2-1-2"><title>Step 2: Identify Dependencies Between Prediction Errors, Race, and Ethnicity</title><p>To achieve error parity, that is, mean errors being independent of racial and ethnic population distributions, we determine the relationship between errors and race and ethnic distributions. For that purpose, DemOpts fits a regression model between the prediction losses PBL (y<sub>ip</sub>,y<sub>i</sub>) across data points <italic>i</italic> and their corresponding county race and ethnicity distribution for each race D<sub>i</sub>:</p><disp-formula id="equWL1"><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>P</mml:mi><mml:mi>B</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>&#x2005;</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>&#x2005;</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>&#x2005;</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>&#x2005;</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>&#x2005;</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>&#x2005;</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">D</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mrow><mml:mo>&#x2008;</mml:mo></mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>o</mml:mi><mml:mi>k</mml:mi><mml:mi>a</mml:mi><mml:mi>h</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where d<sub>i</sub> are the corresponding county demographic features extracted from the US census data (represented as the percentage of each racial and ethnic group of the county for datapoint i), and lookahead refers to the number of days into the future the COVID-19 case prediction was generated for. In matrix representation:</p><disp-formula id="equWL2"><mml:math id="eqn2"><mml:mi>P</mml:mi><mml:mi>B</mml:mi><mml:mi>L</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mo>=</mml:mo><mml:mi mathvariant="normal">&#x03B2;</mml:mi><mml:mi>*</mml:mi><mml:mi>D</mml:mi><mml:mo>+</mml:mo><mml:mi> </mml:mi><mml:mi>&#x03B1;</mml:mi></mml:math></disp-formula><p>Once the regression model is fit, both regression coefficients (&#x03B2;) and their statistical significance (<italic>P</italic> value) are passed on to Step 3 to modify the adjusted loss and attempt to decouple race from the errors (loss).</p></sec><sec id="s2-1-3"><title>Step 3: Adjust the Loss</title><p>DemOpts modifies the conventional loss of deep learning models by adjusting for racial or ethnic bias in the error, that is, the loss is increased whenever a statistically significant regression coefficient for a race or ethnicity is found in Step 2 (with <italic>P</italic> value threshold=.005). By increasing the loss, DemOpts attempts to reduce the association between errors and race. Specifically, the loss is adjusted by the product of the original loss PBL (y<sub>ip</sub>,y<sub>i</sub>), the percentage race or ethnicity D<sub>j</sub> that holds a significant relationship with the error, and its coefficient &#x03B2;<sub>j</sub> in absolute value:</p><disp-formula id="equWL3"><mml:math id="eqn3"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mi>B</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:munder><mml:mi>H</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>p</mml:mi><mml:mi>v</mml:mi><mml:mi>a</mml:mi><mml:msub><mml:mi>l</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>|</mml:mo></mml:mrow><mml:mo>&#x2217;</mml:mo><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2217;</mml:mo><mml:mi>L</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mi>w</mml:mi><mml:mi>h</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="equWL4"><mml:math id="eqn4"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>H</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mtable columnalign="left left" rowspacing=".2em" columnspacing="1em" displaystyle="false"><mml:mtr><mml:mtd><mml:mn>1</mml:mn><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>x</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x003C;</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mn>0.005</mml:mn></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mn>0</mml:mn><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mi>x</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x2265;</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mn>0.005</mml:mn></mml:mtd></mml:mtr></mml:mtable><mml:mo fence="true" stretchy="true" symmetric="true"/></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula></sec></sec><sec id="s2-2"><title>Evaluation Protocol</title><p>In this section, we present a novel evaluation protocol to assess changes in fairness for TFT forecasting models when debiasing approaches, including DemOpts, are applied. We first describe the TFT COVID-19 county case prediction model we use, and the different debiasing approaches we evaluate on that prediction model. Next, we describe the error parity metrics we use to evaluate the fairness of each prediction model, and finally, we present the approach to analyze whether DemOpts improves the error parity metrics when compared to other state-of-the-art debiasing approaches for regression settings.</p><sec id="s2-2-1"><title>Predictive Model and Debiasing Approaches</title><p>We use the TFT with the conventional PBL function (PBL is the standard metric for reporting model performance in CDC Forecast Hub [<xref ref-type="bibr" rid="ref2">2</xref>]) as our baseline model (TFT<sub>Baseline</sub>) to predict the number of COVID-19 county cases for a given day.</p><p>Input data to the TFT model includes past COVID-19 cases per county, mobility data from SafeGraph, and race and ethnicity data for the county. We also train and test another TFT enhanced with the DemOpts debiasing method, TFT<sub>DemOpts</sub>, that adjusts the loss computation to attempt to eliminate or reduce the dependencies between error and race to achieve error parity. In addition, we train and test 3 more TFTs enhanced with state-of-the-art debiasing methods for regression settings, namely, individual fairness TFT<sub>Individual</sub> [<xref ref-type="bibr" rid="ref70">70</xref>], group fairness TFT<sub>Group</sub> [<xref ref-type="bibr" rid="ref70">70</xref>], and the sufficiency-based regularizer TFT<sub>Sufficiency</sub> [<xref ref-type="bibr" rid="ref71">71</xref>]. Individual and group fairness methods calculate penalties by determining overestimations across different groups and weighting the loss by a factor proportional to the overestimations, while the sufficiency-based regularizer trains a joint model and group-specific networks to achieve fair predictions. We replicate their methodology and adapt it to the forecasting setting by keeping TFT as the common network.</p></sec><sec id="s2-2-2"><title>Measuring Model Fairness</title><p>We choose error parity as our fairness metric [<xref ref-type="bibr" rid="ref17">17</xref>], with a focus on evaluating whether the distribution of predictive errors at the county level is independent of county majority race and ethnicity, that is, prediction errors are not statistically significantly different across racial and ethnic groups. To measure the fairness of each of the models TFT<sub>Baseline</sub>, TFT<sub>DemOpts</sub>, TFT<sub>Individual</sub>, TFT<sub>Group</sub>and TFT<sub>Sufficiency</sub>, we propose a 2-step process.</p><sec id="s2-2-2-1"><title>Step 1: Associate Errors With County Race or Ethnicity</title><p>To carry out the fairness analysis, we need to associate the PBL error of each county with race and ethnicity labels. However, that would require access to race-stratified COVID-19 case data at the county level, which is unfortunately not available due to systemic data collection failures during the pandemic [<xref ref-type="bibr" rid="ref73">73</xref>]. Hence, we propose to associate each county and its error with the majority race, that is, we label each county with the race or ethnicity that has the highest population percentage in that county. During the fairness analysis, we refer to majority White counties as the unprotected group and majority minority counties, such as Black or Hispanic, as the protected groups (details about the racial and ethnic groups considered in the evaluation are provided in the &#x201C;Datasets&#x201D; section).</p><p>In addition, we normalize each county&#x2019;s PBL error by county population size. The normalization by county population allows us to scale the errors appropriately, since higher-population counties will have higher case counts and thus, higher-magnitude errors. Normalizing by population fairly compares the error per unit population of one county with another:</p><disp-formula id="equWL5"><mml:math id="eqn5"><mml:mi mathvariant="normal">N</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:mi mathvariant="normal">m</mml:mi><mml:mi mathvariant="normal">P</mml:mi><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">L</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">p</mml:mi><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">t</mml:mi><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1000</mml:mn><mml:mo>&#x22C5;</mml:mo><mml:mi mathvariant="normal">P</mml:mi><mml:mi mathvariant="normal">B</mml:mi><mml:mi mathvariant="normal">L</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">p</mml:mi><mml:mo>,</mml:mo><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">t</mml:mi><mml:mo>,</mml:mo><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi mathvariant="normal">p</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:msub><mml:mrow><mml:mi mathvariant="normal">p</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:math></disp-formula><p>where y<sub>ti</sub> is the ground truth, y<sub>pi</sub> is the predicted value, and pop<sub>i</sub> is the county population.</p><p>We then calculate the average normalized PBL for each racial or ethnic group:</p><disp-formula id="equWL6"><mml:math id="eqn6"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi mathvariant="normal">A</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">v</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">g</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">N</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">o</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">r</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">m</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">P</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">B</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">L</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">p</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">t</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mrow><mml:mi mathvariant="normal">g</mml:mi></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">c</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">g</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:mi mathvariant="normal">N</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">o</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">r</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">m</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">P</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">B</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">L</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">p</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">y</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">t</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">c</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">g</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>|</mml:mo></mml:mrow></mml:mfrac></mml:mstyle></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where g represents the racial or ethnic group and c<sub>g</sub> is the set of all counties with as the majority group. This gives us the average normalized PBL for each demographic group.</p></sec><sec id="s2-2-2-2"><title>Step 2: Compute Fairness Metric</title><p>Once PBLs have been calculated for each racial and ethnic group in the United States, we can compute the error parity, that is, the fairness metric focused on evaluating whether the prediction errors are different across race and ethnicity. We propose 2 metrics to measure the error parity of COVID-19 county case predictions: hard error parity and soft error parity.</p></sec></sec></sec><sec id="s2-3"><title>Hard Error Parity Metric</title><p>Model predictions exhibit hard error parity when no statistically significant differences exist between normalized mean case prediction errors (AvgNormPBL) across racial or ethnic groups. In other words, normalized mean PBL errors across counties of different racial and ethnic groups are similar and hence, not biased by race or ethnicity. To test for the hard error parity of a prediction model, we propose to run one-way ANOVA followed by post hoc Tukey honestly significant difference (HSD) tests between the normalized mean error distributions of all racial and ethnic groups. ANOVA tests are an adequate choice even in violation of normality for large sample sizes, and in the presence of unequal sample sizes with homogeneous variance; thus, we choose this parametric test due to its superior strength [<xref ref-type="bibr" rid="ref74">74</xref>,<xref ref-type="bibr" rid="ref75">75</xref>].</p><p>Rejecting the null hypothesis for ANOVA would point to significantly different mean error values across some racial or ethnic groups and to a lack of perfect hard error parity. The subsequent analysis of the post hoc Tukey HSD test would reveal the pairs of racial and ethnic groups whose mean error values are significantly different and the numerical difference. The Tukey test also highlights the pairs of racial and ethnic groups for which the mean error is not statistically significantly different, pointing to instances where hard error parity exists for that model.</p></sec><sec id="s2-4"><title>Soft Error Parity Metric</title><p>Instead of measuring the statistical significance of the relationship between county race labels and county errors, we propose to use the Accuracy Equity Ratio (AER) metric [<xref ref-type="bibr" rid="ref76">76</xref>]. AER computes the ratio between the errors of the protected and unprotected groups as follows:</p><disp-formula id="equWL7"><mml:math id="eqn7"><mml:mi>A</mml:mi><mml:mi>E</mml:mi><mml:msub><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>A</mml:mi><mml:mi>v</mml:mi><mml:mi>g</mml:mi><mml:mi>N</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi><mml:mi>P</mml:mi><mml:mi>B</mml:mi><mml:mi>L</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>p</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mi>A</mml:mi><mml:mi>v</mml:mi><mml:mi>g</mml:mi><mml:mi>N</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi><mml:mi>P</mml:mi><mml:mi>B</mml:mi><mml:mi>L</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>u</mml:mi><mml:mi>n</mml:mi><mml:mi>p</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:mfenced></mml:mrow></mml:mfrac></mml:math></disp-formula><p>where subscript pg indicates counties labeled as the protected group (majority minority counties). unpg indicates counties labeled as the unprotected group (White), and AvgNormPBL is the average of the normalized PBL across counties for a given racial group g (pg or unpg).</p><p>As defined, the AER metric goes from 0 to &#x221E;. AER values in the range [0, 1] indicate comparatively lower normalized PBL for protected groups, which means the model predictions could be biased&#x2014;have higher errors&#x2014;for White majority counties; while AER values larger than one indicate that the model could be biased against the protected group, that is, the prediction errors are larger for counties with majority-minority groups. Values close to 1 indicate parity in error distribution between the protected group counties and the majority White counties. We claim that a predictive model achieves soft error parity for a given protected group when the AER value is close to 1, that is, the mean predictive error between that protected group and the White race is similar.</p><p>An alternative approach to assigning majority race or ethnicity would be to explore the associations between PBL errors and the distribution of racial and ethnic groups in a county (independent of COVID-19 cases, since that data are not available). Using a quantile regression, we can explore whether DemOpts eliminates significant associations between racial or ethnic percentages and the PBL errors, or at least reduces their magnitude. This approach removes the majority race requirement, but does not allow us to perform analyses with well-established fairness metrics in the literature, such as AER. Results are provided in the <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec><sec id="s2-5"><title>DemOpts Over State-of-the-Art</title><p>To assess whether DemOpts is a better debiasing approach than state-of-the-art methods, we need to compare the error parity metrics of the COVID-19 county case prediction model enhanced with the DemOpts method, TFT<sub>DemOpts,</sub> against the error parity metrics of the same prediction model enhanced with the other debiasing approaches (individual TFT<sub>Individual</sub>, group TFT<sub>Group</sub>, or sufficiency TFT<sub>Isufficiency</sub>), as well as with the baseline COVID-19 county case prediction model without any debiasing approach, TFT<sub>Baseline</sub>. Next, we describe how we carry out this analysis for the hard and soft error parity metrics.</p></sec><sec id="s2-6"><title>Hard Error Parity</title><p>We computed the hard error parity metric for each of the COVID-19 county case prediction models, using one-way ANOVA and the post hoc Tukey HSD test. An exploration of the statistical significance of the mean error difference for each pair of racial and ethnic groups will reveal whether applying DemOpts to the COVID-19 case prediction model produces fewer instances of significant mean error differences than any of the other debiasing methods or the baseline. In other words, a decrease in the number of significantly different mean PBL errors between races would point to an achievement of hard error parity for more racial and ethnic groups than other state-of-the-art debiasing approaches or the baseline.</p></sec><sec id="s2-7"><title>Soft Error Parity</title><p>To assess whether DemOpts applied to a COVID-19 case prediction model has higher soft error parity than any of the other state-of-the-art debiasing approaches, we propose to compare the AER values for each protected race and ethnic group across the 5 models: TFT<sub>DemOpts</sub>, TFT<sub>Individual</sub>, TFT<sub>Group</sub>, TFT<sub>Sufficiency</sub>, and TFT<sub>Baseline</sub>. Since AER values represent the quotient between the normalized mean prediction errors of a protected race or ethnicity vs White counties, the model with AER values closer to 1 will be the approach with the highest soft error parity. To measure AER&#x2019;s distance to 1, we compute the distance=|1-AER<sub>race</sub>| for each race and ethnic group, which represents the distance to a perfect soft parity error of 1. Distances closer to zero reveal better soft error parities.</p></sec><sec id="s2-8"><title>Datasets</title><p>In this section, we discuss the datasets we use in the DemOpts evaluation in the &#x201C;Results&#x201D; section. We train COVID-19 county case prediction models for the United States using COVID-19 case data, as well as mobility and demographic data. Mobility data has been used by previous work to inform case predictions via human mobility behaviors, under the assumption that the way people move might have an impact on the spreading of the epidemic. On the other hand, demographic data, either raw from the census or combined in different types of vulnerability indices, has also been shown to help predict COVID-19 prevalence, given the fact that COVID-19 has heavily affected vulnerable populations [<xref ref-type="bibr" rid="ref59">59</xref>].</p><sec id="s2-8-1"><title>COVID-19 Case Data</title><p>We use the COVID-19 case data compiled by the NYT at the county level [<xref ref-type="bibr" rid="ref12">12</xref>]. We account for delayed reporting by using the 7-day daily rolling average of COVID-19 cases (computed as the average of its current value and 6 previous days) instead of raw counts. <xref ref-type="fig" rid="figure3">Figure 3</xref> charts the daily COVID-19 reported cases throughout the data collection period.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>COVID-19 reported case counts per 1000 population across the United States.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ojphi_v18i1e78235_fig03.png"/></fig></sec><sec id="s2-8-2"><title>Mobility Data</title><p>SafeGraph open-sourced the mobility patterns of smartphone app users at the onset of the pandemic. These data points are curated by tracking the movements of millions of pseudonymized users via mobile app Software Development Kits (SafeGraph). Based on the data available, we use the daily origin-destination (OD) county-to-county flows [<xref ref-type="bibr" rid="ref77">77</xref>]. OD flows represent the volume of trips between pairs of counties across the United States for each day. For OD flows, we only use SafeGraph inflow (ie, mobility into the county). The inflow mobility is measured as changes in volumes of flows with respect to a baseline of normal behavior computed by SafeGraph using mobility data from February 17, 2020, to March 7, 2020.</p><p>Previous work has shown sampling bias in mobility datasets, revealing that not all races and ethnicities are equally represented due to variations in smartphone penetration rates [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref78">78</xref>]. It has also been shown that sampling bias in mobility data can negatively impact downstream tasks such as COVID-19 forecasting [<xref ref-type="bibr" rid="ref56">56</xref>]. While the addition of mobility data could potentially help improve prediction accuracy and support better decision-making, it also introduces bias. Our empirical analysis of DemOpts aims to understand whether the debiasing method proposed in this paper can improve the fairness of COVID-19 county case predictive models when mobility data is used as input to the predictive model. <xref ref-type="fig" rid="figure4">Figure 4</xref> shows the aggregate mobility data across the country. We see an initial drop in mobility in April (2020&#x2010;04), which corresponds to the first lockdown period. We then observed an increase in mobility a month later, which partially stabilizes after April.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Mobility for all ethnic and racial groups.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ojphi_v18i1e78235_fig04.png"/></fig></sec><sec id="s2-8-3"><title>Race and Ethnicity Data</title><p>We retrieve the race and ethnicity data from each county in the United States from the 2019 5-year American Community Survey. This survey collects data annually from all 50 states, Puerto Rico, and Washington, DC. As described in Step 1 of the evaluation protocol, we associate each county and its errors with the majority race (ie, we label each county with the race or ethnicity that has the highest population percentage in that county). Following this procedure identifies 4 racial and ethnic groups for the majority of counties: Asian, Black, Hispanic, and White. <xref ref-type="table" rid="table1">Table 1</xref> shows the distribution of US counties into these 4 racial and ethnic groups, and <xref ref-type="fig" rid="figure5">Figure 5</xref> show color-coded maps with the majority racial or ethnic group for each county.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Majority label counts.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Majority label</td><td align="left" valign="bottom">Count, n (%)</td></tr></thead><tbody><tr><td align="left" valign="top">Asian</td><td align="left" valign="top">6 (0.194)</td></tr><tr><td align="left" valign="top">Black</td><td align="left" valign="top">127 (4.118)</td></tr><tr><td align="left" valign="top">Hispanic</td><td align="left" valign="top">126 (4.085)</td></tr><tr><td align="left" valign="top">White</td><td align="left" valign="top">2825 (91.601)</td></tr></tbody></table></table-wrap><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Counties and majority-based label. (Mapbox and OpenStreetMap were used to create this map [<xref ref-type="bibr" rid="ref79">79</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref81">81</xref>].)</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ojphi_v18i1e78235_fig05.png"/></fig></sec></sec><sec id="s2-9"><title>Model Training</title><p>For evaluation purposes, we use COVID-19 case and SafeGraph mobility data from March 18, 2020, to November 30, 2020, for the training (207 days) and testing (49 days) of the TFT COVID-19 county case prediction models. The forecast task is the prediction of the number of COVID-19 cases for a given county from day X+1 to X+50, that is, the following 2 months (long-term forecasting with lookahead values from 1 to 50). Specifically, we train and test (1) the TFT<sub>Baseline</sub>, a TFT prediction model without a debiasing method; (2) the TFT<sub>Individual</sub>, TFT<sub>Group</sub>, and TFT<sub>Sufficiency</sub>, TFT prediction models with state-of-the-art debiasing methods; and (3) TFT<sub>DemOpts</sub>, a TFT prediction model enhanced with our proposed debiasing method. All 5 models are trained and tested for the same temporal range, and all are implemented using the PyTorch Forecasting library [<xref ref-type="bibr" rid="ref82">82</xref>]. We limit the period of analysis to a time before COVID-19 vaccines were available, given that after that event, research has revealed a less clear relationship between mobility data and postvaccines COVID-19 case volumes [<xref ref-type="bibr" rid="ref83">83</xref>]. We use the prediction errors (PBL) per racial and ethnic group to analyze and compare the hard and soft error parity of all trained models.</p></sec><sec id="s2-10"><title>Ethical Considerations</title><p>We used openly available datasets for mobility data (SafeGraph), COVID-19 case count (NYT), and demographic data (American Communities Survey). There was no human participant recruitment in this study, and thus we did not require institutional review board approval. All the datasets were aggregated at the county level and do not pose the risk of deanonymization.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Hard Error Parity Results</title><p>ANOVA tests of the normalized mean PBL error distributions across racial and ethnic groups for each debiasing approach were all significant, pointing to a dependency between race and the normalized prediction errors.</p><p><xref ref-type="table" rid="table2">Table 2</xref> shows the <italic>F</italic> statistic and test significance for each of the prediction models with and without debiasing approaches. The significant ANOVA tests reveal that perfect hard error parity is not achieved by any of the debiasing methods. In other words, for some racial and ethnic groups, there exist statistically significant differences between their mean PBL prediction errors of different racial and ethnic groups; this effect occurs for the TFT<sub>Baseline</sub> model as well as across all the other predictive models enhanced with a debiasing approach.</p><p>Nevertheless, post hoc Tukey HSD tests revealed interesting, nuanced results, showing significant differences in errors only between specific pairs of racial and ethnic groups. <xref ref-type="table" rid="table3">Table 3</xref> shows the post hoc Tukey HSD test results for each COVID-19 case predictive model: the baseline, the baseline enhanced with 1 of the 3 state-of-the-art debiasing approaches, and the baseline enhanced with our proposed method (DemOpts). Each row represents the output of the post hoc test, that is, the difference between the normalized mean PBL error of Group 1 and Group 2 (NormPBL<sub>Group1</sub> - NormPBL<sub>Group2</sub>). If the difference is positive, it means that the normalized mean predictive error is higher for Group 1; if the difference is negative, the normalized PBL error is higher for Group 2 (superscript b indicates statistically significant differences).</p><p>The first relevant observation when examining the table is that the baseline model, focused on predicting COVID-19 county cases with no debiasing approach is highly biased, with statistically significant differences between the mean normalized errors across all pairs of races, except for the comparison between Asian and Black counties as well as Hispanic and White counties, for which there is no statistically significant difference. These results reveal that there is no racial or ethnic group that achieves hard error parity and motivate our exploration of whether state-of-the-art debiasing methods or our proposed DemOpts can improve the hard error parity results of the baseline model.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>ANOVA <italic>F</italic> test statistics comparing mean prediction errors.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Fairness method</td><td align="left" valign="bottom">F statistic (<italic>df</italic>)</td></tr></thead><tbody><tr><td align="left" valign="top">Baseline</td><td align="left" valign="top">1195.398<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> (3080)</td></tr><tr><td align="left" valign="top">Group</td><td align="left" valign="top">1455.528<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> (3080)</td></tr><tr><td align="left" valign="top">Individual</td><td align="left" valign="top">1469.698<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> (3080)</td></tr><tr><td align="left" valign="top">Sufficiency</td><td align="left" valign="top">1195.651<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> (3080)</td></tr><tr><td align="left" valign="top">DemOpts<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></td><td align="left" valign="top">668.769<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> (3080)</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup><italic>P</italic>&#x003C;.001.</p></fn><fn id="table2fn2"><p><sup>b</sup>DemOpts: Demographic Optimization.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Hard error parity analysis. Each value represents the difference between the mean normalized pinball loss for each pair of racial and ethnic groups and indicates whether the difference is statistically significant.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Groups 1 and 2</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Group</td><td align="left" valign="top">Individual</td><td align="left" valign="top">Sufficiency</td><td align="left" valign="top">DemOpts<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td></tr></thead><tbody><tr><td align="left" valign="top" colspan="6">Asian</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Black</td><td align="left" valign="top">&#x2212;0.11</td><td align="left" valign="top">&#x2212;0.20</td><td align="left" valign="top">&#x2212;0.12</td><td align="left" valign="top">&#x2212;0.11</td><td align="left" valign="top">1.32</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Hispanic</td><td align="left" valign="top">&#x2212;2.30<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.65<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.50<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.29<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;0.77<bold><sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></bold></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>White</td><td align="left" valign="top">&#x2212;2.06<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.51<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.51<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.06<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;0.96<bold><sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></bold></td></tr><tr><td align="left" valign="top" colspan="6">Black</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Hispanic</td><td align="left" valign="top">&#x2212;2.18<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.45<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.38<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.17<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.09<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>White</td><td align="left" valign="top">&#x2212;1.94<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.31<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.39<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;1.94<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">&#x2212;2.29<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td></tr><tr><td align="left" valign="top" colspan="6">Hispanic</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>White</td><td align="left" valign="top">0.23</td><td align="left" valign="top">0.14</td><td align="left" valign="top">&#x2212;0.01</td><td align="left" valign="top">0.23</td><td align="left" valign="top">&#x2212;0.19</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>DemOpts: Demographic Optimization.</p></fn><fn id="table3fn2"><p><sup>b</sup><italic>P</italic>&#x003C;.001.</p></fn><fn id="table3fn3"><p><sup>c</sup>These values denote no significant difference between the prediction errors of Asian and White counties and of Asian and Hispanic counties.</p></fn></table-wrap-foot></table-wrap><p>When examining <xref ref-type="table" rid="table3">Table 3</xref>, we can observe that predictive models enhanced with the individual, group, or sufficiency debiasing methods do not improve the hard error parity over the baseline. On the one hand, similarly to the baseline model, the state-of-the-art debiasing methods (TFT<sub>Individual</sub>, TFT<sub>Group</sub>, and TFT<sub>Sufficiency</sub>) achieve hard error parity between Asian and Black counties and between Hispanic and White counties, that is, the mean error difference between these counties is not significant, pointing to a fair distribution of errors. On the other hand, for each pair of racial and ethnic groups whose prediction error distributions are significantly different for the baseline (rows with asterisks in the Baseline column), they remain significantly different for the individual, group, and sufficiency debiasing methods (rows with superscript b in the individual, group, and sufficiency columns).</p><p>When examining the significant mean PBL differences between racial and ethnic groups for the baseline and the state of the art debiasing models, we observe that all coefficients have similar values, signaling similar significant mean PBL differences between racial and ethnic groups (with values between 1.942 and 2.659 error cases per 1000 population). The sign of the coefficients reveals higher mean PBL errors for Hispanic and White counties when compared to Asian or Black counties, and higher mean PBL errors for White counties when compared to Hispanic counties across all models. For example, Hispanic and White counties have mean prediction errors 2.302 and 2.064 cases higher, respectively, when compared to Asian counties and while using the baseline model; and Hispanic and White counties have errors 2.457 and 2.313 cases higher, respectively, when compared to Black counties using the baseline model enhanced with the Group debiasing approach.</p><p>Moving on to DemOpts, the table shows that our proposed approach is the only debiasing method that achieves hard error parity in more cases than the baseline, effectively removing some of the associations between race and ethnicity and the normalized mean error distribution (PBL). Specifically, DemOpts removes the significant difference between the prediction errors of Asian and White counties and of Asian and Hispanic counties (refer to values with supercript c in <xref ref-type="table" rid="table3">Table 3</xref>), effectively achieving hard error parity for Asian counties, that is, the mean PBL in Asian counties is always similar to the mean error in counties of all the other racial and ethnic groups. These improvements occur additionally to hard error parity already seen in TFT<sub>Baseline</sub> (hard error parity between Asian and Black counties and between Hispanic and White counties), which are also present in the other 3 debiasing methods. In other words, DemOpts improves the hard error parity of case predictions for 2 additional racial and ethnic pairs compared with any of the other debiasing methods.</p><p>Finally, when looking specifically at the hard error parity between protected (Asian, Black, and Hispanic) and unprotected groups (White), DemOpts achieves hard error parity for Asian and Hispanic groups; that is, their mean prediction errors are not significantly different from those of White counties, while the baseline and the other 3 debiasing methods only achieve hard error parity for the Hispanic group when compared to White counties. These findings with respect to White counties motivate the evaluation of the soft error parity of the different models to determine, for example, whether DemOpts achieves the best soft error parity for the Black group (since hard error parity was not achieved), or to see if DemOpts has better soft error parity than other debiasing methods for Asian or Hispanic groups. Next, we explore the soft error parity metric for the TFT baseline and for all TFT models enhanced with debiasing approaches.</p></sec><sec id="s3-2"><title>Soft Error Parity Results</title><p><xref ref-type="table" rid="table4">Table 4</xref> shows the distance to the perfect soft error parity for each of the debiasing approaches across all protected racial and ethnic groups. As we can observe, DemOpts has the smallest values&#x2014;closest distances to perfect soft error parity&#x2014;for Asian and Black counties, while the individual debiasing method almost achieves perfect soft error parity for the Hispanic counties. In other words, DemOpts is the debiasing approach that produces the most similar errors between Asian and White counties and between Black and White counties, thereby achieving the largest reduction in predictive bias. On the other hand, the Individual debiasing method achieves errors for Hispanic counties that are closest to the White group. In addition, it is important to highlight that the Group and Sufficiency debiasing methods achieve soft error parities that are close to the TFT<sub>Baseline</sub>, which is not enhanced with any debiasing method.</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Soft error parity analysis. Each value represents the distance (|1-AER<sub>race</sub>|) for each protected group and debiasing method. TFT<sub>DemOpts</sub> achieves the highest soft error parity for 2 of the 3 protected races under study.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Group</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Group</td><td align="left" valign="top">Individual</td><td align="left" valign="top">Sufficiency</td><td align="left" valign="top">DemOpts<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup></td></tr></thead><tbody><tr><td align="left" valign="top">Asian</td><td align="left" valign="top">0.811</td><td align="left" valign="top">0.842</td><td align="left" valign="top">0.850</td><td align="left" valign="top">0.811</td><td align="left" valign="top">0.454<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td></tr><tr><td align="left" valign="top">Black</td><td align="left" valign="top">0.764</td><td align="left" valign="top">0.774</td><td align="left" valign="top">0.807</td><td align="left" valign="top">0.764</td><td align="left" valign="top">0.681<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td></tr><tr><td align="left" valign="top">Hispanic</td><td align="left" valign="top">0.093</td><td align="left" valign="top">0.048</td><td align="left" valign="top">0.003<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="left" valign="top">0.093</td><td align="left" valign="top">0.12</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>DemOpts: Demographic Optimization.</p></fn><fn id="table4fn2"><p><sup>b</sup>Smallest error parity for the particular group</p></fn></table-wrap-foot></table-wrap><p>Overall, these results reveal that DemOpts is the debiasing approach that improves the soft error parity of case prediction models, with errors for Asian and Black counties being the closest to errors in White counties. When accounting for additional factors, DemOpts outperforms the other methods by reducing the racial associations of model error.</p><p>In Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>, we provide and discuss the results for the quantile regression analysis in detail. Overall, the results confirm our findings with majority race labels, with DemOpts consistently outperforming other methods, showing the smallest coefficient magnitude for associations between the percentage of Asian, Black, and Hispanic populations and model error.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>Through our comparison of model performance for COVID-19 case prediction across counties of differing racial demographics, we showed that DemOpts outperforms other baselines for debiasing predictions. In our analysis of hard error parity, we found that DemOpts was the only debiasing method to eliminate statistically significant relationships between prediction error and racial demographics when compared with the baseline. While some significant associations remained, DemOpts achieved hard error parity for Asian vs White counties and Asian vs Hispanic counties. In the soft error parity analysis, DemOps substantially outperformed the baselines for Asian and Black counties, with a 69.4% reduction and 23% reduction, respectively, compared with the next closest method.</p></sec><sec id="s4-2"><title>Why is DemOpts Better?</title><p>The results showed that DemOpts is the only debiasing approach to achieve both hard and soft error parity for all 3 racial minority groups when compared with White counties.</p><p>In an attempt to understand why DemOpts succeeds in increasing both hard and soft error parity in the context of COVID-19 county case predictions, and compared with other debiasing methods, we computed the average PBL for each racial and ethnic group and for each predictive model enhanced, or not, with a debiasing method (refer to <xref ref-type="table" rid="table5">Table 5</xref>). We observed that DemOpts achieves better hard and soft error parity metrics because it considerably increases the errors for Asian and Black counties with respect to the baseline, until the differences with Hispanic and White are made not statistically significant (hard error parity) or closer to the White mean errors (soft error parity). Comparing <xref ref-type="table" rid="table4">Tables 4 and 5</xref>, we observed that DemOpts achieves considerably higher fairness for the Hispanic group (when compared to White) than for the Asian and Black groups (0.12 vs 0.454 and 0.681 in <xref ref-type="table" rid="table4">Table 4</xref>). As a result, the average PBL error for the Hispanic group (3.59 in <xref ref-type="table" rid="table5">Table 5</xref>) is considerably higher than the Asian and Black racial groups (1.7 and 1, respectively). We hypothesize that the differences in average errors and performance across racial and ethnic groups could be due to differences in the bias present in the training data, that is, mobility data or COVID-19 case counts could be more biased for Asian or Black groups, thus making it harder to achieve fair predictions when compared to White, and, in turn, due to the fairness-accuracy trade-off, making them more accurate (lower errors).</p><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>Group-wise pinball loss for each model. Demographic Optimization (DemOpts) has higher average pinball loss compared to the other models. The fairness-accuracy tradeoff leads to slightly larger pinball loss values for DemOpts compared to other methods.</p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Group</td><td align="left" valign="bottom">Baseline</td><td align="left" valign="bottom">Group</td><td align="left" valign="bottom">Individual</td><td align="left" valign="bottom">Sufficiency</td><td align="left" valign="bottom">DemOpts<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></td></tr></thead><tbody><tr><td align="left" valign="top">Asian</td><td align="left" valign="top">0.482</td><td align="left" valign="top">0.472</td><td align="left" valign="top">0.444</td><td align="left" valign="top">0.479</td><td align="left" valign="top">1.741</td></tr><tr><td align="left" valign="top">Black</td><td align="left" valign="top">0.600</td><td align="left" valign="top">0.674</td><td align="left" valign="top">0.570</td><td align="left" valign="top">0.598</td><td align="left" valign="top">1.015</td></tr><tr><td align="left" valign="top">Hispanic</td><td align="left" valign="top">2.784</td><td align="left" valign="top">3.131</td><td align="left" valign="top">2.951</td><td align="left" valign="top">2.776</td><td align="left" valign="top">3.597</td></tr><tr><td align="left" valign="top">White</td><td align="left" valign="top">2.546</td><td align="left" valign="top">2.987</td><td align="left" valign="top">2.961</td><td align="left" valign="top">2.540</td><td align="left" valign="top">3.192</td></tr></tbody></table><table-wrap-foot><fn id="table5fn1"><p><sup>a</sup>DemOpts: Demographic Optimization.</p></fn></table-wrap-foot></table-wrap><p>These results show that DemOpts&#x2019; optimization could not decrease prediction errors while trying to improve fairness, showing a fairness-accuracy trade-off that has been reported previously in the literature [<xref ref-type="bibr" rid="ref84">84</xref>]. To further clarify this finding, <xref ref-type="fig" rid="figure6">Figure 6</xref> shows both the average PBL and soft parity across all the models considered in this paper. As shown, DemOpts has the lowest soft error parity, but the highest PBL (top-left corner in the plot), while the other models decrease the PBL by sacrificing fairness (higher error parity in the bottom-right corner).</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Fairness-accuracy tradeoff. Model error (average pinball loss) vs average soft error parity (|1-AER|) for each model.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ojphi_v18i1e78235_fig06.png"/></fig></sec><sec id="s4-3"><title>Limitations</title><p>While DemOpts outperforms other state-of-the-art approaches in debiasing COVID-19 predictions, there are some limitations to DemOpts and our evaluation. First, DemOpts is unable to remove all statistical associations for the hard parity analysis, potentially because doing so would impose further reductions in model performance. For the soft parity analysis, the individual fairness approach is best for Hispanic counties, but the difference in parity levels is small. Regarding evaluation, our focus is exclusively on COVID-19 county case prediction&#x2014;while evaluation on other datasets and prediction tasks would be helpful for future work, our current evaluation provides sufficient evidence to show its applicability to other contexts. In addition, we compare DemOpts to baselines only on error parity metrics. Other fairness metrics may apply to the COVID-19 context and should be evaluated in future work, but we focus on error parity because DemOpts is specifically designed to mitigate it. Finally, we only compare DemOpts and baseline debiasing approaches within TFT models&#x2014;future work should compare with other commonly used models for COVID-19 case prediction.</p><p>Regardless, our novel debiasing approach shows that hard and soft error parity across protected and unprotected racial and ethnic groups can improve relative to other state-of-the-art approaches.</p><p>Finally, it is important to clarify that although in this paper, DemOpts focuses on bias mitigation in COVID-19 forecasting, it could also be applied to other health forecasting tasks where sampling bias in data collection can lead to bias in downstream tasks, for example, forecasting flu cases. These forecasts, when done at the county level and when using mobility data to model human spread, could benefit from the DemOpts method by reducing the effect of mobility bias or case count bias on other infectious diseases.</p></sec><sec id="s4-4"><title>Conclusion</title><p>Researchers have worked tirelessly on the creation of accurate COVID-19 case prediction models to support resource allocation and decision-making. However, sampling and underreporting biases in the data used to train these models have resulted in worse prediction performance for certain protected attributes, pointing to a lack of COVID-19 predictive fairness that could affect decision-making. In this paper, we show that state-of-the-art architectures in COVID-19 case predictions (TFT models) incur unfair prediction error distributions, and we design a novel debiasing approach and evaluation method to increase the fairness of predictions in the context of COVID-19 county case forecasts. DemOpts modifies the loss function in deep learning models to reduce the dependencies between error distributions and racial and ethnic labels. Our results show that DemOpts improves both the hard and soft error parity of COVID-19 county case predictions when compared with state-of-the-art debiasing methods.</p></sec></sec></body><back><notes><sec><title>Funding</title><p>This research is based on work supported by the National Science Foundation Award NO 2210572 (Bringing Transparency and Interpretability to Bias Mitigation Approaches in Place-based Mobility-centric Prediction Models for Decision Making in High-Stakes Settings) to the University of Maryland. There was no direct involvement of the funding institution in the design, analysis, and interpretation of the results.</p></sec><sec><title>Data Availability</title><p>We use open source data: (1) SafeGraph mobility, (2) New York Times COVID-19 case count, and demographic information from the American Communities Survey (ACS).</p></sec></notes><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AER</term><def><p>Accuracy Equity Ratio</p></def></def-item><def-item><term id="abb2">CDC</term><def><p>Centers for Disease Control and Prevention</p></def></def-item><def-item><term id="abb3">HSD</term><def><p>honestly significant difference</p></def></def-item><def-item><term id="abb4">NYT</term><def><p>New York Times</p></def></def-item><def-item><term id="abb5">OD</term><def><p>origin-destination</p></def></def-item><def-item><term id="abb6">PBL</term><def><p>pinball loss</p></def></def-item><def-item><term id="abb7">TFT</term><def><p>temporal fusion transformer</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="web"><article-title>COVID-19 forecasting and mathematical modeling</article-title><source>Centers for Disease Control and Prevention (CDC)</source><year>2023</year><access-date>2025-12-27</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.cdc.gov/coronavirus/2019-ncov/science/forecasting/forecasting-math-modeling.html">https://www.cdc.gov/coronavirus/2019-ncov/science/forecasting/forecasting-math-modeling.html</ext-link></comment></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="web"><source>COVID-19 Forecast Hub</source><year>2020</year><access-date>2025-12-27</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://covid19forecasthub.org/">https://covid19forecasthub.org/</ext-link></comment></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Zou</surname><given-names>D</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>P</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>W</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>Q</given-names> </name></person-group><article-title>Epidemic model guided machine learning for COVID-19 forecasts in the United States</article-title><source>medRxiv</source><comment>Preprint posted online on  May 24, 2020</comment><pub-id pub-id-type="doi">10.1101/2020.05.24.20111989</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Pei</surname><given-names>S</given-names> </name><name name-style="western"><surname>Shaman</surname><given-names>J</given-names> </name></person-group><article-title>Initial simulation of SARS-CoV2 spread and intervention effects in the continental US</article-title><source>medRxiv</source><comment>Preprint posted online on  Mar 21, 2020</comment><pub-id pub-id-type="doi">10.1101/2020.03.21.20040303</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Chiang</surname><given-names>WH</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Mohler</surname><given-names>G</given-names> </name></person-group><article-title>Hawkes process modeling of COVID-19 with mobility leading indicators and spatial covariates</article-title><source>medRxiv</source><comment>Preprint posted online on  Jun 6, 2020</comment><pub-id pub-id-type="doi">10.1101/2020.06.06.20124149</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Galasso</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Hochberg</surname><given-names>R</given-names> </name></person-group><article-title>A random forest model for forecasting regional COVID-19 cases utilizing reproduction number estimates and demographic data</article-title><source>Chaos Solitons Fractals</source><year>2022</year><month>03</month><volume>156</volume><fpage>111779</fpage><pub-id pub-id-type="doi">10.1016/j.chaos.2021.111779</pub-id><pub-id pub-id-type="medline">35013654</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Arik</surname><given-names>S</given-names> </name><name name-style="western"><surname>Li</surname><given-names>CL</given-names> </name><name name-style="western"><surname>Yoon</surname><given-names>J</given-names> </name><name name-style="western"><surname>Sinha</surname><given-names>R</given-names> </name><name name-style="western"><surname>Epshteyn</surname><given-names>A</given-names> </name><name name-style="western"><surname>Le</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Interpretable sequence learning for COVID-19 forecasting</article-title><source>arXiv</source><comment>Preprint posted online on  Aug 3, 2020</comment><pub-id pub-id-type="doi">10.48550/arXiv.2008.00646</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang-James</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Hess</surname><given-names>J</given-names> </name><name name-style="western"><surname>Salkin</surname><given-names>A</given-names> </name><etal/></person-group><article-title>A seq2seq model to forecast the COVID-19 cases, deaths and reproductive R numbers in US counties</article-title><source>Res Sq</source><year>2021</year><month>04</month><day>26</day><fpage>rs.3.rs-456641</fpage><pub-id pub-id-type="doi">10.21203/rs.3.rs-456641/v1</pub-id><pub-id pub-id-type="medline">34013251</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Le</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ibrahim</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sagun</surname><given-names>L</given-names> </name><name name-style="western"><surname>Lacroix</surname><given-names>T</given-names> </name><name name-style="western"><surname>Nickel</surname><given-names>M</given-names> </name></person-group><article-title>Neural relational autoregression for high-resolution COVID-19 forecasting</article-title><year>2021</year><access-date>2026-01-14</access-date><conf-name>epiDAMIK 2021: 4th International Workshop on Epidemiology meets Data Mining and Knowledge Discovery, ACM</conf-name><conf-date>Aug 14-18, 2021</conf-date><comment><ext-link ext-link-type="uri" xlink:href="https://epidamik.github.io/2021/papers/epiDAMIK_40_paper_12.pdf">https://epidamik.github.io/2021/papers/epiDAMIK_40_paper_12.pdf</ext-link></comment></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lucas</surname><given-names>B</given-names> </name><name name-style="western"><surname>Vahedi</surname><given-names>B</given-names> </name><name name-style="western"><surname>Karimzadeh</surname><given-names>M</given-names> </name></person-group><article-title>A spatiotemporal machine learning approach to forecasting COVID-19 incidence at the county level in the USA</article-title><source>Int J Data Sci Anal</source><year>2023</year><volume>15</volume><issue>3</issue><fpage>247</fpage><lpage>266</lpage><pub-id pub-id-type="doi">10.1007/s41060-021-00295-9</pub-id><pub-id pub-id-type="medline">35071733</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Angulo</surname><given-names>FJ</given-names> </name><name name-style="western"><surname>Finelli</surname><given-names>L</given-names> </name><name name-style="western"><surname>Swerdlow</surname><given-names>DL</given-names> </name></person-group><article-title>Estimation of US SARS-CoV-2 infections, symptomatic infections, hospitalizations, and deaths using seroprevalence surveys</article-title><source>JAMA Netw Open</source><year>2021</year><month>01</month><day>4</day><volume>4</volume><issue>1</issue><fpage>e2033706</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2020.33706</pub-id><pub-id pub-id-type="medline">33399860</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Times</surname><given-names>TNY</given-names> </name></person-group><article-title>Coronavirus (covid-19) data in the United States</article-title><source>GitHub</source><year>2020</year><access-date>2025-12-27</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/nytimes/covid-19-data">https://github.com/nytimes/covid-19-data</ext-link></comment></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dong</surname><given-names>E</given-names> </name><name name-style="western"><surname>Du</surname><given-names>H</given-names> </name><name name-style="western"><surname>Gardner</surname><given-names>L</given-names> </name></person-group><article-title>An interactive web-based dashboard to track COVID-19 in real time</article-title><source>Lancet Infect Dis</source><year>2020</year><month>05</month><volume>20</volume><issue>5</issue><fpage>533</fpage><lpage>534</lpage><pub-id pub-id-type="doi">10.1016/S1473-3099(20)30120-1</pub-id><pub-id pub-id-type="medline">32087114</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Smittenaar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Stewart</surname><given-names>N</given-names> </name><name name-style="western"><surname>Sutermaster</surname><given-names>S</given-names> </name><etal/></person-group><article-title>A COVID-19 community vulnerability index to drive precision policy in the US</article-title><source>medRxiv</source><comment>Preprint posted online on  May 19, 2021</comment><pub-id pub-id-type="doi">10.1101/2021.05.19.21257455</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="report"><article-title>COVID-19 community mobility reports</article-title><year>2022</year><access-date>2025-12-27</access-date><publisher-name>Google</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.google.com/covid19/mobility/">https://www.google.com/covid19/mobility/</ext-link></comment></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="report"><article-title>Mobility trends reports</article-title><year>2022</year><access-date>2025-12-27</access-date><publisher-name>Apple</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://covid19.apple.com/mobility">https://covid19.apple.com/mobility</ext-link></comment></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Gursoy</surname><given-names>F</given-names> </name><name name-style="western"><surname>Kakadiaris</surname><given-names>IA</given-names> </name></person-group><article-title>Error parity fairness: testing for group fairness in regression tasks</article-title><source>arXiv</source><access-date>2026-01-16</access-date><comment>Preprint posted online on 2022</comment><comment><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.48550/arXiv.2208.08279">https://doi.org/10.48550/arXiv.2208.08279</ext-link></comment></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Erfani</surname><given-names>A</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>A fairness assessment of mobility-based COVID-19 case prediction models</article-title><source>PLoS ONE</source><year>2023</year><volume>18</volume><issue>10</issue><fpage>e0292090</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0292090</pub-id><pub-id pub-id-type="medline">37851681</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Del Rios</surname><given-names>M</given-names> </name><name name-style="western"><surname>Puente</surname><given-names>S</given-names> </name><name name-style="western"><surname>Vergara-Rodriguez</surname><given-names>P</given-names> </name><name name-style="western"><surname>Sugrue</surname><given-names>N</given-names> </name></person-group><article-title>Latino invisibility in the pandemic</article-title><source>AMA J Ethics</source><year>2022</year><month>04</month><day>1</day><volume>24</volume><issue>4</issue><fpage>E289</fpage><lpage>295</lpage><pub-id pub-id-type="doi">10.1001/amajethics.2022.289</pub-id><pub-id pub-id-type="medline">35405055</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Coston</surname><given-names>A</given-names> </name><name name-style="western"><surname>Guha</surname><given-names>N</given-names> </name><name name-style="western"><surname>Ouyang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>L</given-names> </name><name name-style="western"><surname>Chouldechova</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ho</surname><given-names>DE</given-names> </name></person-group><article-title>Leveraging administrative data for bias audits: assessing disparate coverage with mobility data for COVID-19 policy</article-title><conf-name>Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency [Internet]</conf-name><conf-date>Mar 3-10, 2021</conf-date><conf-loc>New York, NY</conf-loc><fpage>173</fpage><lpage>184</lpage><pub-id pub-id-type="doi">10.1145/3442188.3445881</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Meraihi</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Gabis</surname><given-names>AB</given-names> </name><name name-style="western"><surname>Mirjalili</surname><given-names>S</given-names> </name><name name-style="western"><surname>Ramdane-Cherif</surname><given-names>A</given-names> </name><name name-style="western"><surname>Alsaadi</surname><given-names>FE</given-names> </name></person-group><article-title>Machine learning&#x2013;based research for COVID-19 detection, diagnosis, and prediction: a survey</article-title><source>SN Comput Sci</source><year>2022</year><volume>3</volume><issue>4</issue><fpage>286</fpage><pub-id pub-id-type="doi">10.1007/s42979-022-01184-z</pub-id><pub-id pub-id-type="medline">35578678</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lim</surname><given-names>B</given-names> </name><name name-style="western"><surname>Ar&#x0131;k</surname><given-names>S&#x00D6;</given-names> </name><name name-style="western"><surname>Loeff</surname><given-names>N</given-names> </name><name name-style="western"><surname>Pfister</surname><given-names>T</given-names> </name></person-group><article-title>Temporal fusion transformers for interpretable multi-horizon time series forecasting</article-title><source>Int J Forecast</source><year>2021</year><month>10</month><volume>37</volume><issue>4</issue><fpage>1748</fpage><lpage>1764</lpage><pub-id pub-id-type="doi">10.1016/j.ijforecast.2021.03.012</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hochreiter</surname><given-names>S</given-names> </name><name name-style="western"><surname>Schmidhuber</surname><given-names>J</given-names> </name></person-group><article-title>Long short-term memory</article-title><source>Neural Comput</source><year>1997</year><month>11</month><day>15</day><volume>9</volume><issue>8</issue><fpage>1735</fpage><lpage>1780</lpage><pub-id pub-id-type="doi">10.1162/neco.1997.9.8.1735</pub-id><pub-id pub-id-type="medline">9377276</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Velickovic</surname><given-names>P</given-names> </name><name name-style="western"><surname>Cucurull</surname><given-names>G</given-names> </name><name name-style="western"><surname>Casanova</surname><given-names>A</given-names> </name><name name-style="western"><surname>Romero</surname><given-names>A</given-names> </name><name name-style="western"><surname>Lio</surname><given-names>P</given-names> </name><name name-style="western"><surname>Bengio</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Graph attention networks stat</article-title><source>Stat (Int Stat Inst)</source><year>2017</year><volume>1050</volume><issue>20</issue><fpage>10</fpage><lpage>48550</lpage><pub-id pub-id-type="doi">10.48550/arXiv.1710.10903</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Yu</surname><given-names>B</given-names> </name><name name-style="western"><surname>Yin</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>Z</given-names> </name></person-group><article-title>Spatio-temporal graph convolutional networks: a deep learning framework for traffic forecasting</article-title><access-date>2025-12-27</access-date><conf-name>Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence</conf-name><conf-date>Jul 13-19, 2018</conf-date><pub-id pub-id-type="doi">10.24963/ijcai.2018/505</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Duan</surname><given-names>L</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Cheng</surname><given-names>E</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>C</given-names> </name></person-group><article-title>Deep convolutional neural networks for spatiotemporal crime prediction</article-title><access-date>2026-01-16</access-date><conf-name>Proceedings of the International Conference on Information and Knowledge Engineering (IKE)</conf-name><conf-date>2017</conf-date><fpage>61</fpage><lpage>67</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://www.semanticscholar.org/paper/Deep-Convolutional-Neural-Networks-for-Crime-Duan-Hu/407b84eb7ecfea0557b1a3e38ff0767b9bf25622">https://www.semanticscholar.org/paper/Deep-Convolutional-Neural-Networks-for-Crime-Duan-Hu/407b84eb7ecfea0557b1a3e38ff0767b9bf25622</ext-link></comment></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>S</given-names> </name><name name-style="western"><surname>Jin</surname><given-names>X</given-names> </name><name name-style="western"><surname>Xuan</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>X</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>W</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>YX</given-names> </name><etal/></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Wallach</surname><given-names>H</given-names> </name><name name-style="western"><surname>Larochelle</surname><given-names>H</given-names> </name><name name-style="western"><surname>Beygelzimer</surname><given-names>A</given-names> </name><name name-style="western"><surname>d&#x2019;</surname><given-names>ABF</given-names> </name><name name-style="western"><surname>Fox</surname><given-names>E</given-names> </name><name name-style="western"><surname>Garnett</surname><given-names>R</given-names> </name></person-group><article-title>Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting</article-title><access-date>2025-12-27</access-date><conf-name>Proceedings of the 33rd International Conference on Neural Information Processing Systems</conf-name><conf-date>Dec 8-14, 2019</conf-date><comment><ext-link ext-link-type="uri" xlink:href="https://proceedings.neurips.cc/paper_files/paper/2019/file/6775a0635c302542da2c32aa19d86be0-Paper.pdf">https://proceedings.neurips.cc/paper_files/paper/2019/file/6775a0635c302542da2c32aa19d86be0-Paper.pdf</ext-link></comment></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Peng</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Informer: beyond efficient transformer for long sequence time-series forecasting</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Dec 14, 2020</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2012.07436">https://arxiv.org/abs/2012.07436</ext-link></comment></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Long</surname><given-names>M</given-names> </name></person-group><article-title>Autoformer: decomposition transformers with auto-correlation for long-term series forecasting</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Jun 24, 2021</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2106.13008">https://arxiv.org/abs/2106.13008</ext-link></comment></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>T</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>L</given-names> </name><name name-style="western"><surname>Jin</surname><given-names>R</given-names> </name></person-group><article-title>FEDformer: frequency enhanced decomposed transformer for long-term series forecasting</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Jan 30, 2022</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2201.12740">https://arxiv.org/abs/2201.12740</ext-link></comment></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Liao</surname><given-names>C</given-names> </name><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>W</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>AX</given-names> </name><etal/></person-group><article-title>Pyraformer: low-complexity pyramidal attention for long-range time series modeling and forecasting</article-title><access-date>2025-12-27</access-date><conf-name>International Conference on Learning Representations (ICLR)</conf-name><conf-date>Apr 25-29, 2022</conf-date><comment><ext-link ext-link-type="uri" xlink:href="https://openreview.net/forum?id=0EXmFzUn5I">https://openreview.net/forum?id=0EXmFzUn5I</ext-link></comment></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Nie</surname><given-names>YH</given-names> </name><name name-style="western"><surname>Sinthong</surname><given-names>P</given-names> </name><name name-style="western"><surname>Kalagnanam</surname><given-names>J</given-names> </name><name name-style="western"><surname>Nguyen</surname><given-names>N</given-names> </name></person-group><article-title>A time series is worth 64 words: long-term forecasting with transformers</article-title><source>arXiv</source><comment>Preprint posted online on  Mar 5, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2211.14730</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Vieira</surname><given-names>MR</given-names> </name><name name-style="western"><surname>Fr&#x00ED;as-Mart&#x00ED;nez</surname><given-names>E</given-names> </name><name name-style="western"><surname>Bakalov</surname><given-names>P</given-names> </name><name name-style="western"><surname>Fr&#x00ED;as-Mart&#x00ED;nez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Tsotras</surname><given-names>VJ</given-names> </name></person-group><article-title>Querying spatio-temporal patterns in mobile phone-call databases</article-title><conf-name>2010 Eleventh International Conference on Mobile Data Management</conf-name><conf-date>May 23-26, 2010</conf-date><pub-id pub-id-type="doi">10.1109/MDM.2010.24</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="report"><person-group person-group-type="author"><name name-style="western"><surname>Hernandez</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hong</surname><given-names>L</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Whitby</surname><given-names>A</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name></person-group><article-title>Estimating poverty using cell phone data: evidence from guatemala</article-title><year>2017</year><access-date>2026-01-16</access-date><publisher-name>World Bank Policy Research Working Paper</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://documents.worldbank.org/en/publication/documents-reports/documentdetail/122541487082260120/estimating-poverty-using-cell-phone-data-evidence-from-guatemala">https://documents.worldbank.org/en/publication/documents-reports/documentdetail/122541487082260120/estimating-poverty-using-cell-phone-data-evidence-from-guatemala</ext-link></comment></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Virseda</surname><given-names>J</given-names> </name></person-group><article-title>Cell phone analytics: Scaling human behavior studies into the millions</article-title><source>Information Technologies &#x0026; International Development</source><year>2013</year><access-date>2026-01-16</access-date><volume>9</volume><issue>2</issue><fpage>35</fpage><comment><ext-link ext-link-type="uri" xlink:href="https://itidjournal.org/index.php/itid/article/download/1051/1051-2948-1-PB.pdf">https://itidjournal.org/index.php/itid/article/download/1051/1051-2948-1-PB.pdf</ext-link></comment></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Rubio</surname><given-names>A</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name><name name-style="western"><surname>Oliver</surname><given-names>N</given-names> </name></person-group><article-title>Human mobility in advanced and developing economies: a comparative analysis</article-title><access-date>2026-01-16</access-date><conf-name>2010 AAAI Spring Symposium Series</conf-name><conf-date>Mar 22-24, 2010</conf-date><comment><ext-link ext-link-type="uri" xlink:href="http://www.aaai.org/ocs/index.php/SSS/SSS10/paper/view/1095">http://www.aaai.org/ocs/index.php/SSS/SSS10/paper/view/1095</ext-link></comment></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>Spatial sensitivity analysis for urban hotspots using cell phone traces</article-title><source>Environ Plan B: Urban Anal City Sci</source><year>2021</year><month>11</month><volume>48</volume><issue>9</issue><fpage>2623</fpage><lpage>2639</lpage><pub-id pub-id-type="doi">10.1177/2399808320985843</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Abrar</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Awasthi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>Enhancing short-term crime prediction with human mobility flows and deep learning architectures</article-title><source>EPJ Data Sci</source><year>2022</year><volume>11</volume><issue>1</issue><fpage>53</fpage><pub-id pub-id-type="doi">10.1140/epjds/s13688-022-00366-2</pub-id><pub-id pub-id-type="medline">36406335</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Abrar</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Awasthi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Fr&#x00ED;as-Mart&#x00ED;nez</surname><given-names>V</given-names> </name></person-group><article-title>Auditing the fairness of place-based crime prediction models implemented with deep learning approaches</article-title><source>Comput Environ Urban Syst</source><year>2023</year><month>06</month><volume>102</volume><fpage>101967</fpage><pub-id pub-id-type="doi">10.1016/j.compenvurbsys.2023.101967</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wesolowski</surname><given-names>A</given-names> </name><name name-style="western"><surname>Eagle</surname><given-names>N</given-names> </name><name name-style="western"><surname>Tatem</surname><given-names>AJ</given-names> </name><etal/></person-group><article-title>Quantifying the impact of human mobility on malaria</article-title><source>Science</source><year>2012</year><month>10</month><day>12</day><volume>338</volume><issue>6104</issue><fpage>267</fpage><lpage>270</lpage><pub-id pub-id-type="doi">10.1126/science.1223467</pub-id><pub-id pub-id-type="medline">23066082</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>L</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Torrens</surname><given-names>P</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>Understanding citizens&#x2019; and local governments&#x2019; digital communications during natural disasters</article-title><source>WebSci &#x2019;17: Proceedings of the 2017 ACM on Web Science Conference</source><year>2017</year><fpage>141</fpage><lpage>150</lpage><pub-id pub-id-type="doi">10.1145/3091478.3091502</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Isaacman</surname><given-names>S</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name></person-group><article-title>Modeling human migration patterns during drought conditions in La Guajira, Colombia</article-title><source>COMPASS &#x2019;18: Proceedings of the 1st ACM SIGCAS Conference on Computing and Sustainable Societies</source><year>2018</year><fpage>1</fpage><lpage>9</lpage><pub-id pub-id-type="doi">10.1145/3209811.3209861</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bengtsson</surname><given-names>L</given-names> </name><name name-style="western"><surname>Gaudart</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Using mobile phone data to predict the spatial spread of cholera</article-title><source>Sci Rep</source><year>2015</year><month>03</month><day>9</day><volume>5</volume><issue>1</issue><fpage>8923</fpage><pub-id pub-id-type="doi">10.1038/srep08923</pub-id><pub-id pub-id-type="medline">25747871</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Ghurye</surname><given-names>J</given-names> </name><name name-style="western"><surname>Krings</surname><given-names>G</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>A framework to model human behavior at large scale during natural disasters</article-title><conf-name>2016 17th IEEE International Conference on Mobile Data Management (MDM)</conf-name><conf-date>Jun 13-16, 2016</conf-date><conf-loc>Porto, Portugal</conf-loc><fpage>18</fpage><lpage>27</lpage><pub-id pub-id-type="doi">10.1109/MDM.2016.17</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>L</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>Modeling and predicting evacuation flows during hurricane Irma</article-title><source>EPJ Data Sci</source><year>2020</year><month>12</month><volume>9</volume><issue>1</issue><fpage>29</fpage><pub-id pub-id-type="doi">10.1140/epjds/s13688-020-00247-6</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Virseda</surname><given-names>J</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name></person-group><article-title>Socio-economic levels and human mobility</article-title><source>Qual Meets Quant Workshop-QMQ</source><year>2010</year><access-date>2026-01-19</access-date><fpage>1</fpage><lpage>6</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://enriquefriasm.github.io/files/qmq.3370933.pdf">https://enriquefriasm.github.io/files/qmq.3370933.pdf</ext-link></comment></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fu</surname><given-names>C</given-names> </name><name name-style="western"><surname>McKenzie</surname><given-names>G</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Stewart</surname><given-names>K</given-names> </name></person-group><article-title>Identifying spatiotemporal urban activities through linguistic signatures</article-title><source>Comput Environ Urban Syst</source><year>2018</year><month>11</month><volume>72</volume><fpage>25</fpage><lpage>37</lpage><pub-id pub-id-type="doi">10.1016/j.compenvurbsys.2018.07.003</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Virseda</surname><given-names>J</given-names> </name><name name-style="western"><surname>Gomero</surname><given-names>A</given-names> </name></person-group><article-title>Mobilizing education: evaluation of a mobile learning tool in a low-income school</article-title><source>MobileHCI &#x2019;12: 14th International Conference on Human Computer Interaction with Mobile Devices and Services</source><year>2012</year><fpage>441</fpage><lpage>450</lpage><pub-id pub-id-type="doi">10.1145/2371574.2371642</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>L</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>Topic Models to Infer Socio-Economic Maps</article-title><source>AAAI</source><year>2016</year><volume>30</volume><issue>1</issue><fpage>3835</fpage><lpage>3841</lpage><pub-id pub-id-type="doi">10.1609/aaai.v30i1.9892</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Soto</surname><given-names>V</given-names> </name><name name-style="western"><surname>Virseda</surname><given-names>J</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name></person-group><article-title>Computing cost-effective census maps from cell phone traces</article-title><source>Department of Computer Science, Colombia University</source><year>2012</year><access-date>2026-01-19</access-date><comment><ext-link ext-link-type="uri" xlink:href="http://www.cs.columbia.edu/~vsoto/files/purba12.pdf">http://www.cs.columbia.edu/~vsoto/files/purba12.pdf</ext-link></comment></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kang</surname><given-names>W</given-names> </name><name name-style="western"><surname>Oshan</surname><given-names>T</given-names> </name><name name-style="western"><surname>Wolf</surname><given-names>LJ</given-names> </name><etal/></person-group><article-title>A roundtable discussion: defining urban data science</article-title><source>Environ Plan B Urban Anal City Sci</source><year>2019</year><month>11</month><volume>46</volume><issue>9</issue><fpage>1756</fpage><lpage>1768</lpage><pub-id pub-id-type="doi">10.1177/2399808319882826</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>L</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>E</given-names> </name><name name-style="western"><surname>Villarreal</surname><given-names>A</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>Characterization of internal migrant behavior in the immediate post-migration period using cell phone traces</article-title><conf-name>Proceedings of the Tenth International Conference on Information and Communication Technologies and Development</conf-name><conf-date>Jan 4-7, 2019</conf-date><conf-loc>Ahmedabad, India</conf-loc><fpage>1</fpage><lpage>12</lpage><pub-id pub-id-type="doi">10.1145/3287098.3287119</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name><name name-style="western"><surname>Sloate</surname><given-names>E</given-names> </name><name name-style="western"><surname>Manglunia</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name></person-group><article-title>Causal effect of low-income areas on shared dockless e-scooter use</article-title><source>Transp Res D Trans Environ</source><year>2021</year><month>11</month><volume>100</volume><fpage>103038</fpage><pub-id pub-id-type="doi">10.1016/j.trd.2021.103038</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Badr</surname><given-names>HS</given-names> </name><name name-style="western"><surname>Gardner</surname><given-names>LM</given-names> </name></person-group><article-title>Limitations of using mobile phone data to model COVID-19 transmission in the USA</article-title><source>Lancet Infect Dis</source><year>2021</year><month>05</month><volume>21</volume><issue>5</issue><fpage>e113</fpage><pub-id pub-id-type="doi">10.1016/S1473-3099(20)30861-6</pub-id><pub-id pub-id-type="medline">33152270</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tsai</surname><given-names>TC</given-names> </name><name name-style="western"><surname>Arik</surname><given-names>S</given-names> </name><name name-style="western"><surname>Jacobson</surname><given-names>BH</given-names> </name><etal/></person-group><article-title>Algorithmic fairness in pandemic forecasting: lessons from COVID-19</article-title><source>NPJ Digit Med</source><year>2022</year><month>05</month><day>10</day><volume>5</volume><issue>1</issue><fpage>59</fpage><pub-id pub-id-type="doi">10.1038/s41746-022-00602-z</pub-id><pub-id pub-id-type="medline">35538215</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abrar</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Awasthi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Smolyak</surname><given-names>D</given-names> </name><name name-style="western"><surname>Frias-Martinez</surname><given-names>V</given-names> </name></person-group><article-title>Analysis of performance improvements and bias associated with the use of human mobility data in COVID-19 case prediction models</article-title><source>ACM J Comput Sustain Soc</source><year>2023</year><month>12</month><day>31</day><volume>1</volume><issue>2</issue><fpage>1</fpage><lpage>36</lpage><pub-id pub-id-type="doi">10.1145/3616380</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Douglas</surname><given-names>MD</given-names> </name><name name-style="western"><surname>Respress</surname><given-names>E</given-names> </name><name name-style="western"><surname>Gaglioti</surname><given-names>AH</given-names> </name><etal/></person-group><article-title>Variation in reporting of the race and ethnicity of COVID-19 cases and deaths across US states: April 12, 2020, and November 9, 2020</article-title><source>Am J Public Health</source><year>2021</year><month>06</month><volume>111</volume><issue>6</issue><fpage>1141</fpage><lpage>1148</lpage><pub-id pub-id-type="doi">10.2105/AJPH.2021.306167</pub-id><pub-id pub-id-type="medline">33856884</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Estiri</surname><given-names>H</given-names> </name><name name-style="western"><surname>Strasser</surname><given-names>ZH</given-names> </name><name name-style="western"><surname>Rashidian</surname><given-names>S</given-names> </name><etal/></person-group><article-title>An objective framework for evaluating unrecognized bias in medical AI models predicting COVID-19 outcomes</article-title><source>J Am Med Inform Assoc</source><year>2022</year><month>07</month><day>12</day><volume>29</volume><issue>8</issue><fpage>1334</fpage><lpage>1341</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocac070</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gross</surname><given-names>CP</given-names> </name><name name-style="western"><surname>Essien</surname><given-names>UR</given-names> </name><name name-style="western"><surname>Pasha</surname><given-names>S</given-names> </name><name name-style="western"><surname>Gross</surname><given-names>JR</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Nunez-Smith</surname><given-names>M</given-names> </name></person-group><article-title>Racial and ethnic disparities in population-level Covid-19 mortality</article-title><source>J Gen Intern Med</source><year>2020</year><month>10</month><volume>35</volume><issue>10</issue><fpage>3097</fpage><lpage>3099</lpage><pub-id pub-id-type="doi">10.1007/s11606-020-06081-w</pub-id><pub-id pub-id-type="medline">32754782</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Agarwal</surname><given-names>A</given-names> </name><name name-style="western"><surname>Dud&#x00ED;k</surname><given-names>M</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>ZS</given-names> </name></person-group><article-title>Fair regression: quantitative definitions and reduction-based algorithms</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  May 30, 2019</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1905.12843">https://arxiv.org/abs/1905.12843</ext-link></comment></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fitzsimons</surname><given-names>J</given-names> </name><name name-style="western"><surname>Al Ali</surname><given-names>A</given-names> </name><name name-style="western"><surname>Osborne</surname><given-names>M</given-names> </name><name name-style="western"><surname>Roberts</surname><given-names>S</given-names> </name></person-group><article-title>A general framework for fair regression</article-title><source>Entropy (Basel)</source><year>2019</year><month>07</month><day>29</day><volume>21</volume><issue>8</issue><fpage>741</fpage><pub-id pub-id-type="doi">10.3390/e21080741</pub-id><pub-id pub-id-type="medline">33267455</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Brunet</surname><given-names>ME</given-names> </name><name name-style="western"><surname>Alkalay-Houlihan</surname><given-names>C</given-names> </name><name name-style="western"><surname>Anderson</surname><given-names>A</given-names> </name><name name-style="western"><surname>Zemel</surname><given-names>R</given-names> </name></person-group><article-title>Understanding the origins of bias in word embeddings</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Oct 8, 2018</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1810.03611">https://arxiv.org/abs/1810.03611</ext-link></comment></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Calmon</surname><given-names>FP</given-names> </name><name name-style="western"><surname>Wei</surname><given-names>D</given-names> </name><name name-style="western"><surname>Ramamurthy</surname><given-names>KN</given-names> </name><name name-style="western"><surname>Varshney</surname><given-names>KR</given-names> </name></person-group><article-title>Optimized data pre-processing for discrimination prevention</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Apr 11, 2017</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1704.03354">https://arxiv.org/abs/1704.03354</ext-link></comment></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Yan</surname><given-names>B</given-names> </name><name name-style="western"><surname>Seto</surname><given-names>S</given-names> </name><name name-style="western"><surname>Apostoloff</surname><given-names>N</given-names> </name></person-group><article-title>FORML: learning to reweight data for fairness</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Feb 3, 2022</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2202.01719">https://arxiv.org/abs/2202.01719</ext-link></comment></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>L</given-names> </name></person-group><article-title>Mitigating demographic bias of machine learning models on social media</article-title><year>2023</year><month>10</month><day>30</day><conf-name>EAAMO &#x2019;23: Proceedings of the 3rd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization</conf-name><conf-date>Oct 30 to Nov 1, 2023</conf-date><pub-id pub-id-type="doi">10.1145/3617694.3623244</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Soltan</surname><given-names>AAS</given-names> </name><name name-style="western"><surname>Clifton</surname><given-names>DA</given-names> </name></person-group><article-title>Algorithmic fairness and bias mitigation for clinical machine learning: a new utility for deep reinforcement learning</article-title><source>medRxiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Jun 24, 2022</comment><comment><ext-link ext-link-type="uri" xlink:href="https://www.medrxiv.org/content/early/2022/06/27/2022.06.24.22276853">https://www.medrxiv.org/content/early/2022/06/27/2022.06.24.22276853</ext-link></comment></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Das</surname><given-names>R</given-names> </name><name name-style="western"><surname>Dooley</surname><given-names>S</given-names> </name></person-group><article-title>Fairer and more accurate tabular models through NAS</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Oct 18, 2023</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2310.12145">https://arxiv.org/abs/2310.12145</ext-link></comment><pub-id pub-id-type="medline">37090239</pub-id></nlm-citation></ref><ref id="ref68"><label>68</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Jagodnik</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Ray</surname><given-names>F</given-names> </name><name name-style="western"><surname>Giorgi</surname><given-names>FM</given-names> </name><name name-style="western"><surname>Lachmann</surname><given-names>A</given-names> </name></person-group><article-title>Correcting under-reported COVID-19 case numbers: estimating the true scale of the pandemic</article-title><source>medRxiv</source><comment>Preprint posted online on 2020</comment><pub-id pub-id-type="doi">10.1101/2020.03.14.20036178</pub-id></nlm-citation></ref><ref id="ref69"><label>69</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Albani</surname><given-names>V</given-names> </name><name name-style="western"><surname>Loria</surname><given-names>J</given-names> </name><name name-style="western"><surname>Massad</surname><given-names>E</given-names> </name><name name-style="western"><surname>Zubelli</surname><given-names>J</given-names> </name></person-group><article-title>COVID-19 underreporting and its impact on vaccination strategies</article-title><source>BMC Infect Dis</source><year>2021</year><month>10</month><day>28</day><volume>21</volume><issue>1</issue><fpage>1111</fpage><pub-id pub-id-type="doi">10.1186/s12879-021-06780-7</pub-id><pub-id pub-id-type="medline">34711190</pub-id></nlm-citation></ref><ref id="ref70"><label>70</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Berk</surname><given-names>R</given-names> </name><name name-style="western"><surname>Heidari</surname><given-names>H</given-names> </name><name name-style="western"><surname>Jabbari</surname><given-names>S</given-names> </name><name name-style="western"><surname>Joseph</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kearns</surname><given-names>M</given-names> </name><name name-style="western"><surname>Morgenstern</surname><given-names>J</given-names> </name><etal/></person-group><article-title>A convex framework for fair regression</article-title><source>arXiv</source><comment>Preprint posted online on 2017</comment><pub-id pub-id-type="doi">10.48550/arXiv.1706.02409</pub-id></nlm-citation></ref><ref id="ref71"><label>71</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Shah</surname><given-names>A</given-names> </name><name name-style="western"><surname>Bu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>JK</given-names> </name><name name-style="western"><surname>Das</surname><given-names>S</given-names> </name><name name-style="western"><surname>Panda</surname><given-names>R</given-names> </name><name name-style="western"><surname>Sattigeri</surname><given-names>P</given-names> </name><etal/></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Chaudhuri</surname><given-names>K</given-names> </name><name name-style="western"><surname>Jegelka</surname><given-names>S</given-names> </name><name name-style="western"><surname>Song</surname><given-names>L</given-names> </name><name name-style="western"><surname>Szepesvari</surname><given-names>C</given-names> </name><name name-style="western"><surname>Niu</surname><given-names>G</given-names> </name><name name-style="western"><surname>Sabato</surname><given-names>S</given-names> </name></person-group><article-title>Selective regression under fairness criteria</article-title><year>2022</year><access-date>2025-12-27</access-date><conf-name>Proceedings of the 39th International Conference on Machine Learning, PMLR</conf-name><fpage>19598</fpage><lpage>19615</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://proceedings.mlr.press/v162/shah22a.html">https://proceedings.mlr.press/v162/shah22a.html</ext-link></comment></nlm-citation></ref><ref id="ref72"><label>72</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cramer</surname><given-names>EY</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>The United States COVID-19 Forecast Hub dataset</article-title><source>Sci Data</source><year>2022</year><month>08</month><day>1</day><volume>9</volume><issue>1</issue><fpage>462</fpage><pub-id pub-id-type="doi">10.1038/s41597-022-01517-w</pub-id><pub-id pub-id-type="medline">35915104</pub-id></nlm-citation></ref><ref id="ref73"><label>73</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kader</surname><given-names>F</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>CL</given-names> </name></person-group><article-title>Participatory approaches to addressing missing COVID-19 race and ethnicity data</article-title><source>Int J Environ Res Public Health</source><year>2021</year><month>06</month><day>18</day><volume>18</volume><issue>12</issue><fpage>6559</fpage><pub-id pub-id-type="doi">10.3390/ijerph18126559</pub-id><pub-id pub-id-type="medline">34207130</pub-id></nlm-citation></ref><ref id="ref74"><label>74</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Blanca</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Alarc&#x00F3;n</surname><given-names>R</given-names> </name><name name-style="western"><surname>Arnau</surname><given-names>J</given-names> </name><name name-style="western"><surname>Bono</surname><given-names>R</given-names> </name><name name-style="western"><surname>Bendayan</surname><given-names>R</given-names> </name></person-group><article-title>Non-normal data: is ANOVA still a valid option?</article-title><source>Psicothema</source><year>2017</year><month>11</month><volume>29</volume><issue>4</issue><fpage>552</fpage><lpage>557</lpage><pub-id pub-id-type="doi">10.7334/psicothema2016.383</pub-id><pub-id pub-id-type="medline">29048317</pub-id></nlm-citation></ref><ref id="ref75"><label>75</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zimmerman</surname><given-names>DW</given-names> </name></person-group><article-title>Comparative power of student t test and Mann-Whitney U test for unequal sample sizes and variances</article-title><source>The Journal of Experimental Education</source><year>1987</year><fpage>171</fpage><lpage>174</lpage><pub-id pub-id-type="doi">10.1080/00220973.1987.10806451</pub-id></nlm-citation></ref><ref id="ref76"><label>76</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Castelnovo</surname><given-names>A</given-names> </name><name name-style="western"><surname>Crupi</surname><given-names>R</given-names> </name><name name-style="western"><surname>Greco</surname><given-names>G</given-names> </name><name name-style="western"><surname>Regoli</surname><given-names>D</given-names> </name><name name-style="western"><surname>Penco</surname><given-names>IG</given-names> </name><name name-style="western"><surname>Cosentini</surname><given-names>AC</given-names> </name></person-group><article-title>A clarification of the nuances in the fairness metrics landscape</article-title><source>Sci Rep</source><year>2022</year><month>03</month><day>10</day><volume>12</volume><issue>1</issue><fpage>4209</fpage><pub-id pub-id-type="doi">10.1038/s41598-022-07939-1</pub-id><pub-id pub-id-type="medline">35273279</pub-id></nlm-citation></ref><ref id="ref77"><label>77</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>S</given-names> </name><name name-style="western"><surname>Liang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Li</surname><given-names>M</given-names> </name><name name-style="western"><surname>Rao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Kruse</surname><given-names>J</given-names> </name></person-group><article-title>Multiscale dynamic human mobility flow dataset in the U.S. during the COVID-19 epidemic</article-title><source>Sci Data</source><year>2020</year><month>11</month><day>12</day><volume>7</volume><issue>1</issue><fpage>390</fpage><pub-id pub-id-type="doi">10.1038/s41597-020-00734-5</pub-id><pub-id pub-id-type="medline">33184280</pub-id></nlm-citation></ref><ref id="ref78"><label>78</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Schlosser</surname><given-names>F</given-names> </name><name name-style="western"><surname>Sekara</surname><given-names>V</given-names> </name><name name-style="western"><surname>Brockmann</surname><given-names>D</given-names> </name><name name-style="western"><surname>Garcia-Herranz</surname><given-names>M</given-names> </name></person-group><article-title>Biases in human mobility data impact epidemic modeling</article-title><source>arXiv</source><access-date>2025-12-27</access-date><comment>Preprint posted online on  Dec 23, 2021</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2112.12521">https://arxiv.org/abs/2112.12521</ext-link></comment></nlm-citation></ref><ref id="ref79"><label>79</label><nlm-citation citation-type="web"><article-title>Mapbox</article-title><access-date>2026-01-16</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.mapbox.com/about/maps">https://www.mapbox.com/about/maps</ext-link></comment></nlm-citation></ref><ref id="ref80"><label>80</label><nlm-citation citation-type="web"><article-title>OpenStreetMap</article-title><access-date>2026-01-16</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.openstreetmap.org/about/">https://www.openstreetmap.org/about/</ext-link></comment></nlm-citation></ref><ref id="ref81"><label>81</label><nlm-citation citation-type="web"><article-title>Improve this map</article-title><access-date>2026-01-16</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://apps.mapbox.com/feedback/">https://apps.mapbox.com/feedback/</ext-link></comment></nlm-citation></ref><ref id="ref82"><label>82</label><nlm-citation citation-type="web"><source>PyTorch-based forecasting with sktime integration sktime</source><year>2020</year><access-date>2026-01-16</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/sktime/pytorch-forecasting">https://github.com/sktime/pytorch-forecasting</ext-link></comment></nlm-citation></ref><ref id="ref83"><label>83</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gatalo</surname><given-names>O</given-names> </name><name name-style="western"><surname>Tseng</surname><given-names>K</given-names> </name><name name-style="western"><surname>Hamilton</surname><given-names>A</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>G</given-names> </name><name name-style="western"><surname>Klein</surname><given-names>E</given-names> </name><collab>CDC MInD-Healthcare Program</collab></person-group><article-title>Associations between phone mobility data and COVID-19 cases</article-title><source>Lancet Infect Dis</source><year>2021</year><month>05</month><volume>21</volume><issue>5</issue><fpage>e111</fpage><pub-id pub-id-type="doi">10.1016/S1473-3099(20)30725-8</pub-id><pub-id pub-id-type="medline">32946835</pub-id></nlm-citation></ref><ref id="ref84"><label>84</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>J</given-names> </name><name name-style="western"><surname>Talwalkar</surname><given-names>A</given-names> </name></person-group><article-title>FACT: a diagnostic for group fairness trade-offs</article-title><source>ICML&#x2019;20: Proceedings of the 37th International Conference on Machine Learning</source><year>2020</year><access-date>2026-01-19</access-date><fpage>5264</fpage><lpage>5274</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/10.5555/3524938.3525426">https://dl.acm.org/doi/10.5555/3524938.3525426</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Details on the differentiability and regression analysis of fairness method errors.</p><media xlink:href="ojphi_v18i1e78235_app1.pdf" xlink:title="PDF File, 257 KB"/></supplementary-material></app-group></back></article>