Abstract
Background: The generalizability of clinical research hinges on robust study designs, which include the recruitment and maintenance of a representative study population. This study examines the evolution of the demographic characteristics of 329,038 participants who enrolled and participated in The All of Us Research Program (AoURP), a decentralized study aimed at representing the diversity of the United States.
Objective: The primary objectives of this study were to assess alterations in the demographic composition of the cohort at different protocol stages within AoURP, while analyzing completion rates and timeframes for survey and substudy completion. Additionally, we examined how participant interactions with the program impacted engagement and survey responses.
Methods: We conducted a longitudinal analysis of the AoURP data, tracking changes in demographic composition, completion rates, and completion times for surveys and substudies. Comparative analyses were performed to assess differences in engagement and survey completion based on sociodemographic characteristics of participants involved in postenrollment study components.
Results: The sociodemographic composition of the cohort that participated in the postenrollment study (eg, optional components) differed significantly from that of the recruited population. The proportion of self-identified White participants increased by 21.2%, whereas the proportion of Black or African American participants decreased by 12.18% (P=.02). Participants who identified as White (n=93,614, 52.7%) and NonHispanic (n=109,279, 42.21%) were more engaged compared to those identifying as Black or African American (n=10,887, 15.76%), Asian (n=4274, 38.72%), or Hispanic (n=12,530, 20.7%; P=.006). Participants’ response times to study surveys and completeness varied across all demographic groups (P<.001). Furthermore, those identifying as White skipped fewer survey questions (1.19) compared to those identifying as Black or African American (1.40) or other racial and ethnic identities (P<.001).
Conclusions: The AoURP dataset serves as an exceptional resource for investigating diverse public health concerns. However, the longitudinal analysis of participant-level data underscores a significant skew in population diversity, suggesting the need for targeted strategies to enhance engagement and retention across all groups. Ensuring diversity in the cohort is essential for maintaining the study’s representativeness and the broad applicability of its findings.
doi:10.2196/56803
Keywords
Introduction
There is a history of lack of racial, ethnic, and gender diversity in health studies [
, ]. This lack of diversity can lead to issues related to the generalizability of research findings and equity in health care. However, lack of plurality has decreased in the last three decades, partially due to regulatory and policy efforts within government agencies, such as the National Institutes of Health [ ], Food and Drug Administration [ ], and Department of Health Services [ ], that have sought to enhance minority participation in clinical research as well as identified scientific need [ ]. As the demographic makeup of the United States is becoming more pluralistic, diverse study populations are necessary to have representative and translatable observations. Much of the effort to increase minority representation in research has focused on the recruitment of diverse study populations. However, the recruited population is not always representative of the population that is ultimately studied, due to engagement rates being different across sociodemographic groups. This has been demonstrated in remote observational studies, where there is less direct interaction with participants compared to studies which require in-person visits [ ]. Poor engagement can lead to study failure due to reduced sample sizes, causing loss of power or imbalanced study populations. Prior research has shown that remote-only studies demonstrate different engagement rates based on demographic features including disease presence or absence, age, race or ethnicity, and recruitment methods [ - ]. When the engagement rates vary across different populations, there is a risk that the effective study population (ie, the population with longitudinal data) becomes unrepresentative of the originally recruited cohort.The All of Us Research Program (AoURP) was launched in May 2018, with an aim to recruit more than 1 million participants living in the United States to accelerate health research and precision medicine. AoURP has specifically focused on recruiting demographic categories that have historically been underrepresented in biomedical research (UBRs) and has largely succeeded in this objective through partnership with more than 340 recruitment sites nationwide [
]. During the first year of the AoURP, 80% of recruited participants self-identified as belonging to one or more UBR populations [ ]. In this study, we explore long-term engagement within AoURP by exploring participation in optional components of the AoURP study (eg, surveys and substudies that could be performed post enrollment), the time that it takes participants to complete optional surveys, and survey response completeness. We hypothesize that this information, combined with data on specific interactions of participants with the AoURP can be used to improve and develop strategies that promote sustained engagement across the diverse demographics. This study primarily assesses changes in cohort demographics across AoURP protocol stages, completion rates for surveys and substudies, and the influence of participant interactions on engagement and responses. These insights provide actionable strategies for sustaining diverse, representative cohorts.Methods
Study Design and Participants
This study is a longitudinal study that analyzes data from participants in the AoURP, encompassing individuals aged 18 and older residing in the United States, irrespective of race, ethnicity, sex, gender, or sexual orientation [
, ]. Data from the 329,038 participants in the September 2021 AoURP data freeze were analyzed. The AoURP protocol has evolved since its launch in 2018 ( ). At present, after consenting, participants are given the option to share their electronic health record (EHR) data, provide a biosample, and answer 3 core surveys (ie, The Basics, Overall Health, and Lifestyle) [ ]. Participants can also respond to additional optional health surveys including Healthcare Access (HCA), Personal Medical History (PMH), and Family Medical History (FMH), at any time and in any order after enrollment (Table S1 in ) [ ]. In addition to these components, AoURP included two additional optional substudies, at the time of analysis: (1) a COVID-19 survey and (2) a “bring your own device” (BYOD) Fitbit study, which allows participants to share data from any Fitbit wearable device (Google Inc, Mountain View, CA) owned by them. The AoURP also collects additional data, such as physical measurements; however, our analysis is restricted to recruitment and engagement characteristics based on participant-provided information (eg, health and demographic surveys) and BYOD Fitbit data.The core surveys included information about the participants’ self-reported demographics, which was used to explore differences in participation. Specifically, we used responses to the questions on (1) date of birth; (2) What was your biological sex assigned at birth; (3) Which categories describe you (race); (4) Which categories describe you (ethnicity); and (5) What is your annual household income (from all sources) [
]. Based on the All of Us (AoU) Researcher Workbench, we categorized age into three bands: younger (18‐44 years), middle aged (45‐64 years), and older adults (≥65), and household income into 4 bands (<$50,000, $50,000-$100,000, $100,000-$200,000, >$200,000). For other multiple-choice questions, we included categories with at least 1.8% respondents. This resulted in the following categories: male and female for biological sex at birth; White, Black or African American, and Asian for race; and NonHispanic and Hispanic for ethnicity. These categories were used to assess changes in cohort demographics during the course of the study, with the primary outcomes under investigation being (1) engagement (ie, number of participants who remain engaged by completing optional components after enrollment), (2) response time (ie, time between invitation and completion of optional components), and (3) completeness of response (ie, how often participants chose to answer specific questions). Understanding these changes is essential for examining shifts in participant demographics and engagement over time, which helps identify trends and biases, ensuring the cohort representativeness and development of targeted strategies for improved engagement and retention.
Data Processing
For the primary analysis, we built an engagement cohort consisting of participants who successfully enrolled in the study and responded to the demographic questions using the AoU Researcher Workbench [
]. Since all components of the study were not launched at the same time (eg, BYOD Fitbit and the sharing of electronic health records using Sync4Science [ ] were added later; see for the timeline), we conducted a secondary analysis using enrollment dates extracted from the observation table to evaluate the effect of timing of enrollment on these components. Specifically, this secondary analysis included individuals who enrolled after Sync4Science launch (August, 2018) and after the BYOD Fitbit protocol being launched (November, 2020). For all cohorts, activities where participants completed the surveys were considered as “active/engaged” events. All survey questions included a “prefer not to answer” option. Survey completeness was analyzed across participants by counting the number of questions answered as “prefer not to answer”.Statistical Analysis
Various statistical techniques were used for the three different analyses. Whenever possible we favored the use of nonparametric statistical tests and resampling techniques to avoid relying on distributional assumptions. These included the Kruskal-Wallis test (instead of analysis of variance), Mann-Whitney tests for pairwise group comparisons), bootstrapping for computation of confidence intervals, and permutation tests for calculating statistical significance. For all analyses, a significance level of P<.05 was used with Bonferroni correction for multiple testing [
].To analyze engagement across self-reported age, sex at birth, race, household income, and ethnicity groups, we compared the proportion of participants responding in each demographic group between the core survey and optional components using a χ2 test for homogeneity.
Participants’ response time, as measured by the number of days between joining the study and completing (or joining) an optional component, was analyzed using the nonparametric Kruskal-Wallis one-way analysis of variance for each optional component separately, as the data was not normally distributed [
]. To further understand the variability between each group, pairwise differences between groups were analyzed using the Python language (version number 3.9; Python Software Foundation), including the statannotation package (version 0.4.2) [ ], along with the with Mann-Whitney integrated statistical test and Bonferroni for multiple testing correction [ , ].To better understand the relationship between various self-reported demographic variables, a linear mixed model (LMM) was used to analyze the response time across the 3 surveys concomitantly. To improve the fit of the model, response time data was log-transformed before feeding into the model. All self-reported demographic variables were fitted into the model to evaluate the effect produced by each predictor. Age was treated as a continuous variable, while other variables were fitted as categorical variables. The coefficient of relative change for each variable was reported by the LMM model, and the log-transformed relative change between each group was back-transformed to a percentage change in response time across groups for all the categories, using formula 100(ecoefficient–1). To further evaluate confidence intervals, bootstrapping was used. To account for the repeated measurements of each subject, the bootstrap confidence intervals for the LMMs were constructed based on 1000 iterations of the following couple of steps. First, we obtained a bootstrap sample from the original data by sampling with replacement at the subject level (ie, assuming that our dataset contains data from N subjects, we grouped the repeated measurements of each subject as a “subject block,” and then sampled with replacement N “subject blocks” out of the N blocks in the original dataset.) Second, for each bootstrap sample, we fit a LMM and computed the coefficient of relative change of each variable as described above. Finally, we computed 95% CI for the coefficients of relative change using the percentile interval method, where the lower and upper confidence bounds correspond to the 0.025 and 0.975 quantiles of the distribution of coefficients generated across the 1000 bootstrap iterations.
We independently examined outliers in the response time for optional surveys to check if these outliers were the cause of any systematic differences. Outliers comprised any individual who had a response time less than the first quartile (Q1)–1.5×IQR or greater than Q3+1.5×IQR. To evaluate the differences, we computed the proportion of outlier participants for each demographic group and then conducted a χ2 test (test of homogeneity) for testing the equality of outlier participants proportion across different demographic groups.
Completeness of response was evaluated by counting the number of questions where a participant responded “prefer not to answer.” First, the nonparametric Kruskal-Wallis test was used to evaluate variability across different groups (since the data was not normally distributed). However, due to ties in the completeness of the response variable, we adopted a permutation test to assess differences between the groups [
]. The permutation test was conducted by comparing the observed value of the ANOVA F-statistic computed in the original data (Fobs) against the permutation null distribution of the F-statistic computed on permuted versions of the data (F*), where the group labels were randomly shuffled. The test was based on 1000 data permutations and the permutation P value was computed as the proportion of times the permuted F-statistic was greater or equal to the observed test statistic (ie, P=(1+ sum{i=1..B indicator{F*>=Fobs})/(1+ B), for B=1000). Due to the large sample set, it is possible that we may be detecting very small effect sizes with very high statistical significance across different categories. To assess the effect size, 95% CI were computed using nonparametric bootstrapping for the pairwise group mean difference for each group [ ].Ethical Considerations
This study used deidentified data from the All of Us Research Program, which obtained institutional review board (IRB) approval and informed consent from participants for secondary research [
]. The original data collection was approved by the AoURP IRB [ ], and no additional IRB approval was required for this secondary analysis. According to the All of Us Responsible Conduct of Research training, analyses using deidentified data on the Researcher Workbench do not require a separate IRB review [ ], as the research involves no direct interaction with participants. All analyses adhered to ethical principles, ensuring privacy and confidentiality as outlined by the AoURP.Results
Population Characteristics
Data from 329,078 consented participants were made available as part of the version 5 data release [
] by AoURP. Of the total consented participants, 329,038 (99.98%) completed all core surveys and were considered successfully enrolled in the program. Additionally, 311,253 participants (94.59%) consented to share all their electronic health records; including retrospective medical records from before the AoURP study was launched. Post completion of core surveys, participants had the option to complete three optional health surveys. A significantly smaller proportion completed these three optional surveys: (1) HCA-37.84% (n=124,528/329,078), (2) FMH-117,693 (35.76%), and (3) PMH-113,830 (34.59%) ( ). Both the COVID survey and the BYOD Fitbit data collection were launched at later stages of the study ( ). A total of 99,701 (30.3%) people participated in the first COVID survey and 11,615 (3.52%) participated in BYOD Fitbit data sharing. Participants who enrolled before the BYOD Fitbit data collection protocol was launched were less likely to consent to share their BYOD Fitbit data compared to those who enrolled after the protocol launch (3.51% vs 6.51%, respectively).The AoURP aims to recruit persons from the UBR populations, and recruitment was successful, with 24.36% (n=80,154) of the participants responding to the core surveys with a self-reported race of Black or African American and Asian, while 53.98% (n=177,615) self-reported as White. We compared the distribution of ages and race to the 2020 US Census data [
], which showed the enrichment of self-identified races other than White, as well as middle-aged individuals, in the AoU cohort ( ) at baseline. Additionally, more participants reported being female (Figure S1a in ) and Hispanic ethnicity (Figure S1b in ) compared to the census.(However, this initial enrichment of people self-identifying as belonging to one or more UBR populations diminished as participants engaged with optional components of the protocol. Distributions across demographic variables for each component of the study is shown in . Moreover, a total of 1416 (0.4%) deaths were reported by the end of version 5 data release.
Groups | Core survey | (n=329,038), n (%)HCA | survey (n=124,528), n (%)Family history (n=117,693), n (%) | Medical history (n=113,830), n (%) | EHR | (n=271,421), n (%)Covid survey (n=99,701), n (%) | Fitbit (n=11,615), n (%) |
Age (years) | |||||||
18‐44 | 108,254 (32.9) | 35,740 (28.7) | 33,307 (28.3) | 32,328 (28.4) | 88,212 (32.5) | 23,130 (23.2) | 3,984 (34.3) |
45‐64 | 121,415 (36.9) | 42,090 (33.8) | 39,545 (33.6) | 38,019 (33.4) | 100,697 (37.1) | 33,698 (33.8) | 4,321 (37.2) |
≥65 | 99,369 (30.2) | 46,698 (37.5) | 44,841 (38.1) | 43,483 (38.2) | 82,512 (30.4) | 42,971 (43.0) | 3310 (28.6) |
Sex | |||||||
Male | 125,034 (38) | 42,090 (33.8) | 39,663 (33.7) | 38,475 (33.8) | 104,497 (38.5) | 33,699 (33.8) | 3,357 (28.9) |
Female | 199,726 (60.7) | 81,441 (65.4) | 77,207 (65.6) | 74,559 (65.5) | 163,395 (60.2) | 65,304 (65.5) | 8,200 (70.6) |
Race | |||||||
White | 177,681 (54) | 93,645 (75.2) | 89,094 (75.7) | 86,625 (76.1) | 144,124 (53.1) | 78,963 (79.2) | 9,698 (83.5) |
African American | 69,098 (21) | 10,958 (8.8) | 10,004 (8.5) | 9,448 (8.3) | 58,898 (21.7) | 7,278 (7.3) | 581 (5) |
Asian | 11,187 (3.4) | 4,234 (3.4) | 4,002 (3.4) | 3,870 (3.4) | 8,685 (3.2) | 2,892 (2.9) | 372 (3.2) |
Ethnicity | |||||||
NonHispanic | 258,953 (78.7) | 109,336 (87.8) | 103,570 (88) | 100,398 (88.2) | 212,523 (78.3) | 89,631 (89.9) | 10,674 (91.9) |
Hispanic | 60,543 (18.4) | 12,577 (10.1) | 11,652 (9.9) | 11,041 (9.7) | 51,027 (18.8) | 7,876 (7.9) | 767 (6.6) |
Income (USD) | |||||||
<50,000 | 139,183 (42.3) | 38,106 (30.6) | 35,661 (30.3) | 34,035 (29.9) | 116,982 (43.1) | 27,916 (28) | 2,532 (21.8) |
50,000-100,000 | 59,556 (18.1) | 32,626 (26.2) | 31,071 (26.4) | 30,279 (26.6) | 48,041 (17.7) | 27,518 (27.6) | 3,554 (30.6) |
100,000-200,000 | 46,394 (14.1) | 28,766 (23.1) | 27,422 (23.3) | 26,750 (23.5) | 36,913 (13.6) | 24,526 (24.6) | 3,520 (30.3) |
>200,000 | 20,071 (6.1) | 12,702 (10.2) | 12,005 (10.2) | 11,724 (10.3) | 15,851 (5.84) | 10,469 (10.5) | 1,371 (11.8) |
aCore Survey at enrollment.
bHCA: Health Care Access.
cEHR: electronic health records.
Differences in Engagement
The relative shift in the proportions of participants is reported as percentage changes. We observed an increase of 21.2% in the proportion of self-identified White participants and correspondingly a decrease of Black or African American respondents by 12.18% in the HCA compared to the Core surveys. Similarly, the proportion of participants who identified as NonHispanic (of all races) increased by 9.1%, to 87.8% (n=109,336) of the total respondent pool in the HCA survey, while people self-identifying as Hispanic decreased by 8.33% to 10.1% of the respondent pool (
and ). To explore the engagement further, we compared the percentage of each self-identified demographic group that continued in each optional component of the study (ie, optional surveys, EHR consent, COVID Surveys, and BYOD Fitbit) and observed that the engagement varied across most components except the BYOD Fitbit ( ). The engagement with optional surveys was 3 times higher among people who identified as White, with 52.7% (n=93,614) of White participants who enrolled and also completed the optional surveys (HCA), compared to Black or African American participants (n=10,887, 15.8%), and nearly twice as high as those identifying as Asian (n=12531, 38.7%) (P<.001). Similarly, 42.2% (n=109,279) of those self-identifying as nonHispanic (of all races) engaged with optional surveys as compared to 20.7% (n=12,531) of self-identified Hispanic participants (P<.006). 63.4% (n=12,681) of participants who self-identified with high household income engaged in the optional surveys compared to 27.4% (n=38,190) of self-identified with low household income (P<.001). A total of 47% (n=46,743) of older adults (aged >65 years) engaged with optional surveys compared to 34.64% (n=37,504) and 33.05% (n=40,096)of the middle-aged and younger groups, respectively. Engagement was 40.8% (n=81,460) for participants who self-identified as female at birth engaged in the optional surveys compared to 33.66% (n=42,137) of self-identified male at birth. However, neither self-reported age nor sex at birth were statistically significant indicators of engagement using the χ2 test.The first COVID survey showed large differences in participation across age (P=.02), race (P<.001), household income (P<.001), and ethnicity (P<.001), but not by sex at birth (P=.45), Older individuals participated at higher rates, with 43.2% (n=42,964) of self-reported older adults compared to 27.8% (30,099) of middle-aged and 21.3% (n=25,841) of younger participants completing the first survey. Self-reported White participants (n=79,048, 44.5%) participated at more than 4 times higher levels than self-identified Black or African American participants (n=7281, 10.54%) and nearly twice those identifying as Asian (n=2880, 26.09%). Self-identified high-income participants (n=10,485, 52.4%) participated at more than approximately 2.5 times higher levels than self-identified low household income participants (n=27,876, 20%). Similarly, participants self-identifying as nonHispanic (n=89,552, 34.59%) engaged at almost 3 times higher rates than those identifying as Hispanic (n=7876, 13.01%).
We did not observe any statistically significant differences in engagement by self-reported age, race, ethnicity, or sex in participation rates for the BYOD Fitbit substudy or the consent to share EHR data.
Groups | Core | (n)HCA | survey, (%)P value | Family history (%) | P value | Medical history, n (%) | P value | EHR | , n (%)P value | Covid survey, n (%) | P value | Fitbit | , n (%)P value |
Age groups (years) | .22 | .18 | .19 | .94 | .02 | .99 | |||||||
18‐44 | 121,317 | 40,156 (33.1) | 37,244 (30.7) | 36,274 (29.9) | 96,932 (79.9) | 25,841 (21.3) | 4,489 (3.7) | ||||||
45‐64 | 108,268 | 37,461 (34.6) | 35,295 (32.6) | 33,996 (31.4) | 91,162 (84.2) | 30,099 (27.8) | 3,898 (3.6) | ||||||
>65 | 99,453 | 46,743 (47) | 44,853 (45.1) | 43,461 (43.7) | 82,745 (83.2) | 42,964 (43.2) | 3,282 (3.3) | ||||||
Sex | .41 | .41 | .42 | .81 | .45 | .59 | |||||||
Male | 125,184 | 42,187 (33.7) | 39,683 (31.7) | 38,431 (30.7) | 105,530 (84.3) | 33,674 (26.9) | 3,380 (2.7) | ||||||
Female | 199,658 | 81,460 (40.8) | 77,068 (38.6) | 74,672 (37.4) | 162,122 (81.2) | 65,288 (32.7) | 8,186 (4.1) | ||||||
Race | <.001 | <.001 | <.001 | 0.40 | <.001 | .18 | |||||||
White | 177,637 | 93,615 (52.7) | 89,174 (50.2) | 86,509 (48.7) | 135,359 (76.2) | 79,048 (44.5) | 9,770 (5.5) | ||||||
African American | 69,084 | 10,915 (15.8) | 10,017 (14.5) | 9,395 (13.6) | 63,557 (92) | 7,254 (10.5) | 553 (0.8) | ||||||
Asian | 11,040 | 4,272 (38.7) | 3,974 (36) | 3,875 (35.1) | 8,578 (77.7) | 2,881 (26.1) | 364 (3.3) | ||||||
Ethnicity | .006 | .006 | .006 | .41 | .001 | .23 | |||||||
NonHispanic | 258,895 | 109,254 (42.2) | 103,558 (40) | 100,451 (38.8) | 208,152 (80.4) | 89,578 (34.6) | 10,615 (4.1) | ||||||
Hispanic | 60,535 | 12,531 (20.7) | 11,623 (19.2) | 11,078 (18.3) | 55,147 (91.1) | 7,870 (13) | 787 (1.3) | ||||||
Income (USD) | .001 | .001 | .001 | .98 | <.001 | .31 | |||||||
<50,000 | 139,380 | 38,162 (27.4) | 35,639 (25.6) | 34,092 (24.5) | 134,125 (96.2) | 27,946 (20.1) | 2,537 (1.8) | ||||||
50,000- 100,000 | 59,563 | 32,700 (54.9) | 31,134 (52.3) | 30,222 (50.7) | 55,078 (92.5) | 27,554 (46.3) | 3,556 (6) | ||||||
100,000- 200,000 | 46,381 | 28,728 (61.9) | 27,411 (59.1) | 26,729 (57.6) | 42,346 (91.3) | 24,512 (52.9) | 3,520 (7.6) | ||||||
>200,000 | 20,008 | 12,685 (63.4) | 12,043 (60.2) | 11,785 (58.9) | 18,181 (90.9) | 10,494 (52.5) | 1,377 (6.9) |
aCore survey at enrollment.
bHCA: Health Care Access.
cEHR: electronic health records.
dFitbit was launched later in the study (see
). For a comparison of engagement rates between participants who joined before and after the launch of the Fitbit component (see Table S5 in ).Response Time to Optional Surveys
In addition to differences in engagement, we also analyzed how long it took for participants to participate in optional surveys as measured by the time in days between enrollment and completion of the optional component. Given that only the three optional surveys (ie, HCA, PMH, and FMH) were available, and the COVID Surveys and BYOD Fitbit, were unavailable at the launch of the study, we were only able to explore the delay for these surveys. We found differences in the response time based on self-reported age, sex at birth, race, ethnicity, and household income (all Ps<.001 based on Kruskal-Wallis one way analysis of variance). These results are summarized in Table S2 in
. Overall for the optional surveys, response times ranged from 0 to 1342 days with a mean of 247.9 (SD 233) days and median value was 141 (IQR 314.55) days. Self-reported older adults (median 117, IQR 213) took less time to respond to optional surveys compared to younger (median 135, IQR 251) and middle aged (median 138, IQR259) participants. People who identified as male at birth (median 120, IQR 221) took less time compared to people who self-identified as female at birth (median 133, IQR249). People who self-identified as White completed surveys sooner after enrollment (median 115, IQR 202) in comparison to those who self-identified as Black or African American (median 202, IQR 340), and Asian (median 153, IQR 295). People who identified as nonHispanic of any race (median 123, IQR 225) responded sooner to optional surveys as compared to people who self-identified as Hispanic (of all races) (median 188, IQR 335). People self-identified with high household income (median 119, IQR 195) took less time compared to people with low household income (median 150, IQR 279). We performed a secondary analysis to determine pairwise differences using the nonparametric Mann-Whitney test to illustrate which groups had quicker participation ( ). The linear mixed effect results presented in provided valuable insights into the relationships between various demographic variables and their impact on response time. The table displays the estimated effects in terms of coefficients and relative percentage changes, with the use of bootstrap for accuracy. When examining age groups, individuals aged 45-64 years exhibited a 3.04% increase in response time compared to the reference group (18-44 years), while those aged ≥65 years displayed a decrease of −2.76%. Gender differences were also notable, with women showing a 5.65% increase in response time compared to men. In terms of race, African Americans exhibited a substantial 33.5% increase in response time compared to White individuals, while Asians and other racial groups also displayed significant differences. Ethnicity played a role as well, with Hispanics experiencing an 18.41% increase ( ) . Finally, income levels showed a consistent trend, with higher income brackets associated with decreased response times, highlighting the influence of socioeconomic factors on response behavior. These results offer important insights into the complex interplay between demographics and response times, shedding light on potential areas for further investigation and intervention. We conducted a secondary analysis to check if outliers were responsible for any significant difference and found no difference in the demographics. Distribution across demographic data is shown in Table S3 in . We also explored the interaction effect (see Table S4 in ) between other demographic features and household income. The study found that the effect of age, gender, race, and ethnicity on response time was moderated by income. The response time was slower for people with higher income in all age groups, except 18-44 years. The difference was most pronounced for people aged ≥65 years. The interaction effect between gender and income was only significant for people who identify as neither male nor female. The interaction effect between race and income was significant for all race groups except White. The interaction effect between ethnicity and income was only significant for people who were not in plurality.Variables | Model | |
Coefficient | Percent change (95% CI) | |
Age (years) | ||
18-44 | 0 | 0 |
45-64 | 0.03 | 3.04 (2.99-3.09) |
>65 | –0.028 | –2.76 (–2.78 to –2.69) |
Gender | ||
Male | 0 | 0 |
Female | 0.055 | 5.65 (5.55-5.65) |
Race | ||
White | 0 | 0 |
African American | 0.289 | 33.50 (33.40-33.57) |
Asian | 0.169 | 18.41 (18.30-18.54) |
Ethnicity | ||
NonHispanic | 0 | 0 |
Hispanic | 0.169 | 18.41 (18.26-18.49) |
Income (USD) | ||
<50,000 | 0 | 0 |
50,000-100,000 | –0.045 | –4.79 (–4.91 to –4.56) |
100,000-200,000 | –0.066 | –6.39 (–6.45 to –6.31) |
>200,000 | –0.05 | –4.88 (–4.96 to –4.78) |
Completeness of Responses
To evaluate the frequency of skipped questions in each survey (survey completeness), we counted skipped questions and compared those across self-reported demographic groups (Table S5 in
). Skipped questions were relatively rare, with approximately 124,000 skipped questions out of 34 million answered; on average 329,038 participants skipped a mean of 1.32 (SD 0.74) questions per survey. We reported the results of “completeness of response” based on core surveys only, given that all the questions in the optional surveys were already answered except for one participant, who skipped one question. Using a nonparametric test, we observed differences in completeness for all self-reported demographic groups in the three core surveys. Mean values are reported for missing data (ie, skipped questions). Middle-aged participants skipped more questions ( 1.72 questions) compared to 1.54 for younger and 1.6 for seniors (P<.001). Participants who identified as female at birth skipped fewer questions (1.58) compared to participants who identified male at birth (1.99) (P<.001). Participants who self-reported as Black or African American race (1.86) and people self-reporting nonHispanic ethnicity of any race (1.71) skipped more questions compared to people of self-identified White (1.62) and Asian (1.62) race, and those who self-reported Hispanic (1.57) ethnicity of any race (P<.001). Participants with high household income skipped fewer questions (1.58) compared to participants with low household income (1.61) (P<.001). Also, we conducted pairwise group mean differences to further estimate the effect size which returned significant results .Secondary Analysis on Effect of Timing of Enrollment on Engagement
As all protocols were not launched at the same time, time of enrollment may have affected the engagement behavior of the participants who joined before and after the launch of the specific study components. We conducted a secondary analysis to evaluate if there were any differences in engagement between individuals who enrolled before and after the launch of BYOD Fitbit and Sync4Science separately. We found no significant differences in the demographic groups. Distribution across demographic variables for pre- and postlaunch of BYOD Fitbit (Table S6 in
), and for Sync4Science is shown in Table S7 in .Discussion
Principal Finding
The utility and translatability of studies is dependent on the study being carried out on a representative population. The AoURP has enrolled an extremely large cohort that aims to be representative of the diversity of the broader US population [
]. However, the data collected over time is not as representative as the enrolled population, with differences in engagement across demographic groups. This pattern is observed in both, the demographics that remained engaged and the degree to which participants completed optional substudy components and how quickly they responded [ ]. These variations are driven by socioeconomic, cultural, and structural factors, underscoring the need for tailored strategies to mitigate participation disparities [ ]. Addressing these disparities is essential to maintain the cohort’s representativeness, which directly impacts the validity and generalizability of the findings [ ].Comparison to Prior Work
This study builds on prior research by highlighting demographic engagement disparities in longitudinal cohort studies, an issue that has been well-documented in other population-based research efforts [
, ]. Previous studies have reported similar challenges, such as the underrepresentation of minority groups and the overrepresentation of higher socioeconomic status in voluntary participation metrics [ ]. However, unlike earlier research, this study leverages the unique scale and diversity of the AoURP cohort to provide deeper insights into how engagement patterns vary over time and across demographic groups. By emphasizing socioeconomic and structural drivers of engagement disparities, this work highlights the need for target interventions necessary for retaining participants and ensuring equitable representation [ ].Strengths and Limitations
Our study benefits from a large, diverse cohort and longitudinal data, providing valuable insights into participation patterns. However, there are limitations that must be acknowledged. First, there may be biases due to self-selection and varying dropout rates among demographic groups. Overrepresentation of more engaged participants with higher socioeconomic status in optional surveys could skew health outcome estimates [
]. Second, the high dropout rates observed among underrepresented groups over time pose a challenge to maintaining sample diversity, which could limit the generalizability of findings to the broader US population [ ]. Third, reliance on self-reported data, which may introduce reporting biases that affect the accuracy of the findings [ ]. To address these limitations, efforts should focus on implementing strategies such as over-recruiting populations with higher attrition rates, developing tailored retention programs, and validating self-reported data through objective measures such as linked medical records [ ]. These steps can enhance the reliability and applicability of the study’s findings, ensuring their broader impact.Future Directions
Based on our findings, decentralized studies such as AoURP should prioritize interventions that address engagement disparities. Over-recruiting populations with higher attrition rates and implementing targeted outreach strategies for underrepresented groups can enhance participation. Importantly, these strategies should be codesigned with input from participants to address specific barriers effectively [
]. Addressing challenges such as the digital divide, cultural differences, and logistical barriers is critical to fostering equity in research participation [ , ]. Innovative approaches such as bidirectional engagement where participants actively contribute to study design and receive meaningful feedback could further strengthen retention. Additionally, leveraging digital health tools and adaptive technologies can facilitate participant engagement while addressing socioeconomic barriers [ ].Conclusion
This study underscores the importance of maintaining representativeness in large, decentralized cohort studies such as AoURP. By identifying participation disparities and their underlying drivers, we provide actionable insights to improve cohort retention and engagement strategies. Future research should continue to explore and validate these interventions to ensure equitable and impactful scientific discoveries that benefit all segments of the population.
Acknowledgments
Research reported here was supported by the Office of the Director of the National Institutes of Health under award number U24OD023176. This work would not be possible without the partnership of the participants in the AoURP. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
Data Availability
Data are available only to researchers through the All of Us cloud-based Research Workbench. The complete code used for the analysis is available through a GitHub code repository [
].Conflicts of Interest
All authors listed are employed at Sage Bionetworks. AP is an employee of Boehringer Ingelheim, USA.
References
- Ford JG, Howerton MW, Lai GY, et al. Barriers to recruiting underrepresented populations to cancer clinical trials: a systematic review. Cancer. Jan 15, 2008;112(2):228-242. [CrossRef] [Medline]
- Bolen S, Tilburt J, Baffi C, et al. Defining “success” in recruitment of underrepresented populations to cancer clinical trials: moving toward a more consistent approach. Cancer. Mar 15, 2006;106(6):1197-1204. [CrossRef] [Medline]
- NIH policy and guidelines on the inclusion of women and minorities as subjects in clinical research. NIH Grants and Funding. Oct 30, 2023. URL: http://grants.nih.gov/grants/funding/women_min/guidelines_amended_10_2001.htm [Accessed 2025-02-28]
- Food and drug administration modernization act (FDAMA) of 1997. US Food and Drug Administration. Nov 3, 2018. URL: https://www.fda.gov/regulatory-information/selected-amendments-fdc-act/food-and-drug-administration-modernization-act-fdama-1997 [Accessed 2023-10-30]
- Medicare clinical trial policies. Centers for Medicare & Medicaid Services. Oct 30, 2023. URL: https://www.cms.gov/medicare/coverage/approved-facilities-trials-registries/clinical-policies?redirect=/ClinicalTrialPolicies [Accessed 2025-02-28]
- Yancey AK, Ortega AN, Kumanyika SK. Effective recruitment and retention of minority research participants. Annu Rev Public Health. 2006;27:1-28. [CrossRef] [Medline]
- Pratap A, Neto EC, Snyder P, et al. Indicators of retention in remote digital health studies: a cross-study evaluation of 100,000 participants. NPJ Digit Med. 2020;3:21. [CrossRef] [Medline]
- Kolovson S, Pratap A, Duffy J, Allred R, Munson SA, Areán PA. Understanding participant needs for engagement and attitudes towards passive sensing in remote digital health studies. Int Conf Pervasive Comput Technol Healthc. May 2020;2020:347-362. [CrossRef] [Medline]
- The All of Us Research Program Investigators. The “All of Us” Research Program. N Engl J Med. 2019:668-676. [CrossRef]
- Who can join? All of Us Research Program. Jul 15, 2024. URL: https://www.joinallofus.org/who-can-join [Accessed 2025-02-28]
- All of us research program protocol summary. All of Us Research Program. Jul 15, 2024. URL: https://allofus.nih.gov/sites/default/files/All_of_Us_Protocol_Overview_Mar_2022.pdf [Accessed 2025-02-28]
- Survey explorer. All of Us Research Hub. Apr 25, 2022. URL: https://www.researchallofus.org/data-tools/survey-explorer/ [Accessed 2025-02-28]
- Researcher workbench. All of Us Research Hub. Feb 14, 2022. URL: https://www.researchallofus.org/data-tools/workbench/ [Accessed 2025-02-28]
- Helping patients share EHR data with researchers. Sync For Science. Apr 25, 2024. URL: http://syncfor.science/ [Accessed 2025-02-28]
- Bland JM, Altman DG. Multiple significance tests: the Bonferroni method. BMJ. Jan 21, 1995;310(6973):170. [CrossRef] [Medline]
- Kruskal WH, Wallis WA. Use of ranks in one-criterion variance analysis. J Am Stat Assoc. Dec 1952;47(260):583-621. [CrossRef]
- Statannotations 0.7.1. PYPI. URL: https://pypi.org/project/statannotations/ [Accessed 2025-02-28]
- Mann HB, Whitney DR. On a test of whether one of two random variables is stochastically larger than the other. Ann Math Statist. Mar 1947;18(1):50-60. [CrossRef]
- Wilcoxon F. Individual comparisons by ranking methods. Biometrics Bulletin. Dec 1945;1(6):80. [CrossRef]
- Efron B, Tibshirani RJ. An introduction to the bootstrap. CRC Press; 1994. ISBN: 9780429246593
- All of Us institutional review board (IRB). All of Us Research Program | NIH. Apr 8, 2022. URL: https://allofus.nih.gov/about/who-we-are/institutional-review-board-irb-of-all-of-us-research-program [Accessed 2025-11-25]
- Do I need my project reviewed by the all of us institutional review board (IRB) in order to access this data using the researcher workbench? All of Us Research Hub. Nov 25, 2024. URL: https://www.researchallofus.org/faq/do-i-need-my-project-reviewed-by-the-all-of-us-institutional-review-board-irb-in-order-to-access-this-data-using-the-researcher-workbench/ [Accessed 2025-02-25]
- Release notes. All of Us Research Program. URL: https://support.researchallofus.org/hc/en-us/sections/360006773231-Release-Notes [Accessed 2025-02-25]
- Data: try out our new way to explore data. US census bureau. Apr 29, 2022. URL: https://www.census.gov/data [Accessed 2025-02-28]
- All of us research program. All of Us Research Program | NIH. Jun 1, 2020. URL: https://allofus.nih.gov/future-health-begins-all-us [Accessed 2024-11-25]
- Diversity and inclusion. All of Us Research Program | NIH. Jun 24, 2020. URL: https://allofus.nih.gov/about/diversity-and-inclusion [Accessed 2025-11-25]
- Ochieng CA, Minion JT, Turner A, Blell M, Murtagh MJ. What does engagement mean to participants in longitudinal cohort studies? A qualitative study. BMC Med Ethics. Jun 24, 2021;22(1):77. [CrossRef] [Medline]
- Mapes BM, Foster CS, Kusnoor SV, et al. Diversity and inclusion for the All of Us research program: a scoping review. PLoS ONE. 2020;15(7):e0234962. [CrossRef] [Medline]
- Teague S, Youssef GJ, Macdonald JA, et al. Retention strategies in longitudinal cohort studies: a systematic review and meta-analysis. BMC Med Res Methodol. Nov 26, 2018;18(1):151. [CrossRef] [Medline]
- Costello L, Dare J, Dontje M, Lambert C, Straker L. Applying the 4Ps of social marketing to retain and engage participants in longitudinal cohort studies: generation 2 Raine study participant perspectives. BMC Med Res Methodol. Nov 5, 2022;22(1):288. [CrossRef] [Medline]
- Murray AL, Xie T. Engaging adolescents in contemporary longitudinal health research: strategies for promoting participation and retention. J Adolesc Health. Jan 2024;74(1):9-17. [CrossRef] [Medline]
- Ferris EB, Wyka K, Evenson KR, et al. Recruitment and retention strategies for community-based longitudinal studies in diverse urban neighborhoods. JMIR Form Res. Mar 24, 2021;5(3):e18591. [CrossRef] [Medline]
- Population studies: diverse cohorts (milestone 1.C). NIH Institute of Aging. URL: https://www.nia.nih.gov/research/milestones/epidemiology-population-studies/population-studies-diverse-cohorts [Accessed 2025-02-28]
- Durocher K, Boparai N, Jankowicz D, Strudwick G. Identifying technology industry-led initiatives to address digital health equity. Digit Health. 2021;7:20552076211056156. [CrossRef] [Medline]
- Participant engagement and health equity workshop. National Institutes of Health (NIH). Sep 3, 2015. URL: https://www.nih.gov/allofus-research-program/participant-engagement-health-equity-workshop [Accessed 2024-11-25]
- Sage bionetworks. Github. URL: https://github.com/Sage-Bionetworks/allofus_engagement [Accessed 2025-02-25]
Abbreviations
AoU: All of Us |
AoURP: All of Us Research Program |
BYOD: bring your own device |
EHR: electronic health record |
FMH: Family Medical History |
HCA: Health Care Access |
IRB: institutional review board |
LMM: linear mixed model |
PMH: Personal Medical History |
UBR: underrepresented in biomedical research |
Edited by Taiane de Azevedo Cardoso; submitted 26.01.24; peer-reviewed by Ahmed Hassan, Faisal Albalwy, Vaishnavi Nagesh; final revised version received 02.01.25; accepted 17.01.25; published 19.03.25.
Copyright© Vijay Yadav, Elias Chaibub Neto, Megan Doerr, Abhishek Pratap, Larsson Omberg. Originally published in the Interactive Journal of Medical Research (https://www.i-jmr.org/), 19.3.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Interactive Journal of Medical Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.i-jmr.org/, as well as this copyright and license information must be included.