key: cord-0722332-3uxbt8cg authors: Nelson, Lyndsay A; Pennings, Jacquelyn S; Sommer, Evan C; Popescu, Filoteia; Barkin, Shari L title: A 3-Item Measure of Digital Health Care Literacy: Development and Validation Study date: 2022-04-29 journal: JMIR Form Res DOI: 10.2196/36043 sha: 3a45d9c47b621007f4469f605ba8a462944a78be doc_id: 722332 cord_uid: 3uxbt8cg BACKGROUND: With increased reliance on digital health care, including telehealth, efficient and effective ways are needed to assess patients’ comfort and confidence with using these services. OBJECTIVE: The goal of this study was to develop and validate a brief scale that assesses digital health care literacy. METHODS: We first developed an item pool using existing literature and expert review. We then administered the items to participants as part of a larger study. Participants were caregivers of children receiving care at a pediatric clinic who completed a survey either on the web or over the telephone. We randomized participants into development and confirmatory samples, stratifying by language so that exploratory factor analysis and confirmatory factor analysis could be performed with separate samples of participants. We assessed the scale’s validity by examining its associations with participants’ demographics, digital access, and prior digital health care use. RESULTS: Participants (N=508) were, on average, aged 34.7 (SD 7.7) years, and 89.4% (454/508) were women. Of the 508 participants, 280 (55.1%) preferred English as their primary language, 157 (30.9%) preferred Spanish, and 71 (14%) preferred Arabic; 228 (45%) had a high school degree or less; and 230 (45.3%) had an annual household income of 1, factor loadings >0.4, Cronbach α>.75, and variance explained >0.40 were used as the criteria to evaluate the items retained in the full factors and to attempt to create a reduced factor [25] [26] [27] . Once the full and reduced factors were finalized, CFA was performed on the confirmatory sample using the items suggested by the EFA. The CFA was conducted using robust maximum likelihood estimation to test the goodness of fit between the theorized factor structure suggested by the EFA and the confirmatory sample data set. The robust estimation was used because of the Likert-type ordinal responses of the items and does not assume multivariate normality of the items. A constraint value of 1 was placed on 1 item in the factor as is common in modeling analyses with a defined scale. Goodness of fit for the CFA was assessed evaluating the absolute fit, incremental fit, and parsimonious fit of the full and reduced factors [28] . The absolute fit criteria to conclude good fit between the proposed factor structure and the data included nonsignificant chi-square values, root mean square error of approximation, and standardized root mean square residual <0.08 [29] . Incremental fit criteria included the comparative fit index and nonnormed fit index >0.95. Parsimonious fit was indicated by adjusted chi-square (c2/df)<3.0 [30] . To assess the reliability of the full and reduced factors, Cronbach α was computed for the EFA and CFA. Composite reliabilities, calculated according to the weighted Ω formula from McDonald [31] , were also calculated for the CFA because of concerns that Cronbach α may be inappropriate for use in structural equation modeling [32] . The variance explained was also reported for the EFA, and the average variance explained (AVE) values were calculated for the CFA with the recommended critical value >0.50 indicating that the factors explained enough of the variance in the construct [33] . We assessed the scale's validity by examining its associations with participants' demographics (ie, gender, race, ethnicity, language, education, and income), digital access, and prior digital health care use. We used the Spearman ρ for continuous variables and the Kruskal-Wallis test for categorical variables. Participants were, on average, aged 34 (Table 2) . Chi-square and independent samples 2-tailed t tests revealed no significant differences between the development and confirmatory samples for any demographic characteristics. b Separate databases were used for English-, Spanish-, and Arabic-speaking participants, and the data were combined for analysis. Race and ethnicity were collected with a single item; the response options included White, Black, Asian, Middle Eastern, Hispanic, Native American, Native Hawaiian, other race or ethnicity, and prefer not to answer. In the English and Arabic database, participants could select all options that applied. Participants were coded as multiple if they selected more than one race and/or ethnicity, except for Middle Eastern+White, which was coded as Middle Eastern. Because of incorrect configuration, the race and ethnicity item was not enabled as a check-all item in the Spanish-speaking database (ie, these participants could only check 1 race or ethnicity). Of the 151 participants who completed the Spanish survey, 144 (95.4%) selected Hispanic, 6 (4%) selected White, and 1 (0.6%) selected Black. c GED: General Educational Development. To evaluate the full digital health care literacy score, all 6 items were initially entered into the EFA, and 1 factor was extracted with an eigenvalue >1 and factor loadings between 0.45 and 0.96. Cronbach α was excellent at .89, and the variance explained was high at 62% (Table 3) . Next, a reduced factor was created by eliminating items from the full factor one at a time and evaluating the resulting factor loadings and variance explained. Items were eliminated based on correlations with other items >0.90 with conceptual overlap (1 item; I can install applications/programs... overlapped with I can use applications/programs...) and the lowest factor loadings (2 items). On the basis of these criteria, 3 items were retained in the reduced scale (Table 3) , with a resulting excellent reliability (Cronbach α=.90) and a high variance explained (78%). Both the full and reduced factors were evaluated with CFA ( Table 3 ). The factor loadings for the full factor ranged from 0.38 to 0.95, with model statistics of χ 2 9 =16.0; P=.07. All the fit statistics exceeded the criteria for good fit between the proposed factor structure and the data. The reliability was excellent, with coefficient Ω=0.88, and the AVE was high at 0.62. The reduced scale also had excellent CFA fit, with factor loadings between 0.82 and 0.94. The chi-square value was 0, meaning the model was saturated (equal number of parameters and df). This also means that the factor was perfectly parsimonious (adjusted chi-square value of 0). All the fit statistics well exceeded the criteria for good fit between the proposed factor structure and the data. The reliability was excellent, with coefficient Ω=0.90, and the AVE was high at 0.75. Because of the excellent fit of the reduced factor, we focus on this version of the scale and its validity in the following sections. The DHLS score was negatively associated with age (ρ=-0.164; P<.001), and positively associated with both education (ρ=0.139; P=.005) and income (ρ=0.379; P<.001). There was not a significant association with gender, H 1 =1.267; P=.26. The overall model for language was significant, H 2 =117.115; P<.001. Arabic speakers had lower scores than English and Spanish speakers, and Spanish speakers had lower scores than English speakers (Figure 2 ). The model for race was also significant, H 5 =93.167; P<001. Middle Eastern participants had lower scores than all other racial groups, and Hispanic participants had lower scores than all groups, except Middle Eastern (Figure 2 ). Among the 508 participants in the study, 25 (4.9%) did not own a smartphone, 191 (37.6%) did not own a laptop computer, and 401 (78.9%) did not own a desktop computer; in addition, 30 (5.9%) did not have internet access at home and 43 (8.5%) said that their internet connection was not good. We found significant associations between most of our digital access items and the DHLS score such that participants who did not have digital tool access had lower scores than those who did. Specifically, participants who did not own a smartphone or a laptop computer had lower digital literacy scores ( Figure 2 ). However, there was not an association between desktop computer ownership and scores (Figure 2 ). Having a more stable network connection to use the internet at home (ρ=0.343; P<.001) and to use a cell phone data plan (ρ=0.312; P<.001) were both associated with higher scores. Nearly half of the participants (211/508, 41.5%) had never accessed a health app, and 35.8% (182/508) were not signed up for the patient portal. Most (341/508, 67.1%) had not used video telehealth to obtain care for their children. Participants who had never used a health app had lower digital health care literacy scores than those who had (Figure 2 ). In addition, participants who were not signed up for the patient portal had lower scores than those who were signed up. Participants who had not used video telehealth to obtain care for their children had lower literacy scores than those who had (Figure 2 ). Among those who had, there was a positive association between the ease of scheduling the visit and their DHLS score (ρ=0.279; P=.001). Among those who had not used video telehealth, perceived difficulty of scheduling a visit was associated with lower scores (ρ=0.459; P<.001). Given the increased reliance on digital technologies during the COVID-19 pandemic, it is critical that we understand which patients are and are not equipped for this shift in health care delivery. Without gauging patients' confidence in skills for using telehealth and similar health care technologies, we risk exacerbating health disparities [14, 17] . We developed the DHLS, a scale designed to measure an individual's digital health care literacy, and validated it among a diverse sample of caregivers of young children. Overall, the scale had strong psychometric properties, and the reduced version of the scale performed just as well as the full version, supporting its continued and more efficient use. Participants with lower digital health care literacy had less experience with digital health care and were less likely to own digital tools. In addition, those with less education, with lower income, and people of color had lower digital health care literacy. To our knowledge, this is one of the first tools intended to measure confidence with the skills necessary for using digital health care services, including telehealth. The Digital Health Literacy Instrument is another scale designed to measure digital health literacy; however, the items are complex and highly specific (eg, When typing a message [e.g., to your doctor, on a forum, or on social media such as Facebook or Twitter] how easy or difficult is it for you to clearly formulate your question or health-related worry); furthermore, the scale is long (21 items), which could lead to attrition among users with less education or literacy. The DHLS is a brief, 3-item assessment developed among a racially and socioeconomically diverse sample, and it measures the basic skills necessary for using digital health services. Of note, we focused our application of the scale in this paper on telehealth; however, it may have application to other types of digital tools. This is supported by our study, which validated the scale against the use of similar technologies (eg, whether patients had used a health app and whether they were signed up for a patient portal). Although the reduced 3-item scale is easier to administer, we encourage other researchers to use either the reduced or full scale (the latter includes additional items about digital skills, more broadly, beyond video chat) to explore other applications. Overall, we found similar associations between participants' characteristics and DHLS scores as other studies reporting on similar digital literacy tools. For example, having less education and lower income has previously been associated with lower eHealth Literacy Scale scores [34] . Although lower telehealth literacy was associated with older age, aligning with other studies examining digital literacy [35, 36] , the effect was very small. This is likely due to the limited variation in age among our sample: all participants were caregivers of children aged <13 years, with the average caregiver age being only 34.7 (SD 7.7) years. In our study, we found that Hispanic and Middle Eastern participants had lower digital health care literacy than White and Black participants, and Middle Eastern participants had significantly lower scores than Hispanic participants. A similar pattern emerged when looking at language such that Arabic speakers had the lowest digital health care literacy, followed by Spanish speakers, and then English speakers. The findings highlight the importance of examining differences in race and language by unique groups rather than collapsing groups into non-White or non-English. Our scale could be applied as a brief assessment in clinical settings when assessing individuals' ability to use telehealth. If a participant identifies as more digitally fluent, they may be a strong candidate for telehealth and can receive subsequent instructions for setting up a visit. However, if they identify as being less digitally fluent, resources can be provided to help that individual be better equipped for a visit. Several organizations are exploring solutions to help those with lower technology literacy prepare for telehealth appointments. For example, at VUMC, a medical student-led volunteer initiative was started to help patients set up and test devices for their telehealth appointments [37] . Students used a standardized telephone script to guide patients with downloading the proper software and understanding what to expect for the visit [37] . Another approach in Harris County, Texas, included a nonphysician staff member reaching out to ensure that patients had the proper technology and had resolved issues before the appointment [38] . Primary care practices at University of California San Francisco started an outreach program to all patients aged >65 years with scheduled visits to walk them through setting up and using the video platform app [39] . Although such initiatives have had success with preparing patients for telehealth, they are extremely time and resource intensive; a screening tool such as the DHLS could help identify only those who are most in need of assistance, thereby increasing efficiency and effectiveness. Another approach could be to simply ask patients whether they need extra help setting up a telehealth visit; however, this may have the opposite effect and lead to missing patients who do require help. That is, it is possible that some individuals may not know they need the help, especially if they have never had a telehealth visit. By using items that target the basic skills necessary to use digital tools, the scale could help to accurately identify patients who are unaware that they need assistance. Moreover, some patients may feel uncomfortable communicating that they need help. We hope that this tool provides a respectful approach for identifying those patients who require assistance. With respect to research, the DHLS could be used as a way to help describe the digital literacy of the sample and determine whether there was representation from low digital literacy communities. It could also be useful to assess whether the use rates or efficacy of a digital technology or program were related to digital literacy. In general, we hope that the scale is included in other studies, whether for descriptive purposes, as a predictor, or as a covariate, to broaden our understanding of its applications and how it functions. This study includes several limitations. First, these data were collected cross-sectionally; therefore, we cannot draw conclusions regarding causality. It is possible that having lower digital health care literacy leads to a lower likelihood of accessing digital health care services or vice versa. Similarly, as part of a cross-sectional study, we are limited in our ability to propose a cutoff score for determining who requires additional assistance with digital health care; however, certain study designs can effectively answer this question. For example, a future study might administer the DHLS and then attempt to conduct a telehealth visit with all participants. By examining the difference in scores between those who were and were not successful with completing the visit, we could determine a cutoff score that helps identify the likelihood of being able to successfully carry out a telehealth visit in typical circumstances. In this study, one of our goals was to explore associations between the scale and a variety of barriers to telehealth, of which scheduling a visit was one; however, scheduling a visit is likely reflecting both clinic-level and patient-level characteristics and therefore we recommend interpreting this association with some degree of caution. All participants were caregivers of children and recruited from a clinic in Middle Tennessee, which limits generalizability to other populations and other regions; however, we enrolled a racially, ethnically, and socioeconomically diverse sample of participants. We developed the items such that they can theoretically be used widely with different types of individuals, and we encourage researchers to use and validate the scale in other populations. Although the DHLS was negatively correlated with age, the sample was, on average, of younger age (mean age 34.7, SD 7.7 years), and it will be especially important to see how the scale functions with older populations who tend to experience more barriers to digital health [40] [41] [42] . In addition, although we included participants who spoke English, Spanish, and Arabic, there were likely confounding differences among the groups, and we did not use a sample-matching approach to ensure comparability of participant characteristics among languages. To consider the scale validated for all languages, a future study would need to include large numbers of individuals who spoke each language with sufficient heterogeneity and representation of participant characteristics. Relatedly, because our scale items were originally written and derived by English speakers, it is possible that the lower mean scores observed within the Spanish and Arabic groups could have been at least partially caused by intrinsic bias. Full validation within each language would help to confirm whether intrinsic bias was present. Finally, patients were not included in the development of the scale; it is possible that the inclusion of patient input could have strengthened it. Widespread adoption of telehealth by clinicians and patients alike has the potential to revolutionize health care delivery, improving both quality of life and clinical outcomes. However, as part of this quest, we must consider those patients who may not have the digital access or skills to use telehealth-in many cases, these are the same patients who tend to have worse outcomes. A screening tool such as the DHLS can be a useful resource to identify patients who require additional assistance to effectively engage with telehealth. Validating the scale among other patient populations and in other settings will support the scale's ultimate utility to reduce health care inequities. Covid-19 and health care's digital revolution Digital technologies in the public-health response to COVID-19 The role of telehealth during COVID-19 outbreak: a systematic review based on current evidence Experts say the 'new normal' in 2025 will be far more tech-driven, presenting more big challenges. Pew Research Center The clinical effectiveness of telehealth: a systematic review of meta-analyses from 2010 to 2019 Telemedicine in the OECD: an umbrella review of clinical and cost-effectiveness, patient experience and implementation A systematic review of economic analyses of telehealth services using real time video communication | iss. 4 | e36043 Home broadband adoption, computer ownership vary by race, ethnicity in the U. S. Pew Research Center Digital divide persists even as lower-income Americans make gains in tech adoption Patient characteristics associated with telemedicine access for primary and specialty ambulatory care during the COVID-19 pandemic Disparities in telemedicine access: a cross-sectional study of a newly established infrastructure during the COVID-19 pandemic Socioeconomic and demographic disparities in the use of telemedicine for ophthalmic care during the COVID-19 pandemic RESEARCHRacial and socioeconomic disparities in access to telehealth Impact of the digital divide in the age of COVID-19 Good intentions are not enough: how informatics interventions can worsen inequality How states are expanding broadband access. Pew Trust Strategies for digital care of vulnerable patients in a COVID-19 world-keeping in touch Evaluating the need to address digital literacy among hospitalized patients: cross-sectional observational study Development of the digital health literacy instrument: measuring a broad spectrum of health 1.0 and health 2.0 skills eHEALS: the eHealth literacy scale Assessing competencies needed to engage with digital health services: development of the eHealth literacy assessment toolkit Validation of the Computer Literacy Scale (CLS) Testing reliability and validity of the eHealth Literacy Scale (eHEALS) for older adults recruited online lavaan: an R package for structural equation modeling What is coefficient alpha? An examination of theory and applications Applied Multivariate Statistics for the Social Sciences Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Structural Equation Modeling Multidisciplinary J Alternative ways of assessing model fit Application of confirmatory factor analysis to the study of self-concept: first-and higher order factor models and their invariance across groups Composite reliability in structural equations modeling Evaluating structural equation models with unobservable variables and measurement error | iss. 4 | e36043 eHealth literacy: predictors in a population with moderate-to-high cardiovascular risk CRIC Study Investigators. Mobile health (mHealth) technology: assessment of availability, acceptability, and use in CKD Mobile health usage, preferences, barriers, and eHealth literacy in rheumatology: patient survey study Technology literacy as a barrier to telehealth during COVID-19 How to make telemedicine more equitable Addressing equity in telemedicine for chronic disease management during the Covid-19 pandemic. NEJM Catalyst Innovations in Care Delivery Aging barriers influencing mobile health usability for older adults: a literature based framework (MOLD-US) Older adults' views on eHealth services: a systematic review of scientific journal articles PERSPECTIVE: older adults' adoption of technology: an integrated approach to identifying determinants and barriers This work was supported by the National Institute on Minority Health and Health Disparities (3U54MD010722-05S1 and COEUS M2003140). LAN was supported by a career development award from the National Heart, Lung, and Blood Institute (K12HL137943). The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. The authors thank the Vanderbilt Pediatric Primary Care Clinic at Vanderbilt University Medical Center and the participants for their contributions to this research. None declared. ©Lyndsay A Nelson, Jacquelyn S Pennings, Evan C Sommer, Filoteia Popescu, Shari L Barkin. Originally published in JMIR Formative Research (https://formative.jmir.org), 29.04.2022. This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.