key: cord-0426750-i5z8mvtf authors: Chasukwa, M.; Choko, A. T.; Muthema, F.; Nkhalamba, M. M.; Saikolo, J.; Tlhajoane, M.; Reniers, G.; Dulani, B.; Helleringer, S. title: Collecting mortality data via mobile phone surveys: a non-inferiority randomized trial in Malawi date: 2022-03-02 journal: nan DOI: 10.1101/2022.03.02.22271441 sha: 3570cc28328d5f1255cd3d78129c97ce2a588871 doc_id: 426750 cord_uid: i5z8mvtf Introduction: Despite the urgent need for timely mortality data in low-income and lower-middle-income countries, mobile phone surveys rarely include questions about recent deaths. There are concerns that such questions might a) be too sensitive, b) prompt negative/adverse reactions among respondents, c) take too long to ask and/or d) generate unreliable data. We assessed the feasibility of mortality data collection during mobile phone surveys. Methods: We conducted a non-inferiority trial among a random sample of mobile phone users in Malawi. Participants were allocated 3:1 to an interview about recent deaths in their family (treatment group) or about their economic activity (control group). In the treatment group, half of the respondents completed a short mortality questionnaire, focused on information necessary to calculate recent mortality rates, whereas the other half completed an extended questionnaire that also included questions about symptoms and healthcare use. The primary trial outcome was the cooperation rate. Secondary outcomes included the completion rate, self-reports of negative feelings and stated intentions to participate in future interviews. We also documented the amount of time required to collect mortality data, and we explored the quality of death reports. Results: The difference in cooperation rates between treatment and control groups was 0.9 percentage points (95% CI = -2.3, 4.1), which satisfied the non-inferiority criterion. Similarly, the mortality questionnaire was non-inferior to the control questionnaire on all secondary outcomes. Collecting mortality data required approximately 2 to 4 additional minutes per reported death, depending on the inclusion of questions about symptoms and healthcare use. More than half of recent deaths elicited during mobile phone interviews had not been reported to the national civil registration system. Conclusion: Including mortality-related questions in mobile phone surveys appears acceptable and feasible. It might help strengthen the surveillance of mortality trends in low-income and lower-middle-income countries with limited civil registration systems. What is already known? • In many low-income and lower-middle-income countries, civil registration systems only record a fraction of all deaths. The excess mortality associated with health crises is thus not known in near real-time. • Mobile phone surveys are increasingly common in low-income and lower-middle-income countries. They could help fill mortality-related data gaps, but there are concerns that asking questions about recent deaths over the phone might be too sensitive, might take too long, and/or might generate unreliable data. What are the new findings? • In a randomized trial conducted with mobile phone users in Malawi, asking questions about recent deaths was not less acceptable than asking questions about economic activity and household livelihoods. • Few participants reported experiencing negative feelings during the interview, and these feelings were temporary. • More than half of the deaths reported during mobile phone interviews had not been previously registered with the national civil registration system. • Including questions about recent deaths in mobile phone surveys appears feasible and acceptable. • It might help strengthen the surveillance of mortality trends in low-income and lowermiddle-income countries with limited civil registration systems. In many low-income and lower-middle-income countries (LLMICs), few deaths are registered with competent administrative authorities in a timely manner [1] [2] [3] . In such settings, mortality statistics are only updated every few years, after periodic household surveys or censuses are conducted. The retrospective data these inquiries generate allow estimating mortality levels for periods stretching a few months to a few years prior to data collection [5] , but they do not allow monitoring mortality in near real-time. Household surveys an censuses are also often postponed or canceled during epidemics or other crises, due to elevated risks of disease transmission or heightened safety concerns [6] . In most LLMICs, the data available to track the mortality impact of health crises are obtained from more partial data collection systems. For example, the counts of deaths routinely reported during epidemics such as Ebola or COVID-19 often only include the deaths that occur among those who have been diagnosed with the disease. Yet, the coverage of testing services is very limited in LLMICs [7] . Some COVID-19 cases may be lost to follow-up after diagnosis, and surveillance systems predominantly record deaths that occur at health facilities [8], even though many deaths occur at home [9, 10] . Epidemics also indirectly affect mortality, for example by disrupting health services [11, 12] . As a result, reported counts of deaths only include a fraction of the excess mortality caused by an epidemic [15] . This might foster perceptions that global health crises (e.g., have "spared" LLMICs, and it might preclude countries from effectively advocating for resources required to mitigate the impact of such crises [16] . In the medium to long-term, strengthening civil registration systems is the key intervention required to address this data gap [17, 18] . Increasing death registration is thus a key indicator of . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint progress towards the sustainable development goals (SDGs). In the short term however, interim data sources are needed to better understand how ongoing crises affect population health in LLMICs. The World Health Organization (WHO) has recommended setting up rapid mortality surveillance systems to monitor the death toll of the COVID-19 pandemic [19] , for example by extracting data about deaths from registers maintained by health facilities, or by tallying the number of burials taking place at local cemeteries. New data sources (e.g., satellite images, social media) might also help keep track of increasing numbers of deaths [20, 21] . Rapid mortality surveillance systems have documented the previously unknown extent of excess mortality caused by the COVID-19 pandemic in several LLMICs [16] . Unfortunately, setting up such surveillance is complex. It requires significant investments in data acquisition and management. Baseline data that pre-date health crises might be difficult to obtain, for example if historical records were not properly kept at a health facility or cemetery. Behaviors related to the management of deaths (e.g., where and when to carry out a burial) might also change during epidemics and other crises, due to lockdowns and restrictions on mobility or access to health facilities. In such contexts, short-term fluctuations in mortality documented by surveillance systems might be due to behavioral changes as well as excess mortality. Finally, rapid mortality surveillance systems have been predominantly established in large urban areas [22] . Small towns and rural areas are less often included in such initiatives. Additional tools are needed to better document the total impact of health crises on mortality in LLMICs with limited death registration. We investigated the use of mobile phone surveys (MPS) to collect mortality data. MPS are surveys in which participants are recruited and interviewed entirely by phone [23] . They are increasingly conducted in LLMICs, for example to monitor risk factors for non-communicable diseases [24] or to document the effects short-term fluctuations in economic activity, schooling or healthcare use [25, 26] . They present several advantages over . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint other modes of data collection. Because they are implemented remotely, they remove the need for physical interactions between data collectors and participants, and can be sustained during epidemics [27] . They can also be repeated more frequently than household surveys or censuses, since they require less complicated logistics. Finally, owing to the rapid penetration of mobile phones in rural areas of LLMICs, they might allow monitoring mortality trends outside of large cities more conveniently than other rapid surveillance systems. Several MPS have already included questions about mortality. During the Ebola epidemic in west Africa, a survey in Monrovia (Liberia) asked randomly selected phone owners about recent deaths in their households [6] . Since the beginning of the COVID-19 pandemic, surveys in India and Bangladesh have monitored recent household deaths [28, 29] . In particular, a question about recent COVID deaths was included in a large phone survey, which allowed estimating that the death toll related to the COVID-19 pandemic in India might be 6-7 times larger than officially reported [30] . Such examples of mortality-related MPS remain however isolated. There are concerns that questions about recent deaths might be too sensitive to ask during phone interviews, thus prompting high levels of refusals to participate, or to answer specific questions. In some instances, questions about deaths might upset respondents or trigger distress. Whereas such reactions can be addressed when interviews occur in person (e.g., through signs of empathy), they may be more difficult to mitigate remotely. Questions about deaths might also take too long to administer. There are strict recommendations to keep the duration of MPS short [31] , and eliciting mortality data about deaths with sufficient detail requires time. Existing mortality-related MPS have thus only included limited ascertainments of deaths and their circumstances. Finally, MPS interviewers might not be able to probe and cross-check answers provided by respondents . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint as thoroughly as during in-person interviews. As a result, the mortality data generated by MPS might be of lesser quality. We report results from a randomized trial of the collection of mortality data via MPS recently conducted in Malawi. Trial design: we compared responses to mobile phone interviews about recent deaths (treatment group), to those obtained in mobile phone interviews about economic activity (control group). The control questionnaire has been used in multiple LLMICs since the beginning of the COVID-19 pandemic, with high levels of reported acceptability [25] . We conducted a noninferiority randomized trial [32] , in which we tested whether the treatment questionnaire did not lead to unacceptably worse response patterns than the control questionnaire. Within the treatment group, participants were randomly assigned to either a short or an extended questionnaire on deaths. The short mortality questionnaire only ascertained information required to measure and triangulate recent mortality rates. The extended mortality questionnaire also included questions about symptoms and circumstances of reported deaths. Study setting: the trial was conducted in Malawi, a low-income country in southeastern Africa, with a population of approximately 18 million inhabitants. Despite a legal obligation to register vital events, and the possibility to register events in decentralized offices and locations, few deaths are reported to the National Registration Bureau, i.e., the authority in charge of vital records. As a result, data series based on death registration, typically used to monitor shortterm fluctuations in mortality in high-income countries, are not available in Malawi [33] . The most recent household surveys that collected mortality data in Malawi occurred between 2015 . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. We recruited study participants among users of Malawi's two major mobile networks through random digit dialing (RDD). We worked with Sample Solutions, a firm specializing in the provision of RDD samples. Sample Solutions first generated a list of phone numbers at random using the country's numbering scheme. They then matched this list to a global registry of authorized network subscribers, and excluded numbers that could not be located. Finally, a team of 15 interviewers contacted the selected numbers to introduce the study, assess the eligibility of mobile phone users who were reached, and ask for their consent to participate in interviews. We implemented sampling quotas based on age, gender and regional residence. We formed 18 sampling strata based on these characteristics, and enrollment continued in each stratum until the quota was filled or until progress towards this quota stopped. All interviews were conducted in local languages (Chichewa, Chitumbuka, ChiYao) or in English, depending on mobile users' . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint preferences. If a mobile user did not speak any of these languages, he/she was not included in the trial. Randomization: after interviewers successfully contacted a mobile phone number, they introduced themselves to the user and stated that they were conducting a survey about the impact of the COVID-19 pandemic on several aspects of the lives of Malawians. They asked the phone user if they were interested in learning more about the study. If so, they assessed the user's eligibility by asking 4 questions about gender, age and residence. Mobile users who did not meet the age criterion and those whose sampling stratum was already filled were told that they were not eligible for the study. Eligible mobile users were randomized into the treatment and control groups in a 3:1 ratio. At that time, interviewers read consent scripts and sought oral consent from mobile users. In both study groups, the consent scripts included similar explanations of how phone numbers were selected, and statements about that the interview would take approximately 15 minutes to complete. Consent scripts in both groups also mentioned that participants would receive a small amount of airtime (1,200 Malawian Kwachas, or 1.5 US Dollar) as a token of appreciation upon completion of the interview. Since the main themes of the interview differed between treatment and control groups, the description of other study procedures included differed between the two groups (see supporting documents). Study interviewers were unaware of the mobile users' assignment to the treatment or control group until the users' sampling stratum had been determined and their eligibility had been confirmed. Randomization was stratified by sampling stratum. It was conducted using random numbers generated in Stata 15.1. Within the treatment group, randomized allocation to the short vs. extended mortality questionnaires was conducted in a similar manner in a 1:1 ratio. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint 1 0 Outcomes: The primary outcome of the trial was the cooperation rate [39] . It was calculated among mobile users randomized to either the treatment or control group, and it was defined as the number of completed interviews divided by the sum of completed interviews, call-backs, partial interviews and refusals. Secondary outcomes included a) the completion rate, i.e. the number of completed interviews divided by the sum of completed interviews and partial interviews and b) the proportion of respondents who stated their willingness to participate in future interviews of varying durations. We also measured c) the proportion of respondents who self-reported that some of the survey questions made them upset. Sample size: We used standard formulae for the determination of sample size in non-inferiority trials [40] . Based on preliminary pilot data, we assumed that a cooperation rate of 86 per cent in the control group and we set a non-inferiority margin of 5 percentage points. Our null hypothesis was that the cooperation rate in the treatment group is lower than in the control group by more than 5 percentage points. Our alternative hypothesis was that the cooperation rate in the treatment group is lower than in the control group by less than 5 percentage points. To test this hypothesis with 80% power and alpha = 0.05, we required at least 1,194 mobile subscribers in the treatment group and 398 mobile subscribers in the control group. Sample size calculations were performed using the SampleSize4ClinicalTrials package in R. socio-economic characteristics such as their age, marital status and educational level, as well as their household's access to water and electricity. In the treatment group, respondents were asked to list deaths that had occurred among members of their households since the beginning of 2021, and to indicate the survival status of their parents and (maternal) siblings. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) 1 1 Additional questions were included for respondents who had reported that one of their parents or siblings had died since 2019. Respondents allocated to the short mortality questionnaire were solely asked to state the age at death, date of death (month and year) and status of their relative's death in the national civil registration system. Respondents allocated to the long mortality questionnaire were also asked if their deceased relative(s) had experienced symptoms commonly observed in COVID-19 patients (e.g., cough, fatigue, loss of taste/smell), and whether they had sought healthcare in the weeks prior to death. Finally, they were asked to indicate the place of death (e.g., at the hospital or at home) and place of burial of their deceased relatives. In the control group, instead of questions about deaths, respondents were asked to report their recent economic activities, to list the sources of livelihood of their household, and to describe how they manage their finance (e.g., ownership of bank accounts). In both study groups, we asked respondents about their reactions to the interview. We adapted debriefing questions used in post-disaster surveys (e.g., terror attacks) to identify respondents who might have experienced negative feelings during the interview [41] . Participants who selfreported negative feelings were asked to classify these feelings on a 3-point scale ranging from "a little upset" to "very upset". They were also asked if they were still upset by these questions at the end of the interview, or if they were "okay now". Finally, we asked interviewers to indicate if they noticed signs of emotional distress during the interview (e.g., crying, long silences, voice alterations). We offered psychological support to all respondents who self-reported ongoing negative feelings, or who were identified by interviewers as having displayed signs of emotional distress. If a participant indicated being interested in such a service, the study interviewer transmitted the participant's phone number to an on-call clinical psychologist. This practician then called the . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint referred participant to a) assess their psychological well-being, b) provide phone-based counseling, and c) if needed, provide required referrals for further follow-up. Finally, respondents in both groups were provided details about who to contact in case of questions, and were asked if they wanted to receive information about SARS-CoV-2. If so, study interviewers read a short script that included details about modes of transmission, symptoms and strategies to prevent infection/transmission. They also indicated how to get additional information through resources provided by the Ministry of Health (e.g., hotline, social media). Data collection: Interviewers were trained for six days. This included a review of recruitment, screening and consent procedures, and explanations of survey questions in each study group. Training also included sessions about the use of tablets for mobile data collection, and the development of skills to conduct of sensitive interviews and detect potential signs of distress among mobile respondents. After mock interviews, interviewers conducted a 3-day pilot. The trial was then conducted between September 21 st , 2021 and October 12 th , 2021. All data were collected on tablets using surveyCTO. We recorded how interviewers administered key parts of a randomly selected subset of interviews, including the consent statement, the ascertainment of deaths in the treatment group, and the assessment of negative reactions to the interview. Study supervisors listened to these audio-files to monitor compliance with study instructions, and provided feedback to study interviewers on the basis of these recordings. Study supervisors also placed follow-up calls to a 1 in 15 sample, which included a) all participants who experienced negative feelings during the interview, and b) a random sample of respondents who did not experience such reactions. During these follow-up calls, supervisors re-assessed participants' reactions to the interview. They also offered psychological support to . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. Statistical methods: We conducted pre-specified intent-to-treat analyses of trial data. We measured primary and secondary outcomes according to participants' assigned study groups, then we computed the differences in those outcomes between treatment and control groups. We calculated two-sided confidence intervals around these differences [42] , and we assessed the position of these confidence intervals relative to the pre-specified non-inferiority margin (-5 percentage points). In computing confidence intervals, we assumed unequal variances in estimates of proportions between the two study groups. We explored whether respondents with recent deaths among their relatives experienced negative feelings more frequently. We reported the number of participants who asked to obtain counseling and support from our on-call psychologist. In the treatment group, we described the amount of time required to collect data on mortality during mobile phone interviews. We estimated a linear regression model in which interview duration (in minutes) was the dependent variable, and predictors included a binary variable denoting the type of mortality questionnaire . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint (short vs. long), a categorical variable indicating how many recent deaths were reported by the respondent (none, one or two or more deaths) and an interaction term between those two variables. We investigated the quality of data on the characteristics of deaths reported during the study. We measured the proportion of reported deaths with missing data on age at death and month of death. We assessed whether these proportions varied by source of the death report (e.g., household deaths, parental deaths or siblings' deaths). Due to small sample sizes in some categories, we used exact methods to calculate confidence intervals [43] . We described the time series of deaths reported during the survey, by source of the death report. Finally, we measured the proportion of reported deaths that were also registered in the national civil registration system, by year of death and source of the death report. Study interviewers dialed more than 7,000 unique mobile numbers (figure 1). They reached 3,054 mobile users. However, 698 mobile users (22.8%) immediately indicated that they were not interested in the study. Only 5 mobile users (0.2%) were excluded due to language-related reasons. Among mobile users whose eligibility was assessed (n = 2,318), 69 did not meet the age-related inclusion criteria (2.9%), and 566 were excluded because their quota had already been filled (24.4%). In total, 1,683 mobile users were randomized to the treatment or control groups. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The distribution of interview results among randomized mobile users is presented in table A1. The overall cooperation rate was 92.2% (1552/1683). Seventy-eight mobile users refused to provide consent (4.6%). Thirty-two mobile users (1.9%) consented but opted to be called at another time to complete the interview ("call-backs"). Study interviewers could not reach them again despite multiple attempts. Finally, 21 mobile users discontinued their interview (1.2%). . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. members of the treatment who completed the interview did not self-report that they were upset by some of the interview questions (96.6%) vs. 367 out of 378 in the control group (97.1%). Finally, 94.7% of participants in the treatment group and 93.7% in the control group stated their intention to participate again in (hypothetical) similar interviews in the near future. The noninferiority criterion was met for all pre-specified secondary outcomes (figure 2). Referrals to psychological counseling: Fifty-one participants self-reported being upset by (some of) the interview questions. In the treatment group, 12 out of 40 participants who self-reported negative feelings indicated that they were "very upset" by some of the interview questions. Participants with a recent death among their relatives reported being "very upset" more frequently than other participants (figure A1). In the control group, all participants who selfreported negative feelings (n = 11) indicated being either "a little upset" or "moderately upset". Among respondents who self-reported negative feelings, 3 stated that they were still upset at the end of the interview (all in the treatment group). Study interviewers identified 6 additional participants who displayed signs of distress during the interview, but who did not self-report being upset (3 in the treatment group, and 3 in the control group). In follow-up calls with study supervisors that took place 1-2 days after the interview, a few respondents who did not initially self-report negative feelings indicated that some of the questions asked during the interview had made them upset (table A2, n=7). None of these respondents however reported that these negative feelings persisted at the time of their call with study supervisors. In total, 9 respondents were informed about the possibility to talk to an on-call clinical psychologist who would provide support and information. Three participants (2 in the treatment group, and 1 in the control group) . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint In this randomized trial in Malawi, a mortality-related questionnaire administered by mobile phone was non-inferior to a questionnaire about economic activity previously administered in several LLMICs, and recognized as highly acceptable. Cooperation and completion rates were high among respondents asked to answer questions about recent deaths in their households and families. The frequency of self-reported negative feelings was low (<3%), and it was not heightened in respondents who were asked mortality-related questions. Respondents who reported a recent family or household death during the interview experienced stronger negative feelings, but these feelings were transitory. The few respondents who opted to talk to an on-call clinical psychologist did not require further follow-up. In this trial, concerns that questions related to mortality might be too sensitive to ask during MPS were thus not realized. We included questions about mortality that emulated those asked during household surveys and censuses. Our questionnaires were thus more detailed than instruments included in previous MPS, which focused on specific types of deaths (e.g., Ebola or COVID deaths), or only ascertained deaths among members of the respondent's household. Asking questions about mortality required an additional 2-4 minutes per reported death, depending on the inclusion of questions about circumstances, symptoms and healthcare use around the time of death. This time investment concerns only a small group of survey participants, since >80% of respondents did not report any recent death in our survey. Mortality-related phone interviews generated data that appeared of good quality. Missing data on age at death, and date of death, were limited. They were more frequent in reported deaths of siblings, and least frequent in reported deaths of household members. The levels of missing data we observed were comparable to those observed in many household surveys conducted . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint in-person [44] . In this sample, the mortality questions generated reports of approximately 8.4 deaths per quarter over the past 3 years, with marked increases in the most recent time period (i.e., the first 3 quarters of 2021) when questions about household deaths were applicable. Future work should investigate the sample sizes required in mobile surveys to allow detecting short-term fluctuations in mortality. Our data indicate that 50-65% of the deaths that were reported during the mobile interview had not been previously registered with the National Registration Bureau, i.e., the administrative unit in charge of civil registration in Malawi. Conducting mortality-related MPS might thus supplement existing data collection systems. It might also document short-term fluctuations in mortality among population groups that might otherwise be frequently excluded from data collection (e.g., small towns and rural villages). Our trial has several limitations. First, our sample size was too small to assess data quality in more detail. We could not measure heaping in reported ages of siblings [45] , nor could we investigate the reporting of deaths on shorter time scales (e.g., weeks). Second, some of the outcomes we considered (e.g., negative feelings, death registration) were based on selfreported data. They might have been affected by social desirability biases. If the extent of such biases did not differ by trial group, our assessments of the non-inferiority of the mortality questionnaire are however unaffected. In addition, we implemented robustness checks (e.g., audio-recordings, supervisor follow-ups) to enhance the reliability of these data. Finally, we did not explore the acceptability of more extensive mortality questionnaires that might enable the attribution of causes of deaths [46] , nor did we investigate the measurement of child mortality through the collection of birth or pregnancy histories [47]. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The development of MPS as a reliable tool for mortality surveillance in LLMICs also requires investigating the selectivity of samples recruited by mobile phone. These samples exclude population members who do not have access to a phone. They disproportionately include younger and more urban respondents. In some settings, they predominantly include males, who are more likely to be phone owners. Patterns of phone ownership and utilization might also change over the course of an epidemic, for example if poor households are forced to sell mobile phones due to hardship, or are less able to afford charging their phone (and may not be reachable). Population in the rural areas may not be reachable because their phones would be off as a result of not charging the battery associated with erratic power supply. Statistical models that account for such selectivity should be developed. Despite limitations and ongoing research needs, our work in Malawi suggests that MPS are a potentially useful tool for mortality surveillance in LLMICs with limited civil registration systems. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint Notes: Values of the x-axis are expressed in percentage points. They are calculated as % treatment group minus % control group. Error bars represent two-sided 95% confidence intervals around the difference in proportions between study groups. The non-inferiority criterion is met when the confidence interval remains to the right of the non-inferiority margin (red vertical line). Similar results were obtained when calculating one-sided confidence intervals. NI = Non-inferiority. 7 n al . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint Figure 3 : Duration of treatment group interviews, by type of mortality questionnaire and recent deaths among respondents' relatives. Notes: error bars represent 95% confidence intervals obtained from linear regression models. 8 . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint Figure 4 : Exploratory assessments of the quality of reported data on recent deaths. Notes: In Panel A), error bars represent 95% exact confidence intervals. The difference in the likelihood of missing data on age at death by source of report was significant at the p<0.05 level. In Panel B), Q1 refers to the first quarter of a year, i.e., January, February and March. Two deaths reported to have occurred in October 2021 by respondents interviewed after October 1 st , 2021 were omitted from the plot. Household deaths were only elicited for the period between the start of 2021 and the survey data. In Panel C), the width of each bar is proportional to the number of deaths reported in each year. 9 , he r. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) Table A1 : distribution of study results, by study group Notes: "Busy/call-back" refers to respondents who consented to being interviewed, indicated that they would prefer being called-back at a later time, and could not be reached again before the completion of the study. Table A2 : concordance of self-reported data on negative feeling Notes: supervisor follow-ups were conducted 1-2 days after the initial interview. The supervisors were not aware of the answers provided about self-reported feelings during the initial interview; nor were they aware of the respondents' assignment to the control or treatment groups. . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint Figure A1 : Strength of self-reported negative feelings experienced during the interview, by reports of recent deaths and study group. Notes: respondents were first asked if any of the interview questions made them upset, and if so, they were asked to indicate the severity of their negative feelings. In the treatment group, the strength of negative feelings was associated with recent reports of deaths (p=0.02). In the treatment group, the width of each bar is proportional to the number of respondents reporting different numbers of deaths during the interview. 1 . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. Interview duration (in minutes) Type of mortality questionnaire: Extended Short . CC-BY-NC-ND 4.0 International license It is made available under a is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. (which was not certified by peer review) The copyright holder for this preprint this version posted March 2, 2022. ; https://doi.org/10.1101/2022.03.02.22271441 doi: medRxiv preprint Adult death registration in Matlab, rural Bangladesh: completeness, correlates, and obstacles A global assessment of civil registration and vital statistics systems: monitoring data quality and progress A scandal of invisibility: making everyone count by counting everyone Unconventional approaches to mortality estimation Divergences in trends in child and adult mortality in sub-Saharan Africa: survey evidence on the survival of children and siblings Morbidity and Health-Seeking Behaviour during the Ebola Epidemic 2014-2015 in Monrovia Results from a Mobile Phone Survey A cross-country database of COVID-19 testing Utilization of non-Ebola health care services during Ebola outbreaks: a systematic review and meta-analysis Public health impact of the 2014-2015 Ebola outbreak in West Africa: seizing opportunities for the future Childhood immunization during the COVID-19 pandemic: experiences in Haiti Malaria morbidity and mortality in Ebola-affected countries caused by decreased health-care capacity, and the potential effect of mitigation strategies: a modelling analysis Rapid mortality surveillance using a national population register to monitor excess deaths during SARS-CoV-2 pandemic in South Africa Under-reporting of deaths limits our understanding of true burden of covid-19 The COVID-19 Pandemic: Effects on Civil Registration of Births and Deaths and on Availability and Utility of Vital Events Data Subnational mortality estimates for India in 2019: a baseline for evaluating excess deaths due to the COVID-19 pandemic Mortality surveillance during the COVID-19 pandemic Excess mortality during the COVID-19 pandemic: a geospatial and statistical analysis in Aden governorate Excess mortality during the COVID-19 pandemic: a geospatial and statistical analysis in Covid-19 deaths in Africa: prospective systematic postmortem surveillance study Noncommunicable Disease Risk Factors and Mobile Phones: A Proposed Research Agenda Falling living standards during the COVID-19 crisis: Quantitative evidence from nine developing countries Patterns of demand for non-Ebola health services during and after the Ebola outbreak: panel survey evidence from Remote data collection for public health research in a COVID-19 era: ethical implications, challenges and opportunities Survey evidence of excess mortality in Bihar in the second COVID-19 surge No excess mortality detected in rural Bangladesh in 2020 from repeated surveys of a population of 81,000 COVID mortality in India: National survey data and health facility deaths Viewpoint: High-frequency phone surveys on COVID-19: Good practices, open questions Noninferiority Trials: Is a New Treatment Almost as Effective as Another? Malawi and Millennium Development Goal 4: a Countdown to 2015 country case study The ABCs of Covid-19 prevention in Malawi: Authority, benefits, and costs of compliance Knowledge, risk perceptions, and behaviors related to the COVID-19 pandemic in Malawi Monitoring COVID-19 Impacts on Households in Malawi: Findings from the First Round of the High-Frequency Phone Survey Demographic and health surveys: a profile Population-Based HIV Impact Assessments Survey Methods, Response, and Quality in Zimbabwe, Malawi, and Zambia Effect of airtime incentives on response and cooperation rates in non-communicable disease interactive voice response surveys: randomised controlled trials in Bangladesh and Uganda Through the looking glass: understanding non-inferiority Participant reactions to survey research in the general population after terrorist attacks Superiority and non-inferiority: two sides of the same coin? Confidence intervals for a binomial proportion An assessment of DHS maternal mortality data and estimates Trends in Demographic and Health Survey data quality: an analysis of age heaping over time in 34 countries in Sub Saharan Africa between Rapid Mortality Surveillance of COVID-19 Using Verbal Autopsy