FormalPara Key Summary Points

Sleep-related disorders are common in primary care practice and incorporating sleep tracker data may help in improving patient care.

Our pilot study assessed the feasibility of using a Fitbit Charge 2 device and SleepLife application to improve communication in primary care settings between patients and their providers.

This prospective, randomized, parallel group pilot study was conducted in 20 primary care clinics, amongst primary care providers and their patients with insomnia diagnoses, on prescription sleep aid medications.

Only one physician logged into the SleepLife portal. The lack of physician engagement was a significant limitation of the study.

At the end of the 6-week intervention, patients’ composite general satisfaction scores with sleep health management decreased significantly in the intervention arm when compared to controls. Their satisfaction with communication also decreased significantly.

Physician engagement can be improved by integrating sleep software into the electronic medical record (EMR), providing specific therapy/management suggestions based on sleep disturbances patients may discuss with their physicians about and by providing physicians extra education opportunities regarding sleep.

A study in the future utilizing the learning points and improving on the limitations from this pilot can lead to a big step in improving sleep communication between primary care physicians and their patients and integrating health technology into primary care for all age groups.

Introduction

On the basis of a National US survey in 2012, 69% of adults track at least one health indicator using either a tracking device or some other means [1]. The main health indicators tracked were diet, weight, and exercise. Compared to the aforementioned health indicators, studies evaluating sleep indicators through the use of trackers are few. Although one study did show that physicians do not consider sleep data tracking for patients to be useful, there is general lack of availability of reliable data discussing the use of sleep trackers in informing health decisions [2]. On the basis of our literature review, none of the available studies have looked at an intervention designed to utilize tracker-based data to improve physician–patient communication regarding sleep [3].

Commercially available fitness and sleep trackers are prevalent and consumer use is growing rapidly. Primary care physicians seldom use subjective patient-generated sleep data such as sleep diaries and have been slow to integrate objective sleep data collected from commercial sleep trackers. Education about sleep and sleep disorders has been inadequate, leading to a low rate of recognizing these disorders and discomfort with discussing them [4, 5].

The National Sleep Foundation (NSF) has led recent efforts to establish normative data (i.e., appropriate ranges) for sleep duration and sleep quality. NSF has also partnered with the Consumer Technology Association to establish a workgroup involving over 40 sleep tracking technology companies, which is working to standardize sleep tracking data collection and reporting. Additionally, NSF has developed a novel and innovative tool, the SleepLife® application and portal that translates data retrieved from commercially available sleep trackers into a personal sleep tracking record. Communication about sleep between patients and physicians is low at about 31–37% based on the Sleep and Healthy Activity Diet Environment and Socialization (SHADES) study [6]. This information presented the timely opportunity to test a new paradigm for patient and physician communication using objective patient sleep data obtained using the Fitbit Charge 2™ device and SleepLife® application.

We designed this pilot study with the primary objective to examine whether an application utilizing commercially available sleep tracker (Fitbit) data with continuous flow of sleep information to physicians and patients can be used to improve physician–patient communication regarding sleep and sleep-related behaviors and habits. The other objectives included improvement in physician and patient satisfaction with sleep counseling and effects on total sleep time. The program included a Health Insurance Portability and Accountability Act (HIPAA)-compliant, carefully controlled physician and patient interface (SleepLife®) and physician access to patient objective data under daily living environments collected via commercially available wearables.

Methods

Design

This prospective, randomized, parallel group observational pilot study was conducted in Indianapolis, Indiana. Patients and physicians were enrolled from primary care clinics in the greater Indianapolis area. The study protocols were approved by the Indiana University Institutional Review Board (IRB), IRB Protocol Number 1802077293 and Merck & Co. DRC. The IRB (Protocol Number: 1802077293) provided approval coverage for the 20 primary care clinics where the study was conducted. This study was performed in accordance with the Helsinki Declaration of 164, and its later amendments. All subjects provided informed consent to participate in the study in addition to consent for publication if any identifying information is included in the manuscript.

Participants The DataCore team at the Regenstrief Institute used the Regenstrief Medical Record System (RMRS) to identify patients and physicians at primary care clinics in Indianapolis, Indiana, USA. RMRS is the primary instrument for processing data and monitoring patient and physician activity for the hospital. The RMRS is a modular system, composed of Registration and Scheduling, Laboratory, Pharmacy, and Database modules; and maintains several other databases including vital signs, results of laboratory tests and diagnostic tests, discharge summaries, and inpatient and outpatient charges. The sample patient population at primary care clinics is representative of the population of a mid-sized city in America. The primary care clinics are dispersed throughout the greater Indianapolis area and represent a diverse group of subjects.

Inclusion Criteria

  1. 1.

    Age 18 years or older

  2. 2.

    Have a diagnosis of insomnia identified by electronic medical record (EMR) and/or a validated questionnaire

  3. 3.

    Taking one or more of the following prescription medications for insomnia: zolpidem (Ambien, Ambien CR, Edluar, Intermezzo, Zolpimist), suvorexant (Belsomra), butabarbital (Butisol), quazepam (Doral), estazolam (Prosom), flurazepam (Dalmane), triazolam (Halcion), tasimelteon (Heltioz), eszopiclone (Lunesta), temazepam (Restoril), ramelteon (Rozerem), secobarbital (Seconal), doxepin (Silenor), zaleplon (Sonata), and trazodone (Oleptro, Desyrel)

  4. 4.

    English speaking

  5. 5.

    Ability to consent in person

  6. 6.

    Have access to a telephone with smartphone capabilities (iOS/Android)

  7. 7.

    Have visited their primary care physician in the last 6 months or have an upcoming visit within 2 weeks

Exclusion Criteria

  1. 1.

    Have ischemic or cerebrovascular disease affecting collection of study outcomes (via ICD codes I6*, 43*)

  2. 2.

    History of dementia (via ICD codes F0*, 290*)

  3. 3.

    History of bipolar/schizophrenia/depression (via ICD codes F2*, F31*, F32*, F33*, 296*, 295*, 311*)

  4. 4.

    History of alcohol or substance abuse (via ICD codes F1*, 304*, 303*)

  5. 5.

    Be incarcerated or reside in long-term acute care/ skilled nursing facility

  6. 6.

    Pregnancy

  7. 7.

    Be unable to complete study questionnaires because of hearing loss or visual impairment

These criteria were chosen because they could either affect a person’s sleep quality and habits or affect their ability to participate in this particular study. Each participant had to be taking a sleep aide to ensure that the patient’s sleep problems were persistent enough to require a sleep aide.

Study Procedures

Each primary care clinic was randomized to either the SleepLife® intervention or control arm. The study flow is depicted in Fig. 1. The Regenstrief Data Core team used the Regenstrief Medical Record System (RMRS) to extract data on how many potential patients existed at each Indiana University Health Primary Care clinic in the greater Indianapolis area. The ICD-9 codes of 327.*, 780.5*, 347.* and ICD-10 code of G47* were used to screen for insomnia and other related sleep disorders. The research team including both the principal investigator and research staff, visited clinics to present the study protocol at individual clinics prior to randomization. Enrollment for physicians was completed at the Primary Care clinic. The details of the study were discussed with each physician and a signed copy of their informed consent was obtained before any study-related procedures. Once a physician was enrolled, his or her potential patients as originally identified by the RMRS were screened by the research staff to ensure they met all inclusion and no exclusion criteria. Patients with a diagnosis of insomnia ascribed by a health care provider were approached for enrollment. If a patient was on a prescription sleep aid and did not have a diagnosis of insomnia, they were contacted by research staff and were screened through the insomnia severity index questionnaire to be eligible for enrollment. Patients were enrolled either at their primary care clinic or at a public place that they found comfortable and convenient, like a coffee shop or a library.

Fig. 1
figure 1

Study flow

Intervention

Intervention components for this study included a Fitbit Charge 2™ as the sleep tracker and the smartphone application SleepLife®, developed by the National Sleep Foundation, as well as a website portal for physicians.

Our consent forms outlined the objectives of the study, noting that the primary objective was to improve physician–patient communication regarding sleep. For physicians, a short talk was given at each study site presenting the study to the clinic faculty, allowing for questions and concerns to be addressed. We then went to each physician individually, consented for the study, and if applicable (i.e., for the study site in the intervention group) downloaded the SleepLife portal onto their work computers, provided instructions on how to use it as well as a handout outlining these instructions. After completing the informed consent, physicians completed the demographic questionnaire and physician satisfaction questionnaire, which referred to their satisfaction with their communication about sleep with their patients. The SleepLife® web-portal was the primary method of intervention for the physicians. The SleepLife® web-portal was designed to connect to the SleepLife® application on the smartphones of patients. The SleepLife® web-portal imported and organized the patients’ sleep data from the SleepLife® application on the patients’ smartphone into an easy to view format for physicians and allows physicians the ability to send messages to their patients through the portal. After enrollment, physicians in the intervention arm were provided with access and training to use the SleepLife® portal. A shortcut link to this portal was saved on their desktop in the clinic, or was bookmarked on their browser, according to their preference. In the control arm, physicians did not receive access to the SleepLife® portal. The physicians in the intervention arm received sleep data on their enrolled patients on a weekly basis, as compared to the physicians in the control group who did not receive any sleep data from their patients. On the basis of the sleep data, physicians in the intervention arm could provide, via the web portal, weekly feedback to patients on their sleep. Intervention physicians were also provided reminders about checking the portal.

Each physician was enrolled in the study for the entire duration of the study to accommodate for differential rates of patients’ enrollment. Patients were enrolled in a staggered manner from various clinics instead of focusing on one physician or clinic at a time as a result of the variable response times from clinic staff, patients, and physicians.

All enrolled patients, after providing informed consent, completed a demographic questionnaire, a medication form, Katz and Lawton scales, patient satisfaction questionnaire, and a sleep health survey [6,7,8,9,10]. Fitbit Charge 2™ devices were provided to all patients at the time of enrollment. All patients downloaded the Fitbit application.

Patients enrolled in the intervention arm were provided with access to the SleepLife® application and research staff helped them download it onto their smartphone at the time of enrollment. They received detailed information about how to use the application including a one-page instructional document designed by the National Sleep Foundation. The SleepLife® application synced with the Fitbit application to obtain and organize the patients’ sleep data. Through the SleepLife® application, patients could view their sleep data, input the time they got into bed, when they took their sleep medication, and send messages to their primary care provider. The SleepLife® application was not loaded onto the control group’s smartphones.

The intervention period for patients was 6 weeks, beginning from the date of enrollment. Through this time, data points and questionnaires were collected via email from patients in the control arm, and through the SleepLife® application for patients in the intervention arm. The sleep health survey was collected from patients at the time of enrollment and then again at the end of 6 weeks in the study. The patient satisfaction questionnaire was collected at the time of enrollment, and afterwards at 2, 4, and 6 weeks from the date of enrollment.

In the event that a patient did not log any data via the tracker or app for 3 days in a row, research staff contacted the patient to determine if any adjustments or corrections were required to the tracker or the application.

Measures

Main outcomes collected were physician–patient communication, patient and physician satisfaction, sleep outcomes of total sleep time and satisfaction with sleep. We used a combination of different questionnaires and forms to obtain our outcomes and covariate information: a demographic questionnaire and a physician satisfaction questionnaire for physicians and a demographic questionnaire, Katz Scale, Lawton scale, medications form, patient satisfaction survey, and sleep health survey for patients [6,7,8,9,10]. The following questionnaires were used either as an entire collection instrument itself, or adapted to be a part of collecting study-specific outcomes:

Katz Scale Instrument to assess functional status, based on the patient’s ability to perform activities of daily living independently. It includes items like bathing, dressing, toileting, transferring, and feeding.

Lawton Scale Instrument to assess patient’s ability to perform activities like communication via telephone, laundry and finance management independently, helping assess their functional status.

Charlson Comorbidity Index Scale predicting the 1-year mortality for a patient on the basis of their comorbid conditions.

Patient Satisfaction Questionnaire A scale validated to assess patient satisfaction.

Campbell Assessment Validated tool to assess both doctor and patient perceptions of the communication.

Physician demographic details included gender, age, preferred method of contact, medical specialty, location of practice, number of patients seen per day, time allocated to each patient, and a question about whether they were satisfied with the amount of time they were allocated per patient. Physicians also filled out the physician satisfaction questionnaire at enrollment, and again at the end of the study.

Patient demographics included age, race, gender, and years of education. Height, weight, body mass index, heart rate, blood pressure, and Charlson Comorbidity Index were also included [11, 12]. Activities and instrumental activities of daily living (ADL/IADLs) were obtained through Katz and Lawton scales, and the medication form was filled out through information from the patients themselves and their medical records [6, 7]. The medication information was updated at the end of the study again, via the patient’s medical chart, to note if there had been any changes to the patient’s medication regime. Patients also filled out the sleep health survey and patient satisfaction questionnaire at enrollment and then again periodically throughout the study. The sleep health survey was filled out at the end of 6 weeks, and the patient satisfaction questionnaire was filled out at 2 weeks, 4 weeks, and then at the end of 6 weeks. We collected data about patients’ objective sleep times automatically throughout the study duration through the Fitbit and SleepLife® applications. Questions from the “National Sleep Foundation: Sleep Health Index”, Patient Satisfaction Questionnaire (PSQ), and the Campbell Assessment were adapted for the study purposes [8,9,10]. They were used to create the sleep health survey, patient satisfaction questionnaire, and the physician satisfaction questionnaire. Scores on the individual questions were tabulated to generate overall general satisfaction (GS) and communication (C) scores. The adapted questions, mapped to their outcomes, are included as Online Appendix A. Brief descriptions of these have been given earlier.

Our direct tool of measurement was the Fitbit. Fitbits were chosen because a systematic review showed that sleep-staging Fitbit models compared to a polysomnogram were accurate in determining total sleep time and wake after sleep onset even though Fitbit data tend to underestimate sleep-onset latency [13]. It was also one of the two commercial sleep trackers that had been validated by independent testing [14, 15]. In particular, Fitbit Charge 2™ in a validation study had shown 0.96 sensitivity in detecting sleep and a 0.61 specificity in detecting wake [16]. Fitbit models have their inherent limitations and tend to overestimate total sleep time and sleep efficiency, and underestimate wakefulness after sleep onset (WASO) [13]. Sleep stage transition is difficult to assess using the Fitbit technology [2]. Fitbit sleep data has less agreement with polysomnography findings in patients with insomnia compared to those who are good sleepers [3]. However a systematic review article concluded that they have been found to be appropriate for patients to judge their own sleep hygiene and data, and for physicians to be able to gain a superficial perspective on how a patient is doing in terms of their sleep habits [5]. We utilized the Fitbit Charge 2 as a device to measure patient’s sleep for this study.

Statistical Analysis

Physician Data

Physician Satisfaction/Communication Survey

For each individual at each week, composite scores for physician general satisfaction (GS) and communication (C) were calculated as means of the scores for the relevant questions being answered. Questions 6, 7, 10, 11, and 13–16 were used to calculate GS; and questions 1–5, 8, 9, and 12 for the C score (Online Appendix A). Each question had a scale of 1–5. No reverse coding was performed here. Higher scores meant worse GS and C.

Demographic Variables

Medical specialty was reclassified as “Family medicine” and “Internal medicine”. Number of years in practice was reclassified as < 1, 1–5, 6–10, and > 10. Descriptive statistics were summarized. This included count and frequency for categorical variables, and mean, standard deviation (SD), median, and quantiles for continuous variables.

Statistical Tests

For demographic variables, two-sample t tests were performed to compare the difference of means of continuous demographic variables between intervention and control groups; chi-squared tests were performed to examine the association between categorical demographic variables and study groups.

For individual questions and composite scores, the difference between the scores was calculated by subtracting the baseline scores from the scores obtained at week 6. The raw means (with standard deviation) of each difference in score at week 6 were calculated for intervention and control group, respectively. The significance of the difference in the difference of scores previously calculated was evaluated using linear regression models, adjusting for age, gender, and race. The p values less than 0.05 were considered as statistically significant. A positive difference in scores meant worse GS and C.

Sample size and power calculations were completed using PASS. Statistical significance threshold is set as 0.05; intra-cluster correlation is set at 0.3, and effect size 0.5. For five clusters in each arm, with 20 patients in each cluster, we have statistical power of 0.85.

Patient Data

Patient Satisfaction Survey

Patient satisfaction surveys were obtained at baseline, week 2, week 4, and week 6 for each patient. Participants who chose “don’t know” or “refuse to answer” were coded as missing. Questions 1–4 and 10–12 were reversely coded weekly to reduce bias. For each individual at each week, composite scores for “general satisfaction (GS)”, “communication (C)”, and “time with doctors (TD)” were calculated as means of the scores for the relevant questions being answered respectively. Questions 1, 3, 6–8, 10, 12, and 13 were used to calculate GS; 2, 4, 5, 9, and 11 for C; and 8 and 12 for TD (Online Appendix A). Each question has a scale of 1–5. Here, when calculating a composite score, patients that answered fewer than two questions in a category (GS, C, or TD) were coded as missing. After reverse coding, higher scores meant a higher level of GS, better C, and longer TD.

Patient Sleep Survey

Patient sleep surveys were obtained at baseline and week 6 for each patient. Participants who chose “don’t know” or “refuse to answer” were coded as missing. Question 1 was rescaled as excellent = 100, very good = 75, good = 50, only fair = 25, and poor = 0. Question 2 was inversely coded. Questions 2–7 were rescaled as 7d = 100, 6d = 86, 5d = 71, 4d = 57, 3d = 43, 2d = 29, 1d = 14, 0d = 0. Thus, each question had a scale of 0 to 100. The mean of questions 1–6 was calculated for each survey the patients answered and provided the sleep quality index (SQ) score. The mean of questions 1–7 was calculated for each survey the patient answered and provided the sleep outcome (SL) score. Higher scores meant worse sleep quality and sleep outcome.

Demographic Variables

Race was reclassified as Black/African American, White/Caucasian, and other. Education was reclassified as high school, GED or vocational school, 13 years—bachelor’s degree, and higher degree. Age was calculated from each patient’s birth date, with reference to 30 November 2018. “Not available” in heart rate was coded as missing. Descriptive statistics including count and frequency for categorical variables, and mean, standard deviation (SD), median, and quantiles for continuous variables were summarized.

Statistical Tests

For demographic variables, two-sample t tests were performed to compare the difference between the means of the continuous demographic variables of the intervention and control groups. Chi-squared tests were performed to examine the association between categorical demographic variables and the study groups.

For individual questions and composite scores, the difference between the scores was obtained by subtracting the baseline scores from week 2, 4, and 6, respectively. The raw means (with standard deviation) of each difference in scores at weeks 2, 4, and 6 were calculated for the intervention and control group separately. The significance of the difference in the difference of scores previously calculated was evaluated using generalized estimating equation (GEE) models, adjusting for age, gender, and race. The GEE model could account for potential correlation of patients seeing the same physician, and physician is hence used as the clustering variable. A p value less than 0.05 was considered as statistically significant. A positive difference in scores meant a higher level of GS, better C, and longer TD.

The sample size and power calculation was conducted prior to participant enrollment. We were planning to recruit 20 patients in each of the five clusters. However, the actual enrollment of the patients and physicians was less than what we expected. While we hoped to recruit the same number of patients and physicians in the intervention and control groups, we believe that our eventual numbers of 28 (control) vs 21 (intervention) physicians and 41 (control) vs 34 (intervention) patients are not highly unbalanced. Hence, we did not consider adjustment of this issue. There are demographic differences between the control and intervention groups. Given this, we have adjusted for age, gender, and race in the GEE-based analysis, for both patient and physician satisfaction data.

Results

Study Results

Physicians and patients were recruited between June 2018 and February 2019 at Indiana University Health primary care clinics in the greater Indianapolis area. Out of the 20 enrolled primary care clinics, 12 were randomized to the control arm and eight to the intervention arm. Forty-nine physicians were enrolled with 28 physicians in the control arm and 21 in the intervention arm. We enrolled 75 patients, 41 in the control arm and 34 in the intervention arm.

Figure 2 provides the number of patients that participated and their intervention allocation status. Two patients were lost to follow-up and one patient withdrew from the study.

Fig. 2
figure 2

Patient study sample and allocation

Physician demographics are reported in Table 1 for both the intervention and control arms. The mean age of the physicians was 42 years in the control arm compared to 53 years in the intervention arm. Patient demographic characteristics are displayed in Table 2 for both the intervention and control arms. A higher number of patients in the control group had bachelors or a higher degree [intervention, n = 23 (67.6%); controls, n = 37 (90.25%)].

Table 1 Physician demographics: comparison between intervention and control groups
Table 2 Patient demographics: comparison between intervention and control groups

The survey completion rate amongst physicians was 55%. Only one physician logged onto the SleepLife portal during the course of the study to check their patients’ sleep status. Physician communication scores did not change significantly between the intervention and control group (Table 2). Physician general satisfaction (GS) scores were also not different between the groups at the end of the study (Table 3). Patients showed high rates of involvement in the program with 83% completing all survey questions. At the end of the 6 weeks, patients’ composite general satisfaction scores in the intervention group decreased (i.e., worsening satisfaction with care) significantly as compared to the change in the control group (p = 0.03) (Table 4). Patients’ perception of physician–patient communication also decreased in the intervention arm as compared to the change in the control arm, at the end of 6 weeks. This was indicated by the lower “communication” scores for the intervention versus control group at the end of the 6 weeks (p = 0.01). The sleep outcomes on the sleep health questionnaire improved significantly in the intervention group as compared to the control group (p = 0.04) (Table 4).

Table 3 Comparison between general satisfaction and communication scores between the physicians in the intervention and control groups
Table 4 Comparison between the patients’ satisfaction and sleep composite scores between intervention and control groups

Fitbit Results for Sleep Time

The SleepLife® application collected sleep-related measures directly through the Fitbit device. Fitbit data contained the measurements for seven variables, which were available for 43 patients over the duration of the study. These seven variables included sleep duration, total time the patient spent in bed, latency to sleep onset, number of times they woke up after falling asleep, time they spent awake after falling asleep, percentage of time they spent awake in bed, and sleep efficiency (Online Appendix B Figs. B.1–B.7). These data demonstrate the feasibility of capturing sleep data through our program using the SleepLife® application and Fitbit.

Safety

No adverse events occurred during the course of the study.

Discussion

Physician–patient communication and patient satisfaction regarding sleep and sleep-related behaviors and habits can be collected through the SleepLife® application. Patients demonstrated very high engagement levels with the program, while physicians had very poor engagement. Patients in the intervention group showed worsening satisfaction with their sleep care and physician–patient communication, possibly highlighting a perception gap with the expectation that they would be receiving enhanced attention from their primary care provider regarding their sleep complaints. Sleep health outcomes improved in both groups but there was a significant increase in the intervention group. This discrepancy could be a result of patients paying more attention to their sleep habits and taking steps that may improve their sleep regardless of their physicians’ input. However, since both groups received a Fitbit and were able to view their own sleep health data, this difference could reflect the impact that the thought of more engagement, time and focus spent on sleep health could affect their actual sleep. The use of a symptom diary to self-monitor and provide data to both patients and physicians has been well studied and shown to be beneficial and have a positive impact on outcomes when both sets of participants are engaged in the activity [17]. Sleep diaries have been prescribed by physicians for a long time to help patients self-monitor and identify sleep patterns. Given the popularity of sleep trackers, as mentioned earlier, many have started monitoring their sleep similarly but using technology. If physicians were able to directly view their patients’ sleep outcomes, this could help to appropriately prescribe medications for patients depending on which sleep behavior they are having trouble with and to help them with better sleep hygiene and non-pharmacological measures. They may be able to have objective outcomes to work with once patients implement these measures that can provide positive feedback.

As stated earlier, physicians were not engaged in the project and did not close the feedback loop as we had expected. This could be a possible reason for the significant decrease in general patient satisfaction and physician–patient communication scores in the intervention arm compared to the control arm. Patients likely expressed less satisfaction with their doctors when data pertaining to their sleep was sent from their wearable devices to their primary care physicians, but no action was taken. This observation could be due to multiple reasons. Firstly, it may appear as if the physicians are not spending adequate time on communication regarding sleep. During the process of enrollment, a couple of patients commented that they felt as though their sleep disturbances were not a priority for their physicians and were glad to hear about the study because it felt like it was finally given significance. Many patients expressed frustration over their sleeping problems. Secondly, the patients themselves may also simply be more aware of their sleeping habits because of using the Fitbit devices, the SleepLife® application, and being a part of the study itself. Additionally, an absence of strong bidirectional communication around the patient’s disordered sleep habits could foster feelings of decreased satisfaction with respect to sleep care. There was no post-participation debriefing regarding the lack of communication between physicians and patients that was supervised by the research team, although this would be useful for patients in future studies.

Research has shown that primary care physicians and advanced practice registered nurses perceive that wearable devices are helpful in managing patients’ health and that patients are interested in these technologies. However, the studies do not find wearable devices useful for tracking patients’ sleep, compared to patients’ physical activity and diet [2]. This was attributed to most likely be due to sleep being a secondary tracking feature on most wearable devices. Although our study did not focus on this in particular, it may have played a role in the low rates of physician engagement. The SHADES study showed that patients or physicians were not discussing sleep more than a third of the time over the course of a patient’s care [6]. This is consistent with our study which showed generally low levels of general satisfaction and communication in the patient satisfaction questionnaire which were about their communication with their primary care physicians about sleep. The work in the SHADES study also indicated that patients that were depressed were more likely to raise their concerns about sleep with their physicians [6]. This was something that we could not assess for in our study given that we excluded patients with the diagnosis of depression, but would be interesting to discuss further; it is possible that the presence of a mental health condition may prompt the idea of disordered sleep to come up more in conversation, and improving discussions about sleep would require us to think about why this would happen.

One of the major barriers in the diagnosis of sleep disorders is the fact that patients tend not to volunteer information pertaining to their sleep to their primary care physicians [18]. Physician education in sleep and sleep disorders has been studied as being inadequate as well [4]. This has led to a low rate of recognizing, and consequently managing, sleep disorders in primary care at both academic and community centers [5]. Primary care doctors are often uncertain regarding their diagnosis, or do not assess for insomnia at all either as a result of feeling uncomfortable with the diagnosis, treatment, or because of time constraints [18]. Our study did not adequately assess for either of these barriers. It highlights the need for increased physician awareness and engagement to address the needs of patients suffering from insomnia. Only one physician actually accessed their patients’ sleep data via the SleepLife portal. This was a major limitation of our study, indicating that the current SleepLife portal and application may not be a feasible mode of delivery/presentation of objective patient sleep data to physicians. We performed a small focus group with four physicians from one study site after the study had been completed to evaluate reasons why physician engagement was low in the study. The reasons for non-engagement included lack of time, not having the motivation or not remembering to log into a patient portal when that patient was not scheduled for a visit, and not feeling comfortable interpreting and acting upon the data received. The lack of physician engagement in our study meant that patients and physicians were not communicating about sleep, leading to results not accurately reflecting satisfaction with that communication. In fact, patients may have been more frustrated having been told that they would be able to discuss their sleep concerns with their physicians and then not getting a response from them. If the study had gone on for more than 6 weeks and would have included an in person visit with the physician where the patient presented their Fitbit results to the physician and they discussed them together, this may have meant significantly improved communication. In short, wearable devices allow patients to view their own sleep data, communicate with their physicians, and may instill a greater sense of independence. However, in the absence of reciprocity from their physicians, patient satisfaction is adversely affected. This supports the theory that collaborative physician–patient communication, in which the patients share responsibility and control over their health along with their physicians, could result in better care and outcomes.

Physicians’ perception of physician–patient communication and physician satisfaction scores did not demonstrate significant differences between the intervention and control arms. These results are not robust as there was minimal to no physician engagement during the course of the study with only one physician logging into the SleepLife® portal. One of the major discussion points with physicians as we were ending the study was about how to integrate the SleepLife® portal into their workflow. Physicians reported that if they had the time, or possibly an additional incentive, they would be more open to review patient data via the SleepLife® portal. Of note, studies have attempted to formulate care coordination plans that eased time constraints and engaged physicians in patient communication more regularly [19,20,21,22]. This needs to be explored further in a study with a larger sample size of physicians, a more robust system of obtaining physician outcome assessments or questionnaires, and utilizing behavioral nudges incorporating the principles of behavioral economics to improve engagement and collaboration.

Our pilot study provides insight into what can be taken forward to design a future study. One of the strengths was the randomized design which minimized bias and possible confounding factors to a minimum in both groups. The SleepLife® portal and application were easy to navigate and had a well-designed user interface. Troubleshooting of the SleepLife® application was performed quickly and efficiently. We received this feedback from physicians themselves and by patients who called in for assistance with the application. Patient enrollment was streamlined as the study moved forward. Phone calls were found to be an acceptable and effective method of gaining assent from patients. Utilization of the SleepLife® application to collect questionnaire data was efficient and led to over 80% of patients completing all questionnaires. We were able to enroll several patients over the age of 60 years, allowing us to view and learn from their interaction with a wearable device and technology generally.

Limitations

There were several limitations to this study. The major one was the lack of physician engagement. They did not check the SleepLife® portal and sending them reminders was both not acceptable to them and did not work. It had a significant impact on our results. Physicians would prefer to have sleep data brought to them by a patient at their individual clinic visit, or have an incentive, financial or otherwise, to look at the portal.

One way to improve physician engagement is to create a tool that automatically sends a physician their patient’s sleep data via EMR message every couple of weeks in an easy-to-read format. This kind of software would eliminate the need for physicians to access a separate website, remember to log in to the website, and thus save time. It would integrate with the patient’s EMR and allow physicians to access their information in a timely manner without them having to learn new software or check a different interface themselves. We could also add a page on specific therapy/management suggestions based on what difficulties patients may face with sleep, e.g., trouble initiating sleep, staying asleep, etc., on sleep hygiene, and on medications that may be appropriate for particular age groups. As part of a study, another way to do so would be to provide an incentive for physicians like CME credits or meal gift cards with every message they send to their patients regarding sleep. When applying the study results to the real world, as stated earlier, a tool that is integrated with the EMR and provides the physicians with data instead of them having to seek it out is the best way to improve their engagement. Other ways include providing physicians in training, and already trained physicians with easy access to resources on sleep education, allowing them to be more comfortable with diagnosing, discussing, and managing sleep disorders.

Physicians were also not enthused at having to complete physician satisfaction questionnaires. Many therefore may have been in a rush and not answered the questionnaires thoughtfully. Primary care physicians may benefit from additional sleep-based educational modules and discussions around how to integrate technology into primary care and how helpful it can be. Another major limitation was that we did not include patients with mental health conditions or with a history of alcohol or substance abuse. This decreased our pool of patients greatly because of the high percentage of overlap between mental health conditions and sleep disorders. The study team felt that including mental health disorders may affect a patient’s sleep independently and make it difficult to see a real difference due to the intervention itself. The small sample size of the study precluded us from making any meaningful inference on findings amongst the different sleep disorders as well. It was also very time intensive to enroll clinics and physicians. The response time for clinics was slow, leading to the study going on for longer than planned. Although we were able to enroll many patients over the age of 60, it was quite time intensive to enroll and teach them how to use the Fitbit and the SleepLife® application. Encouraging them to use the application and the device may be difficult.

Conclusion

The use of wearable devices to measure patients’ sleep measures and to provide data to primary care physicians is an area that has the potential for a significant amount of research and discussion. Patient engagement levels were high; however, physician engagement was extremely poor. The absence of engagement of physicians in the program could explain the decrease in physician–patient communication and patient general satisfaction scores in the intervention arm. Future studies that consider and work around the limitations reported in our pilot project are required to further consumer health technology use in primary care, which could potentially lead to improved physician–patient communication and sleep-related health outcomes.