Increasing Patient Engagement in Pharmacovigilance Through Online Community Outreach and Mobile Reporting Applications: An Analysis of Adverse Event Reporting for the Essure Device in the US



Preparing and submitting a voluntary adverse event (AE) report to the US Food and Drug Administration (FDA) for a medical device typically takes 40 min. User-friendly Web and mobile reporting apps may increase efficiency. Further, coupled with strategies for direct patient involvement, patient engagement in AE reporting may be improved. In 2012, the FDA Center for Devices and Radiologic Health (CDRH) launched a free, public mobile AE reporting app, MedWatcher, for patients and clinicians. During the same year, a patient community on Facebook adopted the app to submit reports involving a hysteroscopic sterilization device, brand name Essure®.


Patient community outreach was conducted to administrators of the group “Essure Problems” (approximately 18,000 members as of June 2015) to gather individual case safety reports (ICSRs). After agreeing on key reporting principles, group administrators encouraged members to report via the app. Semi-structured forms in the app mirrored fields of the MedWatch 3500 form. ICSRs were transmitted to CDRH via an electronic gateway, and anonymized versions were posted in the app. Data collected from May 11, 2013 to December 7, 2014 were analyzed. Narrative texts were coded by trained and certified MedDRA coders (version 17). Descriptive statistics and metrics, including VigiGrade completeness scores, were analyzed. Various incentives and motivations to report in the Facebook group were observed.


The average Essure AE report took 11.4 min (±10) to complete. Submissions from 1349 women, average age 34 years, were analyzed. Serious events, including hospitalization, disability, and permanent damage after implantation, were reported by 1047 women (77.6 %). A total of 13,135 product–event pairs were reported, comprising 327 unique preferred terms, most frequently fatigue (n = 491), back pain (468), and pelvic pain (459). Important medical events (IMEs), most frequently mental impairment (142), device dislocation (108), and salpingectomy (62), were reported by 598 women (44.3 %). Other events of interest included loss of libido (n = 115); allergy to metals (109), primarily nickel; and alopecia (252). VigiGrade completeness scores were high, averaging 0.80 (±0.15). Reports received via the mobile app were considered “well documented” 55.9 % of the time, compared with an international average of 13 % for all medical products. On average, there were 15 times more reports submitted per month via the app with patient community support versus traditional pharmacovigilance portals.


Outreach via an online patient community, coupled with an easy-to-use app, allowed for rapid and detailed ICSRs to be submitted, with gains in efficiency. Two-way communication and public posting of narratives led to successful engagement within a Motivation-Incentive-Activation-Behavior framework, a conceptual model for successful crowdsourcing. Reports submitted by patients were considerably more complete than those submitted by physicians in routine spontaneous reports. Further research is needed to understand how biases operate differently from those of traditional pharmacovigilance.

FormalPara Key Points
Spontaneous adverse event reporting to the US FDA was encouraged using an easy-to-use Web and mobile app along with engagement of a Facebook patient group, specifically for Essure, a hysteroscopic sterilization device.
A total of 1349 valid reports were received through the app over approximately 19 months, equivalent to 15 times more reports than through traditional channels, with high completeness scores.
The reports were characterized including symptoms and outcomes reported, and the motivations and incentives in this engagement model for pharmacovigilance are discussed.


It is widely acknowledged that once medical products are on the market, adverse events (AEs) are severely underreported [1, 2]. While legally binding reporting from manufacturers to regulatory agencies and formal post-marketing studies are conducted, capturing complete information continues to be a challenge. The challenge is exacerbated for implantable medical devices, midst pharmacoepidemiology practice developed largely for drugs or biologics [3]. For example, in the USA, nearly a third of mandatory reports from device manufacturers are delivered later than the 5-day requirement, creating bottlenecks in which information becomes lost, misinterpreted, or delayed [4].

Meanwhile, patient-reported outcomes are now accepted for clinical trials with new medical products [5], and we have seen renewed focus on patient-reported outcomes in comparative effectiveness research [6, 7]. However, only 2 % of post-marketing reports about medical devices received by the US Food and Drug Administration (FDA) come from patients [8]. Part of the reason for low patient participation is likely due to reporting burden; the FDA estimates it takes from 40 to 73 min to complete a report for a medical device using the MedWatch 3500 (voluntary) and MedWatch 3500A (mandatory) forms, respectively [9]. While patient reporting for drugs has been shown to be a valuable addition, with many countries encouraging this practice [1013], the same has not been seen for medical devices; whether patient reporting for implantable medical devices can provide high-quality, complete, and novel information remains an open question.

Despite limited patient engagement in device AE reporting, online tools have served to expand participation in public health reporting, generally termed “digital disease detection.” These tools have been applied in monitoring infectious diseases [14], and “crowdsourcing” or “participatory epidemiology” efforts in this domain have proven especially successful [15]. Crowdsourcing consists of systematic efforts to collect information from a wide audience, particularly through the use of online tools, that are mutually beneficial to the participants and activity sponsors [16]. Keating and Furberg’s conceptual framework for crowdsourcing, known as the Motivation-Incentive-Activation-Behavior (MIAB) model, deconstructs crowdsourcing into components that are required for its successful implementation [17]. While successful crowdsourcing requires open and active communication between participants and sponsors, existing pharmacovigilance practices typically yield limited feedback and communication.

The objective of this study was to assess the potential for participatory epidemiology in post-marketing medical device surveillance, specifically by engaging an online patient community to encourage submission of individual case safety reports (ICSRs) through an online tool. We evaluated the quality of data collected through a Web and mobile app called MedWatcher and applied the MIAB framework to characterize successful patient engagement. The device of interest is Essure® (Bayer HealthCare Pharmaceuticals, Inc., Parsippany, NJ, USA), the first hysteroscopic sterilization device approved in the USA [18, 19]. This class III device contains an inner coil of stainless steel with polyethylene fibers and an outer coil of titanium-nickel [20]. The coil is placed into each fallopian tube during an outpatient visit, and is promoted as “permanent birth control.” Subsequent tissue growth around the coil occludes the fallopian tubes, with a confirmatory test 3 months post-procedure to confirm correct placement and blockage [21].


Data Sources

MedWatcher is a Web and mobile app developed by Epidemico (Boston, MA, USA) and launched in September 2012, freely available to the US public for streamlined and user-friendly AE reporting to the FDA. MedWatcher was developed in partnership with the FDA Center for Devices and Radiologic Health (CDRH) to overcome the limitations posed by traditional reporting methods. It is available in English language, on iOS [22] or Android [23] devices as well as on a mobile-optimized website.

Using MedWatcher, patients and physicians can submit AEs for medical devices, drugs, vaccines, and biologics. The app’s report form corresponds to the fields of the MedWatch 3500 form and requires an event description and email address. Optional fields include sex, age, event outcome, and image file. Users receive an email confirmation with the content of their report, formatted to allow printing and sharing with their care providers. In parallel, the system automatically prepares the ICSR in an E2B format [24]. ICSRs are processed in a secure cloud computing environment, manually reviewed to remove spam and test submissions, and transmitted electronically to CDRH using a dedicated voluntary reporting gateway, where automatic consistency, formatting, and completeness checks are conducted before each report is entered into the Manufacturer and User Facility Device Experience (MAUDE) database.

Two-way communication through the gateway allows for FDA case report numbers to be provided back to the patient via the app, enabling consolidation of follow-up reports. As stated in the terms of service for the app, the narrative text of each report is de-identified and shared publicly with other app users, fostering a sense of community and creating a source for safety information. For this analysis, we obtained public, redacted narratives involving Essure from the MedWatcher website [25], starting from the first submitted report (May 11, 2013) through to December 7, 2014.

Product Selection

We selected Essure as the product of analysis because it was the most frequently reported product via the MedWatcher app. Of the 3290 MedWatcher reports received as of June 2015, 2600 involve Essure. We also had significant and successful patient engagement with Essure users through social media.

Essure was originally approved in the USA in 2002. The “Summary of Safety and Effectiveness” reported AEs during pivotal trials, differentiating between those that happened on the day of implantation and those that occurred during 1 year of follow-up [26]. The top five events on the day were cramping, pain, nausea/vomiting, dizziness/light headedness, and bleeding/spotting. The top five events in the first year of follow-up were (lower) back pain, abdominal pain/cramps, dyspareunia (painful sexual intercourse), pain/discomfort, and dysmenorrhea (menstrual cramps). While allergy to metals was not observed in clinical trials, it was mentioned as a possibility in the “Warnings” section of the label.

A more recent FDA review of the 943 Essure AEs received from November 4, 2002 to October 25, 2013 found: “The most frequently reported adverse events were pain (606), haemorrhage [bleeding] (140), headache (130), menstrual irregularities (95), fatigue (88), and weight fluctuations (77). The most frequent device problems reported were the migration of the device or device component (116), patient device incompatibility (113) (e.g., possible nickel allergy), device operating differently than expected (73), malposition of the device (46), and device breakage (37)” [27]. The FDA received one report from a physician about a death due to necrotizing Streptococcus spp. infection associated with the device, although the manufacturer explains that “the medical opinion of the attending physician was that the cause of death was not directly related to the essure [sic] inserts or procedure” [28]. Academic researchers noted tubal perforation [29], pain [30, 31], and placement failures [32] in clinical routine practice.

Patient Community Outreach

“Essure Problems” is a Facebook group [33] launched in March 2011 by Angie Firmalino, a patient experiencing severe AEs following Essure implantation. Since launch, patients organically joined the Facebook group, and as of June 2015, there were 17,850 members, managed by 11 volunteer administrators. The group provides an environment where patients can share information and experiences regarding Essure, including an organized directory of files such as doctors offering device removal, a list of symptoms experienced by members, and a collection of publications and articles about the product. Discussions of benefits also occur, but the group was formed largely in the context of harm. In October 2013, a representative from the MedWatcher app development team (co-author CYB) joined the group to provide technical support to patients filing AE reports to the FDA.

Through active engagement in the patient community, factors that contributed to participation in reporting were observed by applying the MIAB model. In this model, “motivation” is the reason for interest, and “incentive” is what leads someone to act. “Activation” is the set of factors that lead to actual participation, and “behavior” is the activity of interest and outcome, in this case, submitting an AE report [17]. Specifically during engagement of the Facebook patient group, factors that acted as motivation and incentive that were distinct from traditional reporting channels were noted.

In preparation for the present study, a series of discussions were initiated with the Facebook group administrators, starting February 2014, to explain the intent for a research publication. Two administrators of the Facebook group were elected to participate in the research process (co-authors MG and KD), and additional discussions clarified the use of a regulatory coding ontology, appropriate interpretation of spontaneous data, and expectations for the peer-review process. Concepts of legitimacy and integrity were clarified on both sides, and there was agreement that the results would be prepared for publication regardless of the content of the data received. The outcomes of discussions were communicated back to the Facebook group administrators or the entire group.

Coding Adverse Events (AEs)

AE symptoms reported in mobile app reports were tagged by two certified MedDRA coders (co-authors CYB and CEP) using the MedDRA (Medical Dictionary for Regulatory Activities) version 17 at the preferred term (PT) level [34]. All reports were in English. The two coders jointly tagged the first 20 reports to establish coding guidelines then proceeded to code the remaining reports independently, maintaining a living document of codes and coding guidelines, which were iteratively discussed and updated.

VigiGrade Completeness Scores

VigiGrade completeness scores, developed by the Uppsala Monitoring Centre of the World Health Organization (WHO-UMC), were calculated based on rules outlined by Bergvall et al. [35]. VigiGrade completeness scores have been used routinely in the WHO’s safety report database, VigiBase, since 2010. Dimensions accounted for in the VigiGrade completeness score include time-to-onset, indication, outcome, sex, age, dose, country, primary reporter’s occupation, report type, and the presence of informative free-text information. Possible scores range from 0.07 to 1.0, with each report starting at 1.0 and then subsequently penalized for each dimension lacking or containing limited information.

Since Essure is a device with only one indication (permanent female birth control) and no off-label use was noted, there was limited opportunity to penalize for indication or dose. Because the MedWatcher app is intended for submission to the US FDA, the assumption was made that the patients resided in the USA unless otherwise stated (respondents providing non-US addresses were notified of a failure to submit to the FDA and then guided to report to their national authorities). Free-text information was required on the app’s report form. With these assumptions and requirements, the lowest score possible for Essure reports submitted through MedWatcher was 0.139.

Data Analysis

Consistency and logic checks were used to identify input and coding errors. Narrative fields were cross-checked for consistency with structured data elements to correct errors where possible by directly emailing the reporter for clarification. Events requiring medical care were coded as those that were life threatening, resulted in a hospital stay or prolongation of one, or resulted in a visit to the emergency department. The PT “nonspecific reaction” was not included in analyses. Important medical events (IMEs) were identified using the European Medicines Agency list for MedDRA 17 [36]; the IME list contained 7605 PTs. Summary and descriptive statistics were calculated in Stata version 13 (College Station, TX, USA) or visualized in DataGraph 4 beta (Chapel Hill, NC, USA).


A total of 1354 user submissions were received from May 11, 2013 through to December 7, 2014. Four reports were removed from analysis because of a lack of an identifiable AE in the narrative, and one report from a male was removed, for a final sample size of 1349 unique user reports. Of these, 99 users (7.3 %) submitted a collection of 147 follow-up reports. Many of these ICSRs reported AEs that occurred before the date of report, and from the launch of the product in 2002 through 2014, the reported occurrence of AEs steadily increased, with most events occurring in 2013 and 2014 (cut off December 7), totaling 614 events over those 2 years (Fig. 1).

Fig. 1

Date of Essure adverse events reported via MedWatcher app. Most women considered their adverse events to be severe (grey bars). Asterisk Data for 2014 are complete only through December 7

App Usage

The average Essure report took 11.4 min to complete [standard deviation (SD) 10.0 min, range 0.3–60 min]. Time spent on submission was available for 64 % of reports, as this metric was not originally captured by the app; more than 60 min on app was considered an abandoned attempt. A total of 40 submissions resulted in follow-up by MedWatcher project staff to solicit additional information or clarification, in accordance with predetermined rules of engagement, via email (n = 32; 2.4 %) or private Facebook Messenger (n = 18; 1.3 %). Facebook Messenger follow-up was required because of non-responsiveness to email solicitations; it was later revealed that these users checked email less frequently than Facebook Messenger or they had limited access to the email addresses used, which belonged to spouses or children. Photos were submitted in 12 reports (0.1 %). These photos depicted rashes, x-rays, endoscopy images, sonograms, product packaging, medical records, and letters from medical providers (Fig. 2).

Fig. 2

Selected images for Essure reports submitted via app. Images for Essure-related adverse events showing a device migration and b allergic reaction to nickel. Arrow indicates improper location of device outside of right fallopian tube

Nature of Reported AEs

Statutorily defined serious events [37] were reported in 1047 cases (77.6 %). The events were serious enough to require hospitalization and other medical attention in 475 cases (35.2 %), and 382 reports (28.3 %) indicated the patient had experienced disability or permanent damage after implantation of the device.

A total of 13,135 product–event pairs were reported, comprising 327 unique MedDRA PTs. The average reported age was 34.0 years (SD 6.1 years), ranging from 21 to 56 years, with serious and non-serious events distributed consistently over age (Fig. 3).

Fig. 3

Ages of reporters by seriousness of event. The average age reported for Essure events was 34.0 years (standard deviation 6.1 years), ranging from 21 to 56 years

A total of 327 unique PTs were reported. The number of unique PTs per submission ranged from 1 to 84, averaging 10.1 (95 % confidence interval 9.6–10.6) per submission. The most frequently reported PTs included fatigue, back pain, and pelvic pain (Fig. 4). Of the 1349 reports, 598 (44.3 %) included mention of at least one regulatory-defined IME. The most frequently reported IMEs including mental impairment and device dislocation are shown in Fig. 5. Suicidal ideation was reported by 14 women; two suicide attempts were reported. Other events of interest include loss of libido (n = 115); allergy to metals (109), primarily nickel; and alopecia (hair loss) (252).

Fig. 4

Most frequently reported adverse events for Essure. Average age and 95 % confidence intervals of the 15 most frequently reported adverse events, by MedDRA preferred term. Sample sizes given immediately after preferred term. Dotted line represents the average age of all reporters (33.9 years)

Fig. 5

Most frequently reported important medical events for Essure. Average age and 95 % confidence intervals of the 15 most frequently reported important medical events, by MedDRA preferred term. Sample sizes are presented immediately after preferred term. Dotted line represents the average age of all reporters (34.0 years)

Some variation in age distribution was observed among the more common AEs and IMEs (Figs. 4, 5). Average age was 33.9 years. PTs occurring in women older than average were device dislocation (35.0 years), arthralgia (joint pain) (35.2 years), uterine perforation (35.2 years), and endometrial ablation (scarring of uterus to stop bleeding) (36.0 years). Younger women more frequently reported post-procedural haemorrhage (bleeding) (31.9 years) and spontaneous abortions (miscarriage) (29.5 years).

Completeness of Reporting

The average length of the free-text narrative field was 104 words, ranging from 3 to 1557; no restrictions on length were imposed by the app. VigiGrade completeness scores ranged from 0.2 to 1.0, with an average of 0.80 (SD 0.15). Time-to-onset information was provided in 858 reports (63.6 %), with high precision (less than 1-month uncertainty) for 33.0 %. Using the WHO’s threshold for “well-documented reports” (0.80) [35], 55.9 % (n = 754) of reports were well documented. Completeness scores did not vary by age (data not shown). However, average narrative completeness scores were higher among women who also reported serious AEs using checkboxes (0.83, SD 0.13) compared with those who did not use this feature (0.61, SD 0.13), suggesting the need to balance structured and semi-structured fields.


This analysis presents an approach for encouraging patient AE reporting via a crowdsourcing tool in collaboration with online patient community outreach. The inverse relationship between survey length and response rate is well studied [38, 39], and it is likely that the efficiency of reporting, from 40 min via traditional routes to 11.4 min via the MedWatcher app, contributed to higher volumes of reports being submitted. During the 132 months after marketing authorization, CDRH received 943 reports in MAUDE for Essure (an average of seven per month) [40]. By comparison, there were 1349 reports received via the app during the 19 months (103 per month) of the study period (ratio 14.7:1), acknowledging that some reports may have been submitted via both channels, and that some events may have occurred years before. For drug AEs in general, Hoffman et al. have found increasing reports during the first three quarters after approval and relatively constant counts after that [41]. While it is unclear how this pattern may apply for consumer-oriented medical devices, the sudden increase in reports submitted via the app that occurred a decade after initial marketing is likely related to the outreach conducted by the Facebook group administrators and use of the MedWatcher tool.

The MIAB model, Fig. 6, illustrates three key differences between traditional pharmacovigilance and reporting via the app with support from a patient community. One factor that encouraged reporting was the sense of community within the Facebook group, in which members validated each other’s experiences and rallied group members to submit their reports for common goals. Studies have shown that altruism, such as the desire to prevent harm to others and to contribute to research and knowledge, is a motivator for AE reporting [42]. Two-way communication creates the opportunity to include an intrinsic motivation—the validation of patient experience—by leveraging peer comments as potential reinforcement.

Fig. 6

Motivation-Incentive-Activation-Behavior models for adverse event reporting. “Motivation” is the reason for interest, and “incentive” is what leads someone to act. “Activation” is the set of factors that lead to actual participation. “Behavior” is the activity of interest and outcome, in this case, submitting an adverse event report. Dotted lines represent feedback loops

Second, new incentives arise from positive communications within the patient community, endorsing the app with reassurance that a group member’s participation will help other women. This is evidenced by the following posts in the Facebook group: “We have an obligation to our daughters to report it, if we don’t, this could be a viable birth control option for them” and “If you have had any problems after being implanted with Essure, please file a report with the FDA […] That is the only way the FDA knows what is going on […]. File here. [URL to MedWatcher app]”. The incentive of receiving external validation and empathy for one’s experience is supported by public encouragement in the patient community. For example, when a group member posted a screen capture of the email confirmation received from the app, others responded with many Facebook “likes,” a mechanism of quick, positive response to a post, or comments such as “Wootwoot! AWESOME!!”

The third difference with traditional pharmacovigilance is the presence of feedback loops based on bidirectional communication. One feedback loop resulting from a positive user experience reinforces the community’s endorsement of the technology. A second feedback loop operates between the end user and the patient community when anonymized and redacted reports are publicly posted in the app. In summary, the collaboration between the mobile app and Facebook group yielded new incentive structures and feedback mechanisms by allowing patients to communicate confidentially with each other and by making anonymized reports public.

Leverage-salience theory [43] further corroborates that individual participation in survey research is greater when community involvement is present. As Keating and Furberg point out, motivation “is greater if members of an individual’s social network indicate the importance of participating in an event” [17]. The motivational support pathways emerged naturally without interference by app developers or the FDA; deliberate alignment of incentives is an area for consideration in future efforts. Another potential enhancement is providing personalized feedback. Previous research has shown that personalized feedback from healthcare providers is a key factor in submitting another AE in the future [44]. While the publicly posted reports may go some distance in generating feedback content, further steps can be taken to customize this information for the reporter.

The reports received via the app were more complete on average than reports received by regulatory agencies worldwide. The average VigiGrade completeness score is 0.45 for the 7.0 million reports in VigiBase through to January 2012 [35], while the reports received via the app averaged 0.80. Reports received via the app were considered “well documented” 55.9 % of the time, while the international average is 13 % [35], in other words, app reports were more than four times as likely to be well documented. Further, reports from patients via the app were more complete than reports completed by physicians worldwide (for all medical products); only 24 % of reports from physicians are considered well documented [35]. Since VigiGrade scores vary by country, it is worth noting that US physicians rank near the bottom of the list globally in completeness of reports [35]. The information provided by patients, in 11.4 min on average via the app, was of much higher quality than anticipated.

While this study highlights the usefulness and value of online tools in patient reporting, its greatest limitations are generalizability and replicability. Spontaneous report data are limited in general by the lack of a patient exposure denominator. Further, it is unclear what biases may exist in reports submitted by a particularly active and motivated group of patients and how their access to technology may play a role in these biases. Traditional spontaneous report data originating from patients versus healthcare providers have typically involved a different breadth of body organ systems, yet provide similar overall nature of drug problems [45]. Similar patterns may be present in data reported via consumer-oriented apps. It will be important in the future to explore what kinds of biases may operate due to social stigma that may lead to differential submissions of certain types of medical events (e.g., those less stigmatized) or by certain types of patients (e.g., with more normative patient identities) at the intersection of social media and apps. Until these studies are undertaken, there will be limited direction for how to incorporate stimulated data into quantitative signal detection methods, especially in regard to signal-to-noise determination. It also remains to be seen whether a co-promotion model with patient communities is a sustainable and scalable enterprise. With the MedWatcher app, similar efforts are already emerging with other Facebook patient groups.

In addition, the nature of the product (narrow indication, no notable off-label use) allowed us to make assumptions that led to higher VigiGrade completeness scores. Due to the lack of a device-specific completeness metric, VigiGrade was used as the best alternative. While VigiGrade scores are intended for multi-national comparisons, global data on medical device reports are not currently available, making relative completeness of the reports somewhat difficult to interpret.

Despite these limitations, the FDA has been responsive to this reporting population. In the June 2014 report, FDA stated that they “reviewed Essure patient reports of problems (including Web-based testimonials) and reports of problems submitted to the FDA from other sources, including doctors, patients, and the manufacturer” [40]. We applaud the Agency’s efforts to extend the conversational space for collecting AEs. The Agency went as far as to meet with representatives of the Facebook group in February 2014, as a venue to communicate and express the group’s concerns. Most recently, on June 24, 2015, the Agency announced that an Advisory Committee Meeting would be held in September 2015 to discuss risks and benefits of Essure, citing that “the majority of reports received since 2013 have been voluntary reports, mostly from women who received Essure implants” [46].


Creating a mobile app that mimics ICSRs provides a technological solution to the greater medical problem of underreported AEs. A more holistic solution demonstrated in the present study is the engagement of patient groups to responsibly promote use of the app, grounded equally in social theory and medical informatics. At the heart of the user experience in this coupled engagement approach is the sharing of information, a traditionally sensitive subject in pharmacovigilance. Impending technological advances such as automated anonymization can support broader liberation of ICSR narratives, bringing pharmacovigilance closer to the “lively, engaging, dynamic, collaborative, humane enterprise” that it has the potential to be [47].


  1. 1.

    Tubert P, Bégaud B, Péré JC, et al. Power and weakness of spontaneous reporting: a probabilistic approach. J Clin Epidemiol. 1992;45(3):283–6.

    CAS  Article  PubMed  Google Scholar 

  2. 2.

    Bates DW, Evans RS, Murff H, et al. Detecting adverse events using information technology. J Am Med Inform Assoc. 2003;10(2):115–28.

    PubMed Central  Article  PubMed  Google Scholar 

  3. 3.

    Jalbert JJ, Ritchey ME, Mi X, et al. Methodological considerations in observational comparative effectiveness research for implantable medical devices: an epidemiologic perspective. Am J Epidemiol. 2014;180(9):949–58.

    Article  PubMed  Google Scholar 

  4. 4.

    United States. Department of Health and Human Services. Office of Inspector General, Semiannual report. 2009. p. volumes.

  5. 5.

    Guidance for industry: patient-reported outcome measures: use in medical product development to support labeling claims. 2009 [cited December 2009]; Available from:

  6. 6.

    Frank L, Forsythe L, Ellis L, et al. Conceptual and practical foundations of patient engagement in research at the patient-centered outcomes research institute. Qual Life Res. 2015;24(5):1033–41.

    PubMed Central  Article  PubMed  Google Scholar 

  7. 7.

    Costlow MR, Landsittel DP, James III AE, et al. Model for a patient-centered comparative effectiveness research center. Clin Transl Sci. 2015;8(2):155–9.

    Article  PubMed  Google Scholar 

  8. 8.

    Duggirala HJ, Herz ND, Canos DA, et al. Disproportionality analysis for signal detection of implantable cardioverter-defibrillator-related adverse events in the Food and Drug Administration Medical Device Reporting System. Pharmacoepidemiol Drug Saf. 2012;21(1):87–93.

    Article  PubMed  Google Scholar 

  9. 9.

    Kux L. Agency Information Collection Activities; Proposed Collection; Comment Request; MedWatch: The Food and Drug Administration Medical Products Reporting Program. 2014 12/11/2014 [cited 2015 3/26]; Available from:

  10. 10.

    Blenkinsopp A, Wilkie P, Wang M, et al. Patient reporting of suspected adverse drug reactions: a review of published literature and international experience. Br J Clin Pharmacol. 2007;63(2):148–56.

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  11. 11.

    Hazell L, Cornelius V, Hannaford P, et al. How do patients contribute to signal detection? A retrospective analysis of spontaneous reporting of adverse drug reactions in the UK’s Yellow Card Scheme. Drug Saf. 2013;36(3):199–206.

    Article  PubMed  Google Scholar 

  12. 12.

    Banerjee AK, Okun S, Edwards IR, et al. Patient-Reported Outcome Measures in Safety Event Reporting: PROSPER Consortium guidance. Drug Saf. 2013;36(12):1129–49.

    PubMed Central  Article  PubMed  Google Scholar 

  13. 13.

    van Hunsel F, Talsma A, van Puijenbroek E, et al. The proportion of patient reports of suspected ADRs to signal detection in the Netherlands: case-control study. Pharmacoepidemiol Drug Saf. 2011;20(3):286–91.

    Article  PubMed  Google Scholar 

  14. 14.

    Hartley DM, Nelson NP, Arthur RR, et al. An overview of internet biosurveillance. Clin Microbiol Infect. 2013;19(11):1006–13.

    CAS  Article  PubMed  Google Scholar 

  15. 15.

    Freifeld CC, Chunara R, Mekaru SR, et al. Participatory epidemiology: use of mobile phones for community-based health reporting. PLoS Med. 2010;7(12):e1000376.

    PubMed Central  Article  PubMed  Google Scholar 

  16. 16.

    Estellés-Arolas E, González-Ladrón-de-Guevara F. Towards an integrated crowdsourcing definition. J Inf Sci. 2012;38(2):189–200.

    Article  Google Scholar 

  17. 17.

    von Rosenstiel L, Grundlagen-und Bezugsdisziplinen der Arbeits-und Organisationspsychologie. Handbuch der Arbeits-und Organisationspsychologie; 2007. p. 17.

  18. 18.

    Kerin JF, Cooper JM, Price T, et al. Hysteroscopic sterilization using a micro-insert device: results of a multicentre phase II study*. Hum Reprod. 2003;18(6):1223–30.

    Article  PubMed  Google Scholar 

  19. 19.

    Cooper JM, Carignan CS, Cher D, et al. Microinsert nonincisional hysteroscopic sterilization. Obstet Gynecol. 2003;102(1):59–67.

    Article  PubMed  Google Scholar 

  20. 20.

    Valle RF, Carignan CS, Wright TC. Tissue response to the STOP microcoil transcervical permanent contraceptive device: results from a prehysterectomy study. Fertil Steril. 2001;76(5):974–80.

    CAS  Article  PubMed  Google Scholar 

  21. 21.

    Your complete guide to the Essure procedure. 2013 [cited 2015 3/26]; Available from:

  22. 22.

    MedWatcher for drugs, vaccines and medical devices. [cited 2015 2/25]; Available from:

  23. 23.

    MedWatcher drug/device/vaccine. [cited 2015 2/25]; Available from:

  24. 24.

    MedWatcher Essure. [cited 2014 12/20]; Available from:

  25. 25.

    Guidance for Industry. E2B(R3) Electronic transmission of individual case safety reports (ICSRs) implementation guide. Feb 2014 [cited 2015 6/24]; Available from:

  26. 26.

    Summary of safety and effectiveness data. 2002 [cited 2015 3/26]; Available from:

  27. 27.

    FDA’s review of reported problems. 2014 [cited 2015 2/25]; Available from:

  28. 28.

    MAUDE adverse event report: Bayer Healthcare LLC Essure device, occlusion, tubal, contraceptive. 2013 [cited 2015 2/25]; Available from:

  29. 29.

    Langenveld J, Sebastiaan V, Bongers MY, et al. Tubal perforation by Essure: three different clinical presentations. Fertil Steril. 2008;90(5):2011. e5–2011. e10.

  30. 30.

    Syed R, Levy J, Childers ME. Pain associated with hysteroscopic sterilization. JSLS. 2007;11(1):63–5.

    PubMed Central  PubMed  Google Scholar 

  31. 31.

    Conover MM, Howell JO, Wu JM, et al. Incidence of opioid-managed pelvic pain after hysteroscopic sterilization versus laproscopic sterilization, US 2005-2012. Pharmacoepidemiol Drug Saf. 2015. doi:10.1002/pds.3766.

    PubMed  Google Scholar 

  32. 32.

    Panel P, Grosdemouge I. Predictive factors of Essure implant placement failure: prospective, multicenter study of 495 patients. Fertil Steril. 2010;93(1):29–34.

    Article  PubMed  Google Scholar 

  33. 33.

    Essure Problems. Facebook group. [cited 2015 3/26]; Available from:

  34. 34.

    Brown EG, Wood L, Wood S. The Medical Dictionary for Regulatory Activities (MedDRA). Drug Saf. 1999;20(2):109–17.

    CAS  Article  PubMed  Google Scholar 

  35. 35.

    Bergvall T, Noren GN, Lindquist M. VigiGrade: a tool to identify well-documented individual case reports and highlight systematic data quality issues. Drug Saf. 2014;37(1):65–77.

    Article  PubMed  Google Scholar 

  36. 36.

    EudraVigilance Human. [cited 2014 12/8]; Available from:

  37. 37.

    Federal Food, Drug, and Cosmetic Act (FD&C Act). 2006 [cited 2015 6/24]; Available from:

  38. 38.

    Sahlqvist S, Song Y, Bull F, et al. Effect of questionnaire length, personalisation and reminder type on response rate to a complex postal survey: randomised controlled trial. BMC Med Res Methodol. 2011;11(1):62.

    PubMed Central  Article  PubMed  Google Scholar 

  39. 39.

    Jepson C, Asch DA, Hershey JC, et al. In a mailed physician survey, questionnaire length had a threshold effect on response rate. J Clin Epidemiol. 2005;58(1):103–5.

    Article  PubMed  Google Scholar 

  40. 40.

    Essure permanent birth control. 2014 [cited 2015 2/25]; Available from:

  41. 41.

    Hoffman KB, Dimbil M, Erdman CB, et al. The Weber effect and the United States Food and Drug Administration’s Adverse Event Reporting System (FAERS): analysis of sixty-two drugs Approved from 2006 to 2010. Drug Saf. 2014;37(4):283–94.

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  42. 42.

    van Hunsel F, van der Welle C, Passier A, et al. Motives for reporting adverse drug reactions by patient-reporters in the Netherlands. Eur J Clin Pharmacol. 2010;66(11):1143–50.

    PubMed Central  Article  PubMed  Google Scholar 

  43. 43.

    Groves RM, Presser S, Dipko S. The role of topic interest in survey participation decisions. Public Opin Q. 2004;68(1):2–31.

    Article  Google Scholar 

  44. 44.

    Oosterhuis I, van Hunsel FP, van Puijenbroek EP. Expectations for feedback in adverse drug reporting by healthcare professionals in the Netherlands. Drug Saf. 2012;35(3):221–32.

    PubMed  Google Scholar 

  45. 45.

    Inch J, Watson MC, Anakwe-Umeh S. Patient versus healthcare professional spontaneous adverse drug reaction reporting: a systematic review. Drug Saf. 2012;35(10):807–18.

    Article  PubMed  Google Scholar 

  46. 46.

    Essure permanent birth control. 2015 [cited 2015 6/30]; Available from:

  47. 47.

    Hugman B. The fatal love of forms. Drug Saf. 2011;34(8):705.

    Article  PubMed  Google Scholar 

Download references


The authors thank all the women who shared their experiences via MedWatcher. The views expressed in this paper are solely those of the authors and not of the funding source. The authors acknowledge past and present CDRH collaborators: Fei Wu, John-William Declaris, Doug Wood, Mary Beth Ritchey, Isaac Chang, Benjamin Eloff, and Allison Huffman, as well as FDA scientists Bob Ball and Skip Francis. The authors appreciate reviews of portions of this manuscript by Kristen Bibeau and Roxanne Saucier. The authors thank the mobile development team for improving the app: Katelynn O’Brien, Kyra McKenna, and Lucas Baptista. The authors thank Eric Xu and Carly Winokur for assistance in preparing this manuscript for submission.

Author information



Corresponding author

Correspondence to Chi Y. Bahk.

Ethics declarations


This research was funded in part by the Center for Devices and Radiologic Health of the US Food and Drug Administration (HHSF223201210016C).

Conflict of interest

CYB, CCF, CMM, CEP, HR, and ND are employees, and JSB is a consultant of Epidemico, Inc., a company attempting to commercialize the information technology aspects of this research. MG, KD, and RB have no conflict of interest to report. The manufacturer(s) of Essure® are not clients of Epidemico, nor were they contacted about this research prior to submission. Epidemico is a wholly owned subsidiary of Booz Allen Hamilton.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (, which permits any noncommercial use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Bahk, C.Y., Goshgarian, M., Donahue, K. et al. Increasing Patient Engagement in Pharmacovigilance Through Online Community Outreach and Mobile Reporting Applications: An Analysis of Adverse Event Reporting for the Essure Device in the US. Pharm Med 29, 331–340 (2015).

Download citation


  • Adverse Event Reporting
  • Patient Community
  • Implantable Medical Device
  • Individual Case Safety Report
  • Important Medical Event