1 Introduction

Prior to the Coronavirus pandemic, much of medical and health professions education was delivered face-to-face with limited use of online teaching and learning methodologies (Enoch & Williams, 2021). The pandemic forced higher education institutions to transition quickly from face-to-face to remote teaching to comply with pandemic protocols (Almarzooq et al., 2020; Seah et al., 2021), resulting in challenges for all educational institutions. The challenges were exceptionally critical for health professions educators (HPEs) as students were prohibited from participating in labs and clinical field placements (Stokes, 2020).

Even under ideal conditions, online courses require careful planning and design, and effective online education requires a variety of instructional strategies to allow students to interact meaningfully with content, the instructor, and their classmates (Means et al., 2014). When educators responded to pandemic restrictions, they did not have time for careful design and development of online learning environments. Aspects of online learning, including technology policies and training, asynchronously-focused pedagogical decisions, robust assessment strategies, and quality assurance, were absent (Shisley, 2020). Instead, didactic learning was rapidly transitioned to a remote, synchronous environment, often referred to as online or blended (Schultz & DeMers, 2020), and clinical experiences transitioned to telemedicine, which led to complex barriers to learning for health professions students. As the pandemic progressed, faculty considered ways to transition from remote to deep online learning directed at students which included consideration for presence and engagement (Schultz & DeMers, 2020). To capture the range of experiences inclusive of remote learning, transition to deep online learning, and thoughtfully planned online learning, we use the term online learning.

Overall, students faced many challenges in the rapid transition to online learning that included obstacles to technology access, insufficient digital learning competencies, and difficulty navigating online academic honesty, privacy, and confidentiality (Turnbull et al., 2021). The move to online learning required students to establish new study routines without the structure and support of the campus setting (von Keyserlingk et al., 2022). These barriers and stressors required support to develop preparedness, access, and transferability between online learning and clinical skill performance (Shawaqfeh et al., 2020; Van Doren et al., 2020).

The rapid transition to online learning may have impacted student perceptions of their learning experience and online learning self-efficacy. Although researchers have explored self-efficacy in online learning environments, little research exists on the impact of transitions from face-to-face to online learning. This research project expands on the existing literature related to self-efficacy and student perception of meaningful learning in a collaborative environment through development and validation of a tool that measures health professions students’ self-efficacy with online learning while exploring how cognitive, social, and teaching presence is experienced by students who transition from one learning environment to another. Ultimately, students may experience social, teaching, and cognitive presence differently and have various levels of self-efficacy for online modalities as the learning environment was not chosen, but forced, due to the pandemic. Understanding the impact of learning environment transitions on student perceptions of their learning experience and self-efficacy can inform how educators approach online curriculum design and learning environments in health professions education.

1.1 Health professions students and online learning

Most health professions education programs (e.g., medicine, physician assistant, and nursing programs) are offered primarily in a face-to-face environment; however, institutions are beginning to use online education platforms to expand the reach of their programs (Stuart & Triola, 2015; Taylor et al., 2019). These programs vary widely in structure and how technology supports their pedagogical approach (Cook et al., 2010). While some health professions programs have moved completely online or adopted hybrid approaches to education such as the flipped classroom model, other institutions have not embraced online learning as a modality of delivering instruction, particularly in the clinical context (Jones, 2015; Londgren et al., 2021).

Effective online teaching incorporates several basic concepts organized around types of interaction between the learner and the content, the teacher, and other learners (Means et al., 2014). Good teaching practices have been adapted for online instruction, often based on Chickering and Gamson’s (1999) principles of good practice, including encouraging interaction and higher-order thinking, providing opportunities for self-directed learning, formative and summative assessment, and effectively communicating task completion, high expectations, and diversity (Saiyad et al., 2020; Vyas et al., 2010). Adapting face-to-face instruction to an online learning environment requires faculty competencies in teaching and learning that represent social, pedagogical, managerial, and technical skills (Grant & Thornton, 2007; Saiyad et al., 2020; Tekian & Harris, 2012).

Transitioning to online learning in health professions education has been met with some resistance. Faculty report lack of technical skills, time, and institutional infrastructures that inhibit online learning development (Dyrbye et al., 2009; Niebuhr et al., 2014, Perlman et al., 2014). Poor student motivation, high anxiety, and poor interaction between learners and facilitators can hinder online learning, leading to concerns about the depth and breadth of learning and group communication skills (Regmi & Jones, 2020). Although each of these challenges has a solution rooted in communication, collaboration, and culture (O’Doherty et al., 2018), the pandemic forced educators to confront the issue with little to no time to implement these foundational solutions.

Despite hesitations about online health professions education, the outcomes of effective online learning are well documented. When comparing online learning experiences with traditional face-to-face experiences, researchers have not found significant differences in learning outcomes and student satisfaction (George et al., 2014). Moreover, when comparing interactive online learning environments that use discussion forums and other learning technologies to passive approaches, the interactive approach was found to improve knowledge and skills and student satisfaction (George et al., 2014). Although the rapid transition to online learning during the pandemic was challenging, it also provided an opportunity for educators to use what they already knew about online education to enhance their teaching with new online instructional strategies and skills to integrate well-planned online learning opportunities in their curriculum.

1.2 Community of inquiry and self-efficacy

The Community of Inquiry (CoI) theoretical framework provides a lens for examining meaningful learning experiences through three interdependent elements – social presence, cognitive presence, and teaching presence (Garrison et al., 1999). Cognitive presence occurs with a triggering event or exploration in which information is exchanged and ideas are connected or applied. Social presence is the ability of participants in the learning environment to present themselves as real through emotional expression, open communication, and collaboration. Finally, teaching presence is understood as a function of course facilitation and course design which includes the “selection, organization, and primary presentation of course content, as well as the design and development of learning activities and assessment” (Garrison et al., 1999, p. 90). There are notable relationships between the constructs, with teaching presence demonstrating significant prediction of both cognitive and social presence (Garrison et al., 2001; Gutiérrez-Santiuste et al., 2015; Stenbom, 2018).

Self-efficacy considers students’ beliefs about what they can do with the skills and abilities that they have (Bong & Skaalvik, 2003) and refers to their beliefs in their capabilities to design and implement a course of action that leads to goal attainment (Bandura, 1977). Self-efficacy then influences choice of action, amount of effort, and length of perseverance in the face of obstacles. A strong sense of self-efficacy can lead to greater confidence in the ability to take on difficult tasks and challenges like online learning in order to develop skills (Alqurashi, 2016). Persistence in activities that may seem threatening but are safe can lead to mastery and enhancement of self-efficacy (Bandura, 1977). Students who have a strong sense of self-efficacy in their capacity to achieve tasks are motivated to take actions that make success more likely.

Much of the research on self-efficacy in online learning environments has been conducted in higher education but has focused on computer and internet self-efficacy (Alqurashi, 2016). However, connections have been made between computer self-efficacy and student satisfaction (Lee & Hwang, 2007; Lim, 2001) and intent to take online courses in the future (Lim, 2001). In addition, students with high self-efficacy in internet usage for information seeking exhibited higher self-efficacy for online learning (Tang & Tseng, 2013). Student readiness to learn online impacted satisfaction and is correlated with perceived effectiveness of e-learning (Almuwais et al., 2021). Students who persisted academically likely developed enhanced self-efficacy in the online learning environment. Self-efficacy motivates learner choice to initiate and persist with self-regulation and is essential to explain successful learner behaviors (Bandura, 1977; Bong & Skaalvik, 2003). Therefore, self-efficacy is a potentially important factor in an online learning environment where learner agency substitutes for traditional classroom structure (Shea & Bidjerano, 2010).

Connections have been made between the Community of Inquiry framework and learner self-efficacy with teaching presence positively predicting self-efficacy, and self-efficacy mediating the effect between social and cognitive presence (Lin et al., 2015). We used this study to further explore the connections between self-efficacy and the CoI framework by developing and validating a tool that measures self-efficacy with online learning while exploring how cognitive, social, and teaching presence is experienced by students who transition from one learning environment to another.

2 Method

An interinstitutional team of health professions education (HPE) researchers from six U.S. universities and academic health centers developed a survey tool to investigate the impact of the rapid transition from face-to-face to online learning on students’ self-efficacy with online learning, attitudes toward online learning, and the factors impacting learning throughout the pandemic. The survey included 31 items on a six-point Likert scale from ‘strongly disagree’ (1) to ‘strongly agree’ (6). Each item asked health professions students to rate their agreement with statements about their experiences with online instruction and learning during the pandemic. The survey items were developed and reviewed by a group of researchers in health professions programs. The measurement constructs (self-efficacy, attitudes towards online learning, teaching presence, social presence, and cognitive presence) were developed from a thorough literature review. Based on a comprehensive literature review, important constructs and corresponding items related to student online learning were synthesized and revised to fit for measuring modality change purposes. The research team then ranked the items based on the necessity of each item and construct. Debriefing discussions and iterative revisions were made to ensure content and face validity prior to achieving consensus on the final 14 CoI and 13 self-efficacy items.

2.1 Study procedures

Researchers distributed the student survey in fall 2021 to deans and health professions program directors at their respective institutions. Recruitment emails asked deans and directors to share the survey via email with students enrolled in health professions programs at their institutions from fall 2019 to fall 2021. The recruitment email contained a link to a Qualtrics online consent form which was followed by the 31 survey items. A series of demographic questions, including gender, race, age, field of study, and prior online learning experiences preceded the 31 survey items statements. The study was deemed exempt by the Institutional Review Boards at participating institutions where data collection and analysis required human subjects research review.

2.2 Participants

A sample of 205 students from 5 institutions participated in the study. The sample consisted of students from a wide range of programs with the majority enrolled in Nursing (18.5%), Doctor of Medicine (MD) (17.6%), and Physician Assistant (12.2%) programs. Demographic data showed that 74.6% of participants were female and approximately 82% were under 35 years old. Seventy five percent of participants had online course experience prior to the pandemic; however, 64% indicated that they had “some” or “minimal experience” with online learning while only 22% indicated “quite a bit” or “a great deal” of experience. Demographic data are presented in Table 1.

Table 1 Demographic data

2.3 Measures

2.3.1 Community of inquiry

The Community of Inquiry (CoI) framework has been widely used to assess learners’ perceptions in the online and distance educational environments (Arbaugh et al., 2008; Garrison & Kanuka, 2004; Vaughan et al., 2013). The original CoI Survey includes 34 items based on the three interrelated components for successful learning: cognitive presence (12 items), social presence (9 items), and teaching presence (13 items). The reliability and validity of the original CoI survey and framework are well established in the literature (Stenbom, 2018), and high correlations have been found among the three components in previous research with various samples of students (Arbaugh et al., 2008; Bangert, 2009; Díaz et al., 2010; Kozan & Richardson, 2014; Shea et al., 2012; Stenbom, 2018; Swan et al., 2008). In the present study, the original CoI survey items were revised to create a new instrument with 14 items (4 cognitive presence, 5 social presence, and 5 teaching presence items) most relevant to the purpose of the study. The CoI items were adapted from the original CoI scale items and a Q-methodology study conducted by Ramlo (2021). Ramlo distilled 36 statements from a variety of sources including social media, student-led newspapers, and The Chronicle of Higher Education, and categorized them within seven themes related to the rapid, pandemic-related transition from face-to-face to online classes. The ideas represented in these statements filled in gaps where the student experience during the transition was not adequately reflected by the original CoI scale. The research team members, who have extensive knowledge in health professions and online education, modified the survey items to reflect the changes in the perceived impacts of online learning with traditional face-to-face classes. A common stem introduced the items for the CoI portion of the survey, which read as follows; "After my courses went online due to the pandemic..."

2.3.2 Self-efficacy

Self-efficacy, based primarily on Social Cognitive Theory (SCT), indicates one’s personal beliefs and perceived capacity to perform an action or behavior to complete a task successfully (Bandura, 1977). To assess perceived self-efficacy, 12 items were adopted from two existing self-efficacy scales. The 31-item Online Learning Self-Efficacy Survey (OLSS) is a reliable and valid tool to measure students’ preparedness, concerns, and learning needs for online learning as reflected by their self-efficacy, with an overall reliability of 0.95 (Sun & Rogers, 2021). Ten items from the OLSS were adapted for the present study (3 items from online learning task self-efficacy, 3 items from instructor and peer interaction and communication self-efficacy, and 4 items from self-regulation and motivation efficacy). Two additional items from Aguilera-Hermida’s (2020) self-efficacy scale were revised and added to the survey to cover aspects of self-efficacy not adequately covered by the OLSS: confidence in ability to be successful in online classes and to discuss topics with classmates and/or professors. One additional item was developed by the research team: confidence in learning new materials to achieve course objectives. A common stem introduced survey items gauging student self-efficacy after a learning modality change: “After experiencing a change in course delivery/learning modality as a result of the COVID-19 pandemic...”

2.4 Statistical analysis

As a preliminary step, data were screened for missing data and univariate outliers across each item using IBM SPSS software, Version 28, RRID:SCR_016479. There were no significant outliers, and missing data ranged from 0.5% to 6.8% across items. Skewness and kurtosis were within acceptable range for all items (skewness ranging from –1.63 to 0.82; kurtosis < 3.21). Cronbach’s alpha reliability coefficients were estimated for each subscale of the new Learning Modality Change Community of Inquiry (CoI) and Self-Efficacy scales in SPSS. The Learning Modality Change CoI and Self-Efficacy scales were validated separately as two independent scales to create two instruments of reasonable length and provide flexibility for future researchers to choose to use one scale or both scales in a single survey. To assess the validity of the Learning Modality Change CoI and Self-Efficacy scales, exploratory structural equation modeling (ESEM) was performed using Mplus, Version 7.3, RRID:SCR_015578 with maximum likelihood estimation with robust standard errors (MLR) because this approach is robust to non-normal and missing data (Muthén & Muthén, 1998–2012). ESEM integrates the advantages of exploratory and confirmatory factor analysis and structural equation modeling by overcoming poor item-level factor structure fit and discriminant validity as well as biased structural parameter estimates (Marsh et al., 2009, 2014). ESEM provides “confirmatory tests of a priori factor structures, relations between latent factors and multigroup/multi-occasion tests of full (mean structure) measurement invariance” (Marsh et al., 2014, p. 85).

ESEM is known to be unbiased and comparable with confirmatory factor analysis in producing factor loadings and correlations without specifying the factor loading pattern (Asparouhov & Muthén, 2009; Sass & Schmitt, 2010; Schmitt, 2011). A goemin rotated solution was used for all models by allowing the correlations between the factors. On the basis of the recommendation of Kline (2016) and Hu and Bentler (1999), goodness of fit was assessed by multiple fit indices: chi-square (χ2) goodness-of-fit index, the comparative fit index (CFI) and the Tucker–Lewis index (TLI) ≥ 0.95 and 0.90. indicating excellent and acceptable fit; the root mean square error of approximation (RMSEA) and standardized root mean square residual (SRMR) ≤ 0.08, indicating a reasonable fit. Relative fit of different models was also considered, and a more parsimonious model is supported if there is a change in CFI of less than 0.01 (Chen, 2007) or a change in RMSEA of less than 0.015 (Chen, 2007). Additionally, factor loadings less than 0.4 (Stevens, 1992) and cross-loadings (Tabachnick & Fidell, 2001) were used in determining which items to drop from each revised scale.

3 Results

Several ESEM models were compared to find the best fitting model for the Learning Modality Change CoI and Self-Efficacy scales. Model fit indices comparisons are presented in Table 2. The ESEM of three correlated factors of the Learning Modality Change CoI scale with 14 items achieved an acceptable model fit (χ2(52) = 116.98, p < 0.001, CFI = 0.96, TLI = 0.93, RMSEA = 0.08, SRMR = 0.02). Two of the items (11 and 14) had significant cross-loadings on Factors 2 and 3. However, Item 14 loaded much higher on Factor 3 (> 0.6) than on Factor 2 (< 0.4), which was not a concern. Since Item 11 had relatively closer cross-loadings at two factors, another ESEM analysis was performed after removing Item 11, resulting in a poorer model fit than the initial model with 14 items. We retained Item 11 in Factor 3 because that item addresses motivation related to online learning, which is an important element to consider in measurement of cognitive presence. Although the initial model with 14 items had cross-loadings on Item 11, retaining it on Factor 3 was deemed acceptable because it demonstrated a higher loading there (> 0.4) than on the other Factor (< 0.4).

Table 2 ESEM fit indices with MLR estimator

A similar analysis procedure was applied to the Learning Modality Change Self-Efficacy scale. The ESEM of the three correlated factors in the 13-item self-efficacy scale failed to achieve a good model fit as evidenced by the RMSEA and TLI not falling within the recommended limits of the fit indices (χ2(42) = 130.16, p < 0.001, CFI = 0.94, TLI = 0.88, RMSEA = 0.10, SRMR = 0.03). Item 1, I feel confident in my ability to be successful in online classes, and Item 5, I am able to learn new material to achieve course objectives, showed significant and moderate cross-loadings on Factors 1 and 3. After removing Item 1, model fit significantly improved (χ2(33) = 54.57, p < 0.001, RMSEA = 0.06, CFI = 0.98, TLI = 0.96, SRMR = 0.02); however, Item 5 had a factor loading smaller than 0.3 at the planned Factor 1 but a significant factor loading at Factor 3. Alternatively, after removing Item 5, the model produced an acceptable model fit, but Item 2 had very close and significant cross-loadings at Factor 2 and Factor 3. After looking at the fit indices and factor loadings as well as reexamining the meaning of the factor, Items 1 and 5 were removed from the model, yielding a favorable final model with 11 items (χ2(25) = 48.00, p < 0.001, CFI = 0.98, TLI = 0.95, RMSEA = 0.07, SRMR = 0.02). No significant and close cross-loadings were present in the final model.

Standardized factor loadings and factor correlations for the final ESEM of the Learning Modality Change CoI and Self-Efficacy scales in the student sample are presented in Tables 3 and 4, respectively. The internal consistency (Cronbach’s alpha) of the Learning Modality Change CoI and Self-Efficacy scales were 0.92 and 0.95, respectively. The internal reliability for each of the three subscales of Learning Modality Change CoI scale ranged from 0.89 to 0.92. The internal reliability for each of the three subscales of the Learning Modality Change Self-Efficacy scale ranged 0.78 to 0.92. Factor correlations for the self-efficacy scale ranged from 0.45 to 0.66. Factor correlations for the CoI scale ranged from 0.59 to 0.71. Additionally, all subscales of the Learning Modality Change CoI scale were found to be positively associated with subscales of Learning Modality Change Self-Efficacy scale, with correlations ranging from 0.34 to 0.72, all p values < 0.001, and evidence of good concurrent validity.

Table 3 ESEM solution: three factors based on 14 community of inquiry items
Table 4 ESEM solution: three factors based on 11 self-efficacy items

4 Discussion

4.1 Results summary

The ESEM confirmed three factors for the final Learning Modality Change CoI scale: cognitive presence with 4 items, social presence with 5 items, and teaching presence with 5 items. All items loaded as we originally conceptualized on their respective factors with all loadings greater than 0.4 (Table 3). Additionally, consistent with Lin et al. (2015), all subscales of the Learning Modality Change CoI scale were found to be positively associated with self-efficacy subscales. This supports results from a study by Shea and Bidjerano (2010) who found that teaching presence and social presence were significantly correlated with student self-efficacy. Compared to the original 34 item CoI survey (Abbitt & Boone, 2021; Arbaugh et al., 2008), the new 14-item instrument developed in the present study provides strong validity and reliability in measurement of cognitive, social, and teaching presence after changes in learning modality in health professions education programs. While CoI scales have been used to measure the experiences of students studying in a pre-existing online environment, this research lends support for the use of the CoI framework to study effects of learning modality changes on health professions students’ learning experiences.

The ESEM confirmed three factors for the final Learning Modality Change Self-Efficacy scale: online learning task self-efficacy with 3 items, interaction and communication self-efficacy with 4 items, and self-regulation and motivation with 4 items. All 11 items loaded as we originally conceptualized, with all loadings greater than 0.45 as presented in Table 4. Items aligned well with the original subscales of the OLSS (Sun & Rogers, 2021). The new, shorter version of the scale that resulted from this study demonstrated favorable construct validity and internal reliability in measuring health professions students’ online learning self-efficacy after a shift in learning modalities.

4.2 Strengths and limitations

Overall, the two new instruments Learning Modality Change CoI and Self-Efficacy scales produced in this study demonstrated substantial validity and reliability. These scales provide a reliable, shorter, and more efficient measurement of the CoI and self-efficacy of students who experience a change in learning modality. The Learning Modality Change CoI Scale can be used to measure the impacts of learning modality changes on students’ perceptions of cognitive, social, and teaching presence. The Online Learning Self-Efficacy Scale can be used to measure changes in self-efficacy of students transitioning from one learning modality to another. The two scales can be used separately or together, depending on the research context and survey requirements. The wide range of health professions programs and institutions represented by the student sample in this study supports the external validity of the findings. The new scales are efficient yet comprehensive measures of the constructs they represent and are ready to be used by researchers in health professions education. In addition, these scales can be used separately or together in other educational fields with adjustments to survey stems and further validation.

Moreover, the application of ESEM technique showed a promising approach to validate a modified measurement tool. As evidenced by moderate to high correlation among the three factors of CoI, especially social presence with teaching and cognitive presence, there is potential overlap between the three factors. Not surprisingly, social presence is more difficult to achieve in an online environment than in a face-to-face setting, and measurement of the social presence construct typically overlaps with teaching and cognitive presence (Garrison et al., 2001; Gutiérrez-Santiuste et al., 2015; Stenbom, 2018). Further work with another student sample and in other educational contexts might help clarify the issue. Future work might include a larger sample of students to conduct measurement invariance analyses across groups such as gender, race, field of study, or online experience.

4.3 Implications and conclusion

It is important to differentiate between carefully planned online learning that occurs under ideal conditions and the rapid transition to remote learning that many students experienced at the start of the COVID-19 pandemic. Online course environments require extensive planning and design prior to the start of the course, including considerations for cognitive, social, and teaching presence (Garrison et al., 1999). Online learning calls for a variety of teaching and technology tools that lead to meaningful interactions with content, instructors, and classmates (Means et al., 2014). The rapid transition to online learning during the pandemic did not allow educators to carefully plan their online learning environments, nor did it account for instructors’ and students’ lack of comfort and self-efficacy with online learning environments and tools. Much of the instruction that occurred was conducted via remote, synchronous learning environments where instructors attempted to replicate their face-to-face teaching practices amid challenges with access to technology, students experienced increased temptation to engage in academic dishonesty, and students and instructors expressed concern about privacy and confidentiality, (Turnbull et al., 2021).

Although faculty and students faced challenges in the transition, the move to online learning environments in higher education provided opportunities for institutions to reconsider their curricular designs and approaches to delivery of instruction. Higher education is under ongoing pressure to meet the dynamic needs of society, and the pandemic made a significant impact on current and future educational practices (Mbhiza, 2021). The rapid transition to online learning ultimately led to a paradigm shift where online learning environments gained value as student and faculty comfort evolved. Online learning widens access to education for student populations historically underrepresented in higher education including those from low socioeconomic backgrounds, students with disabilities, regional and remote students, indigenous students, and first-generation students (Stone, 2017). The increased value and comfort with online learning provides an opportunity for health professions education programs who are tasked with increasing diversity in the health occupations workforce (Bouye et al., 2016; Jackson & Gracia, 2014) to increase access to underrepresented student populations (Gumport, 2016; Letizia, 2017). Increasing access to health professions education through expansion of online learning opportunities demands an understanding of the impact of transitions in learning modalities on student self-efficacy and learning experiences and outcomes.

This study expands on the existing literature about self-efficacy with online learning and the Community of Inquiry Framework through development of the Learning Modality Change CoI and Self-Efficacy Scales. As higher education institutions evaluate future use of online learning environments, this study provides a tool for examining the impacts of transitions to online learning on student self-efficacy for learning and cognitive, social, and teaching factors that influence their learning experience and outcomes. While this tool was developed for health professions education, it can be used to measure changes in all student populations as they transition from one learning environment to another.