Dental education increasingly emphasises the need to obtain complex theoretical knowledge, to integrate this knowledge and skills and to apply these to new clinical scenarios and situations. Although it is essential to offer students diverse clinical practice opportunities, this is not always possible due to a lack of sufficient suitable patients, the existing gap between theoretical knowledge and practical skills, an increased focus on patient safety, a lack of representation of thinking processes in the reasoning process, time-intensive tutor guidance and difficult planning (Kononowicz et al., 2016; Postma & White, 2015; Ray et al., 2017; Wang et al., 2020). Therefore, alternative teaching tools and methods such as virtual patients (VPs) have become increasingly important in dental education and they have been accepted by now as a viable supplementary method (Kleinert et al., 2007; Weiner et al., 2016).

Using VPs in dental education can afford students the opportunity to practise clinical skills in a safe learning situation at their own pace, whilst also receiving individual feedback (Kleinert et al., 2007). The aim of VPs is to provide novices with clinical training on how to deliver cases, but without the risks to patients that clinical rotations may involve (Clark et al., 2012). They allow students to practise real simulated interactions with patients and help them to learn complex interviewing skills, make risk assessments, and develop strategies in treatment planning before their first real encounter with patients in the clinic (Boynton et al., 2007; Weiner et al., 2016). As such, VPs can bridge the gap between the preclinical and clinical setting, and between theoretical knowledge and real patient practice (Boynton et al., 2007). Their major benefit is that they allow learners to repetitively and consciously practise diverse clinical skills at anytime, anywhere and at any point in the curriculum (Weiner et al., 2016).

Despite these benefits, Kononowicz et al. (2016) expressed several concerns in their systematic review about the use of VPs. For instance, rather than being used as a supplementary tool, VPs might sometimes replace real patients altogether, learners may be less empathic as VPs are text-intensive, students have little control over the interaction with a VP and VP contents may not be adequately integrated with the overall intended learning outcomes. The authors went on to argue that VP practices should therefore be adjusted to these outcomes and be incorporated within educational programmes and the curriculum; new opportunities might otherwise be ignored, rejected or cause a learning overload (Kononowicz et al., 2016). In a similar vein, previous literature reviews (Chiniara et al., 2013; Cook et al., 2013; Fries et al., 2021) have concluded that there are substantial gaps in the empirical evidence regarding the value, methodologies and outcomes of simulation in healthcare. Hence, we need more evidence on how to design simulations and integrate them into curricula or training programmes (Chiniara et al., 2013).

In the literature (Cook et al., 2013; Ellaway et al., 2014; Issenberg et al., 2005), we have identified different approaches to describe the research on VP design and implementation in medical and dental education. Ellaway (2014), for instance, focused on embedding VPs and concluded that medical education must find the proper ways to use them because their effectiveness depends on how they are designed and the educational activities around them. Few studies, however, have focused on the extent to which existing instructional design (ID) principles and effective use of VPs have been considered. Issenberg et al. (2005) identified 10 ID features of high-fidelity medical simulations that lead to most effective learning, which were: ‘feedback, repetitive practice, curriculum integration, range of difficulty level, multiple learning strategies, capture clinical variation, controlled environment, individualised learning, defined outcomes and simulator validity’. In another comprehensive review, Cook et al. (2013) evaluated and compared the effectiveness of ID features and concluded that the empirical evidence supports the following features: ‘range of difficulty, repetitive practice, distributed practice, cognitive interactivity, multiple learning strategies, individualised learning, mastery learning, feedback, longer time, and clinical variation’.

Although these reviews connected specific ID principles to effective simulation design, they are not linked to any existing, integrated ID model and are therefore of limited use in building future VPs. This is important because the quality and effectiveness of a learning experience are the product of sound instructional design. Essentially, ID theories offer a framework for the use of proper strategies and predict their effectiveness. ID models are used in effective learning environments to foster ID strategies that elicit appropriate cognitive processes. It is with good reason, therefore, that Cook et al. (2013) called for future research that directly compares alternate IDs in order to shed more light on the mechanisms of effective simulation-based instruction.

Chiniara et al. (2013) established a conceptual framework for simulation in healthcare education. Although spanning four levels, the proposed framework is not very specific about how a simulation should be designed, nor does it investigate whether and how it helps. When education is effective, students are able to transfer their knowledge to real clinical practice. Early education, however, often focuses on theoretical aspects and science-oriented presentations of knowledge without making ample connections to clinical practice (Kononowicz et al., 2016). New educational approaches have emphasised instructional methods that ensure the integration and application of students' knowledge and skills in effective clinical practice (Allaire, 2015; van Merriënboer & Kester, 2014). One of these approaches is the four-component instructional design (4C/ID) model that targets the development of complex skills and the subsequent transfer of learning from the classroom to new and real situations. The model is accompanied by a ten-step process (van Merriënboer & Kirschner, 2018) detailing how to design instruction for complex learning. The assumption underlying this model is that the design of complex learning environments should encompass four basic components, specifically: (a) learning tasks that are based on authentic, whole tasks, (b) supportive information, (c) procedural information and (d) part-task practice (van Merriënboer & Kirschner, 2018).

Few literature reviews have gone beyond a set of ID principles by selecting an ID model and scrutinising its application in the selected studies. The present study aimed to determine whether or not and to what extent the principles of a given, proven and closely connected ID model have been applied in VP design in dental education and what we can learn from this in terms of strengths and weaknesses. The 4C/ID model aims to help instructional designers develop educational programmes aimed to teach complex skills or professional competencies. We selected this model, not only because it targets the development of complex skills and the transfer of learning from theory to practice, but also because it has been widely researched and used for various types of medical simulations (e.g. de Melo et al., 2017; Maggio et al., 2015; Vandewaetere et al., 2015). Moreover, the model synthesises many of the points that were considered in other reviews. In the following paragraphs, we will first introduce the 4C/ID model and connect it with the features identified in the aforementioned reviews (Cook et al., 2013; Issenberg et al., 2005). Based on this knowledge, we will subsequently review a set of studies on VPs in dental education with the aim to understand how much they considered these ID principles to make effective use of VPs in education.

Theoretical framework: the 4C/ID model and key ID features

As stated above, the underlying framework of this study is the 4C/ID model. The assumption underlying this model is that the design of complex learning environments should encompass four basic components. To focus our study, we will introduce each of these components and connect them with the features identified in previous reviews:

  1. (a)

    Learning tasks. These tasks should be authentic and based on real-life situations so as to facilitate the integration of knowledge, skills and attitudes as well as the transfer of learning. Moreover, sets of learning tasks should exhibit a high degree of variability and have decreasing levels of learner support and guidance within each task class (van Merriënboer & Kirschner, 2018). That is to say, tasks should warrant clinical/task variation that refers to the diversity of patients' clinical scenarios (Cook et al., 2013; Issenberg et al., 2005) and the use of different instructional strategies to scaffold learning by means of tasks that vary from worked example to patient case (Cook et al., 2013). The design of these learning tasks, moreover, should take account of the learning mode. That is, will the task be presented to an individual or a group of learners? This decision should be based on their learning needs or the complexity of the task (Issenberg et al., 2005; Kirschner et al., 2009), which brings us to another ID feature: task difficulty. According to this principle, tasks should not only vary content-wise, but also in terms of their complexity, which depends on the number of elements included in the task and their degree of interaction (van Merriënboer & Kirschner, 2018). Previous studies have indeed flagged the range of task complexity as a significant variable in simulation-based medical education and concluded that the practice of medical skills across a range of difficulty levels can improve learning (Issenberg et al., 2005). Finally, instructional designers should also ensure that the learning tasks be integrated with the remaining three components (supportive and procedural information and part-task practice). Such integration is essential to facilitate learners’ understanding of key concepts by connecting theory and practice through learning tasks (Mora & Coto, 2014). As Issenberg et al. (2005) argued, rather than being offered as an exceptional or stand-alone activity, simulation-based education should be part of the standard curriculum and be spread over a period of time in the form of distributed practice.

  2. (b)

    Supportive information. Defined per task class, this information offers learners a bridge between knowledge and practice, allowing them to learn successfully and carry out the non-routine aspects of learning tasks. The information explains how a domain is organised, includes strategies to approach problems in a systematic problem-solving process (e.g. diagnose – treatment– follow-up) and introduces the general or exploratory principles that help learners to complete each phase successfully (van Merriënboer & Kirschner, 2018). In addition, it includes cognitive interactivity/feedback, which is defined as ‘the receipt, processing, and formulation of the learner's internal thoughts and other cognitive processes in response to events in the instructional environment’ (Song et al., 2014, p. 2). This interactivity should also encompass reflection strategies such as comparison, elaboration, and entrapment to encourage learners' cognitive engagement. Through prompts, cues and questions, cognitive feedback induces students to reflect on the quality of problem-solving processes and solutions in order to create more effective cognitive schemas and improve future performance (van Merriënboer & Kirschner, 2018).

  3. (c)

    Procedural information. Necessary to perform routine aspects of a learning task, this information is best presented just in time when learners are working on the task to facilitate schema automation. It quickly fades as the students gain more expertise. At this point, learners may receive corrective feedback pertinent to the outcomes of their performance. Based on this as well as the cognitive feedback, learners are able to self-assess and monitor their progress in the learning process. Unlike cognitive feedback, however, the procedural information helps learners to identify and correct errors relevant to the task at hand (van Merriënboer & Kirschner, 2018).

  4. (d)

    Part-task practice. This final component calls for extensive repetitive practice of routine aspects of a task in order to reach a high level of automaticity. As such, it is typically required when learners must attain a determined standard for routine aspects of a task to qualify for a task level. This is the case when the learning tasks do not provide the required amount of practice. The aim of part-task practice is to prompt skills acquisition and the transfer of skilled behaviour from simulator settings to patient care settings (Cook et al., 2013).

Questions

Using the 4C/ID model, we attempted to structure key ID features that apply to the design of virtual patients in dental education. Such an overview might aid the systematic analysis of VP design. The results of such an analysis will not only reveal the strengths and weaknesses of a specific design under scrutiny, but it will also give an indication as to whether there is sufficient coverage of the model to recommend its use for future VP design. The adoption of a shared ID model looks beyond mere ID principles and will facilitate future research that compares alternate IDs.

In this study, we therefore, addressed the following research questions:

  • To what extent have the principles of a given, proven and task-focused ID model been applied in VP design in dental education?

  • What can we learn from this application in terms of strengths and weaknesses?

To this end, we reviewed studies on VPs in dental education to find out how much they considered the ID aspects from the perspective of the 4C/ID model.

Methods

Protocol

We conducted the narrative review at hand according to the PRISMA guideline (as far as applicable) which includes preferred reporting items for systematic reviews and meta-analyses (Page et al., 2021). Figure 1 gives an overview of the main flow of the study.

Fig. 1
figure 1

Study flow

Databases and search strategies

In the period spanning 2000 through to 23 February 2022, we systematically searched Embase, ERIC, PubMed, MEDLINE, CINAHL, Scopus and PsycInfo for studies that used VPs in dental education. For this search, we used a single tool: the LibSearch engine. We restricted our search to peer-reviewed articles in English only. The search strategy included terms that addressed VP simulation as follows:

(kw:(“virtual patient” OR “Patient Simulation” OR “Computer Simulation” OR “Computeri?ed Model”) OR ab:(“virtual patient” OR “Patient Simulation” OR “Computer Simulation” OR “Computeri?ed Model”))

To zoom in on dental education, we narrowed down the search by adding the following terms:

AND (kw:(dentistry) OR ab:(dentistry) OR kw:(“dental”) OR ab:(“dental”)) AND (kw:(“ educ*” OR “teach*” OR “train*” OR “learn*” OR “instruct*” OR “stud*”) OR ab: (“educ*” OR “teach *” OR “train*” OR “learn*” OR “instruct*” OR “stud*”))

Consequently, the final search string read:

(kw:(“virtual patient” OR “Patient Simulation” OR “Computer Simulation” OR “Computeri?ed Model”) OR ab:(“virtual patient” OR “Patient Simulation” OR “Computer Simulation” OR “Computeri?ed Model”)) AND (kw:(dentistry) OR ab:(dentistry) OR kw:(“dental”) OR ab:(“dental”)) AND (kw:(“educ*” OR “teach*” OR “train*” OR “learn*” OR “instruct*” OR “stud*”) OR ab:(“educ*” OR “teach*” OR “train*” OR “learn*” OR “instruct*” OR “stud*”))

Study inclusion and eligibility criteria

We included peer-reviewed studies published in English that investigated the use of screen-based (computer-based) VPs representing patient scenarios with the aim to teach dental students at any stage of training or practice. We excluded studies in which VPs were combined with other kinds of simulation, such as immersive virtual-augmented or mixed-reality environments, mannequins and standardised patients, as well as studies in which VPs represented individual organs or focused on psychomotor skills training, without presenting the case in a clinically relevant scenario. Finally, we also excluded studies that were not original research such as other reviews and meta-analyses, and studies that focused on postgraduate participants or other types of professional training.

Study selection process

The study selection process comprised the following steps: First, we removed duplicates, using EndNote X8 reference manager software. Second, the first author (FJ) scanned the abstracts for eligibility, screening an article in full if the respective abstract lacked information. She subsequently discussed the results in the research team where articles were accepted for further selection by mutual agreement. FJ then followed the same procedure to review and discuss the articles hitherto selected. As a final step, we identified additional papers, using the snowballing technique.

Data collection and analysis

Before analysing the extent to which ID principles were considered, we reviewed the selected studies in terms of their design approach and research methods used, including their study goals, participants, scenario path type, field of dentistry and their main findings (see Appendix for more information). Subsequent review of the studies under scrutiny was guided by the 4C/ID model elements. Figure 2 visualises the analysis framework which we explained in the introduction, the upper part presenting the 4C/ID model components and the lower part listing the associated ID features identified in previous studies. To find out which ID features were applied in each study, the first author subjected the studies to a thorough scrutiny. In case one or more of the features seemed ambiguous, the second author analysed the respective study as well. All authors subsequently discussed the results and finalised the analysis by mutual agreement. By means of descriptive statistics, we brought into focus the frequency with which ID features were applied across studies. For each study listed in the appendix, we generated codes that related to each ID feature of which we will provide descriptions in the Results section.

Fig. 2
figure 2

The 4C/ID model and key ID features identified in other studies

Results

Study selection

Our initial search returned 6744 studies, of which 5,805 remained after duplicates were removed (see Fig. 3). An initial screening of titles and abstracts led to the exclusion of another 5653 studies, either because they were not original research articles or because they combined various types of simulation; whilst some did not use screen-based VPs, yet others used dissimilar types of simulation or mixed these up with augmented or virtual reality (see Fig. 3). This preliminary screening eventually led to the selection of 152 studies for full-text review. After reviewing these texts in full, we excluded those studies that focused on postgraduate learners or mixed training levels, because undergraduates were our focus. Once we had also excluded the studies with vague study designs and studies focusing on psychomotor skills training, 21 studies were ultimately found eligible for inclusion in the present review.

Fig. 3
figure 3

Study selection flow

General study characteristics

The studies applied different approaches, using comparative (cross-sectional, single-group, case–control, cohort-study, and trial) or non-comparative (focus-group, case-study, survey, narrative) designs. More specifically, 15 studies used a comparative approach, one a non-comparative and five a mixed-methods approach. The study goals were mainly to introduce VPs, compare VPs together or with the regular curriculum (Table 1). The number of participants varied from 9 to 171 undergraduate students. In terms of scenario path type, 12 studies used a linear (passive/active) approach involving a single path from introduction through to history taking, examination, diagnosis and treatment planning in which the learner had to finish one section before moving to the next. Six studies, by contrast, had a branching design in which the pathway depended on the learner’s decision at every strategic node in the VP, creating multiple potential paths that could lead to a common or multiple end points. In 17 studies, the focus was on teaching undergraduate students clinical management or clinical reasoning skills such as history taking, clinical examination and treatment planning (Clark et al., 2012; Weiner et al., 2016; Zary et al., 2009, 2012). Four other studies (Boynton et al., 2007; Kleinert et al., 2007; Papadopoulos et al., 2013; Sanders et al., 2008), on the other hand, focused on training students to apply communication and nonclinical skills, for instance when dealing with developmentally disabled paediatric patients. Appendix 1 gives a complete overview of the articles we reviewed and their study features.

Table 1 Goals of the studies under review

Learning task principles

With respect to the first component of the 4C/ID model (Table 2), we found that 14 studies used a variety of tasks ranging from two VP cases (Allaire, 2015; Zary et al., 2012) to train critical thinking and reasoning skills to eight VPs (Clark et al., 2012; Marei et al., 2017, 2018a, 2018b; Weiner et al., 2016) based on complex situations to prepare students for stressful moments in the clinic and to examine different study settings. One study (Zary et al., 2009) even used 24 VP cases for diagnostics and therapy planning in general dentistry during two educational semesters. Another two studies focused specifically on creating different levels of task complexity: Woelber et al. (2012) developed two VPs on localised aggressive periodontitis that differed in their level of complexity from easy to complex. Littlefield et al. (2003), moreover, designed three VPs for undergraduate dental students and general dentists that ranged from low to moderate difficulty. The studies also used different task types or instructional strategies such as patient cases or worked examples (Kleinert et al., 2007; Sanders et al., 2008) to facilitate learning and provide a bridge between knowledge and practice, thereby facilitating the transfer of learning.

Table 2 Instructional design principles and the number of studies that used them

During the VP programme, students can take responsibility for their learning progress either individually or in a group, depending on the learning mode. In the majority of studies (n = 16), VPs were employed for individual use. This means that, during the simulation programme, VPs gave instructions and assessed students’ learning outcomes or their general use based on the following information: the number of questions asked, tests ordered, examinations performed, students’ final scores, the total time spent during the programme (Clark et al., 2012; Schittek Janda et al., 2004; Zary et al., 2012), their knowledge and perceived difficulty level (Boynton et al., 2007; Kleinert et al., 2007; Papadopoulos et al., 2013; Sanders et al., 2008; Woelber et al., 2012; Yoshida et al., 2012), their self-assessment of the learning process and general acceptance (Zary et al., 2009), outcomes of the group discussion at the end of the individual sessions (Allaire, 2015; Lund et al., 2011; Marei et al., 2018a, 2018b), information from alternative tools for teaching and assessing students’ diagnostic skills (Littlefield et al., 2003) and the transfer of knowledge (Marei et al., 2017). In seven studies, by contrast, VPs served as group training activities that involved two or more learners and were followed by reflection sessions (Allaire, 2015; Lund et al., 2011; Marei et al., 2019, 2018a, 2018b; Seifert et al., 2019; Weiner et al., 2016). Of these studies, two employed both individual and collaborative group training, by randomly assigning students to each group (Marei et al., 2019, 2018a, 2018b) to compare their effectiveness. The latter study by Marei et al., (2018a, 2018b) found that a collaborative activity could enhance students’ knowledge acquisition and retention, especially when the activity concerned difficult topics that involved multiple, interacting elements.

According to 4C/ID, curriculum or course integration is a prerequisite for the transfer of learning and filling the gap between theory and practice. In most studies (n = 15), the simulation programme was a formal, integral part of the curriculum that served to help students bridge the gap between what they had learnt in the classroom and the provision of patient care. More specifically, VPs were embedded in the standard curriculum before the actual encounters with patients to prepare students for clinical training. In most cases (n = 13), simulation training was distributed over a period of time ranging from one day (Antoniou et al., 2014; Clark et al., 2012; Lund et al., 2011; Weiner et al., 2016), through one week (Papadopoulos et al., 2013; Schittek Janda et al., 2004), to one or more educational semesters with planned/unlimited access to VPs (Allaire, 2015; Zary et al., 2009).

Supportive information principles

As for the second component of the 4C/ID model (Table 2), only a few studies (n = 2) provided students with access to case-specific background information to support schema formation. Allaire (2015), for instance, did so by offering additional sources such as journal articles and videos. Each section of the case included guided questions that could be answered individually or in pairs, and then discussed with a faculty facilitator in a small group. In the study by Clark et al. (2012), moreover, each participant received a guide sheet outlining the common features of a medical interview and physical examination of the head and neck region. As mentioned before, cognitive interactivity/feedback is another important part of the supportive information in the 4C/ID model. It facilitates the transfer of learning from the classroom to a clinical setting and assists learners in creating cognitive schemas and problem-solving strategies. The studies we reviewed employed a variety of reflection and interaction strategies to promote learners' cognitive engagement. To help learners reflect on the quality of problem-solving processes, some studies used debriefing, summarisation and group discussion after each VP or at the end of the session (Allaire, 2015; Littlefield et al., 2003; Lund et al., 2011; Marei et al., 2019, 2018a, 2018b; Weiner et al., 2016). Other studies provided cognitive feedback from the instructor or facilitator (Allaire, 2015; Antoniou et al., 2014; Lund et al., 2011; Seifert et al., 2019) or during tutor-aided group discussion (Marei et al., 2019, 2018a, 2018b; Weiner et al., 2016) to encourage reflection. In a similar fashion, Zary et al. (2009) provided constructive feedback in the form of an automatically generated checklist that matched student recommendations and compared them to expert recommendations to promote reflection. The strategy used to reinforce learning in the study by Sanders et al. (2008) was to demonstrate the point of teaching. That is to say, after students had entered their responses, they were shown a video clip giving them immediate feedback on their decisions. Schittek Janda et al. (2004), by contrast, described an approach in which learners, rather than being presented with pre-selected options, were forced to mobilise their own cognitive processes in order to make a decision about a particular case. Finally, Woelber et al. (2012) used cognitive interactivity, by providing collaborative discussion between students at decision points and giving feedback on their actions.

Procedural information principles

According to the third component of the 4C/ID model (Table 2), learners must receive feedback at the procedural level to help them perform routine aspects of the learning tasks. The majority of the studies under scrutiny did provide different types of feedback during or after the simulation activity. We already mentioned the cognitive feedback that served to promote reflection on tasks in some studies. At this level, however, we considered corrective feedback that helped students to monitor their performance and to identify and correct their errors. We found that some studies (n = 9) provided immediate feedback after each information and decision point during the programme and presented the correct or incorrect answer depending on students’ responses (Abbey et al., 2003; Boynton et al., 2007; Kleinert et al., 2007; Lund et al., 2011; Mardani et al., 2020; Marei et al., 2019, 2018a, 2018b; Seifert et al., 2019; Zary et al., 2009). Kleinert et al. (2007), for instance, provided immediate and corrective feedback on students’ response after each decision point. In some other studies (Boynton et al., 2007; Marei et al., 2019, 2018a, 2018b; Papadopoulos et al., 2013; Seifert et al., 2019; Woelber et al., 2012), students immediately received pertinent information that included a text-based description of the consequences of their choices. The purpose of this approach was to ensure that students could not proceed to the next step until the correct option was chosen.

Some studies (n = 5) provided just-in-time information during practice in the form of detailed instructions via the chat window before students entered their choices. Seifert et al. (2019), for instance, presented additional information via text boxes (glossary) to compensate for the lack of on-site interaction with a supervisor. In Clark et al. (2012)’s study, moreover, students received a guide sheet outlining the common elements of a medical interview and physical examination. Finally, in both Kleinert et al. (2007) and Sanders et al. (2008)’s study, students received a concise summary of information pertinent to the case during so-called ‘information points’.

Part-task practice principles

Our scrutiny of the selected studies did not yield any information on part-task practice, the fourth component of the 4C/ID model.

Discussion

In reviewing the body of empirical studies, we encountered a range of perspectives on VP design and implementation. Previous studies have emphasised the role of specific ID features and associated these with effective VP design (Cook et al., 2013; Issenberg et al., 2005). In doing so, however, they did not link to an existing, integrated ID model, and their recommendations are therefore difficult to apply in future VP design. VPs will be effective only when both the simulation and the educational activities around them have been properly designed (Ellaway, 2014). The aim of the present study was therefore to determine whether or not and to what extent the 4C/ID principles as part of a given, proven and closely connected ID model have been applied in VP design in dental education and what we can learn from this in terms of strengths and weaknesses. Our overall conclusion is that a structured approach to the design and implementation of VPs is lacking. Although all but one of the key ID principles identified were covered in at least part of the studies, none of them used an integrated design.

Learning tasks

According to the 4C/ID model, learning tasks should, in addition to being varied so as to reflect the variety encountered in real life and to promote transfer, have a constant level of complexity within each task class that should, moreover, increase as tasks become more advanced (van Merriënboer & Kirschner, 2018). In general, most VP designers appeared to be well aware that clinical task variation is crucial to promote the transfer of learning.

With regard to the learning mode, the studies by Marei et al., (2018a, 2018b, 2019) confirm that such decision should depend on the complexity of the task (Issenberg et al., 2005; Kirschner et al., 2009). Marei et al., (2018a, 2018b) suggested that when VP cases are complex and involve more interaction between multiple elements, the decision to use collaborative learning with VPs is more obvious. In Marei et al. (2019)’s study, the collaborative group performed significantly better and experienced less intrinsic and extraneous cognitive load than when VPs were used independently. Other studies, by contrast, laid a greater emphasis on the benefits of individual VP use. Yoshida et al. (2012), for instance, concluded that VPs can help students to become self-directed lifelong learners because they can act independently, in a self-paced manner, to solve problems. Most studies, however, did not motivate their choice of learning mode. We therefore welcome additional research on how best to strike a balance between individual and group-based learning with VPs.

As another important feature of ID, curriculum integration has been addressed by previous studies which emphasised that VPs should be aligned with the learning goals and be properly incorporated into educational programmes and curricula (Kononowicz et al., 2016). More specifically, Marei et al. (2019) suggested that VPs should be designed in accordance with their intended use, which should therefore already receive consideration in the early stages of VP development. More than half of the studies we reviewed had embedded simulations as an integral part of their curriculum or training programme, and a few had even investigated how they could achieve such integration effectively (Boynton et al., 2007; Lund et al., 2011; Schittek Janda et al., 2004). Finally, Seifert et al. (2019) concluded in their study that it is, indeed, feasible to integrate VP cases within a curriculum and that this leads to substantial growth of clinical competence.

Supportive information

The second component of the 4C/ID model specifies that learners should receive supportive information explaining how to approach a problem, suggesting how to organise the domain or offering foundational knowledge. When teaching complex topics, the preferred method is to sequence VPs after lectures in a deductive approach as it could facilitate knowledge transfer. Importantly, inductive or deductive approaches to sequencing VPs should differ in the amount of guidance depending on students’ prior knowledge and VP complexity (Marei et al., 2017).

In a minority of the studies, cognitive interactivity/feedback was used to promote reflection on learning processes. The approach to achieve this described in Antoniou et al. (2014) was to offer students constructive feedback in each step of task completion. In other studies, such feedback was delivered by VPs (Boynton et al., 2007) or the instructor (Allaire, 2015; Antoniou et al., 2014; Lund et al., 2011; Seifert et al., 2019), either during the programme or after it in a tutor-aided group discussion (Marei et al., 2019, 2018a, 2018b; Weiner et al., 2016). According to Lund et al. (2011), the successful application of VPs depended on the amount and type of feedback, as well as on the feedback given in the teacher-aided discussion. Lack of attention to cognitive feedback could undermine students’ ability to establish effective cognitive schemas that improve their future understanding and performance (van Merriënboer & Kirschner, 2018).

Procedural information

At procedural level, in a narrow majority of the studies, learners received feedback in a corrective form, although this feedback was not always clearly distinguishable from cognitive feedback. In these cases, learners received immediate feedback after each decision point and, depending on their responses, were shown the correct or incorrect answer (Abbey et al., 2003; Boynton et al., 2007; Kleinert et al., 2007; Lund et al., 2011; Marei et al., 2019, 2018a, 2018b; Seifert et al., 2019; Zary et al., 2009). Sometimes, this just-in-time information took the form of detailed instructions with a brief explanation of the consequences of their choices (Marei et al., 2019, 2018a, 2018b; Papadopoulos et al., 2013; Seifert et al., 2019; Woelber et al., 2012).

Part-task practice

As the final component of the 4C/ID model, part-task practice is only needed when a routine aspect of a complex skill requires some degree of automation. As such, repetitive practice did not receive any attention in the studies under review. This may be attributable to the nature of the subjects and type of practices we included. Since we only considered full patient cases about clinical reasoning and management skills to the exclusion of psychomotor skills training, our inclusion criteria may have obviated the need for automation.

Conclusion

VP training affords students the opportunity to experience early clinical exposure by practising real patient scenarios in a safe environment. This study attempted to structure applied ID principles with the aim to design VPs in dental education based on the 4C/ID model. At a general level, we found that the key ID principles were well represented in our studies. This suggests that it is feasible to proceed to the next step in designing VPs, that is, by using a structured ID approach. An integrated design approach such as 4C/ID will improve the effectiveness of VPs (Ellaway, 2014) and will enable us to compare alternate IDs in order to shed more light on the mechanisms of effective simulation-based instruction (Cook et al, 2013). In the next sentences, we will discuss our key findings and propose future design applications based on the strengths and weaknesses of the features applied. During VP training, students received immediate or delayed feedback from the system and sometimes also participated in an instructor-guided reflection session after the programme. Lack of attention to cognitive feedback might undermine students’ ability to establish effective cognitive schemas that improve their future understanding and performance. The 4C/ID model emphasises the provision of cognitive feedback precisely to create such schemas and reflect on the learning process. Another factor that is essential to facilitate learning and connect theory to practice is curriculum integration. When properly incorporated into curricula, VP training has the potential to promote self-directed learning in students, helping them to formulate learning objectives and to improve the quality of their knowledge structure in the process. The choice of learning mode, moreover, should depend on the aims and complexity of the respective VP case; nevertheless, we need additional research on how best to strike a balance between individual and group-based learning with VPs. Finally, from our review findings we may infer that most VP designers generally appreciate the importance of clinical task variation to improving the transfer of learning.

Lessons learnt from strengths and weaknesses

As a matter of conclusion, below we will summarise the key lessons learnt from this review with respect to the strengths and weaknesses of applying the 4C/ID model in VP design in dental education:

  • Most VP designers generally appreciate the importance of clinical task variation to improving the transfer of learning.

  • Collaborative VPs are clearly beneficial in complex cases where multiple elements interact. However, we need additional research on how best to strike a balance between individual and group-based learning with VPs.

  • VPs should be aligned with the learning goals and be properly incorporated into curricula.

  • Lack of attention to cognitive feedback might undermine students’ ability to establish effective cognitive schemas that improve their future understanding and performance.

Finally, to further improve the design of VPs and to enable the comparison and study of their effectiveness, we recommend a structured ID approach such as 4C/ID.

Prospects and limitations

This review addressed the substantial gaps in a design and curriculum integration of VPs that is based on the ID model/principles. It also underscored the need for more research that directly compares alternate ID principles and the interaction effects among them, to better understand the role of specific ID features in designing effective simulation-based instruction.

In this study, we only focused on screen-based, whole VP scenarios designed for clinical reasoning and management skills training in dentistry. Consequently, our inclusion criteria (we did not consider psychomotor skills) affected the attention that the studies under review paid to certain ID features.

Finally, we only had access to the study reports and not to the VPs themselves or the information on how they were actually incorporated into the curriculum.