Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

The effect of observing novice and expert performance on acquisition of surgical skills on a robotic platform

  • David J. Harris,

    Roles Data curation, Formal analysis, Investigation, Visualization, Writing – original draft

    Affiliation Department of Sport and Health Sciences, University of Exeter, Exeter, United Kingdom

  • Samuel J. Vine,

    Roles Conceptualization, Funding acquisition, Methodology, Project administration, Resources, Software, Supervision, Writing – review & editing

    Affiliation Department of Sport and Health Sciences, University of Exeter, Exeter, United Kingdom

  • Mark R. Wilson,

    Roles Conceptualization, Funding acquisition, Investigation, Project administration, Resources, Software, Supervision, Writing – review & editing

    Affiliation Department of Sport and Health Sciences, University of Exeter, Exeter, United Kingdom

  • John S. McGrath,

    Roles Conceptualization, Funding acquisition, Methodology, Writing – review & editing

    Affiliations Exeter Surgical Health Services Research Unit, RD&E Hospital, Exeter, United Kingdom, University of Exeter Medical School, Exeter, United Kingdom

  • Marie-Eve LeBel,

    Roles Conceptualization, Funding acquisition, Methodology, Supervision, Writing – review & editing

    Affiliation Division of Orthopaedic Surgery, University of Western Ontario, London, Canada

  • Gavin Buckingham

    Roles Conceptualization, Funding acquisition, Methodology, Project administration, Resources, Supervision, Writing – review & editing

    G.Buckingham@exeter.ac.uk

    Affiliation Department of Sport and Health Sciences, University of Exeter, Exeter, United Kingdom

Abstract

Background

Observational learning plays an important role in surgical skills training, following the traditional model of learning from expertise. Recent findings have, however, highlighted the benefit of observing not only expert performance but also error-strewn performance. The aim of this study was to determine which model (novice vs. expert) would lead to the greatest benefits when learning robotically assisted surgical skills.

Methods

120 medical students with no prior experience of robotically-assisted surgery completed a ring-carrying training task on three occasions; baseline, post-intervention and at one-week follow-up. The observation intervention consisted of a video model performing the ring-carrying task, with participants randomly assigned to view an expert model, a novice model, a mixed expert/novice model or no observation (control group). Participants were assessed for task performance and surgical instrument control.

Results

There were significant group differences post-intervention, with expert and novice observation groups outperforming the control group, but there were no clear group differences at a retention test one week later. There was no difference in performance between the expert-observing and error-observing groups.

Conclusions

Similar benefits were found when observing the traditional expert model or the error-strewn model, suggesting that viewing poor performance may be as beneficial as viewing expertise in the early acquisition of robotic surgical skills. Further work is required to understand, then inform, the optimal curriculum design when utilising observational learning in surgical training.

Introduction

Despite the shifting emphasis within surgical education towards competency-based training, current curricula still depend upon didactic learning, doing under guidance, and repetition [1], with observational learning—the process of passively watching another individual perform a task—maintaining a core role in developing surgical expertise [2]. Whilst the value of observing an expert model is well established in a wide range of contexts [3,4], there is increasing support for the benefits of observing error-strewn performance [5,6]. Given the role of observational learning in effective and economical surgical training [2], this study aimed to determine the style of visual information leading to the most effective early acquisition of surgical skills on a robotic system.

Robotically-assisted devices are now widely used in minimally invasive surgery due to their benefits for patient outcomes and increased levels of surgeon satisfaction [79]. There is, however, limited research on how best to instruct trainees during robotic surgical training, when the user must establish a new range of sensorimotor contingencies [10]. Observational learning, the process of observing an actor perform a task prior to physical practice, provides a way of accelerating this type of motor learning [4,11]. Modelling of actions is an effective method for transmitting information, particularly when complex movements are hard to verbalise. While observation is unlikely to lead to optimal learning on its own [3], it is particularly effective as a supplement to physical practice [11]. This benefit is often attributed to the observer developing a model of the coordination patterns required, or a ‘perceptual blueprint’ [12,13], that serves as a standard of reference. This may occur through overlapping neural circuits for action observation and action production [14,15].

The efficacy of observational learning has previously been demonstrated for surgical skill acquisition [16]. For example, in a minimally invasive surgical task, Snyder et al. [17] found simply observing an instance of expert performance to benefit medical students already trained to proficiency on a virtual reality (VR) simulator. Growing evidence, however, suggests that the observed model need not be an expert, and that in fact, viewing errors may contribute to the development of mechanisms for error detection and correction [3,18,19]. For instance, Buckingham et al. [6] found object lifting errors to be reduced in observers who viewed the lifting errors of others. Evidence for the benefit of error observation in a more complex task has also been found in simulated orthopaedic surgery [20]. Following baseline assessment, participants observed either an expert surgeon or an untrained novice perform a simple localization task, where the novice model displayed more erratic instrument control, slower task progression and poorer target visualisation. LeBel et al. found that the two groups improved to a similar degree in the majority of metrics immediately post-observation, but at a one-week follow-up the novice observation group showed significant improvements over their expert-observing counterparts. Given that observing expertise is suggested to promote a model of coordination patterns, and that observing errors may enable models of error detection and correction, the combined benefits of the two approaches could maximise this style of learning. The efficacy of viewing a ‘learning model’, developing from high-error to low-error performance, indicates a mix of performance may aid learning [18]. Indeed, in a sequential timing task Rohbanfard and Proteau [21] found a mixed novice/expert schedule to be most effective for learning [see also [22,23]. Nonetheless, there is only preliminary evidence for the benefits of mixed model observation. Therefore we aim to evaluate the effectiveness of a mixed model in comparison to novice and expert models alone.

In summary, prior research indicates that learning from errors may provide significant benefits for skill acquisition over sole observation of expertise [3,20,24]. In order to examine observational learning in the context of the acquisition of early robotic-surgical skill learning, we examined performance on a daVinci surgical system (Intuitive Surgical Ltd.) before and after observing novice, expert and mixed models. Furthermore, in order to determine the role played by attention in facilitating improvement, participants’ eye movements were recorded during model observation. It is expected, following LeBel et al. [20], that observing errors will be as beneficial as observing expertise immediately post-intervention, but will facilitate improved retention at a one week follow-up.

Hypotheses

  • All observational learning groups will outperform the control group post-intervention and at a retention test.
  • Participants viewing the error-strewn model will outperform the expert model at retention.
  • The mixed observation model (novice then expert) will outperform either model alone at retention.

Methods

Participants

120 undergraduate medical students (48 males) with no previous experience of robotically-assisted surgery were recruited through advertisement during lectures (mean age = 20.2 years, standard deviation = 2.5). A power analysis using G*Power [25] indicated that to find a group difference in line with LeBel et al. (20) of η2 = .10 with power = 0.95, a minimum of 104 participants were required. Ethical approval from the University of Exeter research ethics board was obtained prior to data collection. Participants gave written informed consent at the start of testing and were compensated £15.00 for participation.

Stimuli and equipment

The experimental tasks were performed on the daVinci S robotically-assisted surgical system (Intuitive Surgical Ltd., Sunnyvale, CA). This system consists of a remote control/viewing console connected to a separate operating cart with three robotic arms (see Fig 1), two of which were connected to a laparoscopic tool (a needle driver with articulated wrist) and one connected to an endoscope. The endoscope records two high definition videos (16:9 aspect ratio) that are played, stereoscopically, in real time to the control console, providing a 3D view of the surgery. The instruments are fingertip controlled with a motion scaling algorithm allowing mapping of hand to instrument movements. The articulated wrist of the laparoscopic instrument allows movements across the same degrees of freedom as the human hand (see [26]).

thumbnail
Fig 1. The DaVinci S robotic platform with surgeon console (left) and operating cart (centre).

https://doi.org/10.1371/journal.pone.0188233.g001

We examined performance across one primary task and two separate transfer tasks. The primary task (see Fig 2a) was a ‘ring-carrying’ exercise adapted from the Fundamentals of Robotic Surgery curriculum [27]. A plastic ring is transported along a 15cm curved wire, with the aim of moving the ring from the base to the end of the wire with the right hand, transferring to the left hand and then returning to the starting position without touching the wire. This task is designed to assess proficiency of instrument control. The first transfer task was a ‘ball pick up and drop’ (see Fig 2b), adapted from the Fundamentals of Laparoscopic Surgery curriculum [28] and used previously to assess robotic and laparoscopic skills [9,29]. Five foam balls sat on raised platforms and had to be picked up and transferred to a receptacle. The second transfer task was a ‘knot tying’ exercise (see Fig 2c) adapted from the Fundamentals of Robotic Surgery curriculum, which requires a simple half-knot to be tied in a piece of string suspended from two wires. This task was designed to mimic basic suturing skills.

thumbnail
Fig 2.

(a) Ring-carrying task (left). (b); Ball pick-up and drop (centre). (c); Knot-tying (right).

https://doi.org/10.1371/journal.pone.0188233.g002

Video stimuli for the observational learning intervention were developed by recording performance of the primary task (ring-carrying) with a GoPro Hero 3+ camera from novice and expert models. The video view displayed the ring-carrying task and the instruments (as in Fig 2a). The novice video was selected from a number of trials collected from novice performers (non-surgeons with no prior experience) during pilot testing that showed typical errors (e.g. repeated touches of the wire, erratic movements). The expert video featured performance by a consultant urologist, an expert with the daVinci system. Each video showed the task being completed twice. The mixed video showed one instance of the novice video followed by one instance of the expert video.

Procedure

Participants were required to attend two testing sessions (approximately 30 minutes each) at the Surgical Health Service Research Unit (HeSRU) of the Royal Devon and Exeter Hospital. Visit one consisted of a baseline test on the primary ring-carrying task, followed by the observation intervention and a post-test on the ring-carrying task. Visit two, which occurred seven days after the first visit (as in [20]), assessed retention of learning on the primary task followed by two transfer tasks.

At the start of testing participants were assessed for stereoscopic acuity using the Randot stereoacuity test (Stereo Optical Co, Inc., Chicago, IL). The test measures participants’ ability to discriminate a stereogram circle from two distractors across increasing levels of difficulty (from 400 to 20 seconds of arc), using a 1–10 scoring system. Participants then watched a short introductory video explaining the experimental procedures. Participants were then given one minute to familiarise themselves with the console controls prior to the baseline test (as in [9]). Task performance was recorded with a GoPro Hero 3+ camera set to the side of the stimuli (out of view), for later offline analysis. Instrument movement parameters were recorded with a pair of GENEActiv accelerometers attached to each of the robot arms, which recorded tri-axial accelerations at 100 Hz.

Following the baseline ring-carrying test, participants were randomly assigned to one of the four observation conditions using random number generation. During the three observational learning conditions (novice, expert and mixed) participants watched a video of the ring-carrying task, whilst the control group were given a word search task for the mean duration of the videos. Participants in the video observation groups had point of gaze recorded using SMI ETG 2.0 eye-tracking glasses (SensoMotoric Instruments, Boston MA) that record onto a customized Samsung Galaxy smartphone. The glasses are lightweight (76 g) and record binocular eye movements and the visual scene at 30 Hz to a spatial resolution of 0.5°. During the videos, participants placed their head in a chin rest to minimise head movement and ensure attention was appropriately focused on the content of the videos. All videos were presented on an HD 24 inch Acer monitor placed 48 cm from the participant.

Data analysis

Performance scores on the three experimental tasks were obtained from video analysis following testing. For the ring-carrying task, time to completion was obtained from the recording, then errors were manually coded (number and type of error). A small touch of the wire scored one point and a drag along the wire or dropping the ring scored two points. The number of errors was divided by the time to completion to create an errors per second score. For the knot tying task, time to completion was similarly recorded, with a performance score obtained from a scoring system for surgical knots adapted from Tytherleigh et al. [30,31]. Points were assigned for completing the subsections of the knot correctly, without fumbling or making repeated attempts. For the ball drop transfer task, time to completion was recorded and errors coded using one point for a missed grab of the ball and two points for dropping the ball.

Surgical instrument proficiency was assessed using movement parameters obtained from accelerometry. Tracking of implement movement is widely used for assessing surgical skill, with path length and number of movements particularly linked to expertise [32,33]. Two movement variables were chosen to assess performance in this task. Firstly, jerk, a measure of change in acceleration over time, has previously been linked to surgical performance [34,35] and indicates smooth instrument movements [36]. Secondly, entropy, a measure of randomness/complexity, has been linked to skill learning and automaticity [37] and has previously been used to quantify surgical performance [38]. Accelerometry data was collated and synchronised using GENEActiv PCSoftware, and then analysed using Matlab. Raw data was filtered using a 2nd order, 3Hz Butterworth low pass filter to reduce noise [39], before converting to the Euclidean Norm Minus One [ENMO], which integrates x, y, and z plane accelerations and removes the net effect of gravity [40]. The ENMO provides a measure of total acceleration in any direction. Matlab was also used to calculate jerk [derivative of accelerations with respect to time] from the ENMO, using the formula Jerk = ΔAcc/ΔTime. Sample entropy of accelerations, which is robust to varying sample lengths, was calculated using the natural logarithm of the conditional probability that a series similar for n points remains similar at the next point [see [41]].

Allocation of attention has been linked to effective observational learning [42], which can be assessed through point of gaze [43]. Gaze behaviour has previously been used to discriminate attentional strategies of novice and expert surgeons [44] and can indicate which areas of the task observers pay attention to [45]. Gaze data was analysed using SMI BeGaze 3.6 software, which provides automatic fixation detection, and allows user defined areas of interest (AOIs) to be assigned to the visual scene. Dynamic AOIs were assigned to the ring and robot end effectors to determine the proportion of time spent fixating relevant/irrelevant information and the frequency of switches between areas. Data that displayed a poor calibration was removed from the analysis (n = 8).

Statistical analysis was performed using JASP [46]. Outliers (+/- 3 standard deviations from the mean) in performance (n = 5) and accelerometry (n = 3) were removed. Dependent variables were analysed using 3 (trial) x 4 (group) mixed ANOVAs, with one-way ANOVA used to examine interactions and Bonferroni-Holm corrected post-hoc tests where appropriate. Violations of sphericity were corrected for using a Greenhouse-Geisser correction factor. In order to determine equivalence of null effects, Bayes Factors (BF10) for main effects and paired comparisons were also obtained using a symmetric Cauchy prior. In addition, to evaluate the extent to which visual attention relates to subsequent learning, Spearman’s correlation coefficients were computed between gaze metrics and performance.

Results

One way ANOVA indicated there to be no group differences in age (p = .51, BF10 = 0.10) or stereovision (p = .40, BF10 = 0.14), showing the groups to be well matched (see Table 1).

To determine the relationship between our objective measures of instrument movement and the subjective coding of error performance, correlational analysis was used across the four groups. There was a moderate relationship between jerk and performance (see Fig 3), indicating increased jerk was related to more errors per second, r(351) = .52 p < .001, BF10 = 5.25e+28. There was, however, no clear relationship between acceleration entropy and errors per second, r(351) = .08, p = .11, BF10 = .30 (see Fig 4).

thumbnail
Fig 3. Relationship of instrument jerk with rate of errors, displaying regression line with 95% confidence intervals.

https://doi.org/10.1371/journal.pone.0188233.g003

thumbnail
Fig 4. Relationship of instrument movement entropy with rate of errors, displaying regression line with 95% confidence intervals.

https://doi.org/10.1371/journal.pone.0188233.g004

To measure group differences in movement parameters a 3 (trial) x 4 (group) mixed ANOVA was conducted on mean jerk (m/s3) (see Fig 5). There was a significant main effect of trial, F(2,232) = 5.02, p = .007, ꙍ2 = .033, BF10 = 2.96, with jerk increasing at retention. There was no main effect of group, F(3,116) = 0.80, p = .50, ꙍ2 = .000, BF10 = .17 and no interaction, F(6,232) = 0.71, p = .64, ꙍ2 = .000, BF10 = .03. Post-hoc tests on the effect of trial showed no difference from baseline to post-intervention (p = .12, BF10 = 0.23), but there were significant differences from baseline to retention (p = .02, BF10 = 2.90), and post-intervention to retention (p < .001, BF10 = 24511.24).

thumbnail
Fig 5. Mean (and standard error) of instrument jerk (m/s3) across experimental trials.

https://doi.org/10.1371/journal.pone.0188233.g005

Similarly, a 3 (trial) x 4 (group) mixed ANOVA was conducted on acceleration entropy (see Fig 6), indicating a significant main effect of trial, F(2,232) = 11.82, p < .001, ꙍ2 = .084, BF10 = 1443.24, with entropy decreasing over trials. There was no main effect of group F(3,116) = 0.49, p = .69, ꙍ2 = .000, BF10 = .09, or interaction effects, F(6,232) = 0.16, p = .99, ꙍ2 = .000, BF10 = .01. Post-hoc tests on the effect of trial showed a significant difference from baseline to post-intervention (p = .009, BF10 = 5.33), baseline to retention (p < .001, BF10 = 521.30) and post-intervention to retention (p = .009, BF10 = 4.69).

thumbnail
Fig 6. Mean (and standard error) of instrument entropy (bits) across experimental trials.

https://doi.org/10.1371/journal.pone.0188233.g006

To evaluate the effect of the intervention on performance a 3 (trial) x 4 (group) mixed ANOVA was conducted on errors/second scores (see Fig 7). There was a significant main effect of trial F(1.80,194.47) = 13.03, p < .001, ꙍ2 = .093, BF10 = 1400.70, indicating general improvement from baseline. There was also a significant main effect of condition, F(3,108) = 2.75, p = .05, ꙍ2 = .045, BF10 = 1.25 and a significant trial by group interaction, F(5.40,194.47) = 2.79, p = .012, ꙍ2 = .041, BF10 = 2.08.

thumbnail
Fig 7. Mean (and standard error) of performance (errors/second) across experimental trials.

https://doi.org/10.1371/journal.pone.0188233.g007

To examine the condition by trial interaction, follow-up one-way ANOVAs were conducted to test for group differences at each time point. At pre-test there were no significant group differences, F(3,115) = 0.74, p = .53, ꙍ2 = .000, with a BF10 = .11 providing strong evidence for group equivalence. At post-intervention, there was a significant effect of group, F(3,112) = 3.88, p = .01, ꙍ2 = .069, BF10 = 3.66, indicating an intervention effect. Post-hoc analysis showed significant differences between Control and Novice (p = .02, BF10 = 4.58) and Control and Expert (p = .04, BF10 = 2.55) groups, but there were no differences between Novice and Expert (p = .69, BF10 = 0.31), Novice and Mixed (p = .49, BF10 = 1.34), Expert and Mixed (p = .61, BF10 = 0.59) or Mixed and Control (p = .40, BF10 = 0.61). This indicates that watching the Novice or Expert video significantly improved performance over the Control condition.

At retention there was also a significant main effect of group, F(3,113) = 3.29, p = .02, ꙍ2 = .055, BF10 = 1.86. After correction for multiple comparisons, however, post-hoc analysis indicated no significant pairwise differences, although there was a trend towards both novice (p = .07, BF10 = 11.44) and expert (p = .07, BF10 = 9.00) groups showing a difference from the mixed condition. A Bayes factor of BF10 = 0.27 (p = .99) suggests that Expert and Novice groups were equivalent, as were Control and Mixed BF10 = 0.30 (p = .99), but there was inconclusive evidence for Control v Expert, (p = .27, BF10 = 0.78) and Control v Novice (p = .27, BF10 = 0.76).

In the knot tying transfer task, there was no effect of group on time to completion, F(3,116) = 0.18, p = .91, ꙍ2 = .000, BF10 = 0.06 or knot tying score, F(3,115) = 0.23, p = .88, ꙍ2 = .000, BF10 = 0.06. For the ball drop task, there was also no effect of group on time to completion, F(3,115) = 1.09, p = .36, ꙍ2 = .002, BF10 = 0.16 or errors, F(3,113) = 1.48, p = .23, ꙍ2 = .012, BF10 = 0.24. For both tasks, Bayes factors provide moderate to strong evidence that performance was equivalent across groups.

In order to evaluate the relationship between gaze measures and performance, Spearman’s correlation coefficient was calculated. There was a significant relationship between irrelevant fixation time (away from instruments or ring) and performance at retention, r(79) = .31, p = .006, but not immediately post intervention, r(79) = .18, p = .13. Irrelevant fixations also related to increased entropy in instrument movements post-intervention, r(79) = .35, p = .002. There was no relationship between proportion of time spent observing either the instrument effectors or the ring and subsequent performance (all p values >.30), and no relationship between average fixation duration and performance (all p values >.37).

Discussion

Surgical skills training has a direct and lasting impact on surgical success and patient outcomes [47,48], but there has been limited investigation of how teaching methods utilised in surgical training curricula can be improved. The acquisition of motor skill in robotic surgery is strongly related to volume of physical practice [49]. Research has shown that observational learning activates the same cortical motor regions as physical practice [50] and allows complex skills to be learnt more quickly [11]. In addition, action observation can contribute to skill learning when physical practice is not possible due to limited training resources, or during online or simulated training programmes. Therefore, this study aimed to build on recent evidence promoting the benefits of observing error-strewn performance [3,6,20] to determine whether viewing errors provided an effective model in a robotically-assisted surgical training task.

Overall, although we observed consistent performance benefits of observation across all of our video-watching groups, the content of what was observed had limited impact. As hypothesised, and in line with previous evidence [19,22,24], observing error-strewn performance provided an immediate learning benefit compared to a non-observation control. This was seen through a reduction in errors per second from baseline to post-intervention. However, unlike Buckingham et al. [6] and Brown et al. [24] there was no benefit of observing errors above that of observing expertise. Comparable learning from novice and expert models has been found previously [19], and LeBel et al. [20] only found group differences after a one-week interval, indicating novice/expert differences may take time to develop in a complex task.

It was also predicted that a mixed observation schedule would afford the greatest learning benefit, as it would provide the ‘perceptual blueprint’ from the expert and the benefits for error detection from the novice. However the results of the mixed group were inconclusive, providing no evidence for the benefit of a mixed schedule over and above the other observation groups, immediately after viewing the videos. Furthermore, there was no difference between the mixed group and the non-observation control groups one week after the intervention, suggesting that observation of a combination of expert and novice performance might impair retention of skills compared to viewing either model in isolation. This finding is in contrast to those of Rohbanfard and Proteau [21] who found a mixed expert/novice model to be more beneficial than observing either model alone. This difference may be partly explained by the very simple sequential timing task used by Rohbanfard and Proteau [21], as findings from simple tasks may not relate well to the acquisition of complex sensorimotor skills [11]. The benefit of a mixed model has also been demonstrated by Domuraki et al. [5] in a central line insertion task. Domuracki et al. [5] found a performance benefit for a mixed model over the expert only model, but only when it was accompanied by specific feedback relating to observed errors. The authors concluded that errors are of benefit when observers know them to be errors, although the results could also be a more general effect of receiving additional feedback. In essence, the present findings may diverge from previous studies due to these task and feedback differences, although the inconclusive benefit of the mixed observation does suggest that there was no additive effect from seeing both novice and expert performance.

Additionally, the current findings differ from the results of LeBel et al. [20] in terms of performance of the task at a one-week follow up retention test. While LeBel et al. found group differences at retention, our results were less clear showing no obvious difference in terms of performance between the novice and expert groups. This apparent similarity between the groups was also evident in terms of performance in either transfer task. As no observation condition outperformed controls at transfer, benefits for learning were likely related to specifics of the ring-carrying task rather than a more general acquisition of skill on the daVinci system.

Our findings may diverge from similar previous studies due to additional factors influencing the mapping of observation to action [15]. For one, unlike the present study, participants in LeBel et al. (20) observed the operator’s hand movements. Hence they saw not only the movement effects [the surgical instruments] but also the effector [the hands of the operator]. Mattar and Gribble [18] suggest that in learning from errors, observers model not only what movements to make, but how to make them. Therefore the benefit of error observation may have been greater in LeBel et al. as observers not only received information of what errors to avoid, but how to avoid them. This is supported by unpublished data from LeBlanc et al. [51], who recorded gaze behaviour whilst untrained participants observed the novice and expert video models used by LeBel et al. LeBlanc and colleagues found that time spent observing the model’s hands was related to reduced task completion time and efficient movement in the subsequent simulated arthroscopic task. This effect was driven by the novice observation group, who showed strong relationships between observing the hands and probe distance, time to completion and a global performance score. These relationships did not hold in the expert-observing group, suggesting that when learning from errors, how the error is made provides crucial information.

Another difference between the current study and previous work might relate to the feedback available to the operator. In the surgical simulator used by LeBel et al. [20], participants received haptic feedback of contact with anatomical structures provided by robotic force-feedback devices. In the current study, by contrast, participants received no haptic feedback from the end-effector. While touch feedback is crucial for completion of many basic manual tasks [52], the evidence for the benefits of haptics in robotic surgical performance is not unanimous [53]. A meta-analysis by Van der Meijden and Schijven [54] shows mixed results for the importance of haptic feedback in surgical performance, but a generally positive effect for early phase skill acquisition. Mattar and Gribble [18] describe how simulation is required when adapting information about movement kinematics from visual information, which is likely supported by haptic information given the benefit of haptic feedback for motor skill learning [55,56]. As such, viewing the hands and receiving haptic feedback may have influenced the way errors were simulated, through increased knowledge of how the errors were made. Future studies may wish to examine whether the effectiveness of learning from errors depends, in part, on receiving information about how to avoid them [18,51].

Consistent with previous work showing that attention can modulate effective observational learning [13,42], we found increased time directing gaze to irrelevant areas of the task (i.e., not on the ring or instruments) was found to be related to increased movement entropy immediately post task, and poorer performance at retention. This relationship illustrates how inattention can be detrimental to learning. Interestingly, there was no obvious benefit of viewing the ring versus the instrument end effectors, in contrast to previous work which has demonstrated benefits of attending to the target of the action in surgical skill learning [57,58]. Thus, although paying attention to relevant aspects of the task is crucial [13], here, attending to the instruments or the ring was equally informative.

An additional outcome of this study is the utility of accelerometry measures for assessing performance in a surgical task. Wrist-worn accelerometers have previously been used for the tracking of gait kinematics [39], but had not been examined in a surgical setting. Motion tracking from surgical simulators has been used previously to calculate measures like path length, force exertion and jerk [20,35,59] that have clear relevance for surgical performance. The current findings, however, suggest that wrist worn accelerometers can provide a simple and cost effective alternative for assessing movement parameters and can contribute to objective assessment of surgical skill. Here, and in line with previous research [36], improved performance was related to reduced instrument jerk. It has been found previously that novices differ from experts in force parameters within laparoscopic surgery, exerting greater force with larger accelerations and jerk [35,36,60]. Trejos et al. [35] found force based metrics like jerk and acceleration to be a better predictor of experience level than task completion time. Assessment of jerk has strong face validity, given that large erratic movements can lead to tissue damage and poor patient outcomes [61]. However, the same relationship was not found for acceleration entropy, a measure of randomness, despite a strong reduction in entropy across trials. While participants’ movements may have become less random with experience and growing confidence [37], absolute measures of force, like jerk, may have more direct relevance for surgical performance.

A limitation of this study may be the restricted physical practice undertaken by participants. Participants may have still obtained major benefits from extra practice that meant group differences were more difficult to observe. Additionally Andrieux and Proteau [22] suggest that the benefits of error-observation are maximised through accompanying physical practice. Future research may wish to intersperse multiple retention tests with periods of physical practice to assess whether group differences emerge. Also, given the differences found between the present study and LeBel et al. [20], future studies could usefully assess the benefit of seeing the kinematics of the operator’s hands for understanding how to avoid errors. Further work shall aim to incorporate these findings to evaluate how the use of varied models and different styles of information might enhance the efficacy of training curricula.

References

  1. 1. Gorman PJ, Meier AH, Rawn C, Krummel TM. The future of medical education is no longer blood and guts, it is bits and bytes. Am J Surg. 2000 Nov 1;180(5):353–6. pmid:11137686
  2. 2. Cordovani L, Cordovani D. A literature review on observational learning for medical motor skills and anesthesia teaching. Adv Health Sci Educ. 2016 Dec 1;21(5):1113–21.
  3. 3. Blandin Y, Proteau L. On the cognitive basis of observational learning: Development of mechanisms for the detection and correction of errors. Q J Exp Psychol Sect A. 2000 Aug 1;53(3):846–67.
  4. 4. Hodges NJ, Williams AM, Hayes SJ, Breslin G. What is modelled during observational learning? J Sports Sci. 2007 Mar 1;25(5):531–45. pmid:17365540
  5. 5. Domuracki K, Wong A, Olivieri L, Grierson LEM. The impacts of observing flawed and flawless demonstrations on clinical skill learning. Med Educ. 2015 Feb 1;49(2):186–92. pmid:25626749
  6. 6. Buckingham G, Wong JD, Tang M, Gribble PL, Goodale MA. Observing object lifting errors modulates cortico-spinal excitability and improves object lifting performance. Cortex. 2014 Jan 1;50:115–24. pmid:23953062
  7. 7. Herron DM, Marohn M, SAGES-MIRA Robotic Surgery Consensus Group. A consensus document on robotic surgery. Surg Endosc. 2008 Feb;22(2):313–325; discussion 311–312. pmid:18163170
  8. 8. Broholm M, Pommergaard H-C, Gögenür I. Possible benefits of robot-assisted rectal cancer surgery regarding urological and sexual dysfunction: a systematic review and meta-analysis. Colorectal Dis. 2015 May 1;17(5):375–81. pmid:25515638
  9. 9. Moore LJ, Wilson MR, Waine E, Masters RSW, McGrath JS, Vine SJ. Robotic technology results in faster and more robust surgical skill acquisition than traditional laparoscopy. J Robot Surg. 2015 Mar 1;9(1):67–73. pmid:26530974
  10. 10. Wolpert DM, Diedrichsen J, Flanagan JR. Principles of sensorimotor learning. Nat Rev Neurosci. 2011 Oct 27;12(12):739–51. pmid:22033537
  11. 11. Wulf G, Shea CH. Principles derived from the study of simple skills do not generalize to complex skill learning. Psychon Bull Rev. 2002 Jun 1;9(2):185–211. pmid:12120783
  12. 12. Sheffield FD. Theoretical considerations in the learning of complex sequential tasks from demonstration and practice. AA Lumsdaine Ed Stud Responses Program Instr Wash DC Natl Acad Sci-Natl Res Counc. 1961;
  13. 13. Bandura A. Principles of behavior modification. Vol. ix. Oxford, England: Holt, Rinehart, & Winston; 1969. 677 p.
  14. 14. Hardwick RM, Caspers S, Eickhoff SB, Swinnen SP. Neural Correlates of Motor Imagery, Action Observation, and Movement Execution: A Comparison Across Quantitative Meta-Analyses. bioRxiv. 2017 Oct 10;198432.
  15. 15. Jeannerod M. Neural Simulation of Action: A Unifying Mechanism for Motor Cognition. NeuroImage. 2001 Jul 1;14(1):S103–9.
  16. 16. Custers EJFM, Regehr G, McCulloch W, Peniston C, Reznick R. The Effects of Modeling on Learning a Simple Surgical Procedure: See One, Do One or See Many, Do One? Adv Health Sci Educ. 1999 May 1;4(2):123–43.
  17. 17. Snyder CW, Vandromme MJ, Tyra SL, Porterfield JR, Clements RH, Hawn MT. Effects of Virtual Reality Simulator Training Method and Observational Learning on Surgical Performance. World J Surg. 2011 Feb 1;35(2):245–52. pmid:21086125
  18. 18. Mattar AAG, Gribble PL. Motor Learning by Observing. Neuron. 2005 Apr 7;46(1):153–60. pmid:15820701
  19. 19. Pollock BJ, Lee TD. Effects of the Model’s Skill Level on Observational Motor Learning. Res Q Exerc Sport. 1992 Mar 1;63(1):25–9. pmid:1574658
  20. 20. LeBel M-E, Haverstock J, Cristancho S, van Eimeren L, Buckingham G. Observational Learning During Simulation-Based Training in Arthroscopy: Is It Useful to Novices? J Surg Educ [Internet]. 2017 Jun 23 [cited 2017 Jul 17]; Available from: http://www.sciencedirect.com/science/article/pii/S1931720417301174
  21. 21. Rohbanfard H, Proteau L. Learning through observation: a combination of expert and novice models favors learning. Exp Brain Res. 2011 Dec 1;215(3–4):183–97. pmid:21986667
  22. 22. Andrieux M, Proteau L. Observation learning of a motor task: who and when? Exp Brain Res. 2013 Aug 1;229(1):125–37. pmid:23748693
  23. 23. Andrieux M, Proteau L. Mixed observation favors motor learning through better estimation of the model’s performance. Exp Brain Res. 2014 Oct 1;232(10):3121–32. pmid:24897947
  24. 24. Brown LE, Wilson ET, Obhi SS, Gribble PL. Effect of Trial Order and Error Magnitude on Motor Learning by Observing. J Neurophysiol. 2010 Sep 1;104(3):1409–16. pmid:20631214
  25. 25. Faul F, Erdfelder E, Lang A-G, Buchner A. G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav Res Methods. 2007 May 1;39(2):175–91. pmid:17695343
  26. 26. Jarc AM, Nisky I. Robot-assisted surgery: an emerging platform for human neuroscience research. Front Hum Neurosci [Internet]. 2015 Jun 4 [cited 2017 Jul 17];9. Available from: http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4455232/
  27. 27. Smith R, Patel V, Satava R. Fundamentals of robotic surgery: a course of basic robotic surgery skills based upon a 14-society consensus template of outcomes measures and curriculum development. Int J Med Robot. 2014 Sep 1;10(3):379–84. pmid:24277315
  28. 28. Ritter EM, Scott DJ. Design of a proficiency-based skills training curriculum for the fundamentals of laparoscopic surgery. Surg Innov. 2007 Jun;14(2):107–12. pmid:17558016
  29. 29. Vine SJ, Chaytor RJ, McGrath JS, Masters RSW, Wilson MR. Gaze training improves the retention and transfer of laparoscopic technical skills in novices. Surg Endosc. 2013 Sep 1;27(9):3205–13. pmid:23479253
  30. 30. Tytherleigh MG, Bhatti TS, Watkins RM, Wilkins DC. The assessment of surgical skills and a simple knot-tying exercise. Ann R Coll Surg Engl. 2001 Jan;83(1):69–73. pmid:11212457
  31. 31. Causer J, Harvey A, Snelgrove R, Arsenault G, Vickers JN. Quiet eye training improves surgical knot tying more than traditional technical training: a randomized controlled study. Am J Surg. 2014 Aug;208(2):171–7. pmid:24881015
  32. 32. Moorthy K, Munz Y, Sarker SK, Darzi A. Objective assessment of technical skills in surgery. BMJ. 2003 Nov 1;327(7422):1032–7. pmid:14593041
  33. 33. Mason JD, Ansell J, Warren N, Torkington J. Is motion analysis a valid tool for assessing laparoscopic skill? Surg Endosc. 2013 May 1;27(5):1468–77. pmid:23233011
  34. 34. Ghasemloonia A, Maddahi Y, Zareinia K, Lama S, Dort JC, Sutherland GR. Surgical Skill Assessment Using Motion Quality and Smoothness. J Surg Educ. 2017 Mar 1;74(2):295–305. pmid:27789192
  35. 35. Trejos AL, Patel RV, Naish MD, Malthaner R, Schlachta C. The application of force sensing to skills assessment in Minimally Invasive Surgery. In: 2013 IEEE International Conference on Robotics and Automation. 2013. p. 4370–5.
  36. 36. Hwang H, Lim J, Kinnaird C, Nagy AG, Panton ONM, Hodgson AJ, et al. Correlating motor performance with surgical error in laparoscopic cholecystectomy. Surg Endosc Interv Tech. 2006 Apr 1;20(4):651–5.
  37. 37. Cordier P, France MM, Pailhous J, Bolon P. Entropy as a global variable of the learning process. Hum Mov Sci. 1994 Dec 1;13(6):745–63.
  38. 38. Watson RA. Computer-Aided Feedback of Surgical Knot Tying Using Optical Tracking. J Surg Educ. 2012 May 1;69(3):306–10. pmid:22483129
  39. 39. Mayagoitia RE, Nene AV, Veltink PH. Accelerometer and rate gyroscope measurement of kinematics: an inexpensive alternative to optical motion analysis systems. J Biomech. 2002 Apr 1;35(4):537–42. pmid:11934425
  40. 40. van Hees VT, Gorzelniak L, León ECD, Eder M, Pias M, Taherian S, et al. Separating Movement and Gravity Components in an Acceleration Signal and Implications for the Assessment of Human Daily Physical Activity. PLOS ONE. 2013 Apr 23;8(4):e61691. pmid:23626718
  41. 41. Richman JS, Moorman JR. Physiological time-series analysis using approximate entropy and sample entropy. Am J Physiol—Heart Circ Physiol. 2000 Jun 1;278(6):H2039–49. pmid:10843903
  42. 42. Janelle CM, Champenoy JD, Coombes SA, Mousseau MB. Mechanisms of attentional cueing during observational learning to facilitate motor skill acquisition. J Sports Sci. 2003 Oct 1;21(10):825–38. pmid:14620026
  43. 43. Corbetta M, Akbudak E, Conturo TE, Snyder AZ, Ollinger JM, Drury HA, et al. A Common Network of Functional Areas for Attention and Eye Movements. Neuron. 1998 Oct 1;21(4):761–73. pmid:9808463
  44. 44. Wilson M, McGrath J, Vine S, Brewer J, Defriend D, Masters R. Psychomotor control in a virtual laparoscopic surgery training environment: gaze control parameters differentiate novices from experts. Surg Endosc. 2010 Oct 1;24(10):2458–64. pmid:20333405
  45. 45. Duchowski AT. Eye Tracking Methodology: Theory and Practice. Springer; 2017. 387 p.
  46. 46. Love J, Selker R, Marsman M, Jamil T, Dropmann D, Verhagen AJ, Wagenmakers EJ. JASP (Version 0.7)[computer software]. Amsterdam, the netherlands: Jasp project. 2015.
  47. 47. Birkmeyer JD, Finks JF, O’Reilly A, Oerline M, Carlin AM, Nunn AR, et al. Surgical Skill and Complication Rates after Bariatric Surgery. N Engl J Med. 2013 Oct 10;369(15):1434–42. pmid:24106936
  48. 48. Sutton DN, Wayman J, Griffin SM. Learning curve for oesophageal cancer surgery. Br J Surg. 1998 Oct 1;85(10):1399–402. pmid:9782024
  49. 49. Wulf G, Shea C, Lewthwaite R. Motor skill learning and performance: a review of influential factors. Med Educ. 2010 Jan 1;44(1):75–84. pmid:20078758
  50. 50. Calvo-Merino B, Grèzes J, Glaser DE, Passingham RE, Haggard P. Seeing or doing? Influence of visual and motor familiarity in action observation. Curr Biol CB. 2006 Oct 10;16(19):1905–10. pmid:17027486
  51. 51. LeBlanc J, Kirschner A, Buckingham G, Harris D, Haverstock J, LeBel M-E. Eye-gaze strategies affect performance on arthroscopic simulators during observational learning. Prep.
  52. 52. Robles-De-La-Torre G. The importance of the sense of touch in virtual and real environments. IEEE Multimed. 2006 Jul;13(3):24–30.
  53. 53. Wagner CR, Howe RD. Force Feedback Benefit Depends on Experience in Multiple Degree of Freedom Robotic Surgery Task. IEEE Trans Robot. 2007 Dec;23(6):1235–40.
  54. 54. van der Meijden OAJ, Schijven MP. The value of haptic feedback in conventional and robot-assisted minimal invasive surgery and virtual reality training: a current review. Surg Endosc. 2009 Jun 1;23(6):1180–90. pmid:19118414
  55. 55. Morris D, Tan H, Barbagli F, Chang T, Salisbury K. Haptic Feedback Enhances Force Skill Learning. In: Second Joint EuroHaptics Conference and Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems (WHC’07). 2007. p. 21–6.
  56. 56. Williams RL, Srivastava M, Conaster R, Howell JN. Implementation and Evaluation of a Haptic Playback System. 2004 May 3: https://digital.lib.washington.edu:443/researchworks/handle/1773/34888
  57. 57. Wilson MR, Vine SJ, Bright E, Masters RSW, Defriend D, McGrath JS. Gaze training enhances laparoscopic technical skill acquisition and multi-tasking performance: a randomized, controlled study. Surg Endosc. 2011 Dec 1;25(12):3731–9. pmid:21671125
  58. 58. Vine SJ, Masters RSW, McGrath JS, Bright E, Wilson MR. Cheating experience: Guiding novices to adopt the gaze strategies of experts expedites the learning of technical laparoscopic skills. Surgery. 2012 Jul 1;152(1):32–40. pmid:22464048
  59. 59. Sickle KRV, Iii DAM, Gallagher AG, Smith CD. Construct validation of the ProMIS simulator using a novel laparoscopic suturing task. Surg Endosc Interv Tech. 2005 Sep 1;19(9):1227–31.
  60. 60. Chmarra MK, Klein S, de Winter JCF, Jansen F-W, Dankelman J. Objective classification of residents based on their psychomotor laparoscopic skills. Surg Endosc. 2010 May 1;24(5):1031–9. pmid:19915915
  61. 61. Rodrigues SP, Horeman T, Dankelman J, van den Dobbelsteen JJ, Jansen F-W. Suturing intraabdominal organs: when do we cause tissue damage? Surg Endosc. 2012 Apr 1;26(4):1005–9. pmid:22028014