Figures
Abstract
Purpose
To evaluate the efficacy of the virtual reality training simulator Eyesi to prepare surgeons for performing pars plana vitrectomies and its potential to predict the surgeons’ performance.
Methods
In a preparation phase, four participating vitreoretinal surgeons performed repeated simulator training with predefined tasks. If a surgeon was assigned to perform a vitrectomy for the management of complex retinal detachment after a surgical break of at least 60 hours it was randomly decided whether a warmup training on the simulator was required (n = 9) or not (n = 12). Performance at the simulator was measured using the built-in scoring metrics. The surgical performance was determined by two blinded observers who analyzed the video-recorded interventions. One of them repeated the analysis to check for intra-observer consistency. The surgical performance of the interventions with and without simulator training was compared. In addition, for the surgeries with simulator training, the simulator performance was compared to the performance in the operating room.
Results
Comparing each surgeon’s performance with and without warmup trainingshowed a significant effect of warmup training onto the final outcome in the operating room. For the surgeries that were preceeded by the warmup procedure, the performance at the simulator was compared with the operating room performance. We found that there is a significant relation. The governing factor of low scores in the simulator were iatrogenic retinal holes, bleedings and lens damage. Surgeons who caused minor damage in the simulation also performed well in the operating room.
Conclusions
Despite the large variation of conditions, the effect of a warmup training as well as a relation between the performance at the simulator and in the operating room was found with statistical significance. Simulator training is able to serve as a warmup to increase the average performance.
Citation: Deuchler S, Wagner C, Singh P, Müller M, Al-Dwairi R, Benjilali R, et al. (2016) Clinical Efficacy of Simulated Vitreoretinal Surgery to Prepare Surgeons for the Upcoming Intervention in the Operating Room. PLoS ONE 11(3): e0150690. https://doi.org/10.1371/journal.pone.0150690
Editor: Gianni Virgili, University of Florence, ITALY
Received: October 15, 2015; Accepted: February 18, 2016; Published: March 10, 2016
Copyright: © 2016 Deuchler et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the paper and its Supporting Information files.
Funding: This work was supported by Adolf Messer Stiftung, http://www.adolf-messer-stiftung.de/, to FK; Gemeinnützige Hertie-Stiftung, http://www.ghst.de/, to FK; and Hessisches Ministerium für Wissenschaft und Kunst, http://www.lehrpreis-hessen.de/, to FK. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. No grant numbers are available. VRmagic provided support in the form of salaries for authors [Markus Schill, Clemens Wagner, and Benjamin Schoene], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.
Competing interests: The authors have the following interests: Markus Schill and Clemens Wagner are shareholders of and employed by VRmagic, a commercial company. Benjamin Schoene is employed by VRmagic. The Eyesi simulator that is manufactured by VRmagic plays an important role in their study and is mentioned accordingly in the paper. There are no additional patents, products in development or marketed products to declare. This does not alter the authors' adherence to all the PLOS ONE policies on sharing data and materials, as detailed online in the guide for authors.
Introduction
Virtual reality education and training are established tools for building up expertise in many specialities worldwide, such as aviation or medicine.
Khalifa et al [1] found that ophthalmology training programs were struggling to find viable methods of assessing and documenting the surgical skills of trainees, and that the role of virtual reality education and training in future curricula was still uncertain in 2006.
Vitreoretinal surgery training with the ophthalmic surgical simulator Eyesi was introduced in June 2003 when two Eyesi simulators were used for the first time in a teaching lab, during the 5th International Vitreoretinal Symposium (VRS) in Frankfurt. The success of this lab led to the introduction of the dry lab concept: simulation in combination with practice eyes, table microscopes and real surgical machines.
Simulator-based training in the anterior as well as the posterior segment of the eye found its way into training and education curricula rapidly [2].
In particular the anterior segment has been well evaluated in numerous studies [3–9] focusing both on construct validity (in the anterior [10–13] and posterior segment [14]) as well as on the transfer of skills from virtual reality to wet lab training and to the operating room. Feudner et al [15] investigated how capsulorhexis training on Eyesi improved wetlab capsulorhexis performance of surgical novices, while McCannel et al [16] took a step forward by comparing the capsulorhexis performance of residents in the operating room before and after the introduction of simulator-based training and found a significant reduction in complications.
The assessment of surgical performance under special conditions such as sleep deprivation [17–18] or beta-blockade [19] has also been discussed. Gill et al [20] investigated the ability of simulation to predict the actual surgical proficiency of residents and attending physicians for anterior segment surgery.
For the posterior segment, there are no known skills-transfer trials from virtual reality environments to the operation room.Our study evaluates in a prospective setup whether preparation at the simulator affects the surgical performance and whether it allows conclusions on the expected actual pars plana vitrectomy (PPV) performance in retinal reattachment surgery.
Material and Methods
Thanks to financial support from the German government and two foundations, the Frankfurt University Eye Clinic was able to start an extensive study focusing on the establishment of standard operation procedures (SOPs), success criteria in the field of retinal reattachment surgery and implementation of simulator-based practice.
The study was approved by the Ethical review committee University Frankfurt/M (IRB decision no. E 190/11, transaction no. 403/11), part of the study plan was the simulator-based warmup and performance test in advance of surgical procedures that is discussed in this article. The Ethical review committee also approved the written participant consent, which was provided by all participants.
This study was conducted in accordance with the Tenets of the Declaration of Helsinki. Patients’ records were pseudonymized and de-identified prior to statistical analysis.
In the preparation stage of the test, four participating surgeons performed several identical simulator courses within a two-week period. In the course of the paper, these surgeons are referred to as “user NN” with NN representing years of practice. The preparation stage was designed to make the participants familiar with the specific properties of the simulated environment.
During the study, the simulator software version 2.7.6 (release date: 24th of October 2011) was used. The simulator was equipped with a widefield viewing system and a vitreoretinal forceps handpiece.
In the surgery stage, we looked at PPVs for management of complex retinal detachment cases [21] performed by surgeons after a break from work of at least 60 hours.
After assigning a case to the best-suited surgeon for the surgery at hand, a starting point was determined with a simple 1:1 randomization scheme: either the surgeon went to the operating room immediately or he performed the simulator training courseware (duration: approximately 20 minutes) prior to going to the operating room. All surgical interventions were video-recorded and subsequently analyzed by two “blinded” observers (surgical vitreoretinal experience of more than 20 respectively more than 5 years). The more experienced observer graded the videos twice to allow a discussion of intraobserver consistency.
Simulator courseware
The simulator training comprised an abstract bimanual task, two peelings (non-dominant and dominant hand) and a simulated retinal detachment surgery.
The following tasks were used:
- Bimanual Scissors Training Level 2 –Trains precise, simultaneous, asynchronuous instrument movements (aspirator, scissors) close to the retina.
- ILM Peeling Level 5 –Advanced task with a slightly stained, brittle membrane, i.e. frequent regrasping is necessary.
- ILM Peeling Level 5 –In this second ILM task with similar tissue properties, the peeling must be done with the non-dominant hand.
- Retinal Detachment Level 1 –Rhegmatogenous detachment with a horseshoe tear in the temporal periphery of a right eye, i.e. for some steps, a left-handed, nasal access is required.
All tasks were selected to match the nature of the real surgery–demanding vitreoretinal, bimanual PVR surgery. We specifically added an abstract task for bimanual skills training because, to our experience, the deliberate practice [22] of an isolated skill enables even experienced surgeons to raise their established performance level in this particular skill.
Simulator scoring metrics
The simulator score is composed of numerous parameters. Each parameter has a specific point range that indicates its maximal influence. The total score is calculated by adding the score of all target parameters and subtracting the faults. The targets add up to 100 points whereas the sum of all faults can be higher. If so, the resulting score is cut off at zero points. Adding the fault scores instead of taking the average has been implemented in the simulator in analogy to the real situation: a surgical performance that is unsatisfying under multiple aspects adds up to an even worse situation. This is important in a training setup where predefined performance levels must be reached. For details see S1 Table.
In the realm of retinal detachment surgery, we identified the following primary targets: (1) amount of peeled ILM in macular area, (2) amount of reattached retina, and (3) amount of tractive tissue removed from the detached retina.The last parameter comprises the removal of enrolled edges of retinal tissue, of vitreous residuals, and of pigment and blood cells around retinal tears.
For the simulator these primary targets were scored according to S2 and S3 Tables.
Scoring metrics of the video analysis
The scoring metrics for the video analysis were derived from the set of parameters measured by the simulator. Some parameters in the simulation (e.g. “fovea/optic disc/vessel hit by laser”) were combined into one parameter for the video analysis (e.g. “retinal injury due to laser energy”) or omitted due to minor surgical relevancy or inaccessibility (e.g. peeled ILM outside macular area, vitrector suction on retina).
GRASIS (Global Rating Assessment of Skills in Intraocular Surgery) is a validated scoring system for ophtalmic surgical competency [23]. As Fig 1 illustrates, our scoring metrics parameters can be categorized according to GRASIS. Note that there are differences: some GRASIS categories are not used at all (e.g. “Preoperative planning“, “Surgical professionalism“). They apply to a surgical novice rather than to a practicing surgeon. The three primary surgical targets were added to the GRASIS “Overall Performance”category, since GRASIS has no “targets”category.
The number of parameters assigned to each category varies considerably (from 7 parameters in the “Treatment of ocular structures”category to 1 parameter in “Microscope use”and “Use of non-dominant hand“). The total score is calculated by averaging over all parameters. Therefore the GRASIS categories do not contribute with equal weight. While this reflects the importance of a category in our setup quite well, it also represents a distortion of the GRASIS metrics. In order to support the validity of our approach, we performed an inter- and intragrader analysis of the video rating.
Data pool and statistical method
The simple randomization resulted in the following distribution (total / with warmup / without warmup): user 02 (2 / 2 / 0), user 03 (6 / 2 / 4), user 07 (6 / 3 / 3), user 25 (7 / 2 / 5), i.e. 9 out of 21 cases were performed with simulator-based warmup whereas 12 cases were not preceeded by simulator trainingFor the examination of the warmup effect cases with warmup were compared to cases without warmup for each surgeon. For the 9 cases with warmup a detailed comparison of simulator and surgical performance was carried out.
Due to the special setup (N surgeons perform M interventions), we used a mixed-model approach for repeated measurements [24]. For illustrating the relation between performance on the simulator and in the operating room, we performed a linear regression after Pearson.For all statistical evaluations a p<0.05 was considered statistically significant. All the analyses were performed using BiAS V10.12 [25] for Windows and the R package V3.1–120 [26].
Results
Inter- and intragrader consistency
The grading of the recorded surgeries was performed by two independent vitreoretinal experts.
The Shrout-Fleiss intraclass correlation with a two-way mixed, single measure model resulted in a correlation coefficient of ICC (3,1) = 0.97 and a 95% confidence interval of 0.872…0.994.
In addition, an intra-individual regrading was performed. Here, the Bland-Altman intraclass correlation resulted in a correlation coefficient of ICC = 0.99 with a 95% confidence interval of 0.981…0.998.
Effect of warmup training
The warmup training increases the final outcome in the operating room by 0.5 to 1 points in the video analysis. The statistical analysis with a mixed-model approach for repeated measurements showed a significant effect (p = 0.0302).
Warmed up or not, the standard deviation (SD) of the three less experienced surgeons (2/3/7 years of experience) is rather high with SD = 0.76 (n = 14) whereas the scores of the expert surgeon (25 years of experience) vary considerably less with SD = 0.21 (n = 7). An f-test showed (p = 0.004) that this difference between the less experienced surgeons and the expert is significant.
In the following it will be investigated if this surgical variability can be predicted by analyzing the simulator scores of the warmup training.
Simulator-based prediction of operating room performance
Fig 2 compares the scores achieved in the simulation and in the operating room. In general, the graph shows that performance varies from day to day for all surgeons, both on the simulator and in the operating room, and that these variations correlate.
Individual surgeons are marked by different symbols. The experience of each surgeon in years is shown as part of his user name. The linear fit was determined by using a Pearson regression.
In order to account for within-surgeon correlation we used a mixed model approach. The repeated measure test showed a significance of p = 0.000342 and a good model fit of R2 = 0.986.
For illustration, a linear regression after Pearson was performed and brought up a consistent result with p = 0.00603.
Table 1 shows the most prominent faults during simulated surgery; Listed are all penalties with more than 10 points off. Both the total amount of deduction (sum of all penalties) and the number of occurrences of a fault are shown in the table. The determinants of low scores on the simulator were lens damage, retinal bleeding and iatrogenic retinal holes.
In Fig 3, the relation between injuries in the simulator and surgical performance is investigated. The figure compares for each user the total penalty for injuries with the performance in the real surgery. It is worth mentioning that a similar relation exists between the overall simulator score and the surgical performance as shown in Fig 2.
Fig 4 shows the operating room score in relation to the speed of the instrument movement in the simulator. The average instrument speed was determined by dividing the Eyesi odometry value (travelled distance of instrument tip) by the time taken. Apparently slower movements in Eyesi, i.e. a more precise, target-oriented approach, result in a better surgical performance.
The most important reasons for score deductions (deductions with more than 10 points off) in the simulator are listed in Table 2. As in Table 1, it is obvious that most deductions are based on tissue injuries–with one exception: the expert “user 25” did not only get less deductions, but, notably, they were limited to areas unrelated to tissue damage such as time and completion of targets.
Discussion
The aim of this study was to investigate whether and to what extent a warmup on the surgical simulator Eyesi is efficient in preparing the surgeon for an upcoming intervention. In addition it was investigated whether the performance parameters of the warmup training have predictive power regarding the outcome of the subsequent surgery.
VR to OR studies are still the exception because trials are difficult to carry out from an ethical standpoint. In our study design we chose the best surgeon available for a given maneuver. After making this choice, the surgeon was randomly assigned to either go straight to the operating room or to warm up at the simulator prior to the surgery.
Of all studies on Eyesi evaluating the role of training in either the anterior or the posterior eye segment, to our knowledge, there is none that investigates the warmup effect of simulator training.
Regarding the second part of the study, only Gill et al [20] used a similar setup and investigated the relationship between simulator performance and surgical performance. However they concentrated on beginners in cataract surgery whereas we focused on the advanced and expert vitreoretinal surgeon.
Despite the limited data set and the large variation of conditions (four surgeons, treating different pathologies, 21 samples), a statistically significant effect of warmup training on the surgical performance was found as well as a significant relation between performance on the simulator and in the operating room.
The average performance level of all surgeons was increased by the warmup training. Surprisingly, we found that even the most experienced surgeons benefit from warmup training.
However, the short training does not have an impact onto the variability of the performance level. We presume that a considerably longer training would be necessary to bring less experienced surgeons to a level of higher reliability that persists longer than the warmup effect that we observed: a 20 minutes warm up module cannot compensate for a lack of systematic training.
The performance values that were taken as the basis for the statistical evaluation were obtained by measuring numerous different parameters and either calculating a weighted sum (for the simulator score), or taking their average (for the video analysis of real surgery).
Parameter analysis
The analysis of individual parameters or groups of parameters allows further insight into the training behavior and the relationship between simulation and real surgery. The ability to map an individual training detail to the corresponding surgical pattern is important for the design of individual simulator courses for training and warmup. In the following, we will discuss the relevance of our primary target parameters (ILM peeling, reattached retina, removed tractive tissue), of iatrogenic damages and, as an example of a parameter that can be measured much easier in simulation than in a video analysis, of the speed of instrument movements.
Injuries.
Eyesi reacts extremely sensitively to undesirable injuries (Fig 3). This is intentional, since avoiding iatrogenic damage of all kinds was considered to positively affect the “over all” surgical performance. As we can see in this study, this is of utmost importance for those in the earlier stages of their surgical education, as they have a higher rate of iatrogenic failures compared to more experienced surgeons.
ILM peeling.
ILM peeling seems to be a fairly standardized procedure. All surgeons peel around 44–50% ILM between the vessel arcades. However, there are significant differences in regard to the peeling quality and the percentage of injuries caused during the peeling. Some curiosities are worth to be mentioned: an increasing level of experience in vitreoretinal surgery might affect the surgeon’s decision to choose training elements they perceive as most needed (this was confirmed by the expert surgeon, who explained why he had not finished the ILM peel in Eyesi: “I shortened the Eyesi training module because I know that once the beginning of a peel works well the rest will flow smoothly anyway”). The expert tends to overrule the simulator‘s operation procedure by making use of individually composed training steps. While this might call into question the measurability of expert performance, at the same time it constitutes a training value in itself, as the artificial environment makes it possible to isolate relevant training steps, thus increasing learning curves and the efficiency of the training (deliberate practice).
Reattached retina.
Both in the simulator and in the real surgery, intraoperative reattachment of the retina could be achieved in 100% of the performed tasks. Therefore this parameter is not suitable for a discriminative analysis of warmup effects.
Removed tractive tissue.
For a standardized treatment of rhegmatogenous retinal detachment with PVR activity, the removal of tissue around retinal tears has been identified as one important surgical step in our hospital. Therefore it was also integrated into the simulated surgery and is the third target criterium in our setup. Working on a tear in the periphery with the appropriate configuration of the surgical machine is a delicate part of the surgery. We see a relationship between the performance at the simulator and in the operating room–with one exception: user 05, who performed well in real surgery, but not on the same level in the simulator.
Instrument movements.
By comparing the score in the operating room and the instrument speed while managing a retinal detachment (Fig 4), we noticed that more expertise correlates with slower movements and less “action” in the vitreous cavity. This seems to be a bit of a surprise first–but is easily explained by the fact that during the vitreoretinal education, every surgeon is told over and over that the slower and more precise she or he moves the instruments, the more efficiently and ultimately faster the procedure will be finished. Eyesi records all movements, classifies them as necessary or unnecessary and scores them with the “odometry” measurement.
Fig 4 shows two exceptional situations (at 1.4/3 and 1.58/3.4): two “non-expert-surgeons” faced two different situations which Eyesi had not prepared them for: the first one was a perfluorocarbon-to-air-to-silicon-oil exchange which was not available in the simulation at that time. The second situation was an inappropriate set of instruments (highly myopic eye, instrument too short) which did not allow an adequate peeling of the ILM as planned, resulting in repeated time-consuming and ultimately unsuccessful approaches (readjusting the microscope, changing dominant and non-dominant hand, etc.).
While the perfluorocarbon-to-air-to-silicon-oil exchange has been added meanwhile, the simulator is not yet able to provide situations with myopic eyes where the right strategies to shorten the actual length of the eye can be trained.
Strength and limitations of this paper
This paper presents true VR-to-OR data about training of several vitreoretinal surgeons with different surgical expertise. This could not be found in literature up to now. Due to ethics considerations (“choose the most suitable surgeon first”), we did not know beforehand which surgeon had to perform how many interventions. In combination with the “weekend situation”, this resulted in a limited number of cases (n = 21). The simple randomization scheme that we used led to one surgeon performing real surgery after warmup only.
Despite these limitations, our mixed model approach showed significant results. For future studies with a similar setup, we suggest to replace the simple randomization scheme by a block randomization with fixed A/B sequences (example: 5 surgeons with 3 A/B sequences, resulting in 30 cases). Although this would prolongate the data acquisition, it would guarantee an equal distribution that ultimately results in a higher statistical power. Note that this design resembles an N-of-1 trial series, especially with regard to basic design considerations, such as randomization and carryover effects [27].
Since this kind of study has not been published before, it is difficult to find a rating system in literature we can refer to: There is limited possibility of comparison between the global rating assessment GRASIS which focusses on surgeons in their early stage of education and the assessment adjusted for this study which looks atthe performance of surgeons in their early and later stages of education including experts. It seems to be an advantage that in this study different skill levels were covered and less experienced surgeons as well as advanced surgeons and experts were included into the data pool.Further questions have to be addressed and answered in consecutive studies. We chose a break of 60 hours to reflect a weekend situation. How do longer or shorter breaks affect the observed drop in performance? Are surgeons in average more (or less) confident after they used the simulator? Does the level of confidence correlate with the measurable skill performance? Could the observed relation between simulator and surgical performance be used to define a threshold for entering the operating room? Should a specific courseware be designed for experts which allow them to choose between different roads and adjust the training to the individual needs under variable conditions, such as different time intervals between surgical interventions, anticipated complications, different stress levels etc.
And finally: what clinically relevant simulations must be added to increase the efficiency of a simulator-based warmup and performance measurement?
Supporting Information
S1 Table. Scoring parameters that all simulated tasks have in common.
Since the targets are task-specific, the table only shows faults and the neutral parameter “Odometer” that measures the travelled distance of the instruments tips. This parameter is needed for an analysis of the surgeons’ instrument speed.
https://doi.org/10.1371/journal.pone.0150690.s001
(PDF)
S2 Table. Specific scoring parameters for simulated ILM peeling.
The primary target is marked with a star (*).
https://doi.org/10.1371/journal.pone.0150690.s002
(PDF)
S3 Table. Specific scoring parameters for simulated retinal detachment surgery.
Target parameters are indicated by a positive, faults by a negative point range. The primary targets are marked with a star (*).
https://doi.org/10.1371/journal.pone.0150690.s003
(PDF)
Author Contributions
Conceived and designed the experiments: SD CW MS FK MK. Performed the experiments: PS MM RA FK. Analyzed the data: SD CW HA DB RB FK. Contributed reagents/materials/analysis tools: CW MS TK. Wrote the paper: SD CW BS FK.
References
- 1. Khalifa Y, Bogorad D, Gibson V, Peifer J, Nussbaum J. Virtual Reality in Ophthalmology Training. Survey of Ophthalmology 2006;51: 259–273. pmid:16644366
- 2. Saleh GM, Lamparter J, Sullivan PM, O'Sullivan F, Hussain B, Athanasiadis I, et al. The international forum of ophthalmic simulation: developing a virtual reality training curriculum for ophthalmology. Br J Ophthalmol 2013;97: 789–92. pmid:23532612
- 3. Baxter JM, Lee R, Sharp JA, Foss AJ, Intensive Cataract Training Study Group. Intensive cataract training: a novel approach. Eye 2013;27: 742–6. pmid:23598673
- 4. Belyea DA, Brown SE, Rajjoub LZ. Influence of surgery simulator training on ophthalmology resident phacoemulsification performance. J Cataract Refract Surg 2011;37: 1756–1761. pmid:21840683
- 5. Bergqvist J, Person A, Vestergaard A, Grauslund J. Establishment of a validated training program on the Eyesi cataract simulator. A prospective randomized study. Acta Ophthalmol 2014;92: 629–34. pmid:24612448
- 6. Jonas JB, Rabethge S, Bender HJ. Computer-assisted training system for pars plana vitrectomy. Acta Ophthalmol Scand 2003;81: 600–4. pmid:14641261
- 7. Lee TD, Adatia FA, Lam WC. Virtual reality ophthalmic surgical simulation as a feasible training and assessment tool: results of a multicentre study. Can J Ophthalmol 2011;46: 56–60. pmid:21283159
- 8. Rogers GM, Oetting TA, Lee AG, Grignon C, Greenlee E, Johnson AT, et al. Impact of a structured surgical curriculum on ophthalmic resident cataract surgery complication rates. J Cataract Refract Surg 2009;35: 1956–60. pmid:19878829
- 9. Spiteri A, Aggarwal R, Kersey T, Benjamin L, Darzi A, Bloom P. Phacoemulsification skills training and assessment. Br J Ophthalmol 2010;94: 536–541. pmid:19628497
- 10. Selvander M, Åsman P. Virtual reality cataract surgery training: learning curves and concurrent validity. Acta Ophthalmol 2012;90: 412–74. pmid:21054818
- 11. Selvander M, Åsman P. Cataract surgeons outperform medical students in Eyesi virtual reality cataract surgery: evidence for construct validity. Acta Ophthalmol 2013;91: 469–74. pmid:22676143
- 12. Privett B, Greenlee E, Rogers G, Oetting TA. Construct validity of a surgical simulator as a valid model for capsulorhexis training. Cataract & Refractive Surg 2010;36: 1835–1838.
- 13. Mahr MA, Hodge DO. Construct validity of anterior segment anti-tremor and forceps surgical simulator training modules: Attending versus resident surgeon performance. J Cataract Refract Surg 2008;34: 980–985. pmid:18499005
- 14. Rossi J, Verma D, Fuji G, Lakhanpal R, Wu S, Humayun M, et al. Virtual vitreoretinal surgical simulator as a training tool. Retina 2004;24:231–236. pmid:15097883
- 15. Feudner E, Engel C, Neuhann I, Petermeier K, Bartz-Schmidt K, Szurman P. Virtual reality training improves wet-lab performance of capsulorhexis. Graefes Arch Clin Exp Ophthalmol. 2009;247: 955–63. pmid:19172289
- 16. McCannel C, Reed DC, Goldman DR. Ophthalmic Surgery Simulator Training Improves Resident Performance of Capsulorhexis in the Operating Room. Ophthalmology 2013;120: 2456–61. pmid:23796766
- 17. Erie EA, Mahr MA, Hodge DO, Erie JC. Effect of sleep deprivation of simulated anterior segment surgical skill. Can J Ophthalmol. 2011;46(1): 61–5. pmid:21283160
- 18. Waqar S, Park J, Kersey TL, Modi N, Ong C, Sleep TJ. Assessment of fatigue in intraocular surgery: analysis using a virtual reality simulator. Graefes Arch Clin Exp Ophthalmol 2011;249: 77–81. pmid:20890612
- 19. Pointdujour R, Ahmad H, Liu M, Smith E, Lazzaro D. Β-blockade affects simulator scores. Ophthalmology 2011;118: 1893–1893e. pmid:21889664
- 20. Gill HS, Sit M, Noble J, Liu E, Ahmed I, Lam WC. An assessment of the ability of surgical simulation to predict actual surgical proficiency in ophthalmology. Open Medicine 2009;3: 3.
- 21.
Deuchler SK, Krüger H, Koss M, Singh P, Koch F. Retina Re-Detachment after Silicone Oil Removal: Destiny or Beat the Enemy? ARVO, 2011 Annual Meeting, Fort Lauderdale, FL United States.
- 22. Ericsson KA. Necessity is the mother of invention: video recording firsthand perspectives of critical medical procedures to make simulated training more effective. Acad Med. 2014; 89: 17–20. pmid:24280862
- 23. Cremers SL, Lora AN, Ferrufino-Ponce ZK. Global Rating Assessment of Skills in Intraocular Surgery (GRASIS). Ophthalmology 2005;112: 1655–1660. pmid:16102834
- 24. Johnson PCD. Extension of Nakagawa & Schielzeth’s R²GLMM to random slopes models. Methods Ecol Evol. 2014;5: 944–946. pmid:25810896
- 25. Ackermann H. A program package for Biometrical Analysis of Samples. Comp Stat & Data Analysis 1991;11: 223–4.
- 26.
Stowell S. Using R for Statistics. 1st ed. New York: Apress; 2014.
- 27. Lilie EO, Patay B, Diamant J, Issel B, Topol EJ, Schork NJ. The N-of-1 clinical trial: the ultimate strategy for individualizing medicine? Per Med 2011; 8: 161–173. pmid:21695041