Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Physician Associate and General Practitioner Consultations: A Comparative Observational Video Study

  • Simon de Lusignan ,

    s.lusignan@surrey.ac.uk

    Affiliations Department of Clinical and Experimental Medicine University of Surrey, Guildford, GU2 7XH, United Kingdom, Division of Population Health Sciences and Education St. George’s University of London, London, SW17 0RE, United Kingdom

  • Andrew P. McGovern,

    Affiliation Department of Clinical and Experimental Medicine University of Surrey, Guildford, GU2 7XH, United Kingdom

  • Mohammad Aumran Tahir,

    Affiliations Department of Clinical and Experimental Medicine University of Surrey, Guildford, GU2 7XH, United Kingdom, AT Medics, St. Charles Hospital, Exmoor Street, London, W10 6DZ, United Kingdom

  • Simon Hassan,

    Affiliation Department of Clinical and Experimental Medicine University of Surrey, Guildford, GU2 7XH, United Kingdom

  • Simon Jones,

    Affiliations Department of Clinical and Experimental Medicine University of Surrey, Guildford, GU2 7XH, United Kingdom, Department of Population Health NYU School of Medicine, 227 East 30th Street, New York, New York, 10016, United States of America

  • Mary Halter,

    Affiliation Faculty of Health, Social Care & Education, Kingston University & St. George’s University of London, London, United Kingdom

  • Louise Joly,

    Affiliation Social Care Workforce Research Unit King’s College London, Strand, London, WC2R 2LS, United Kingdom

  • Vari M. Drennan

    Affiliation Faculty of Health, Social Care & Education, Kingston University & St. George’s University of London, London, United Kingdom

Abstract

Background

Physician associates, known internationally as physician assistants, are a mid-level practitioner, well established in the United States of America but new to the United Kingdom. A small number work in primary care under the supervision of general practitioners, where they most commonly see patients requesting same day appointments for new problems. As an adjunct to larger study, we investigated the quality of the patient consultation of physician associates in comparison to that of general practitioners.

Method

We conducted a comparative observational study using video recordings of consultations by volunteer physician associates and general practitioners with consenting patients in single surgery sessions. Recordings were assessed by experienced general practitioners, blinded to the type of the consulting practitioner, using the Leicester Assessment Package. Assessors were asked to comment on the safety of the recorded consultations and to attempt to identify the type of practitioner. Ratings were compared across practitioner type, alongside the number of presenting complaints discussed in each consultation and the number of these which were acute, minor, or regarding a chronic condition.

Results

We assessed 62 consultations (41 general practitioner and 21 physician associates) from five general practitioners and four physician associates. All consultations were assessed as safe; but general practitioners were rated higher than PAs in all elements of consultation. The general practitioners were more likely than physician associates to see people with multiple presenting complaints (p<0.0001) and with chronic disease related complaints (p = 0.008). Assessors correctly identified general practitioner consultations but not physician associates. The Leicester Assessment Package had limited inter-rater and intra-rater reliability.

Conclusions

The physician associate consultations were with a less complex patient group. They were judged as competent and safe, although general practitioner consultations, unsurprisingly, were rated as more competent. Physician associates offer a complementary addition to the medical workforce in general practice.

Introduction

Physician associates (PAs), known internationally as physician assistants, are a type of mid-level practitioner new to the United Kingdom (UK). The PA is “a new healthcare professional who, while not a doctor, works to the medical model, with the attitudes, skills and knowledge base to deliver holistic care and treatment within the general medical and/or general practice team under defined levels of supervision” by a physician [1]. PAs developed in the United States of America (USA) in the mid-1960s to tackle a national shortage of doctors and to increase access to healthcare [2,3]. In recent years, the profession has begun to spread globally including to the Netherlands, India, Canada and Australia [4].

The first UK-trained PAs graduated in 2009. The PA training programme in the UK is a two year full-time course for graduates, generally with life sciences degrees, and is delivered by an increasing number of higher education institutions, supported by the national health care workforce plans [5]. After graduation PAs are encouraged to register on the PA Managed Voluntary Register (PAMVR), held by the Faculty of Physician Associates at the Royal College of Physicians [6]. It has been suggested that there is a need to develop stronger governance frameworks for the profession [7] and the Faculty has been working for PAs to be included in the state regulatory processes for health professions. One consequence of their exclusion from the regulatory process is that PAs in the UK cannot currently prescribe medications.

Currently, the majority of the PAs’ workload in UK primary care is in providing same day and urgent consultation appointments [8]. This is similar to their role in primary care in the USA and elsewhere [9]. A systematic review [9] highlighted the lack of evidence as to the contribution PAs made to general practice (known as family physician services outside the UK). As an adjunct to a large mixed methods study [10], we investigated the quality of the patient consultation, by PAs compared with general practitioners (GPs) as judged by independent experienced GPs from video recordings.

Video analysis of the consultation is an established part of general practitioner training in the UK, providing insight into the content and complexity of the clinical consultation [11]. Advanced techniques have been used to more precisely observe the interaction between clinician, patient and computer [12] and this approach was used here.

Methods

This was a comparative, observational study of video recorded PA and GP primary care consultations.

The sample

PAs and GPs were recruited from the 12 practices participating in a larger study of the role of PAs (six practices with PAs and six without in the East, South West and South East of England). Details of the parent study and practice recruitment is given elsewhere [10]. PAs and GPs, in the study practices, were invited to participate in the video observation study and written informed consent was gained from each practitioner prior to recording. The volunteer GPs and PAs each identified one specific clinical session in which they saw same day and urgent patient appointments for the video observations to be made. These sessions were in addition to those reported from the parent study [13].

Adult patients attending for same-day appointments in these sessions were invited to participate in the study by the researchers [MH, LJ, WC, JY]. Patients were informed about the study and video recording prior to their consultation, as they arrived for their appointment, and written consent was obtained from all participating patients by the researchers. Patients were informed about their right to withdraw consent for video recording at any stage. Recording started as the patient entered the room and ceased when they left the room. On leaving the consultation, the researchers asked patients whether they continued to consent to the recording being used or if they wished it to be deleted. Practitioners also consented for inclusion of each recording at the end of the clinical session. There was no data collected from the electronic patient record or other sources outside the consultation.

The data

Consultations were recorded in 2012, using an adapted version of the Activity Log Files Aggregation (ALFA) toolkit which is a recording technique designed to facilitate precise observation of the consultation [14].

Recording was restricted to two cameras per consultation, one with a wide-angle view of the consultation and the other focused on the practitioner; this was to comply with the requirements of the NHS research ethics committee. Cameras were positioned so as to not record intimate examinations. The two video feeds were combined to produce a dual-channel video used for assessment. The resulting videos were stored on encrypted and password protected hard drives. Each consultation was stored in full and in an edited version, which removed any evidence which revealed the role of the consulting clinician. These included statements made by the clinician e.g. “I need to get this prescription signed by the doctor”, periods where PAs left the room to get prescriptions signed, and references made by patients e.g. call the clinician “doctor”.

Analysis

The analysis comprised of assessment of the quality of the consultation, whether experienced GPs could correctly identify GPs or PAs from the videos and comments on the safety of the consultations, and an analysis of patient case mix. Each of these components is described below.

Assessment of the quality of the consultation.

Consultation quality was assessed by a group of four GP assessors using the Leicester Assessment Package. The Leicester Assessment Package has been demonstrated to have validity and reliability across levels of clinical competence [1517]. However, the reliability of the package has been questioned by others [18, 19] although no better consultation assessment tool has yet been universally accepted. It has been used for consultation assessment internationally [20], with practice nurses i.e. nurses working in family physician offices [21] and medical students [22].

Assessment of consultation quality is divided into the key areas shown in Table 1.

thumbnail
Table 1. Areas of assessment of the clinical consultation as described by the Leicester Assessment Package.

Record keeping was not assessed in our study therefore an adjusted weighting was used to calculate a global score.

https://doi.org/10.1371/journal.pone.0160902.t001

Each component was graded A to E using the Leicester Assessment Package scoring criteria:

  1. A: Demonstrates mastery of all (or almost all) components consistently and to the highest standard.
  2. B: Demonstrates mastery of all (or almost all) components consistently and to a high standard, and some to the highest standard.
  3. C+: Consistently demonstrates capability in all (or almost all) components to a satisfactory standard—some to a high standard.
  4. C: Demonstrates capability in all (or almost all) components to a satisfactory standard but tends to lack discrimination, organisation and good time management.
  5. D: Demonstrates inadequacies in at least one component. Lacks discrimination and/or organisation. Tends to perform inconsistently.
  6. E: Demonstrates major omissions and/or serious defects. Grossly unacceptable standard overall.

Assessors were experienced GPs from teaching and training primary care practices. Assessing clinicians were required to attend two training sessions (led by SdeL) to familiarise them with the video format and to provide training in use of the Leicester Assessment Package. The two consultation videos used for training purposes were excluded from the final assessment. Each consultation was assessed by a minimum of two clinicians independently. Assessors were not able to comment on record keeping as there was no third video channel (the clinician’s computer screen) incorporated as part of the ALFA recording. This section of the Leicester Assessment Package was therefore excluded from the analysis with global scores adjusted accordingly.

Assessors were blinded to the role of the clinician performing each consultation. Assessors were required to state whether they thought the consulting clinician was a GP or a PA based on the consultations they had viewed by that clinician.

Following assessment of all the videos by the GP assessors, a workshop was undertaken with all the assessors which was recorded, transcribed and analysed. As part of this GP assessors were asked to comment on whether they felt the practitioners they had viewed were safe to practice and for their comments on the use of the Leicester Assessment Package as a method of assessment.

A Mann-Whitney U-test was used to compare differences between PAs and GPs in median scores, for each area of assessment of the Leicester Assessment Package, to identify statistical significance (p>0.05). A global score was calculated from the area of assessment scores using the adjusted weightings in Table 1. Each consultation was assessed by two assessors and these data compared to interrogate inter-rater reliability. Inter-rater reliability was calculated for each area of assessment of the Leicester Assessment Package using quadratic weighted kappa and Spearman’s rank correlation. The assessment process consisted of two rounds of assessment. As part of the second round ten videos were reviewed by the same pair of assessors as in the first round. These data were used to analyse intra-rater reliability (test-retest reliability) assessed using Cronbach’s alpha. Statistical analyses were performed using the software package R.

Case mix analysis.

In order to identify case mix differences between PA and GP consultations each consultation recording was reviewed by a clinician [AMcG] to determine the number of presenting complaints, the category of the presenting complaint (acute, chronic, or minor/symptoms), and the number of relevant chronic conditions.

The number of presenting complaints was defined as the number of unrelated primary complaints raised by the patient during each consultation. The category of the presenting complaint was based on the system devised by de Jong et al [23] and extended by Halter et al [24]. Each presenting complaint was categorised as follows: acute conditions were only those included in the list of acute conditions defined by de Jong et al. [23], chronic conditions were those on the list of chronic conditions defined in the Quality Outcomes Framework or another condition which has lasted for over a year as described by the patient, and minor/symptoms as conditions on the list of minor/symptoms conditions defined by de Jong et al. [23] or another condition which had lasted for less than a year and not classified as an acute condition. Relevant chronic conditions were defined as long-term conditions discussed in the consultation and relevant or related to one or more of the presenting complaints (but not a primary reason for consultation).

The relationship between the patient’s consultation with a PA or GP and the number of presenting complaints, nature of presenting complaint, and number of relevant chronic conditions was explored using a Chi-squared analysis.

Ethical considerations

Research ethics approval was granted by NHS Research Ethics Committee (REC) South East Coast–Surrey (REC reference number: 10/H1109/28).

Results

Consultations were recorded from five GPs and four PAs. Two GPs were female and two PAs female. The average number of years of experience since qualification was 19.2 (±8.9) in GPs and 7.0 (±6.1) in PAs. 86 patients were approached; 73 consented to the video; one subsequently withdrew consent for their consultation to be included following the consultation; no videos were withdrawn by clinicians following filming. A total of 62 patient consultations were used in the final assessment (the remaining videos were randomly selected for use in the training sessions); these comprised 41 GP and 21 PA consultations. One patient requested that only a single camera was used during their consultation. This consultation was included in the analysis with only a single video feed, the recording on the second camera was deleted.

Quality of Consultation

The average global score was C. The individual clinician averaged global scores were GP; C, C+, C+, C+, C+ and PA; C, C, C+, C+. Global scores for individual consultations ranged from D to A (Fig 1) however no PAs were given an overall grade A for any consultation. All consultations were above the minimum standard with no consultation given grade E.

thumbnail
Fig 1. The relative frequency of global consultation scores from 124 assessments of 62 consultations; 41 general practitioner (GP) and 21 physician assistant (PA).

https://doi.org/10.1371/journal.pone.0160902.g001

Median scores were higher for GP consultations than PA consultations in each domain of the Leicester Assessment Package (Table 2).

thumbnail
Table 2. Collated scores for 41 general practitioner (GP) consultations and 21 physician assistant (PA) consultations evaluated using the Leicester Assessment Package.

CI = confidence interval, IQR = Interquartile range.

https://doi.org/10.1371/journal.pone.0160902.t002

Inter-rater reliability of the Leicester Assessment Package was limited (Table 3). In weighted kappa analysis scores for behaviour/relationship with patient demonstrated substantial agreement; interview/history taking and global score, moderate agreement; physical examination, patient management, and problem solving, fair agreement; and anticipatory care no agreement. Spearman’s rank analysis produced comparable results: strong positive correlation between scores for behaviour/relationship with patient, interview/history taking, and global score; weak correlation, for physical examination, patient management, and problem solving; and no correlation, for anticipatory care.

thumbnail
Table 3. Weighted kappa and Spearman’s rank analyses of inter-rater reliability of consultation scores for 41 general practitioner (GP) consultations and 21 physician assistant (PA) consultations evaluated using the Leicester Assessment Package.

https://doi.org/10.1371/journal.pone.0160902.t003

Ten consultations were assessed twice by the same pair of raters giving a total of 20 repeat assessment scores. Intra-rater (rater-rater) reliability, assessed using Cronbach’s alpha, was good for interview/history taking, physical examination, problem solving, behaviour/relationship with patients, and global scores (Table 4). Poor intra-rater reliability was found for patient management. No intra-rater reliability was found for anticipatory care.

thumbnail
Table 4. Cronbach’s alpha analysis of intra-rater reliability of 20 pairs of consultation scores evaluated using the Leicester Assessment Package.

https://doi.org/10.1371/journal.pone.0160902.t004

The assessors were also asked for comments on the Leicester Assessment Package. Assessors highlighted that there was no complexity domain in the assessment and agreed they would like to any future assessment tools to include a measure of consultation complexity.

All four raters attempted to identify whether each clinician was a GP or PA based on the consultations they had observed. GPs were correctly identified 90% (95% CI 75–100; p = <0.001) of the time. PAs were correctly identified 56% (95% CI 31–81; p = 0.227) of the time.

During the workshop with GP assessors they stated that all consultations and practitioners appeared to be practicing safely.

Case mix analysis

From the 62 consultations recorded there were a total of 85 presenting complaints (mean 1.4 per consultation; range 1 to 3 per consultation). Of the 41 GP consultations 18 patients (44%) had two distinct presenting complaints and three (7%) had three distinct presenting complaints. Of the 21 PA consultations one patient (5%) had two presenting complaints and none had three. These differences were statistically significant (χ2 = 14.2, df = 2, p = 0.01).

Of the 85 presenting complaints one (1%) was related to an acute condition, 38 (45%) to chronic conditions, and 46 (54%) to minor/symptom conditions. Patients with presenting complaints relating to chronic conditions were more likely to consult a GP than a PA (χ2 = 14.2, df = 2, p = 0.01).

Fewer than half of the patients observed (28 of 62; 45%) had one or more chronic conditions relevant to their presenting complaint (but not the main reason for presentation); 8 of 21 (38%) of those consulting a PA and 20 of 41 (49%) of those consulting a GP. There was no significant relationship identified between having one or more chronic conditions related to the presenting complaint and the patient consulting a PA or GP (χ2 = 0.95, df = 1, p = 0.33)

Discussion

Principal findings

GPs were assessed to have performed better in all domains of the consultation than the PAs. All clinicians were judged as practising safely. However, the Leicester Assessment Package had limited inter-rater and intra-rater reliability. GP assessors were able to correctly identify GP consultations but failed to correctly identify 2 of 4 of the PAs. PAs in this study saw patients who largely attended with a single presenting complaint, whereas those who attended a GP had two or more presenting complaints. Patients seeing a GP were more likely to have one or more chronic diseases.

Comparison with the literature

To our knowledge, there are no other published studies comparing competence in the consultation between GPs and PAs from video-recorded consultations. We identified only one study in which transcribed audio-recordings were analysed to compare the consultations of GPs and nurse practitioners, another type of mid-level practitioner who substitute for GPs in primary care [25]. It reported on a comparison of utterances rather than an assessment of all the consultation elements. Other comparative studies of nurse practitioners in primary care have analysed data only from patient records and questionnaires [2630].

Patients consulting GPs had more presenting complaints and chronic diseases than those consulting the PAs. This supports evidence reported from England [13] and the USA [31]. The evidence of safety in consultation and also difficulty in determining the difference in the consulting style of some PAs from a GP is supported in our larger study using data from electronic patient records and patient report surveys [13]. PAs were found to be acceptable, effective and efficient in complementing the work of GPs and to generate better records of the consultation than GPs [13]. The GPs were rated more highly in the competency of their consultation which has not been reported before in comparison with PAs or nurse practitioners [31].

Limitations of the method

Our study, through necessity, was small; capturing data from a total of nine clinicians. This limits the generalisability of the results.

In our clinician sample the majority of the GPs had substantially more years training and experience than PAs. Therefore in direct comparison you might expect them to perform better than PAs. In addition we identified differing patient case mix between the two types of practitioner which may limit the comparison of consultation performance.

The impact of video recording may have affected clinician behaviour differentially between PAs and GPs. Some of the GPs in the sample had previously had consultations recorded during their training or practice whereas none of the PAs had experience. One GP scheduled extra breaks into her clinic to prevent delays caused by the videoing process.

We were unable to employ the full ALFA toolkit as ethical restrictions prevented us from capturing a third view; the computer screen. This may have produced some inaccuracies in measuring the clinician-computer interaction time compared with previous studies; however this discrepancy is likely to be small. This prevented us from accurately measuring the record keeping component of the Leicester Assessment Package.

The inter-rater reliability for a number of sections of the Leicester Assessment Package was only fair, namely; physical examination, patient management, and problem solving. There was no inter-rater correlation between scores of anticipatory care. Therefore, the difference between PAs and GPs observed for these domains of the Leicester Assessment Package should be treated with extreme caution. However, good to excellent agreement was observed for behaviour/relationship with patient, interview/history taking, and global scores and we therefore conclude that the differences we observed in these domains are likely to represent genuine quality differences.

Previous concerns have been raised about the reliability of the Leicester Assessment Package 18,19 and the initial analysis of the package demonstrated some scoring inconsistencies [15]. Our findings support these concerns. We suggest that in depth analysis of the Leicester Assessment Package is required to establish which components have a good inter and intra-rater reliability and which components require adaptation.

We cannot exclude rater bias towards GP consultation styles as all the raters were GPs. They may therefore maybe predisposed to rate more highly a consultation style more similar to their own. Additionally blinding of GP raters to clinician type appears to have been incomplete with some clinicians identified correctly by all raters. However despite this incomplete blinding GP raters were not able to correctly identify all clinicians as PAs or GPs.

The ability of GP raters to comment on safe practice of clinicians was limited. Raters were not able to view the consultation notes recorded by practitioner or the prescriptions issued (although in most consultations the prescribed medications were described to the patient by the practitioner).

Implications of the findings

It was interesting that experienced GPs could not differentiate PAs from GPs, and importantly they considered that PAs provide safe consultations. Bearing in mind that a graduate can be trained to be a PA in two years, compared with it taking nine years for a graduate going into medicine to become a GP, it is a credit to their training that they practice safely. However GPs performed consistently better across all domains of assessment than PAs. These differences may, in part, be accounted for by the greater duration of experience of the GPs included in this study, differing case mix, and perhaps partly by rater bias. Additional analysis with adjustment for these potential sources of error is required to see if this difference persists when comparing practitioners with a similar duration of clinical experience.

PAs’ role in the workforce is complimentary to that of GPs. They effectively treated cases that predominantly comprised patients with single presenting problems and were less likely to have a chronic disease. We did not collect any data to know whether this was due to patient or practice staff selection, or a combination of both. However, this suggests how PAs might be complimentary to the primary care workforce.

Conclusions

PAs in the primary care work force, as well as GPs, were regarded as safe. They saw less complex cases. The average competence of PA consultations was rated as lower than those of GPs. There is a place for PAs in the primary health care team, probably best in a group practice setting where they can meet a wider range of patient needs.

Ethical considerations

The research ethical review was undertaken by NHS Research Ethics Committee (REC) South East Coast–Surrey (REC reference number: 10/H1109/28).

Acknowledgments

This paper presents independent research funded by the National Institute for Health Research (NIHR: www.nihr.ac.uk) Health Services and Delivery Research (HS & DR) programme under the management of the National Institute for Health Research Evaluations, Trials and Studies Coordinating Centre (NETSCC) as part of a commissioned project (Grant number: 09/1801/1066). The views and opinions expressed by authors in this publication are those of the authors and do not necessarily reflect those of the National Health Service (NHS), the NIHR, NETSCC, the HS&DR programme or the Department of Health. We would also like to thank Muhammad Ayoub Ali for his contribution to the paper, when employed at University of Surrey; and Wilfred Carneiro (WC) and Jennifer Yiallorous (JY) for recruiting patients to the study. The authors would like to thank all the participating clinicians, practices, and patients.

Author Contributions

  1. Conceived and designed the experiments: SdL MH VMD.
  2. Performed the experiments: MH LJ VMD.
  3. Analyzed the data: APM MT SH SJ MH SdL.
  4. Contributed reagents/materials/analysis tools: SdL.
  5. Wrote the paper: APM VMD MT SH MH SJ LJ SdL.

References

  1. 1. Department of Health. The Competence and Curriculum Framework for the Physician Assistant 2006.
  2. 2. Cawley JF, Cawthon E, Hooker RS. Origins of the physician assistant movement in the United States. JAAPA, 2012. 25(12): p. 36–40, 42. pmid:23600002
  3. 3. Mittman DE, Cawley JF, Fenn WH. Physician assistants in the United States. BMJ, 2002. 325(7362): p. 485–7. pmid:12202333
  4. 4. Bushardt RL, Kuhns DH. The global Physician Assistant movement. Journal of the American Academy of Physician Assistants: 2014 27 8
  5. 5. Health Educaiton England Strategic Framework, Framework 15 2014–2029. Available: https://hee.nhs.uk/sites/default/files/documents/HEE%20Strategic%20Framework%20%202015%20Refresh%20Final%20document.pdf.
  6. 6. Faculty of Physician Associates; Managed Volunary Register. Available: http://www.fparcp.co.uk/pamvr-home/.
  7. 7. Drennan V., Levenson R., Halter M, Tye C. Physician assistants in English general practice: a qualitative study of employers' viewpoints. J Health Serv Res Policy, 2011. 16(2): p. 75–80. pmid:21389060
  8. 8. Drennan VM, Chattopadhyay K, Halter M, Brearley S, de Lusignan S, Gabe J. et al., Physician assistants in English primary care teams: a survey. J Interprof Care, 2012. 26(5): p. 416–8. pmid:22574762
  9. 9. Halter M, Drennan V, Chattopadhyay K, Carneiro W, Yiallouros J, de Lusignan S, et al. The contribution of physician assistants in primary care: a systematic review. BMC Health Serv Res, 2013. 13: p. 223. pmid:23773235
  10. 10. Drennan VM, Halter M, Brearley S, Carneiro W, Gabe J, Gage H, et al. Investigating the contribution of physician assistants to primary care in England: a mixed-methods study. Health Serv Deliv Res, 2014. 2(16).
  11. 11. Salisbury C, Procter S, Stewart K, Bowen L, Purdy S, Ridd M, et al. The content of general practice consultations: cross-sectional study based on video recordings. Br J Gen Pract. 2013 Nov;63(616):e751–9. pmid:24267858
  12. 12. Pearce C, Kumarpeli P, de Lusignan S. Getting seamless care right from the beginning—integrating computers into the human interaction. Stud Health Technol Inform. 2010;155:196–202. pmid:20543329
  13. 13. Drennan VM, Halter M, Joly L, Gage H, Grant RL, Gabe J, et al. Physician associates and GPs in primary care: a comparison.Br J Gen Pract. 2015 May;65(634):e344–50. pmid:25918339
  14. 14. de Lusignan S, Kumarapeli P, Chan T, Pflug B, van Vlymen J, Jones B, et al. The ALFA (Activity Log Files Aggregation) Toolkit: A Method for Precise Observation of the Consultation. J Med Internet Res, 2008. 10(4).
  15. 15. Fraser RC, McKinley RK, Mulholland H. Consultation competence in general practice: testing the reliability of the Leicester assessment package. Br J Gen Pract, 1994. 44(384): p. 293–6. pmid:8068374
  16. 16. Fraser RC, McKinley RK, Mulholland H. Consultation competence in general practice: establishing the face validity of prioritized criteria in the Leicester assessment package. Br J Gen Pract, 1994. 44(380): p. 109–13. pmid:8204317
  17. 17. Jiwa M, McKinley R, O'Shea C, Arnet H, Spilsbury K, Smith M. Investigating the impact of extraneous distractions on consultations in general practice: lessons learned. BMC Med Res Methodol, 2009. 9: p. 8. pmid:19193246
  18. 18. Braunholtz D. Leicester assessment package. Br J Gen Pract, 1995. 45(390): p. 51–2. pmid:7779480
  19. 19. Halkett GK, Jiwa M, O'Shea C, Smith M, Leong E, Jackson M, et al. Management of cases that might benefit from radiotherapy: a standardised patient study in primary care. Eur J Cancer Care 2012;21:259–65. http://dx.doi.org/10.1111/j.1365-2354.2011.01314.x.
  20. 20. Fraser RC, Sarkhou ME, McKinley RK, Van der Vleuten C. Regulatory end-point assessment of the consultation competence of family practice trainees in Kuwait. Eur J Gen Pract 2006;12:100–7. http://dx.doi.org/10.1080/13814780600898353. pmid:17002957
  21. 21. Redsell SA, Hastings AM, Cheater FM, Fraser RC. Devising and establishing the face and content validity of explicit criteria of consultation competence in UK primary care nurses. Nurse Educ Today 2003;23:299–306. http://dx.doi.org/10.1016/S0260-6917(03)00013-3. pmid:12727097
  22. 22. McKinley RK, Fraser RC, van der Vleuten C, Hastings AM. Formative assessment of the consultation performance of medical students in the setting of general practice using a modified version of the Leicester assessment package. Med Educ 2000;34:573–9. http://dx.doi.org/10.1046/j.1365-2923.2000.00490.x. pmid:10886641
  23. 23. De Jong J, Visser MRM, Wieringa-de Waard M. Exploring differences in patient mix in a cohort of GP trainees and their trainers. BMJ Open, 2011. 1(2).
  24. 24. Halter M, Joly L, de Lusignan S, Grant RL, Gage H, Drennan VM. Development and testing of a patient case-mix classification system to differentiate caseloads in primary care: observational data from GPs and Physician Associates. Forthcoming.
  25. 25. Seale C, Anderson E, Kinnersley P. Comparison of GP and nurse practitioner consultations: an observational study.Br J Gen Pract. 2005 Dec;55(521):938–43. pmid:16378563
  26. 26. Shum C, Humphreys A, Wheeler D, Cochrane MA, Skoda S, Clement S. Nurse management of patients with minor illnesses in general practice: multicentre randomised controlled trial. BMJ 2000;320:1038–43. http://dx.doi.org/10.1136/bmj.320.7241.1038. pmid:10764365
  27. 27. Venning P, Durie A, Roland M, Roberts C, Leese B. Randomised controlled trial comparing cost effectiveness of general practitioners and nurse practitioners in primary care. BMJ 2000;320:1048–53. http://dx.doi.org/10.1136/bmj.320.7241.1048. pmid:10764367
  28. 28. Kinnersley P, Anderson E, Parry K, Clement J, Archard L, Turton P, et al. Randomised controlled trial of nurse practitioner versus general practitioner care for patients requesting ‘same day’ consultations in primary care. BMJ 2000;320:1043–8. http://dx.doi.org/10.1136/bmj.320.7241.1043. pmid:10764366
  29. 29. Dierick-van Daele ATM, Metsemakers JFM, Derckx EWCC, Spreeuwenberg C, Vrijhoef HJM. (2009), Nurse practitioners substituting for general practitioners: randomized controlled trial. Journal of Advanced Nursing, 65: 391–401. pmid:19191937
  30. 30. Morgan PA, Abbott DH, McNeil RB, Fisher DA. Characteristics of primary care office visits to nurse practitioners, physician assistants and physicians in United States Veterans Health Administration facilities, 2005 to 2010: a retrospective cross-sectional analysis. Hum Resour Health 2012;10:42. pmid:23148792
  31. 31. Laurant M, Reeves D, Hermens R, Braspenning J, Grol R, Sibbald B. Substitution of doctors by nurses in primary care. Cochrane Database Syst Rev, 2005(2): p. Cd001271. pmid:15846614