-
PDF
- Split View
-
Views
-
Cite
Cite
Stine M D Carstensen, Marie Juul Velander, Lars Konge, Mikkel Østergaard, Mogens Pfeiffer Jensen, Søren Andreas Just, Lene Terslev, Training and assessment of musculoskeletal ultrasound and injection skills—a systematic review, Rheumatology, Volume 61, Issue 10, October 2022, Pages 3889–3901, https://doi.org/10.1093/rheumatology/keac119
- Share Icon Share
Abstract
To examine how residents are trained and assessed in musculoskeletal US (MSUS), MSUS-guided and landmark-guided joint aspiration and injection. Additionally, to present the available assessment tools and examine their supporting validity evidence.
A systematic search of PubMed, Cochrane Library and Embase was conducted in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines and studies published from 1 January 2000 to 31 May 2021 were included. Two independent reviewers performed the search and data extraction. The studies were evaluated using the Medical Education Research Quality Instrument (MERSQI).
A total of 9884 articles were screened, and 43 were included; 3 were randomized studies, 21 pre- and post-test studies, 16 descriptive studies and 3 studies developing assessment tools. The studies used various theoretical training modalities, e.g. lectures, anatomical quizzes and e-learning. The practical training models varied from mannequins and cadavers to healthy volunteers and patients. The majority of studies used subjective ‘comfort level’ as assessment, others used practical examination and/or theoretical examination. All training programs increased trainees’ self-confidence, theoretical knowledge, and/or practical performance, however few used validated assessment tools to measure the effect. Only one study met the MERSQI high methodical quality cut-off score of 14.
The included studies were heterogeneous, and most were of poor methodological quality and not based on contemporary educational theories. This review highlights the need for educational studies using validated theoretical and practical assessment tools to ensure optimal MSUS training and assessment in rheumatology.
Studies examining training and assessment of musculoskeletal US and injection competences have a low level of evidence.
No validated rating scale assessing competencies in US-guided joint injections exists.
International uniform competency-based training programs using validated assessment tools are needed.
Introduction
The use of musculoskeletal US (MSUS) has expanded significantly during recent decades and is widely implemented as a clinical tool for the diagnosis, monitoring and treatment of patients with rheumatological diseases [1–3]. MSUS also provides guidance to invasive procedures such as joint aspirations and injections, which are core procedures for a rheumatologist [4]. Joint aspirations and injections can be performed landmark-guided or MSUS-guided. However, MSUS-guided procedures are more difficult to learn and the quality and value of MSUS, including MSUS-guided aspiration and injection, is highly dependent on the competencies of the individual operator [5, 6].
Mastering MSUS, and MSUS-guided aspiration and injection require theoretical knowledge and extensive hands-on training [7–9]. Additionally, hand–eye coordination training ensures the ability to control the US probe and the needle at the same time [4, 7]. Evidence-based and structured training is essential when residents acquire MSUS competences. MSUS training programs are already available through different scientific societies of rheumatology [10, 11]. The programs typically require both theoretical and practical skills in the competency assessment but differ in the required minimum number of MSUS examinations to obtain a certain competency level [12, 13]. Depending on the specific trainees, a pre-defined number of examinations may not necessarily result in a sufficient competence level and may even exceed the training needed due to different learning curves [5, 14]. A more individualized approach where certification is granted based on an assessment of competences would be efficient and could ensure that all trainees reach the necessary level of proficiency before proceeding to individual practice.
Once training is completed it is crucial to certify resident competences [6, 15] using dedicated assessment tools supported by validity evidence to ensure reliable assessments with minimal bias [16–21].
However, before implementing an assessment tool it is essential to examine different sources of validity, thereby ensuring that the tool is measuring what it is supposed to. The classic validity framework including construct, content and criterion is replaced by Messick’s modern, uniformed framework [22]. Messick’s framework is recommended as the standard by the American Educational Research Association. It is one of the most described and recognized frameworks for validity testing in medical education and procedural training [23]. Messick’s framework consists of five sources of validity: content, response process, internal structure, relations to other variables and consequences [24]. Ideally, all five sources of validity evidence should be examined before using a new assessment tool.
The importance for international harmonization in MSUS education has been highlighted in several studies, including the need for standardized training programs containing both competency-based training, objective assessment and certification [1, 7, 9, 25–30]. A comprehensive work has already been made to promote this harmonization across several European countries [10, 11, 28, 31].
In this review we sought to answer the following research questions:
In residents, is systematic training of MSUS, MSUS-guided and landmark-guided joint aspiration and injection skills in comparison with no formal training associated with improved knowledge, skills or behaviour?
What are the educational content and types of competencies being trained in relation to achieve MSUS, MSUS-guided and landmark-guided joint aspiration and injection skills?
What tools are available to assess competences of residents in MSUS, MSUS-guided and landmark-guided joint aspiration and injection, and what validity evidence supports them?
The overall aim of this systematic literature review was to examine how residents are trained and assessed in MSUS, including MSUS-guided and landmark-guided joint aspiration and injection. Additionally, we aim to present the available assessment tools and examine their supporting validity evidence.
Methods
Design, eligibility criteria and study selection
We have followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines [32] and have illustrated the selection process in a flow diagram according to their recommendations (Fig. 1). We included all original learning or educational studies on how residents are trained and assessed in MSUS, including MSUS-guided and landmark-guided joint aspiration and injection. Moreover, studies reporting the development, evaluation or implementation of assessment tools to assess resident competencies in MSUS, MSUS-guided and landmark-guided joint aspiration and injection were included. Studies were only excluded if they: (i) were published in non-English language or (ii) were abstracts, case reports, case series, letters or review articles. Two authors (S.M.D.C. and M.J.V.) independently identified all relevant articles by screening titles and abstracts and read all potentially eligible studies in full. Any disagreement was resolved by discussion. Additionally, the reference lists of included studies were examined for additional eligible trials.

Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) flow diagram of study selection process
Search strategy
We searched PubMed, Embase and the Cochrane Library using a search strategy developed in collaboration with a research librarian from the medical research library of Copenhagen, Denmark. We included studies from the past two decades, more precisely studies published from January 2000 to May 2021. The search strategy is available as supplementary material (supplementary Fig. S1, available at Rheumatology online). The Covidence systematic review software (Veritas Health Innovation, Melbourne, Australia) was used to manage the identified studies, to search for duplicates and for screening abstracts.
Data collection and extraction
Data extraction was performed independently by the two authors, in a data extraction chart using Microsoft Excel (Microsoft Corp, Redmond, WA, USA) with the following headlines: study design, participants, theoretical and practical training, practical training models, assessment, outcome measures, statistical analysis methods and study conclusion.
Assessment of methodological quality
The studies were evaluated using the Medical Education Research Quality Instrument (MERSQI), which was developed and validated in 2007. The MERSQI was designed to evaluate the methodical quality of medical educational research [33, 34] and includes six domains: study design, sampling, type of data (subjective or objective), validity, data analysis and outcomes. The maximum MERSQI score is 18 points. A mean MERSQI score below 9 of a given study is predicted to be rejected by journal editors, whereas a mean MERSQI score of 10.7 indicates acceptance. Moreover, studies with a score of 14 or above are rated as ‘high quality’ studies [33, 35–37].
Results
After the removal of duplicates, a total of 9884 articles were screened for eligibility, and 76 papers were identified and read in full. The excluded articles did not meet the inclusion criteria and covered topics other than training and assessment of competences in MSUS, MSUS-guided or landmark-guided aspiration and injection. Of the remaining 76 studies, 33 were excluded because they did not meet the inclusion criteria or only the conference abstracts were available (Fig. 1). No further studies were identified based on the reference lists of included studies. Of the 43 included studies, 3 examined MSUS-guided aspiration and injection training [38–40], 14 examined MSUS training [41–54], 23 examined landmark-guided aspiration and injection training [55–77], and 3 described the development of assessment tools [78–80].
The study design, participants, practical training models and assessment are described in Table 1 and more deeply in supplementary Table S1, available at Rheumatology online. Studies developing and validating assessment tools are described in Table 2. Study outcome measures, statistical analysis methods, study conclusion and individual MERSQI evaluation are described in supplementary Table S2 (available at Rheumatology online), whereas a concise MERSQI evaluation comparing MSUS, MSUS-guided and landmark-guided joint aspiration and injection studies is illustrated in Table 3.
Description of included studies examining training of MSUS, MSUS-guided or landmark-guided joint aspiration and injection skills
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studiesa | 14 | 3 | 23 |
Study design | |||
Randomized | 0 | 0 | 3 |
Pre- and post-test | 2 | 2 | 17 |
Descriptive | 12 | 1 | 3 |
Participantsb | |||
Residents | 408 | 38 | 1388 |
Experts | 1 | – | 8 |
Medical students | 149 | ||
Others | 47 | ||
Practical training modelsc | |||
Patients | 6 | 1 | – |
Healthy volunteers | 1 | – | – |
Images from patients | 1 | – | – |
Low fidelity models | – | 1 | – |
Cadavers | – | 1 | 2 |
Mannequins | – | – | 11 |
Mixed | 6 | – | 10 |
Assessmentc | |||
Hands-on skills | 7 | 1 | – |
Theoretical | – | – | 2 |
Questionnaire | 2 | 1 | 13 |
Mixed | 5 | 1 | 8 |
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studiesa | 14 | 3 | 23 |
Study design | |||
Randomized | 0 | 0 | 3 |
Pre- and post-test | 2 | 2 | 17 |
Descriptive | 12 | 1 | 3 |
Participantsb | |||
Residents | 408 | 38 | 1388 |
Experts | 1 | – | 8 |
Medical students | 149 | ||
Others | 47 | ||
Practical training modelsc | |||
Patients | 6 | 1 | – |
Healthy volunteers | 1 | – | – |
Images from patients | 1 | – | – |
Low fidelity models | – | 1 | – |
Cadavers | – | 1 | 2 |
Mannequins | – | – | 11 |
Mixed | 6 | – | 10 |
Assessmentc | |||
Hands-on skills | 7 | 1 | – |
Theoretical | – | – | 2 |
Questionnaire | 2 | 1 | 13 |
Mixed | 5 | 1 | 8 |
Studies developing assessment tools are not included in this table.
Accumulated number of participants enrolled in the studies.
Number of studies. MSUS: musculoskeletal US.
Description of included studies examining training of MSUS, MSUS-guided or landmark-guided joint aspiration and injection skills
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studiesa | 14 | 3 | 23 |
Study design | |||
Randomized | 0 | 0 | 3 |
Pre- and post-test | 2 | 2 | 17 |
Descriptive | 12 | 1 | 3 |
Participantsb | |||
Residents | 408 | 38 | 1388 |
Experts | 1 | – | 8 |
Medical students | 149 | ||
Others | 47 | ||
Practical training modelsc | |||
Patients | 6 | 1 | – |
Healthy volunteers | 1 | – | – |
Images from patients | 1 | – | – |
Low fidelity models | – | 1 | – |
Cadavers | – | 1 | 2 |
Mannequins | – | – | 11 |
Mixed | 6 | – | 10 |
Assessmentc | |||
Hands-on skills | 7 | 1 | – |
Theoretical | – | – | 2 |
Questionnaire | 2 | 1 | 13 |
Mixed | 5 | 1 | 8 |
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studiesa | 14 | 3 | 23 |
Study design | |||
Randomized | 0 | 0 | 3 |
Pre- and post-test | 2 | 2 | 17 |
Descriptive | 12 | 1 | 3 |
Participantsb | |||
Residents | 408 | 38 | 1388 |
Experts | 1 | – | 8 |
Medical students | 149 | ||
Others | 47 | ||
Practical training modelsc | |||
Patients | 6 | 1 | – |
Healthy volunteers | 1 | – | – |
Images from patients | 1 | – | – |
Low fidelity models | – | 1 | – |
Cadavers | – | 1 | 2 |
Mannequins | – | – | 11 |
Mixed | 6 | – | 10 |
Assessmentc | |||
Hands-on skills | 7 | 1 | – |
Theoretical | – | – | 2 |
Questionnaire | 2 | 1 | 13 |
Mixed | 5 | 1 | 8 |
Studies developing assessment tools are not included in this table.
Accumulated number of participants enrolled in the studies.
Number of studies. MSUS: musculoskeletal US.
Description of included studies developing assessment tools to evaluate practical MSUS, MSUS-guided or landmark-guided joint aspiration and injection skills
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studies | 1 | 1 | 1 |
Type of assessment tool | |||
Rating scale | x | – | x |
Checklist | – | x | – |
Sources of validity | |||
Messick’s (5 sources)a | – | – | – |
Other validity sources | |||
Content | x | x | x |
Face | – | – | x |
Concurrent | x | – | x |
Construct | x | – | x |
Predictive | – | – | x |
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studies | 1 | 1 | 1 |
Type of assessment tool | |||
Rating scale | x | – | x |
Checklist | – | x | – |
Sources of validity | |||
Messick’s (5 sources)a | – | – | – |
Other validity sources | |||
Content | x | x | x |
Face | – | – | x |
Concurrent | x | – | x |
Construct | x | – | x |
Predictive | – | – | x |
Messick’s framework is recommended as the standard by the American Educational Research Association and includes five sources of validity: content, response process, internal structure, relations to other variables and consequences. MSUS: musculoskeletal US.
Description of included studies developing assessment tools to evaluate practical MSUS, MSUS-guided or landmark-guided joint aspiration and injection skills
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studies | 1 | 1 | 1 |
Type of assessment tool | |||
Rating scale | x | – | x |
Checklist | – | x | – |
Sources of validity | |||
Messick’s (5 sources)a | – | – | – |
Other validity sources | |||
Content | x | x | x |
Face | – | – | x |
Concurrent | x | – | x |
Construct | x | – | x |
Predictive | – | – | x |
Study characteristics . | MSUS . | MSUS-guided . | Landmark-guided . |
---|---|---|---|
No. of studies | 1 | 1 | 1 |
Type of assessment tool | |||
Rating scale | x | – | x |
Checklist | – | x | – |
Sources of validity | |||
Messick’s (5 sources)a | – | – | – |
Other validity sources | |||
Content | x | x | x |
Face | – | – | x |
Concurrent | x | – | x |
Construct | x | – | x |
Predictive | – | – | x |
Messick’s framework is recommended as the standard by the American Educational Research Association and includes five sources of validity: content, response process, internal structure, relations to other variables and consequences. MSUS: musculoskeletal US.
Assessment of methodological quality of the included studies using the Medical Education Research Study Quality Instrument
Instrument . | Description . | Item . | Options and scoresa . | . | Mean scores . | ||
---|---|---|---|---|---|---|---|
. | . | . | . | . | MSUSb . | MSUS-guidedb . | Landmark-guidedb . |
MERSQI | Developed to appraise the methodological quality of studies of medical education | Study design |
| 1 | |||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.0 | 1.2 | 1.5 | |||
Sampling | Institutions | ||||||
|
| 0.5 | |||||
|
| 1 | |||||
| 1.5 | ||||||
Response rate % | |||||||
| |||||||
| 0.5 | ||||||
| 1 | ||||||
| 1.5 | 1.1 | 0.5 | 1.4 | |||
Type of data |
| 1 | |||||
| 3 | 2.7 | 1.7 | 1.8 | |||
Validity evidence for evaluation instrument scores |
| ||||||
| 1 | ||||||
| 1 | ||||||
| 1 | 0.4 | 0 | 0.2 | |||
Data analysis | Appropriateness of analysis | ||||||
|
| 1 | |||||
| Complexity of analysis | ||||||
| 1 | ||||||
| 2 | 2.4 | 3 | 2.8 | |||
Outcome |
| 1 | |||||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.4 | 1.2 | 1.2 | |||
Total score | 18 | 9.2 | 7.5 | 8.9 |
Instrument . | Description . | Item . | Options and scoresa . | . | Mean scores . | ||
---|---|---|---|---|---|---|---|
. | . | . | . | . | MSUSb . | MSUS-guidedb . | Landmark-guidedb . |
MERSQI | Developed to appraise the methodological quality of studies of medical education | Study design |
| 1 | |||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.0 | 1.2 | 1.5 | |||
Sampling | Institutions | ||||||
|
| 0.5 | |||||
|
| 1 | |||||
| 1.5 | ||||||
Response rate % | |||||||
| |||||||
| 0.5 | ||||||
| 1 | ||||||
| 1.5 | 1.1 | 0.5 | 1.4 | |||
Type of data |
| 1 | |||||
| 3 | 2.7 | 1.7 | 1.8 | |||
Validity evidence for evaluation instrument scores |
| ||||||
| 1 | ||||||
| 1 | ||||||
| 1 | 0.4 | 0 | 0.2 | |||
Data analysis | Appropriateness of analysis | ||||||
|
| 1 | |||||
| Complexity of analysis | ||||||
| 1 | ||||||
| 2 | 2.4 | 3 | 2.8 | |||
Outcome |
| 1 | |||||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.4 | 1.2 | 1.2 | |||
Total score | 18 | 9.2 | 7.5 | 8.9 |
Max score for each item is 3, which gives a total max score of 18.
Mean score for each MERSQI item in studies examining training and assessment of MSUS (n = 14), MSUS-guided (n = 3) or landmark-guided joint aspiration and injection skills (n = 23). MSUS: musculoskeletal US; MERSQI: Medical Education Research Quality Instrument.
Assessment of methodological quality of the included studies using the Medical Education Research Study Quality Instrument
Instrument . | Description . | Item . | Options and scoresa . | . | Mean scores . | ||
---|---|---|---|---|---|---|---|
. | . | . | . | . | MSUSb . | MSUS-guidedb . | Landmark-guidedb . |
MERSQI | Developed to appraise the methodological quality of studies of medical education | Study design |
| 1 | |||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.0 | 1.2 | 1.5 | |||
Sampling | Institutions | ||||||
|
| 0.5 | |||||
|
| 1 | |||||
| 1.5 | ||||||
Response rate % | |||||||
| |||||||
| 0.5 | ||||||
| 1 | ||||||
| 1.5 | 1.1 | 0.5 | 1.4 | |||
Type of data |
| 1 | |||||
| 3 | 2.7 | 1.7 | 1.8 | |||
Validity evidence for evaluation instrument scores |
| ||||||
| 1 | ||||||
| 1 | ||||||
| 1 | 0.4 | 0 | 0.2 | |||
Data analysis | Appropriateness of analysis | ||||||
|
| 1 | |||||
| Complexity of analysis | ||||||
| 1 | ||||||
| 2 | 2.4 | 3 | 2.8 | |||
Outcome |
| 1 | |||||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.4 | 1.2 | 1.2 | |||
Total score | 18 | 9.2 | 7.5 | 8.9 |
Instrument . | Description . | Item . | Options and scoresa . | . | Mean scores . | ||
---|---|---|---|---|---|---|---|
. | . | . | . | . | MSUSb . | MSUS-guidedb . | Landmark-guidedb . |
MERSQI | Developed to appraise the methodological quality of studies of medical education | Study design |
| 1 | |||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.0 | 1.2 | 1.5 | |||
Sampling | Institutions | ||||||
|
| 0.5 | |||||
|
| 1 | |||||
| 1.5 | ||||||
Response rate % | |||||||
| |||||||
| 0.5 | ||||||
| 1 | ||||||
| 1.5 | 1.1 | 0.5 | 1.4 | |||
Type of data |
| 1 | |||||
| 3 | 2.7 | 1.7 | 1.8 | |||
Validity evidence for evaluation instrument scores |
| ||||||
| 1 | ||||||
| 1 | ||||||
| 1 | 0.4 | 0 | 0.2 | |||
Data analysis | Appropriateness of analysis | ||||||
|
| 1 | |||||
| Complexity of analysis | ||||||
| 1 | ||||||
| 2 | 2.4 | 3 | 2.8 | |||
Outcome |
| 1 | |||||
| 1.5 | ||||||
| 2 | ||||||
| 3 | 1.4 | 1.2 | 1.2 | |||
Total score | 18 | 9.2 | 7.5 | 8.9 |
Max score for each item is 3, which gives a total max score of 18.
Mean score for each MERSQI item in studies examining training and assessment of MSUS (n = 14), MSUS-guided (n = 3) or landmark-guided joint aspiration and injection skills (n = 23). MSUS: musculoskeletal US; MERSQI: Medical Education Research Quality Instrument.
Study design and participants
Of the 43 included studies, 16 were descriptive, 21 were pre- and post-test studies, and only 3 were randomized. The three randomized studies all reported findings on landmark-guided aspiration and injection training (Table 1). The remaining three studies developed new assessment tools. The study participants in the included studies were residents in rheumatology, family medicine and internal medicine with varying experience levels. The number of participants included in the studies ranged from 1 to 474 (supplementary Table S1, available at Rheumatology online).
Theoretical skill training
Training of the theoretical skills in MSUS, MSUS-guided and landmark-guided aspiration and injection skills were heterogeneous in the included studies. Within the theoretical part of the training curriculum, there were many different modalities used, e.g. reading material, lectures, videos, anatomical quizzes, interactive games or e-learning (supplementary Table S1, available at Rheumatology online). Furthermore, the time used on the theoretical part of the curriculum varied from 30 min [39] to 5 h [68].
E-learning was used in one study by Filippucci et al. [52]. After a 3-day course followed by 6 months of web-based tutoring, there was an improvement in the agreement of image interpretation (i.e. synovitis yes/no) between the trainees and their trainer.
Practical skill training
Training of the practical skills in MSUS, MSUS-guided and landmark-guided aspiration and injection skills were also heterogeneous in the included studies. Within the practical part of the training curriculum, the modalities varied between direct or indirect supervised training, individual or team-based training, and practical demonstrations on patients, healthy models, mannequins or cadavers. In addition, the time used on the practical part of the training curriculum varied from a 1-day course [38, 57, 67] to 10 months of clinical practice [51].
The practical training models
MSUS-guided aspiration and injection training
The three studies examining MSUS-guided aspiration and injection training used either patient models, fresh frozen cadaver models or a low fidelity model (Table 1). Charnoff et al. presented a low fidelity model [38] made by agar powder dissolved in water, including a chicken leg used to simulate muscle and tendon and two tomatoes simulating blood vessels. The participants were required to identify structures in the agar model using the US equipment and perform an MSUS-guided procedure. Afterwards, they were given a short lecture and instructed to complete the same task again. The residents demonstrated improved MSUS-guided skills; however, no validated assessment tools were used.
MSUS training
Within the 14 studies examining MSUS training, they used either healthy models, patients, US images from patients or a combination (Table 1). One study by Miguel et al. used images from patients with SpA and normal controls in their training program [47]. The residents showed improved image interpretation after the theory session (30 min) and the reading session (60 min).
Landmark-guided aspiration and injection training
Twenty-three studies examined landmark-guided aspiration and injection training, and the majority used mannequins (n = 19, 83%). Three of these studies used a combination of mannequins and cadavers [55, 63, 76], and three used a combination of mannequins and patients in their training programs [68, 70, 74]. Stroud et al. presented a study examining hybrid simulation in knee arthrocentesis training [75]. This study was the only study training the residents in both the technical aspects of the procedure and the communicative aspect using standardized patients. The residents’ performances were video-taped and assessed by two individual physicians using a procedure-specific modified version of the direct operational procedure skills (DOPS) rating form with established evidence of validity [81]. The communicative skills were rated by both the physicians and the standardized patient.
Three randomized studies examined the efficacy of different training curriculum on resident performance and self-confidence. Michels et al. compared three training programs using residents in general practice as participants [55]. The training program included five anatomical sites; glenohumeral, subacromial, lateral epicondyle, carpal tunnel and the knee. Improved skills were found when a theoretical lecture was followed by hands-on training, but no statistical differences in skills were found between training on cadavers as compared with mannequins.
Leopold et al. compared three training programs using residents, nurses and physician assistants as participants [56]. The program only examined one anatomical site: the knee. There was no statistically significant difference in outcome between groups randomized to receive instruction through a printed manual, a video or live hands-on instructions.
Gormley et al. compared two training programs using residents in general practice as participants [57]. The program examined one anatomical site: the shoulder. After a 1-day course, the intervention group got additional training on patients in a joint injection clinic, which improved confidence and increased injection activity 6 months post-course.
Assessment of the theoretical and practical skills
The assessment of participants’ skills varied from study to study, with the majority of studies using subjective comfort level (23), theoretical assessment, e.g. multiple-choice questions (10) and/or practical examination (13) (supplementary Table S1, available at Rheumatology online). Subjective comfort scales ranging from 4 points [39] to 10 points [74] were used as assessment in the 23 studies.
Thirteen studies assessed participants’ practical skills. The studies used a spectrum of different methods to measure and evaluate the participants practical skills e.g. procedure time, structures identified, image interpretation (± synovitis or pathology present/absent), image quality (1–10), Objective structured clinical exam (OSCE) score checklist (yes/no), OSCE score rating scale or DOPS rating form (supplementary Table S1, available at Rheumatology online).
In general, validity evidence supporting the outcomes of the assessment tools was limited. Only three studies reported or discussed the validity of the assessment tool used [45, 51, 75]. Two of them used the DOPS rating form, either in the original version or in a modified procedure-specific version [51, 75] and briefly stated that this assessment tool had been demonstrated to have evidence of validity.
Development of assessment tools
MSUS-guided aspiration and injection skills assessment
Kunz et al. developed a checklist for US-guided arthrocentesis using the Delphi method. However, they only conducted one source of validity evidence: content validity [78] (Table 2). Therefore, they concluded that further validity evidence should be examined before using the tool in the clinical or simulated environment. Furthermore, it is noteworthy that the use of a checklist has shown to have lower sensitivity, reliability and validity measures compared with rating scales [82].
MSUS skills assessment
Kissin et al. developed and gathered validity evidence for written multiple-choice questions and established a passing score [45]. Ten faculty members developed the multiple-choice questions, and the test could discriminate between operators with different levels of experience.
In a more recent study by Kissin et al. the research group focused on gathering validity evidence for a practical examination for assessing MSUS skills, i.e. OSCE [79]. The participants were assessed by trained and blinded raters using the newly developed 5-point rating scale. Their performances were evaluated strictly on image characteristics and quality. A pass/fail score was established using the borderline methodology. The research group used concurrent and construct validity and tested the reliability using interrater reliability (Table 2).
Landmark-guided aspiration and injection skills assessment
Singh et al. developed a procedure specific rating scale to assess resident competences in landmark-guided joint aspiration and injection skills [80]. The 5-point rating scale consisted of nine items in total. The research group used the following sources of validity: construct, content, predictive, face and concurrent validity (Table 2).
Evaluation of the methodological quality
The detailed MERSQI evaluation of the included studies is presented in supplementary Table S2 (available at Rheumatology online) and a concise MERSQI evaluation comparing MSUS, MSUS-guided, and landmark-guided joint aspiration and injection studies are presented in Table 3. Only five studies were above the ‘editor acceptance’ score of 10.7 points and just one of these reached the ‘high-methodological quality’ cut-off of 14 points [45]. Twenty-one studies were below the ‘editor rejection’ score of 9 points. The mean MERSQI quality score was 7.5 points in MSUS-guided aspiration and injection educational studies, 9.2 in studies examining MSUS education and 8.9 in landmark-guided aspiration and injection educational studies (maximum points 18) (Table 3). This translates to an overall low quality of studies with a high risk of bias, especially due to the pre- and post-test study designs, lack of objective competence assessment using validated assessment tools and non-patient-related outcomes.
Discussion
Despite the general agreement on the importance of MSUS, MSUS-guided and landmark-guided joint aspiration and injection skills, no international consensus exists on how to optimally train and assess them. This systematic literature review revealed substantial heterogeneity in the current published studies regarding the theoretical and practical part of the curriculum and the assessment methods, and hence no guidelines can be given. Despite the heterogeneity, all included studies found a favourable effect on confidence or competence regardless of the training intervention or assessment method. However, the majority of the included studies are of low methodological quality, with an associated low MERSQI score. The main methodological problem was study design, where we found a total lack of randomized studies comparing training strategies in MSUS or MSUS-guided aspiration and injection. Moreover, several studies used a single group pre- and post-test study design. This set-up which compares ‘something’ with ‘nothing’ has been discussed and criticized for several years and primarily supports the fact that any training intervention has an effect [83–85]. Additionally, the majority of studies used self-assessment even though there is solid scientific evidence showing that self-assessment is an unreliable measure of competences [86–88]. Validated assessment tools assessing resident practical competences were only used in three studies [45, 51, 75].
When we look deeper into the results of this systematic review, we found three central topics of interest according to curriculum development: the theoretical part, the practical part and the assessment methods. The majority of included studies used a blended training approach with both a theoretical and a practical part. However, the time spent on each part varied greatly between the studies.
For the theoretical part, none of the studies used a strictly web-based learning approach even though similar competences can be obtained in this way when compared with lecture-based learning [89]. One research group used a web-based learning environment for continued learning after a 3-day course with lectures [52]. However, they did not replace the theoretical lecture-based part with a web-based approach before the course. Web-based learning can reduce cost, time and the staff required in lecture-based learning, thereby releasing more time for hands-on training [29, 90, 91]. The flexibility and accessibility of a web-based approach as well as the possibility of repeating the online training after a course are major advantages.
Concerning the practical part of the curriculum it is important to recognize that different hands-on training models can contribute to different aspects of the learning process; healthy models are suitable for introduction to the US equipment and learning the basic anatomy, mannequins and/or cadavers models are suitable to practice MSUS-guided aspiration and injection, and the use of supervised scanning on patients allows the resident to learn the US pathology and to practice documentation. However, there are several challenges when using actual patients in a training curriculum for interventional US: the primary challenge is the higher risk of inaccurate treatment or side effects when a novice performs the procedure, as seen in other medical procedures [92–94]. The secondary challenge is the limited access and time to practice in the clinic under the supervision and the waiting time between relevant procedures for training, which makes the apprenticeship training method less effective [95]. The alternative approach is to practice in a patient-free environment using a mannequin or cadaver. Fresh frozen cadavers maintain sonographic quality and can be manipulated when performing the procedures, and it is possible to access multiple anatomical areas in the same training session [39]. However, it may be more feasible to use a mannequin due to the low cost, the portability and the accessibility, which makes repeated practice possible (i.e. distributed practice). Two randomized studies compared training of landmark-guided aspiration and injection using a cadaver-based program with a mannequin-based program [55, 63]. Michels et al. found that both training on mannequins and cadavers improved skill performance, but training on cadavers revealed better results on the majority of anatomical sites, whereas the study by Berman et al. found no significant difference on residents’ comfort level, however the participants considered the training on cadavers to be the most effective in a survey [63].
In relation to assessment, it is essential to recognize that using subjective confidence level is considered to be obsolete for assessing the effect of educational interventions [87]. Previous studies of self-assessment have shown that self-assessed abilities are uncorrelated with resident performance [77, 96]. Moreover, it would be surprising if the participant did not feel more comfortable after an educational intervention [83, 84]. A substantial body of evidence emphasizes the importance of skill assessment in medical education and procedural training. Numerous studies have proven that assessment drives learning [97, 98]. However, the included studies in this review show that only few use validated assessment tools to measure resident competences.
In order to assess participants theoretical and/or practical competences, development and validation of assessment tools are needed. Three studies included in this review developed assessment tools, however none of the studies examined all five validity sources in the contemporary framework of Messick. The use of outdated frameworks or frameworks where only few sources of validity are examined may result in incorrect trainer decision in relation to assessment of trainees’ competences [99, 100]. Across the other medical specialties using US, two assessment tools have been developed according to the contemporary framework: the Objective Structures Assessment of Ultrasound Skills (OSAUS) [101], which was developed by an expert panel in multiple specialties and are already validated and implemented in several medical fields [17, 102, 103], and the Interventional Ultrasound Skills Evaluation (IUSE) [20], developed by an international expert panel of radiologists. However, the novel IUSE tool has only established content validity and further examination of the other sources of validity needs to be made before implementation. Like Stroud et al. [75], the IUSE-tool also includes assessment of competences in patient communication, which is highly relevant with today’s increased focus on patient-centred treatment. The validation and application of these tools to assess resident competences in MSUS, and MSUS-guided aspiration and injection have yet to be determined.
Because of the aforementioned heterogeneity in the included studies with respect to the study designs, the theoretical and practical part of the curriculum and the assessment methods used, it was not possible to perform a quantitative meta-analysis, which is a limitation of the current review. Another limitation is the poor methodological quality of the majority of included studies. Furthermore, we cannot exclude the possibility of publication bias. It is important to consider that researchers may not desire to publish or distribute studies showing that their intervention (e.g. training program) does not have an effect on outcomes (e.g. resident knowledge, skills or behaviour); moreover negative findings may also be difficult to get published.
In conclusion, we found substantial heterogeneity in the current published studies, most of which were of poor methodological quality and not based on contemporary educational theories. Therefore, there is a need for carefully designed educational studies using validated theoretical and practical assessment tools to define the optimal MSUS training curriculum and competence assessment in rheumatology.
Acknowledgements
We thank Tove Margit Svendsen, research librarian at the Medical Library at Rigshospitalet Denmark, for her assistance with developing the search string.
Funding: No specific funding was received from any bodies in the public, commercial or not-for-profit sectors to carry out the work described in this article.
Disclosure statement: The authors have declared no conflicts of interest.
Data availability statement
All data generated or analysed during this study are included in this article and its supplementary information files, available at Rheumatology online.
Supplementary data
Supplementary data are available at Rheumatology online.
References
- heterogeneity
- cadaver
- education, medical
- internship and residency
- manikins
- rheumatology
- knowledge acquisition
- guidelines
- arthrocentesis
- skills training
- ultrasound examination of musculoskeletal system
- medical residencies
- cochrane collaboration
- assessment scales
- embase
- ultrasound-guided injections
- reviewers
Comments