Back to Projects List
Low-Cost Ultrasound Training
Key Investigators
- David García Mato (Ebatinca S.L., Las Palmas de Gran Canaria, Spain)
- Csaba Pinter (Ebatinca S.L., Las Palmas de Gran Canaria, Spain)
- Rebecca Hisey (Queen’s University, Kingston, ON, Canada)
- Matthew Holden (Carleton University, Ottawa, ON, Canada)
Project Description
Ebatinca S.L. is currently developing a low-cost training platform for ultrasound imaging and ultrasound-guided procedures in low- and middle-income countries. We are developing a 3D Slicer based application to perform training exercises and evaluate participants. The app is called TrainUS and it is available with open-source license.
Some basic features have already been developed: participant/recording management, hardware connection, selection of training exercises,… Currently, we are working on the development of basic exercises to train basic ultrasound skills. The app should be able to evaluate recordings made by users and to provide feedback about their performance.
Objective
Develop specific module to train in-plane needle insertions. The module should display instructions on how to perform the exercise correctly and should enable the recording of ultrasound-guided needle insertions. Then, recordings must be evaluated using the PerkTutor extension, getting overall performance metrics (elapsed time, needle path length, rotations, translations,…). Performance metrics must be displayed in plots and tables to the users, providng specific feedback to users about their performance.
Once integrated, this module could serve as a reference/example for the development of future exercises in TrainUS.
Approach and Plan
- Determine best way to manage recordings of ultrasound images and tracking data. Currently, we are saving the entire sequence browser node as .sqbr files which can then be easily imported into Slicer using the PerkTutor extension.
- Define best methodology to display exercise instructions to users: images vs video
- Discuss the usefulness of assessing performance metric values in “real-time” during sequence playback. Example: distance from needle tip to US plane. Real-time vs overall metrics.
- Integrate exercise into TrainUS app.
- Discuss best strategy to provide specific feedback to users based on recorded data from experts. Deep learning?
Progress and Next Steps
- Saving recordings into a single .sqbr file seems the best option for easy import/export. Current infrastructure enables saving the entire vtkMRMLSequenceBrowserNode into a .sqbr file. Custom app should include updated versions of Slicer and PerkTutor to include the fixes related to the following issues which prevented a correct management of .sqbr files: #6429 and #6435
- It has been proposed that it may be a good idea to show the instructions (images or videos) while the exercise is being performed. This could be integrated as a possible configuration for the exercises. Currently, instructions can be shown as images or videos using slice views integrated into the Slicer layout.
- The new feature to measure and evaluate performance metrics in real-time is considered really useful by the community. This is specially useful to identify specific parts of the recordings where performance drops significantly.
- Feedback for basic skills exercises can be computed with current methodologies integrated into PerkTutor extension. No deep learning is needed unless complexity of the exercise/procedure increases.
- Automatic segmentation (deep learning) could be integrated into specific exercise to compute useful data/metrics from the US image. Example: position of a vessel/mass on the image, position of the needle,…
- Exercise integrated into TrainUS custom app.
Illustrations
Instructions slides displayed to the user before starting the exercise:
Video with instructions displayed to the user before starting the exercise:
Plot showing real-time metric values during recording playback:
Table showing overall performance metrics computed using PerkTutor extension:
Background and References