UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Mixed Reality and Egocentric AI-Assisted Visualisation in Obstetric Ultrasound

Birlo, Manuel; (2025) Mixed Reality and Egocentric AI-Assisted Visualisation in Obstetric Ultrasound. Doctoral thesis (Ph.D), UCL (University College London). Green open access

[thumbnail of Birlo_10216778_Thesis.pdf]
Preview
Text
Birlo_10216778_Thesis.pdf

Download (24MB) | Preview

Abstract

Augmented Reality (AR), particularly in the form of Mixed Reality (MR), holds strong potential to enhance clinical and surgical workflows. However, despite its promise and increasing research interest, demonstrated clinical utility of MR is rare. This thesis investigates specific factors that could contribute to more effective and intuitive AR systems in the future, with the overall goal of supporting real-time AR-assisted decision-making and procedural skill development. First, a systematic literature review of 91 peer-reviewed publications on OST-HMD-based surgical applications was conducted. The review revealed that key technical limitations and human factors remain major barriers to widespread clinical adoption and proposed strategic research directions for improved study design and evaluation. Based on these findings, the focus turned to the underexplored area of AR-assisted training in obstetric ultrasound (US). A novel MR training application (CAL-Tutor) was developed to provide real-time 3D virtual overlay guidance for US probe positioning toward standardised fetal target US planes. A pilot user study demonstrated improved probe navigation and spatial understanding for novel users. Findings from CAL-Tutor led to the development of a new method for egocentric, markerless 3D hand and tool pose estimation. This approach leveraged a scalable synthetic data generation pipeline, combining a generative deep learning model for grasp synthesis with plausible 3D computer graphics-based rendering. The resulting multi-view, multi-modal dataset, HUP-3D, is the first of its kind in the research community. It includes over 31,000 synthetic samples and achieved state-of-the-art single-modality (RGB) accuracy (8.65 mm MPJPE) using the HOPE-Net pose estimation model. A dataset extension (HUP-3D-v2) and initial multi-modal (RGB-D) evaluation finalise the main technical contributions of the thesis. Overall, this thesis contributes to the design of reproducible, scalable and data-driven MR applications for clinical education and interactive procedural training, offering insights for future research at the intersection of MR, egocentric computer vision, and generative deep learning. The findings aim to pave the way for broader adoption of immersive AR- and AI-driven technologies in clinical education and procedural guidance.

Type: Thesis (Doctoral)
Qualification: Ph.D
Title: Mixed Reality and Egocentric AI-Assisted Visualisation in Obstetric Ultrasound
Open access status: An open access version is available from UCL Discovery
Language: English
Additional information: Copyright © The Author 2025. Original content in this thesis is licensed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) Licence (https://creativecommons.org/licenses/by-nc/4.0/). Any third-party copyright material present remains the property of its respective owner(s) and is licensed under its existing terms. Access may initially be restricted at the author’s request.
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science
URI: https://discovery.ucl.ac.uk/id/eprint/10216778
Downloads since deposit
0Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item