Purpose Guidance and quality control in orthopedic surgery increasingly rely on intra-operative fluoroscopy using a mobile C-arm. The accurate acquisition of standardized and anatomy-specific projections is essential in this process. The corresponding iterative positioning of the C-arm is error prone and involves repeated manual acquisitions or even continuous fluoroscopy. To reduce time and radiation exposure for patients and clinical staff and to avoid errors in fracture reduction or implant placement, we aim at guiding—and in the long-run automating—this procedure. Methods In contrast to the state of the art, we tackle this inherently ill-posed problem without requiring patient-individual prior information like preoperative computed tomography (CT) scans, without the need of registration and without requiring additional technical equipment besides the projection images themselves. We propose learning the necessary anatomical hints for efficient C-arm positioning from in silico simulations, leveraging masses of 3D CTs. Specifically, we propose a convolutional neural network regression model that predicts 5 degrees of freedom pose updates directly from a first X-ray image. The method is generalizable to different anatomical regions and standard projections. Results Quantitative and qualitative validation was performed for two clinical applications involving two highly dissimilar anatomies, namely the lumbar spine and the proximal femur. Starting from one initial projection, the mean absolute pose error to the desired standard pose is iteratively reduced across different anatomy-specific standard projections. Acquisitions of both hip joints on 4 cadavers allowed for an evaluation on clinical data, demonstrating that the approach generalizes without retraining. Conclusion Overall, the results suggest the feasibility of an efficient deep learning-based automated positioning procedure, which is trained on simulations. Our proposed 2-stage approach for C-arm positioning significantly improves accuracy on synthetic images. In addition, we demonstrated that learning based on simulations translates to acceptable performance on real X-rays.
During the last two decades we have witnessed great improvements in ICT hardware and software technologies. Three-dimensional content is starting to become commonplace now in many applications. Although for many years 3D technologies have been used in the generation of assets by researchers and experts, nowadays these tools are starting to become commercially available to every citizen. This is especially the case for smartphones, that are powerful enough and sufficiently widespread to perform a huge variety of activities (e.g. paying, calling, communication, photography, navigation, localization, etc.), including just very recently the possibility of running 3D reconstruction pipelines. The REPLICATE project is tackling this particular issue, and it has an ambitious vision to enable ubiquitous 3D creativity via the development of tools for mobile 3D-assets generation on smartphones/tablets. This article presents the REPLICATE project’s concept and some of the ongoing activities, with particular attention being paid to advances made in the first year of work. Thus the article focuses on the system architecture definition, selection of optimal frames for 3D cloud reconstruction, automated generation of sparse and dense point clouds, mesh modelling techniques and post-processing actions. Experiments so far were concentrated on indoor objects and some simple heritage artefacts, however, in the long term we will be targeting a larger variety of scenarios and communities.
Purpose Reduction and osteosynthesis of ankle fractures is a challenging surgical procedure when it comes to the verification of the reduction result. Evaluation is conducted using intra-operative imaging of the injured ankle and depends on the expertise of the surgeon. Studies suggest that intra-individual variance of the ankle bone shape and pose is considerably lower than the inter-individual variance. It stands to reason that the information gain from the healthy contralateral side can help to improve the evaluation. Method In this paper, an assistance system is proposed that provides a side-to-side view of the two ankle joints for visual comparison and instant evaluation using only one 3D C-arm image. Two convolutional neural networks (CNN) are employed to extract the relevant image regions and pose information of each ankle so that they can be aligned with each other. A first U-Net uses a sliding window to predict the location of each ankle. The standard plane estimation is formulated as segmentation problem so that a second U-Net predicts the three viewing planes for alignment. Results Experiments were conducted to assess the accuracy of the individual steps on 218 unilateral ankle datasets as well as the overall performance on 7 bilateral ankle datasets. The experiments on unilateral ankles yield a median position-to-plane error of $$0.73\pm 1.36$$ 0.73 ± 1.36 mm and a median angular error between 2.98$$^\circ $$ ∘ and 3.71$$^\circ $$ ∘ for the plane normals. Conclusion Standard plane estimation via segmentation outperforms direct pose regression. Furthermore, the complete pipeline was evaluated including ankle detection and subsequent plane estimation on bilateral datasets. The proposed pipeline enables a direct contralateral side comparison without additional radiation. This has the potential to ease and improve the intra-operative evaluation for the surgeons in the future and reduce the need for revision surgery.
No abstract
Purpose 3D acquisitions are often acquired to assess the result in orthopedic trauma surgery. With a mobile C-Arm system, these acquisitions can be performed intra-operatively. That reduces the number of required revision surgeries. However, due to the operation room setup, the acquisitions typically cannot be performed such that the acquired volumes are aligned to the anatomical regions. Thus, the multiplanar reconstructed (MPR) planes need to be adjusted manually during the review of the volume. In this paper, we present a detailed study of multi-task learning (MTL) regression networks to estimate the parameters of the MPR planes.Approach First, various mathematical descriptions for rotation, including Euler angle, quaternion, and matrix representation, are revised. Then, three different MTL network architectures based on the PoseNet are compared with a single task learning network.Results Using a matrix description rather than the Euler angle description, the accuracy of the regressed normals improves from 7.7 • to 7.3 • in the mean value for single anatomies. The multi-head approach improves the regression of the plane position from 7.4mm to 6.1mm, while the orientation does not benefit from this approach.Conclusions The results show that a multi-head approach can lead to slightly better results than the individual tasks networks. The most important benefit of the MTL approach is that it is a single network for standard plane regression for all body regions with a reduced number of stored parameters.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.