406 research outputs found
PViT-6D: Overclocking Vision Transformers for 6D Pose Estimation with Confidence-Level Prediction and Pose Tokens
In the current state of 6D pose estimation, top-performing techniques depend
on complex intermediate correspondences, specialized architectures, and
non-end-to-end algorithms. In contrast, our research reframes the problem as a
straightforward regression task by exploring the capabilities of Vision
Transformers for direct 6D pose estimation through a tailored use of
classification tokens. We also introduce a simple method for determining pose
confidence, which can be readily integrated into most 6D pose estimation
frameworks. This involves modifying the transformer architecture by decreasing
the number of query elements based on the network's assessment of the scene
complexity. Our method that we call Pose Vision Transformer or PViT-6D provides
the benefits of simple implementation and being end-to-end learnable while
outperforming current state-of-the-art methods by +0.3% ADD(-S) on
Linemod-Occlusion and +2.7% ADD(-S) on the YCB-V dataset. Moreover, our method
enhances both the model's interpretability and the reliability of its
performance during inference
An image-based method to synchronize cone-beam CT and optical surface tracking
open5siThe integration of in-room X-ray imaging and optical surface tracking has gained increasing importance in the field of image guided radiotherapy (IGRT). An essential step for this integration consists of temporally synchronizing the acquisition of X-ray projections and surface data. We present an image-based method for the synchronization of cone-beam computed tomography (CBCT) and optical surface systems, which does not require the use of additional hardware. The method is based on optically tracking the motion of a component of the CBCT/gantry unit, which rotates during the acquisition of the CBCT scan. A calibration procedure was implemented to relate the position of the rotating component identified by the optical system with the time elapsed since the beginning of the CBCT scan, thus obtaining the temporal correspondence between the acquisition of X-ray projections and surface data. The accuracy of the proposed synchronization method was evaluated on a motorized moving phantom, performing eight simultaneous acquisitions with an Elekta Synergy CBCT machine and the AlignRT optical device. The median time difference between the sinusoidal peaks of phantom motion signals extracted from the synchronized CBCT and AlignRT systems ranged between -3.1 and 12.9 msec, with a maximum interquartile range of 14.4 msec. The method was also applied to clinical data acquired from seven lung cancer patients, demonstrating the potential of the proposed approach in estimating the individual and daily variations in respiratory parameters and motion correlation of internal and external structures. The presented synchronization method can be particularly useful for tumor tracking applications in extracranial radiation treatments, especially in the field of patient-specific breathing models, based on the correlation between internal tumor motion and external surface surrogates.Fassi, Aurora; Schaerer, Joël; Riboldi, Marco; Sarrut, David; Baroni, GuidoFassi, Aurora; Schaerer, Joël; Riboldi, Marco; Sarrut, David; Baroni, Guid
Contactless Sensing of Water Properties for Smart Monitoring of Pipelines
A key milestone for the pervasive diffusion of wireless sensing nodes for smart monitoring of water quality and quantity in distribution networks is the simplification of the installation of sensors. To address this aspect, we demonstrate how two basic contactless sensors, such as piezoelectric transducers and strip electrodes (in a longitudinal interdigitated configuration to sense impedance inside and outside of the pipe with potential for impedimetric leak detection), can be easily clamped on plastic pipes to enable the measurement of multiple parameters without contact with the fluid and, thus, preserving the integrity of the pipe. Here we report the measurement of water flow rate (up to 24 m(3)/s) and temperature with ultrasounds and of the pipe filling fraction (capacitance at 1 MHz with similar to cm(3) resolution) and ionic conductivity (resistance at 20 MHz from 700 to 1400 mu S/cm) by means of impedance. The equivalent impedance model of the sensor is discussed in detail. Numerical finite-element simulations, carried out to optimize the sensing parameters such as the sensing frequency, confirm the lumped models and are matched by experimental results. In fact, a 6 m long, 30 L demonstration hydraulic loop was built to validate the sensors in realistic conditions (water speed of 1 m/s) monitoring a pipe segment of 0.45 m length and 90 mm diameter (one of the largest ever reported in the literature). Tradeoffs in sensors accuracy, deployment, and fabrication, for instance, adopting single-sided flexible PCBs as electrodes protected by Kapton on the external side and experimentally validated, are discussed as well
First experimental verification of prompt gamma imaging with carbon ion irradiation
: Prompt Gamma Imaging (PGI) is a promising technique for range verification in Particle Therapy. This technique was already tested in clinical environment with a knife-edge-collimator camera for proton treatments but remains relatively unexplored for Carbon Ion Radiation Therapy (CIRT). Previous FLUKA simulations suggested that PG profile shifts could be detected in CIRT with a precision of ∼ 4 mm ([Formula: see text]) for a particle statistic equal to [Formula: see text] C-ions using a 10 × 10 cm2 camera. An experimental campaign was carried out at CNAO (Pavia, Italy) to verify these results, using a knife-edge-collimator camera prototype based on a 5 × 5 cm2 pixelated LYSO crystal. PG profiles were measured irradiating a plastic phantom with a C-ion pencil beam at clinical energies and intensities, also moving the detector to extend the FOV to 13 × 5 cm2. The prototype detected Bragg-peak shifts with ∼ 4 mm precision for a statistic of [Formula: see text] C-ions ([Formula: see text] for the extended FOV), slightly larger than expected. Nevertheless, the detector demonstrated significant potential for verifying the precision in dose delivery following a treatment fraction, which remains fundamental in the clinical environment. For the first time to our knowledge, range verification based on PGI was applied to a C-ion beam at clinical energy and intensities
Intra‐frame motion deterioration effects and deep‐learning‐based compensation in MR‐guided radiotherapy
Background
Current commercially available hybrid magnetic resonance linear accelerators (MR-Linac) use 2D+t cine MR imaging to provide intra-fractional motion monitoring. However, given the limited temporal resolution of cine MR imaging, target intra-frame motion deterioration effects, resulting in effective time latency and motion artifacts in the image domain, can be appreciable, especially in the case of fast breathing.
Purpose
The aim of this work is to investigate intra-frame motion deterioration effects in MR-guided radiotherapy (MRgRT) by simulating the motion-corrupted image acquisition, and to explore the feasibility of deep-learning-based compensation approaches, relying on the intra-frame motion information which is spatially and temporally encoded in the raw data (k-space).
Methods
An intra-frame motion model was defined to simulate motion-corrupted MR images, with 4D anthropomorphic digital phantoms being exploited to provide ground truth 2D+t cine MR sequences. A total number of 10 digital phantoms were generated for lung cancer patients, with randomly selected eight patients for training or validation and the remaining two for testing. The simulation code served as the data generator, and a dedicated motion pattern perturbation scheme was proposed to build the intra-frame motion database, where three degrees of freedom were designed to guarantee the diversity of intra-frame motion trajectories, enabling a thorough exploration in the domain of the potential anatomical structure positions. U-Nets with three types of loss functions: L1 or L2 loss defined in image or Fourier domain, referred to as NNImgLoss-L1, NNFloss-L1 and NNL2-Loss were trained to extract information from the motion-corrupted image and used to estimate the ground truth final-position image, corresponding to the end of the acquisition. Images before and after compensation were evaluated in terms of (i) image mean-squared error (MSE) and mean absolute error (MAE), and (ii) accuracy of gross tumor volume (GTV) contouring, based on optical-flow image registration.
Results
Image degradation caused by intra-frame motion was observed: for a linearly and fully acquired Cartesian readout k-space trajectory, intra-frame motion resulted in an imaging latency of approximately 50% of the acquisition time; in comparison, the motion artifacts exhibited only a negligible contribution to the overall geometric errors. All three compensation models led to a decrease in image MSE/MAE and GTV position offset compared to the motion-corrupted image. In the investigated testing dataset for GTV contouring, the average dice similarity coefficients (DSC) improved from 88% to 96%, and the 95th percentile Hausdorff distance (HD95) dropped from 4.8 mm to 2.1 mm. Different models showed slight performance variations across different intra-frame motion amplitude categories: NNImgLoss-L1 excelled for small/medium amplitudes, whereas NNFloss-L1 demonstrated higher DSC median values at larger amplitudes. The saliency maps of the motion-corrupted image highlighted the major contribution of the later acquired k-space data, as well as the edges of the moving anatomical structures at their final positions, during the model inference stage.
Conclusions
Our results demonstrate the deep-learning-based approaches have the potential to compensate for intra-frame motion by utilizing the later acquired data to drive the convergence of the earlier acquired k-space components
- …
