Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Deep iterative vessel segmentation in OCT angiography

Open Access Open Access

Abstract

This paper addresses retinal vessel segmentation on optical coherence tomography angiography (OCT-A) images of the human retina. Our approach is motivated by the need for high precision image-guided delivery of regenerative therapies in vitreo-retinal surgery. OCT-A visualizes macular vasculature, the main landmark of the surgically targeted area, at a level of detail and spatial extent unattainable by other imaging modalities. Thus, automatic extraction of detailed vessel maps can ultimately inform surgical planning. We address the task of delineation of the Superficial Vascular Plexus in 2D Maximum Intensity Projections (MIP) of OCT-A using convolutional neural networks that iteratively refine the quality of the produced vessel segmentations. We demonstrate that the proposed approach compares favourably to alternative network baselines and graph-based methodologies through extensive experimental analysis, using data collected from 50 subjects, including both individuals that underwent surgery for structural macular abnormalities and healthy subjects. Additionally, we demonstrate generalization to 3D segmentation and narrower field-of-view OCT-A. In the future, the extracted vessel maps will be leveraged for surgical planning and semi-automated intraoperative navigation in vitreo-retinal surgery.

Published by The Optical Society under the terms of the Creative Commons Attribution 4.0 License. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

Full Article  |  PDF Article
More Like This
Automated diagnosis and segmentation of choroidal neovascularization in OCT angiography using deep learning

Jie Wang, Tristan T. Hormel, Liqin Gao, Pengxiao Zang, Yukun Guo, Xiaogang Wang, Steven T. Bailey, and Yali Jia
Biomed. Opt. Express 11(2) 927-944 (2020)

MF-AV-Net: an open-source deep learning network with multimodal fusion options for artery-vein segmentation in OCT angiography

Mansour Abtahi, David Le, Jennifer I. Lim, and Xincheng Yao
Biomed. Opt. Express 13(9) 4870-4888 (2022)

Improving foveal avascular zone segmentation in fluorescein angiograms by leveraging manual vessel labels from public color fundus pictures

Dominik Hofer, Ursula Schmidt-Erfurth, José Ignacio Orlando, Felix Goldbach, Bianca S. Gerendas, and Philipp Seeböck
Biomed. Opt. Express 13(5) 2566-2580 (2022)

Supplementary Material (1)

NameDescription
Visualization 1       This video presents several qualitative results of 3D OCT-A vessel segmentation. We obtain those results using a network trained on 2D maximum intensity projections of OCT-A, which we use to segment every axial slice of the 3D volume, without any tra

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (11)

Fig. 1.
Fig. 1. OCT-A vs Preoperative and Intraoperative Imaging: For several subjects with retinal pathology we present: (a) An Intraoperative Video Frame (IVF) captured during vitreo-retinal surgery (b) IVF affinely aligned to OCT-A (c) a preoperative CFP affinely aligned to OCT-A, (d) the OCT-A and (e) the vessel segmentation obtained by our vessel segmentation method. In all cases, OCT-A visualizes the maximum level of vasculature detail around the surgical RoI, the macula, where both CFPs and IVFs tend to provide blurry information and are susceptible to subretinal pathology, such as choroidal pigmentation (last column), that impacts retinal vessel visibility.
Fig. 2.
Fig. 2. Intraoperative vs OCT-A Vessel Visibility: Vessel map annotations by two expert clinicians on intraperative video frames revealed that in the vicinity of the macula (outlined in red) they are unable to detect the level of vasculature details that can be reliably annotated on OCT-A.
Fig. 3.
Fig. 3. OCT-A data overview: (a) Retina cross section: outlined in blue is the volume corresponding to the slices used in the dataset. They span the space from the retinal surface (upper limit of the blue line) to the start of the choroid (outlined in red) where the Superficial and Deep Vascular Plexuses are located. (b) The imaging device produces geometrically flattened slices that correspond to curved slices of the retina’s cross-section. (c) Maximum Intensity Projection is performed on the extracted stack of geometrically flattened slices along the axis vertical to the plane of the slices. Outlined in red is the (approximate) location of the macula around which the scans are centered. The zoomed-in patch depicts (in green) vessels that are considered by our models and areas (in orange) where microvessels are likely to be located, which however cannot be delineated reliably and the models learn to ignore. (d) The imaging device locates the limiting surface between the retinal layers (blue space) and the choroid (red space) allowing us to access geometrically flattened slices thus separating chroroidal and retinal layers.
Fig. 4.
Fig. 4. Schematic representation of the architecture of the base-network as described in Sec. 2.3. The base-network follows the architecture paradigm of UNet. The number below any tensor denotes the number of feature maps at that stage of the network.
Fig. 5.
Fig. 5. Considered CNN architectures: (a) UNet, (b) SHN, and (c) iUNet. For illustration purposes the SHN, and iUNet, are presented with $2$ distinct UNet modules, and $2$ iterations, respectively.
Fig. 6.
Fig. 6. For all models, adding online data augmentation during training (described in 3.4) prevents overfitting by regularizing training while leading to higher validation Quality. The presented curves are computed when training and validating on the same cross-validation fold of the dataset, but this finding was consistent across folds.
Fig. 7.
Fig. 7. Qualitative comparison of results from Imagenet pretrained and fine-tuned baseline DRIU [25] and i-unet-4-topo, trained from scratch. The latter was the top performing model/loss function combination. The two methods achieve similar recall. However, DRIU exhibits noisier predictions with a considerable amount of false positives. Columns $4$ and $6$ present centerline errors (dilated by one pixel to improve visibility) made by the two models, with false and true positives shown in red and green respectively, while missed segments are shown in blue.
Fig. 8.
Fig. 8. Adding iterative refinement to unet-topo: outlined in red are some examples of fine details that are recovered only by i-unet-4-topo. The outermost column depicts zoomed-in regions of interest corresponding to the red bounding boxes, and aids with the comparison of the response of the two models. Columns $3$ and $5$ present centerline errors (dilated by one pixel to improve visibility) made by the two models, with false and true positives shown in red and green respectively, while missed segments are shown in blue.
Fig. 9.
Fig. 9. OCT-A $3$ D segmentation: The 1st row depicts the MIP associated with the raw $3$ D volume which is per-slice segmented by shn-4-topo (the model that gave the best, based on visual inspection, $3$ D results), with the resulting $3$ D segmentation displayed below. The 3rd row displays cross-sections of the segmentation (gray) overlayed on the OCT-A cross-section, the location of which is denoted by the red dashed line. Finally zoomed in cross-sectional details are shown (denoted in upper rows by red dots) which reveal the network mistakenly segments shadowing artefacts (1st, 2nd, 4th columns) below bigger vessels which is normal due to it being unaware of $3$ D context. A video demonstration of the $3$ D segmentations is provided as supplementary material.
Fig. 10.
Fig. 10. Generalizing to $3\,$ mm $\times 3$  mm scans: Using i-unet-4-topo, we can produce plausible segmentations of the narrower FoV scans which reveal more details of the central part of the macula. The 1st and 2nd, and 3rd and 4th columns, demonstrate the correspondence between the two scans and the two segmentations respectively, while the 5th and 6th presents the ground truth centerline and errors with respect to it respectively, with false (red), true (green) positives and missed segments (blue).
Fig. 11.
Fig. 11. Performance as a function of the number of trainable parameters: i-unet-4-topo constitutes the top performing while requiring the minimum number of parameters.

Tables (8)

Tables Icon

Table 1. Distribution of types of pathology in the dataset

Tables Icon

Table 2. Model/loss-function comparisons using 4 -folds cross validation. Mean of metrics on the test set across folds is reported and standard deviation is in parenthesis. Best of each metric in bold, Statistical Significance of difference in Quality between the two top competing methods is indicated.

Tables Icon

Table 3. Effect of iterations (iUNet) and modules (SHN) on quality. Statistically significant differences between top performing model with iterative refinement and top performing model with iterative refinement and topological loss. Best of each model is in bold.

Tables Icon

Table 4. Effect of base network depth on quality. Statistically significant differences between deeper and shallower base networks are indicated. Best of each model is in bold.

Tables Icon

Table 5. Improvements through iterative refinement combined with topological loss.

Tables Icon

Table 6. Quality metric when fractions of the full dataset are considered.

Tables Icon

Table 7. Paired Wilcoxon significance tests for cross validated Quality metric.

Tables Icon

Table 8. Ablation study for the VGG-feature loss on validation sets.

Equations (8)

Equations on this page are rendered with MathJax. Learn more.

L b c e = β i Y + log ( P ( y i = 1 X ; θ ) ( 1 β ) i Y log ( P ( y i = 0 X ; θ ) ,
L t o p o = n = 1 N μ n W n H n C n c = 1 C n F n c ( f ( X ; θ ) ) F n c ( Y ) 2 2 ,
L c o m b = L b c e + L t o p o
L i U n e t = 2 T ( T + 1 ) t = 1 T t L c o m b ( t ) ,
Completeness = μ Y ^ ( Y , τ ) Y ,
Correctness = μ Y ( Y ^ , τ ) Y ^ ,
Quality = μ Y ( Y ^ , τ ) Y ^ μ Y ^ ( Y , τ ) + Y .
p i j = d i j p i l o g ( p i ) + p j ( 1 l o g ( p j ) ) p i p j ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.