Skip to main content

Original Research

Exp. Biol. Med., 23 October 2024
Sec. Bioimaging

Quantitative characterization of retinal features in translated OCTA

Rashadul Hasan BadhonRashadul Hasan Badhon1Atalie Carina ThompsonAtalie Carina Thompson2Jennifer I. LimJennifer I. Lim3Theodore LengTheodore Leng4Minhaj Nur Alam
Minhaj Nur Alam1*
  • 1Department of Electrical and Computer Engineering, University of North Carolina at Charlotte, Charlotte, NC, United States
  • 2Department of Surgical Ophthalmology, Atrium-Health Wake Forest Baptist, Winston-Salem, NC, United States
  • 3Department of Ophthalmology and Visual Science, University of Illinois at Chicago, Chicago, IL, United States
  • 4Department of Ophthalmology, Stanford University School of Medicine, Stanford, CA, United States

Abstract
This study explores the feasibility of quantitative Optical Coherence Tomography Angiography (OCTA) features translated from OCT using generative machine learning (ML) for characterizing vascular changes in retina. A generative adversarial network framework was employed alongside a 2D vascular segmentation and a 2D OCTA image translation model, trained on the OCT-500 public dataset and validated with data from the University of Illinois at Chicago (UIC) retina clinic. Datasets are categorized by scanning range (Field of view) and disease status. Validation involved quality and quantitative metrics, comparing translated OCTA (TR-OCTA) with ground truth OCTAs (GT-OCTA) to assess the feasibility for objective disease diagnosis. In our study, TR-OCTAs showed high image quality in both 3 and 6 mm datasets (high-resolution and contrast quality, moderate structural similarity compared to GT-OCTAs). Vascular features like tortuosity and vessel perimeter index exhibits more consistent trends compared to density features which are affected by local vascular distortions. For the validation dataset (UIC), the metrics show similar trend with a slightly decreased performance since the model training was blind on UIC data, to evaluate inference performance. Overall, this study presents a promising solution to the limitations of OCTA adoption in clinical practice by using vascular features from TR-OCTA for disease detection. By making detailed vascular imaging more widely accessible and reducing reliance on expensive OCTA equipment, this research has the potential to significantly enhance the diagnostic process for retinal diseases.

Impact statement

This study represents a significant advancement in retinal imaging by demonstrating the feasibility of using generative machine learning to translate OCT features into OCTA features, addressing a critical gap in clinical practice. By employing a generative adversarial network framework trained on diverse datasets, the research establishes quantitative features in Translated OCTA. This innovation enhances the ability to objectively diagnose retinal diseases by providing reliable vascular imaging without the need for costly OCTA equipment. The findings reveal that vascular features from TR-OCTA, such as tortuosity and vessel perimeter index, offer more consistent diagnostic trends compared to traditional density features. This new information has the potential to transform retinal disease diagnostics, making detailed vascular imaging more accessible and cost-effective, thereby improving patient outcomes and broadening the adoption of advanced imaging techniques in routine clinical settings.

Introduction

Optical Coherence Tomography (OCT) is a cutting-edge medical imaging technology that has revolutionized our ability to observe and comprehend the complex structures of biological tissues. It is non-invasive and capable of providing highly detailed in-depth retinal pathologies. It generates high-resolution cross-sectional images of tissues using low-coherence light, therefore has been widely adopted in ophthalmic clinical care [1]. As a result, OCT has been demonstrated for early identification and monitoring of various retinal illnesses including diabetic retinopathy (DR), age-related macular degeneration (AMD) and glaucoma that cannot be obtained by any other non-invasive diagnostic technique [28].

The rapid development of OCT, growing interest in this field, and its increasing impact in clinical medicine has contributed to its widespread availability. However, due to its non-dynamic imaging technology, conventional OCT cannot visualize blood flow information such as blood vessel caliber or density and remains only limited to capturing structural information [2, 9]. As a result of this information gap, OCT angiography (OCTA) was developed which can produce volumetric data from choroidal and retinal layers and provide both structural and blood flow information [10, 11]. OCTA provides a high-resolution image of the retinal vasculature at the capillary level, allowing for reliable detection of microvascular anomalies in diabetic eyes and vascular occlusions. It helps to quantify vascular impairment based on the severity of retinal diseases. In recent years, OCTA has been demonstrated to identify, detect, and predict DR [1219], AMD [2022], Glaucoma [23] and several other retinal diseases [2431]. Despite the advantages, widespread deployment of OCTA has been limited due to the high device cost [32, 33]. The additional requirements of hardware and software for an OCTA device pose a financial burden for clinics as well as patients This is one of the major reasons that only a limited number of hospitals and retinal clinics use OCTA for in-depth retinal vascular analysis. Another limitation of OCTA is the process of generating an OCTA scan, which takes longer time and involves repetitive scanning of the retina making the data acquisition harder due to involuntary eye movements and motion artifacts, reducing the quality of OCTA images [33]. Due to the limitation of OCTA data, most studies involving OCTA based imaging biomarkers and involving the use of artificial intelligence (AI) are difficult to validate extensively for future clinical deployment.

From literature, a potential solution to this problem can be the utilization of AI and machine learning (ML) to produce OCTA images from the already available OCT data which has been showing promising outcomes [3439]. Incorporating ML for OCTA translation from OCT offers significant advances in ophthalmic diagnostics by increasing angiographic and functional information in existing OCT data. This transition harnesses ML’s capability to autonomously analyse OCT scans and generate detailed vascular images, traditionally obtained through OCTA, aligned with OCT information. By doing so, it substantially lowers the barriers to accessing high-resolution vascular imaging, which is crucial for diagnosing and monitoring retinal diseases and provides a robust detection system. Furthermore, ML dependent approaches alleviate some of OCTA’s limitations, including its high cost, susceptibility to artifacts from patient movement and the extensive time required for image acquisition.

Different studies have been reported [4042] attempting to leverage ML algorithms for generative-adversarial learning, typically utilizing a UNet for image translation in recent years. However the quality of the translated OCTA (TR-OCTA) is usually sub-optimal and the retinal vascular areas are not refined enough. The first application of this approach was reported by Lee et al., 2019 [34] to train an algorithm to generate retinal flow maps from OCT images avoiding the needs for labelling but it was limited to capture higher density of deep capillary networks. According to some recent studies [3537], incorporating textual information or surrounding pixels, it is possible to improve the OCTA image quality. Le et. Al [39] proposed another approach incorporating spatial speckle variance and generative AI, however, it requires OCT/OCTA data from custom devices. In this paper, we adopt and implement a generative-adversarial learning framework-based algorithm demonstrated by Li et al [36] for translating OCT data into OCTA. The focus of this study is to demonstrate the feasibility of using such TR-OCTA image generated vascular features (Blood Vessel Density (BVD), Blood Vessel Caliber (BVC), Blood Vessel Tortuosity (BVT), Vessel Perimeter Index (VPI)) for disease detection. We compare these OCTA features with ground truth (GT) – OCTAs. The quality of the TR-OCTAs were compared with features such as Structural Similarity Index Measure (SSIM), Fréchet Inception Distance (FID) and patch-based contrast quality index (PCQI). From our observation and statistical analysis, we found that overall, the SSIM values indicate a moderate level of structural similarity between TR-OCTA and GT-OCTA images, with some variability across different patient categories and scan range however PCQI scores are quite close for both dataset and some deviation in FID scores is noticeable. It was observed that the model generally achieved a slightly better performance in depicting normal and pathological retinal features for the 3 mm scans compared to the 6 mm. However, across both field of view (FoV), there were slight discrepancies in quantitative vascular metrics such as BVD, BVC and VPI, highlighting areas where the translation model could be further refined. This analysis underscores the potential of using AI-driven translation models for OCTA image analysis, while also pointing to the need for improvements to enhance the accuracy of vascular feature representation, particularly at varying FoV.

Materials and methods

The overall methodology of our feature extraction pipeline is demonstrated in Figure 1. We first translate OCT data into OCTA (using algorithm demonstrated by Li et. al [36]) and quantify the retinal features in both GT and TR-OCTAs for validation.

Figure 1
www.frontiersin.org

Figure 1. Framework of OCT to OCTA translation and characterization of quantitative features.

Translation algorithm

We adopted and implemented the OCT to OCTA translation algorithm from Li et al [36]. We describe the process here briefly. The process of OCTA translation from OCT images is carried out in 3 steps (Figure 1): (a) generating 3D OCTA volumes from paired 3D OCT volumes using conditional generative adversarial network (GAN), (b) improving image quality by focusing only the vascular regions, utilizing the 2DVSeg model, thorough vascular segmentation, (c) preserving contextual information for better quality translated images through a 2D translation model (2DTR) generating 2D paired OCTA maps. The baseline architecture of the translation model is built upon pix2pix, an image translation model [41]. The aim of the model is primarily to translate OCT volumes to its paired OCTA volume as closely as possible to the original clinical images. [4] The framework includes a 3D GAN where the 3D generator takes a 3D OCT volume as its input and outputs a corresponding TR-OCTA volume. a 3D discriminator is used to effectively distinguish between the original (ground-truth) OCTA volumes and the generated ones. An adversarial loss is used to train both the generator and discriminator. Furthermore, to calculate for each pixel difference between TR-OCTA and GT-OCTA, a distance loss is considered. The framework also uses a 2D vascular segmentation model (Figure 2A) to help with the improved quality of the vascular regions by utilizing OCTA reflected vascular data by focusing on the vascular areas during the 3D volume translation process.

Figure 2
www.frontiersin.org

Figure 2. (A) 2D vascular segmentation model, (B) 2D Translation model.

This model also utilizes a 2D generative translation model (Figure 2B) to build heuristic (suboptimal) 2D OCTA projection maps from their corresponding OCT that can provide heuristic contextual information where output values are affected by the surrounding pixels resulting in outputs with additional contextual information.

Comparative feature analysis

The generated TR-OCTA maps were compared on several quantitative features to the GT projection maps for comparison: BVD, BVC, BVT, and VPI. Also, for qualitative comparison: SSIM, FID and PCQI metrics were used to quantify the translated image quality and similarity to GT OCTA maps. All the metrics evaluation were performed using MATLAB and Python. Feature values were calculated separately for 3 mm and 6 mm across different patient groups and compared between the OCT500 and UIC datasets. A two-tail t-test was carried out for each feature to compare if there is a significant difference between the TR-OCTA and GT-OCTA values with a p value <0.05.

Metrics and features

Similarity metrics

We used three metrics to compare GT and TR-OCTAs, as described below:

SSIM: SSIM or Structural Similarity Index Measure, is a method for measuring the similarity between two images. SSIM is based on the perception of the human visual system and it considers changes in structural information, luminance and contrast. The idea is that pixels have strong inter-dependencies, especially when they are spatially close. These dependencies carry important information about the structure of the objects in the visual scene.

FID: FID score is a metric used to evaluate the quality of images generated by models, such as those produced by GANs. It measures the similarity between two sets of images, typically between a set of generated images and a set of real images, by comparing the statistics of their features extracted by a pre-trained Inception model [43]. The FID score calculates the distance between the feature vectors of the real and generated images, using the Fréchet distance (also known as the Wasserstein-2 distance). A lower FID score indicates that the distribution of the generated images is closer to the distribution of the real images, suggesting higher quality and more realistic images.

PCQI: PCQI is another metric designed to assess the quality of images by focusing on local contrast changes, which are crucial for visual perception, especially in textured regions [44]. Unlike many traditional image quality metrics that evaluate images globally, PCQI operates on small, localized patches of an image, making it particularly effective at capturing and evaluating detailed contrast differences between a reference image and a test image. PCQI calculates the quality score based on three main aspects: patch similarity, contrast distortion, and mean luminance change, within these localized regions. The final score is a weighted sum of these aspects, providing a single quality metric that reflects how perceptually close the test image is to the reference image in terms of local contrast and brightness. A higher PCQI score indicates a better match between the test and reference images, suggesting less contrast distortion and more accurate reproduction of the original image’s visual quality.

Quantitative OCTA features

We characterized three vessel and one density based features (Equations 14), as described below:

BVD: BVD or vessel area density (VD) [45], is the ratio of the blood vessels to the total area measured and can be utilized for identifying early detection of retinal pathologies including DR [46, 47], AMD [48, 49] etc.

BVD=vascularareatotalarea(1)

BVC: BVC, also named as vessel diameter index [50], is calculated as the ratio of vessel area to the vessel length [12]. BVC distortion can be used to quantify retinal vascular shrinkage and is typically observed in different retinopathies such as diabetic retinopathy (DR) [51].

BVC=vascularareavascularlength(2)

BVT: BVT is defined as a measure of degree of vessel distortion [26, 52]. During any retinal pathologies, distorted vessel structures can affect the blood flow efficiency and can be measured as:

BVT=1ngeodesicdistancebetweenendpointsforavesselbrancheuclideandistancebetweenendpointsforavesselbranch(3)

here, n = total number of vessel branches.

VPI: VPI [52] is measured as the ratio of the contour length of the vessel boundaries or vessel perimeter to the total vessel area and has been used for detection of DR and sickle cell retinopathy (SCR) from.

OCTA images:

VPI=overallcontourlengthofbloodvesselboundariestotalbloodvesselarea(4)

Statistical Analysis: We performed statistical analysis based on the selected features to quantify the TR-OCTA and measure the quality of the translation. This analysis will help us improve the accuracy and efficiency of the TR-OCTA translated from GT-OCT and GT-OCTA.

Results

Dataset

We used 2 datasets for our study, a public dataset of 500 patients containing paired 3D OCT and OCTA volumes, OCTA-500 [53] and a dataset of DR patients collected from UIC with 445 scans containing OCT volumes and OCTA projections.

OCT500 dataset is divided into 2 subsets according to their FoV (Field of view), 3 mm and 6 mm. The translation algorithm is applied separately to the two subsets for comparison. The datasets are further divided into different diseased patients and normal patients for quantitative feature comparison. The 3 mm dataset contains 6 AMD patients, 5 Choroidal neovascularization (CNV) patients, 29 DR and 160 Normal patients who are compared statistically after evaluating the feature values. The 3 mm set contains paired OCT and OCTA volumes from 200 patients with a field of view (FOV) 3 mm × 2 mm × 3 mm. Each volume has 304 slices with a size of 640px × 304px. The generated projection map is of 256px × 256px size. The whole dataset is divided into a 70-25-5% split: 140, 10, and 50 volumes for training, validation and test sets respectively. Similarly, the 6 mm set contains paired OCT and OCTA volumes from 300 patients with FOV of 6 mm × 2 mm × 6 mm. Each volume is of size 640px × 400px, containing 400 slices and generated projection maps are of size 256px × 256px. Similar to 3 mm set: 180, 20, and 100 volumes are split as training, validation and test sets. The 6 mm dataset contains 43 AMD, 11 CNV, 14 Central serous chorioretinopathy (CSC), 35 DR, 10 Retinal vein occlusion (RVO), 91 Normal and 96 other retinal pathology-affected patients for which a similar statistical evaluation is carried out and feature values are calculated.

UIC data and image acquisition

The UIC study was approved by the institutional review board of the University of Illinois at Chicago and was in compliance with the ethical standards stated in the Declaration of Helsinki. The patients with DR were recruited from UIC Retinal Clinic. We performed a retrospective study of consecutive diabetic patients (Type II) who underwent OCTA and OCT imaging. The patients are thus representative of a university population of diabetic patients who require imaging for management of diabetic macular edema and DR. OCT/OCTA images of both eyes of every patient were collected. We excluded subjects with macular edema, previous vitreous surgery, and history of other eye diseases. All patients had undergone a complete anterior and dilated posterior segment examination (J.I.L.). The patients were classified by severity of DR (mild, moderate, and severe) according to the Early Treatment Diabetic Retinopathy Study staging system. The grading was done by retina specialist who used a slit-lamp fundus lens for the diagnosis. OCT/OCTA data were acquired using an ANGIOVUE spectral domain OCTA system (Optovue, Fremont, CA), with a 70-kHz A-scan rate, an axial an axial resolution of 5 μm, and a lateral resolution of 15 µm. All the OCTA images were macular scans and had field of view of 6 mm. We exported the OCTA images from the software ReVue (Optovue) and used custom-developed software in Python OpenCV for further image analysis, feature extraction, and image classification.

The UIC dataset contains 445 OCT scans from 41 patients with different DR conditions: control, mild, moderate and severe. The scans were selected based on signal strength Q ≥ 5 for this study. Similar to OCT500, this dataset has both 3 mm (187 scans) and 6 mm (258 scans) scans for different stages of DR: Control, Mild, Moderate and Severe. For 3 mm FOV, we used 35 scans for Control group, 118 for Mild, 37 for Moderate and 97 for Severe. On the other hand, for 6 mm, the set included 59 for Control, 143 Mild, 69 Moderate and 123 Severe scans for comparison. 3 mm slices are of size 640px × 304px and 6 mm slices are mostly of 640px × 400px with some mixed 640px × 304px scans which are used to generate 256px × 256px OCTA slices. Some patients were listed in multiple categories therefore, scans of those patients were included in multiple categories before feature evaluation.

Comparative analysis of similarity and OCTA features

The algorithm for TR-OCTA generation was exclusively trained only on OCT500 and tested on the UIC dataset for feature quantification and comparison. Figures 3A–H depicts GT-OCTA and generated TR-OCTA images at 3 mm and 6 mm scan range from diseased as well as normal patients for OCT500 while Figures 3I–P represent the images for UIC dataset across various patient categories.

Figure 3
www.frontiersin.org

Figure 3. GT and TR-OCTA images for NORMAL (A, E), DR (B, F), CNV (C, G) and AMD (D, H) patients from OCT500. (A–D) images are of 3 mm and (E–H) are of 6 mm. Similarly (I–P) are OCTA images from UIC dataset. (I, M) from Control, (J, N) from Mild, (K, O) from Moderate and (L, P) from Severe group for 3 mm and 6 mm respectively.

OCT500 3 mm dataset has AMD (6), CNV (5), DR (29) and NORMAL (160) patients totalling 200 patients, while 6 mm dataset includes AMD (43), CNV (11), CSC (14), DR (35), RVO (10), OTHER retinopathies (96) and NORMAL (91) patients totalling 300. Two-tail T-tests were carried out (p < 0.05) between GT and TR-OCTAs for BVD, BVC, BVT and VPI (3 mm complete dataset). The results indicate that BVD and BVC have p-values close to 0.5, suggesting that there is no significant difference between these features in the GT and TR-OCTA values (Table 1). This means that these features can be used for effective disease classification. Upon evaluating quality metrics for the datasets (Tables 2, 3), mean SSIM for 3 mm was found to be slightly higher (0.48) than 6 mm dataset showing SSIM ranging from 0.16–0.52 with a mean of 0.42. SSIM values were also calculated for different patient statuses in both datasets. For the 3 mm dataset: AMD patients show a slightly lower mean SSIM, DR dataset on the other hand reveals a higher mean SSIM compared to other patient groups. However, for 6 mm: AMD, CNV, CSC, patients with other retinopathies and Normal group showed a close SSIM mean value except for DR patients with a slightly higher mean SSIM (0.43) and RVO patients, a lower mean SSIM of 0.36 (Table 2). Furthermore, The FID score for the 3 mm dataset was lower (35.88) compared to the 6 mm dataset, which had a higher FID score of 49.06. On the other hand, PCQI scores were comparably high for both datasets with the 3 mm set slightly outperforming the 6 mm (Table 3). All these feature values were also calculated for the complete dataset and separately for different diseased and normal patients for comparative analysis (Table 4). BVD and BVT values from 3 mm show some trend among Normal and AMD, CNV, DR groups which is mimicked by the TR-OCTA (Figures 4A–D). Overall, TR-BVC, TR-VPI, TR-BVT, and TR-BVD values (Figure 5) are concentrated within a specific range and closer to the GT values for each feature respectively. For BVD, some outliers are further away from the lowest value of the BVD range which is consistent with the findings from the categorized distribution of feature values (Supplementary Figures S1A–D).

Table 1
www.frontiersin.org

Table 1. Two-tail t-test between GT-OCTA and TR-OCTA for OCT500 and UIC datasets.

Table 2
www.frontiersin.org

Table 2. SSIM values for 3 mm and 6 mm from both OCT500 and UIC.

Table 3
www.frontiersin.org

Table 3. FID and PCQI scores for the complete datasets of OCT500 and UIC.

Table 4
www.frontiersin.org

Table 4. Statistical analysis of TR-OCTA compared to GT-OCTA for 3 mm dataset from OCT500.

Figure 4
www.frontiersin.org

Figure 4. Feature values for different patient groups from both OCT500 and UIC. (A–D) show feature trend of 3 mm for OCT500, (E–H) for OCT500 6 mm, (I–L) UIC 3 mm and (M–P) UIC 6 mm dataset.

Figure 5
www.frontiersin.org

Figure 5. Feature value distribution of 3 mm and 6 mm scans for OCT500. (A–D) are BVC, VPI, BVT and BVD values for 3 mm. Similarly, (E–H) show values for 6 mm.

Similarly, for the complete 6 mm dataset we performed T-tests (p < .05) for BVD, BVC, VPI and BVT but only BVD was found to have statistically similar values for both TR-OCTA and GT-OCTA images (Table 1). The 6 mm dataset contained central serous chorioretinopathy (CSC), retinal vein occlusion (RVO) and other retinal pathologies that were absent in the 3 mm dataset (Figures 4E–H). In a comparative analysis (Table 5), BVD for RVO and BVT for AMD patients shows a larger deviation compared to other cases when calculated. Also, BVD and BVT values for translated image follow the trend set by the GT-OCTA. However, BVC, BVT, and VPI were measured having closer values in all cases. When plotted, the distribution of the feature values for TR-OCTA and GT-OCTA, similar to the 3 mm dataset, BVC, VPI and BVD show more outliers compared to BVT and the distribution is similar to the 3 mm dataset. Supplementary Figures S1E–K exhibits feature value distribution for AMD, CNV, CSC, DR, RVO, other retinal pathologies and normal patients and a similar trend of BVD feature having more outliers is observed for diseased as well as normal patients in comparison to other features except RVO.

Table 5
www.frontiersin.org

Table 5. Statistical analysis of TR-OCTA compared to GT-OCTA for 6 mm dataset from OCT500.

For UIC dataset Table 2, 3 summarize the qualitative metrics and Tables 6, 7 summarize quantitative feature values for both 3 mm and 6 mm presenting the validity of implementation of automated image-to-image translation. The whole 3 mm dataset show a mean SSIM value of 0.2808 however Control group as well as other DR stages show slight deviation in terms of mean SSIM although their value range stays similar. For the complete 6 mm, SSIM was slightly higher (0.2952) and contrary to the 3 mm set, Control, Mild and Moderate groups show closer values to each other except for Severe with a slight decrease in value: 0.2679. As a quality metrics, FID scores show higher value than OCT500 for both 3 mm and 6 mm: 150.34 and 107.74 respectively however PCQI scores were closer to the ideal value for both 3 mm and 6 mm.

Table 6
www.frontiersin.org

Table 6. Statistical analysis of TR-OCTA compared to GT-OCTA for 3 mm dataset from UIC.

Table 7
www.frontiersin.org

Table 7. Statistical analysis of TR-OCTA compared to GT-OCTA for 6 mm dataset from UIC.

From two-tail T-tests (p < .05) BVD, BVC, BVT and VPI values for TR-OCTA were found to be statistically different from GT-OCTA for 3 mm and 6 mm which is expected due to the fact that UIC data was excluded from the training (Table 1). These feature values were calculated (3 mm and 6 mm) for the complete dataset as well as separately for different patient groups. For 3 mm (Table 6), BVD (Figures 4I–L) shows more deviation from the GT values compared to BVC and VPI except BVT which shows better comparative values. From Figure 6 we can observe the difference between TR and GT images and concentrated values with some outliers for BVC, BVT, and BVD. Supplementary Figures S2A–D shows the range of values for all the features observed in different DR groups and the mean values for TR-OCTA lies below the GT-OCTA most of the cases except for BVD which lies above GT values for Mild, Moderate and Severe group.

Figure 6
www.frontiersin.org

Figure 6. Feature value distribution of 3mm and 6 mm scans for the UIC dataset. (A–D) are BVC, VPI, BVT, and BVD values for 3 mm. Similarly, (E–H) show values for 6 mm scans.

A similar pattern is recognized from the 6 mm analysis (Table 7) having BVD a higher deviation from the GT compared to BVC, BVT, and VPI. BVD shows more outliers for TR-OCTA than other features contributing to the larger difference (Figure 6). All these feature values are calculated separately for Control, Mild, Moderate and Severe groups and in general BVD shows higher difference between TR-OCTA and GT-OCTA among all the features for different patient groups (Figures 4M–P). Supplementary Figures S2E–H represents feature value distribution with outliers for different DR groups. Similar to 3 mm, TR-OCTA BVD feature value distribution shows a mean higher than the GT-OCTA for Control, Mild, Moderate and Severe groups.

Discussion

In this paper, we implemented a recently demonstrated algorithm [36] for OCT-OCTA translation and validated the translated OCTA images to show their utility in quantitatively characterizing retinal features using two datasets OCT500 and UIC. We present a comprehensive analysis comparing the performance of GT- OCTA images with those generated by a TR-OCTA across different patient groups, including those with complete data sets, for both 3 mm and 6 mm FoV. Several qualitative (SSIM, FID, PCQI) and quantitative metrics (BVD, BVC, BVT, and VPI) were considered to validate the comparative performance analysis on a clinical dataset from UIC. We found FID and PCQI scores to be the most reliable qualitative metrics and a combination of BVT, VPI could be considered best for distinguishing diseases specially DR patients in TR-OCTA.

SSIM was utilized as a quality metric to assess the similarity between TR-OCTA and GT-OCTA images, providing insight into the translation model’s ability to replicate key structural features of the retinal vasculature. For OCT500, between 3 mm and 6 mm, mean SSIM from the complete dataset of 3 mm was higher than 6 mm despite both values were far from the ideal value. The reason behind is how the translation algorithm works that focuses on the vasculature rather than the entire image while SSIM considers pixel difference between the reference and the target image for its entirety. This becomes more prominent from the UIC dataset. Both 3 mm and 6 mm sets form UIC shows lower mean SSIM than OCT500 since the algorithm was exclusively trained on OCT500 and as a result the generated TR-OCTAs show less structural similarity overall to the GT-OCTA. SSIM for different patient groups was also considered to measure the response of the TR-OCTAs for different pathologies. UIC dataset mostly show consistent mean SSIM across different pathologies except for RVO (6 mm) indicating the algorithm couldn’t capture the vascular structure from the RVO patient equally as other pathologies which could also be contributed to the lower number of sample available for training (3.34%). Similar observations can be made from UIC 3 mm and 6 mm however the model fails to capture the key structural features of severe stage DR patients. Therefore, from our observations, SSIM is not an ideal metric for GAN generated OCTA images.

Two more quality metrics, FID and PCQI scores were considered as these are more suitable for GAN generated image quality comparison against reference (GT) images. OCT500 3 mm set, having lower FID score, indicate higher similarity to the reference data, suggesting that the 3 mm scans exhibit better image quality compared to the 6 mm scans within this dataset. This trend is consistent with expectations, as higher resolution (smaller scanning area) typically results in finer detail and less distortion. The UIC dataset shows comparatively higher FID scores for both 3 mm and 6 mm scans, indicating lower fidelity compared to the OCT500 dataset which can be explained as the effect of implementing the model to a dataset excluded from the training. Interestingly, the 6 mm scans have a lower FID score than the 3 mm scans, suggesting better relative performance at a larger scanning area for this dataset. This inverse trend can be attributed to variations in image quality for 6 mm image acquisition (image collected after a certain date is of higher quality than the previous scans). Another quality metric, PCQI values for the OCT500 dataset are found to be exceptionally high for both 3 mm and 6 mm. These results indicate minimal variation and high consistency in image quality across different scans. The slight decrease in PCQI for the 6 mm scans is consistent with the increased FID score, further confirming that the 3 mm scans offer superior quality. The PCQI scores for the UIC dataset, while slightly lower than those of the OCT500 dataset, still demonstrate high image quality, for 3 mm and 6 mm sets. The minimal difference in PCQI between 3 mm and 6 mm scans suggests that the image quality is relatively stable across different FoVs, despite the higher FID scores.

To compare the quantitative feature values between TR-OCTA and GT-OCTA, we focus on the performance across different categories since neither OCT500 nor UIC has equal distribution of data across different pathologies and normal or control group. In a comparative analysis from OCT500, BVD shows more deviation from the GT-OCTA, specially for 6 mm, supporting the better FID and PCQI scores for 3 mm, indicating the model can capture the superficial layer vasculature better than the deep layer hence vessel density shows bigger difference between TR-OCTA and GT-OCTA (Tables 4, 5). This feature also has the greatest number of outliers (Figure 5) among all features contributing to larger difference. Similar observations can be made from UIC dataset, showing the models limitation in capturing vasculature from all layers equally. When plotted, the trend set by the GT-OCTA for BVD show higher deviation for CNV in OCT500 3 mm. TR-OCTA also follows the same trend, however, for OCT500 6mm, BVD feature values tend to follow the GT-OCTA apart from RVO patients. On the other hand, for UIC 3 mm and 6 mm, BVD fails to properly exhibit the trend shown by GT-OCTA indicating inclusion of the data in the training process provides better result for BVD (Tables 6, 7). However, for UIC 6mm, BVD shows clear distinction between control and early stage DR as well as severe stage DR, similar to what has been reported in other studies [12]. Overall, when trained, BVD could be considered as a potential biomarker for CNV and RVO patients as evident from the analysis (Figure 4).

BVC, another quantitative feature, unlike BVD, doesn’t show that large of a deviation from the GT values across different pathologies indicating it could be used as a measure of performance even on a dataset not seen before. OCT500 3 mm TR-OCTA tend to show some variations for CNV and RVO patient similar to BVD however the trend goes beyond what is observed from GT-OCTA. From UIC dataset, however, BVC feature values across different pathologies stay in close range across different stages of DR patients without showing any large deviation. However, the TR-OCTA values follow GT-OCTA trend.

BVT feature values show least deviation for all datasets across different patient categories. For OCT500 3 mm, TR-OCTA values follow the trend set by GT-OCTA specially showing a clear distinction among DR and other pathologies as well as normal state (Figure 4C). In consistent with this scenario, UIC BVT feature values stay in close range for (TR-OCTA and GT-OCTA) 3 mm and show a better trend line for 6 mm. From OCT500 6 mm, BVT feature values, while closely following GT-OCTA values, show clear distinction for AMD patients indicating a potential choice for AMD classification at lower FoV.

Finally, VPI feature values for OCT500 3 mm TR-OCTA show an opposite trend compared to GT-OCTA. GT-OCTA show clear distinction among AMD, CNV and DR pathologies however TR-OCTA fails to identify the distinction. However, OCT500 6 mm TR-OCTA shows a much better performance showing the clear deviation of values for CNV, CSC and DR patients indicating the model’s ability to distinguish between normal and AMD, CNV, DR patients in general. A similar picture is depicted by the performance analysis on UIC dataset, specially 3 mm, clearly identifying different DR stages from control groups indicating VPI as a better choice of potential biomarker for DR patients.

Overall, in light of the comparative analysis performed on both datasets, SSIM values were higher for OCT500 rather than UIC indicating the inclusion of the prior dataset affecting the quality of the generated images. Similar scenarios can be observed from FID scores, OCT500 having lower value hence better resemblance to the GT-OCTA images. However, PCQI scores for both datasets (OCT500 and UIC) indicates the TR-OCTA images are almost the same as the GT-OCTAs in terms of contrast and sharpness which is supported by our analysis that some features (BVT, VPI) show better performance across both FoV. Although, feature values for BVC and VPI showed slight variation, they were still in close proximity for both OCT500 and UIC. BVT, however, is found having almost consistent values for both dataset in all cases. BVD, on the other hand, shows bigger variation between TR-OCTA and GT-OCTA for UIC compared to OCT500, an expected outcome similar to SSIM, as density features tend to be affected by local vasculature more than the vessel features.

Despite a better performance from the TR-OCTA across the datasets, our study has some inherent limitations. One limitation being the scarcity of the publicly available data, restricting the ability to perform extensive and varied analyses. Additionally, the dataset used in this research is relatively small, particularly when considering cross-pathological studies. This limited sample size can hinder the generalizability of the findings. Furthermore, our data distribution is unequal across different patient categories and pathologies, potentially introducing bias and affecting the robustness of our conclusions. Another significant challenge is the inconsistency in image quality and signal strength. Not all OCTA images have the same signal strength or quality, which can adversely impact the performance of translated OCTA in both quantitative and qualitative assessments. These limitations highlight the need for more comprehensive datasets and improved image acquisition standards to enhance the reliability and applicability of OCTA feature studies. Although similar studies have been conducted [12, 14] on OCTA features, to our knowledge, this is the first study conducted on quantitative characterization and extensive comparative analysis of retinal features for TR-OCTA images. We tried to establish the idea that while the translation model holds promise in reproducing retinal vasculature across various conditions, there exist minor variations in the accuracy of vascular metrics between TR-OCTA and GT-OCTA images. These discrepancies underscore the necessity for ongoing enhancements to the translation model to achieve higher precision in vascular representation, particularly for pathological conditions where accurate vascular depiction is critical for clinical diagnosis and monitoring.

This study showcases the potential of AI to bridge the gap between OCT’s inability to visualize blood flow information and leveraging generative-adversarial learning frameworks for image translation to capture that information. Our findings suggest that AI-driven translation models can generate high-quality OCTA images from OCT data (demonstrated using SSIM, FID and PCQI metrics) and the quantitative features generated in TR-OCTA follow a similar trend as in GT-OCTA. This has the potential to significantly improve the accuracy and efficiency of diagnosing and monitoring retinal diseases through OCTA imaging, emphasizing the need for further research and development in this area.

In summary, this study demonstrates the potential of generative AI in enhancing OCT imaging for ophthalmic diagnostics. By validating quantitative features to check the viability of TR-OCTA, this research addresses significant limitations in widespread adoption of OCTA in clinical settings. Despite facing challenges such as generalization for different retinal diseases and difficulty in capturing detailed vascular networks, our study lays a solid foundation for future advancements in multi-modal OCT based retinal disease diagnosis and monitoring. The incorporation of AI not only promises to reduce the dependence on costly OCTA devices but also opens new avenues for accessible and accurate retinal healthcare solutions. Moving forward, it is imperative to refine these AI models to improve the resolution and accuracy of translated OCTA images, ensuring they can reliably support clinical decision-making and contribute to the broader understanding of retinal pathologies.

Author contributions

All authors participated in the interpretation of the studies, analysis of the data and review of the manuscript; RB: conceptualization, data curation, formal analysis, investigation, methodology, project administration, resources, software, supervision, validation, visualization, writing–original draft, and writing–review and editing, AT, JL, and TL: resources, result interpretation, and writing–review and editing, MA: conceptualization, data curation, formal analysis, funding acquisition, investigation, methodology, project administration, resources, software, supervision, validation, visualization, writing–original draft, and writing–review and editing. All authors contributed to the article and approved the submitted version.

Data availability

Publicly available datasets were analyzed in this study. The OCT500 dataset can be found here: https://ieee-dataport.org/open-access/octa-500. UIC data can be requested to the corresponding author.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. Supported by NEI R21EY035271 (MA), R15EY035804 (MA); and UNC Charlotte Faculty Research Grant (MA).

Conflict of interest

The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.ebm-journal.org/articles/10.3389/ebm.2024.10333/full#supplementary-material

SUPPLEMENTARY FIGURE S1 | (A–D) show BVC, VPI, BVT, and BVD feature values for the 3 mm dataset (OCT500) with different patient conditions. a(i-iv) are AMD patients, b(i-iv) are CNV patients, c(i-iv) are DR patients. d(i-iv) are Normal patients. (E–K) show feature values for the 6 mm dataset (OCT500) with different patient conditions. e(i-iv) are AMD patients, f(i-iv) are CNV patients, g(i-iv) are CSC patients, h(i-iv) are DR patients, i(i-iv) are RVO patients, j(i-iv) are patients with other retinal pathologies, k(i-iv) are Normal patients.

SUPPLEMENTARY FIGURE S2 | (A–D) show BVC, VPI, BVT and BVD for the UIC 3 mm scans with different patient conditions. a(i-iv) shows Control group, b(i-iv) Mild, c(i-iv) Moderate, d(i-iv) Severe patient scans. Similarly, (E–H) show BVC, VPI, BVT and BVD for the UIC 6 mm scans with different patient conditions. e(i-iv) for Control group, f(i-iv) for Mild, g(i-iv) for Moderate, h(i-iv) for Severe patient scans.

References

1. Bowd, C, Zangwill, LM, Berry, CC, Blumenthal, EZ, Vasile, C, Sanchez-Galeana, C, et al. Detecting early glaucoma by assessment of retinal nerve fiber layer thickness and visual function. Invest Ophthalmol Vis Sci (2001) 42(9):1993–2003.

PubMed Abstract | Google Scholar

2. Fujimoto, J, and Swanson, E. The development, commercialization, and impact of optical coherence tomography. Invest Ophthalmol Vis Sci (2016) 57(9):OCT1–OCT13. doi:10.1167/iovs.16-19963

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Spaide, RF, Fujimoto, JG, Waheed, NK, Sadda, SR, and Staurenghi, G. Optical coherence tomography angiography. Prog Retin Eye Res (2018) 64:1–55. doi:10.1016/j.preteyeres.2017.11.003

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Yang, D, Ran, AR, Nguyen, TX, Lin, TPH, Chen, H, Lai, TYY, et al. Deep learning in optical coherence tomography angiography: current progress, challenges, and future directions. Diagnostics (2023) 13(2):326. doi:10.3390/diagnostics13020326

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Huang, D, Swanson, EA, Lin, CP, Schuman, JS, Stinson, WG, Chang, W, et al. Optical coherence tomography. Science (1991) 254(5035):1178–81. doi:10.1126/science.1957169

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Adhi, M, and Duker, JS. Optical coherence tomography – current and future applications. Curr Opin Ophthalmol (2013) 24(3):213–21. doi:10.1097/ICU.0b013e32835f8bf8

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Sakata, LM, DeLeon-Ortega, J, Sakata, V, and Girkin, CA. Optical coherence tomography of the retina and optic nerve – a review. Clin & Exp Ophthalmol (2009) 37(1):90–9. doi:10.1111/j.1442-9071.2009.02015.x

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Drexler, W, and Fujimoto, J. State-of-the-art retinal optical coherence tomography. Prog Retin Eye Res (2008) 27(1):45–88. doi:10.1016/j.preteyeres.2007.07.005

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Li, X, Chen, H, Qi, X, Dou, Q, Fu, CW, and Heng, PA. H-DenseUNet: hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans Med Imaging (2018) 37(12):2663–74. doi:10.1109/TMI.2018.2845918

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Chalam, K, and Sambhav, K. Optical coherence tomography angiography in retinal diseases. J Ophthalmic Vis Res (2016) 11(1):84. doi:10.4103/2008-322X.180709

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Kim, G, Kim, J, Choi, WJ, Kim, C, and Lee, S. Integrated deep learning framework for accelerated optical coherence tomography angiography. Sci Rep (2022) 12(1):1289. doi:10.1038/s41598-022-05281-0

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Alam, M, Zhang, Y, Lim, JI, Chan, RVP, Yang, M, and Yao, X. Quantitative optical coherence tomography angiography features for objective classification and staging of diabetic retinopathy. Retina (2020) 40(2):322–32. doi:10.1097/IAE.0000000000002373

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Le, D, Alam, M, Yao, CK, Lim, JI, Hsieh, YT, Chan, RVP, et al. Transfer learning for automated OCTA detection of diabetic retinopathy. Translational Vis Sci & Technology (2020) 9(2):35. doi:10.1167/tvst.9.2.35

CrossRef Full Text | Google Scholar

14. Le, D, Alam, M, Miao, BA, Lim, JI, and Yao, X. Fully automated geometric feature analysis in optical coherence tomography angiography for objective classification of diabetic retinopathy. Biomed Opt Express (2019) 10(5):2493. doi:10.1364/BOE.10.002493

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Alam, M, Le, D, Lim, JI, and Yao, X. Vascular complexity analysis in optical coherence tomography angiography of diabetic retinopathy. Retina (2021) 41(3):538–45. doi:10.1097/IAE.0000000000002874

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Le, D, Alam, MN, Lim, JI, Chan, RVP, and Yao, X. Deep learning for objective OCTA detection of diabetic retinopathy. In: F Manns, PG Söderberg, and A Ho, editors. Ophthalmic technologies XXX. SPIE (2020). 60. doi:10.1117/12.2546586

CrossRef Full Text | Google Scholar

17. Alam, MN, Son, T, Toslak, D, Lim, JI, and Yao, X. Quantitative artery-vein analysis in optical coherence tomography angiography of diabetic retinopathy. In: F Manns, PG Söderberg, and A Ho, editors. Ophthalmic technologies XXIX. SPIE (2019). 1. doi:10.1117/12.2510213

CrossRef Full Text | Google Scholar

18. Alam, MN, Zhang, Y, Lim, JI, Chan, RVP, Yang, M, and Yao, X. Quantitative OCT angiography for computer-aided classification of diabetic retinopathy. Invest Ophthalmol Vis Sci (2018) 59(9):1224.

Google Scholar

19. Alam, M, Toslak, D, Lim, JI, and Yao, X. Color fundus image guided artery-vein differentiation in optical coherence tomography angiography. Invest Opthalmology & Vis Sci (2018) 59(12):4953. doi:10.1167/iovs.18-24831

CrossRef Full Text | Google Scholar

20. Palejwala, NV, Jia, Y, Gao, SS, Liu, L, Flaxel, CJ, Hwang, TS, et al. Detection of nonexudative choroidal neovascularization in age-related macular degeneration with optical coherence tomography angiography. Retina (2015) 35(11):2204–11. doi:10.1097/IAE.0000000000000867

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Burlina, PM, Joshi, N, Pekala, M, Pacheco, KD, Freund, DE, and Bressler, NM. Automated grading of age-related macular degeneration from color fundus images using deep convolutional neural networks. JAMA Ophthalmol (2017) 135(11):1170. doi:10.1001/jamaophthalmol.2017.3782

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Gholami, S, Lim, JI, Leng, T, Ong, SSY, Thompson, AC, and Alam, MN. Federated learning for diagnosis of age-related macular degeneration. Front Med (2023) 10:1259017. doi:10.3389/fmed.2023.1259017

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Holló, G. Vessel density calculated from OCT angiography in 3 peripapillary sectors in normal, ocular hypertensive, and glaucoma eyes. Eur J Ophthalmol (2016) 26(3):e42–e45. doi:10.5301/ejo.5000717

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Hsieh, YT, Alam, MN, Le, D, Hsiao, CC, Yang, CH, Chao, DL, et al. OCT angiography biomarkers for predicting visual outcomes after ranibizumab treatment for diabetic macular edema. Ophthalmol Retina (2019) 3(10):826–34. doi:10.1016/j.oret.2019.04.027

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Alam, M, Thapa, D, Lim, JI, Cao, D, and Yao, X. Quantitative characteristics of sickle cell retinopathy in optical coherence tomography angiography. Biomed Opt Express (2017) 8(3):1741. doi:10.1364/BOE.8.001741

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Alam, M, Thapa, D, Lim, JI, Cao, D, and Yao, X. Computer-aided classification of sickle cell retinopathy using quantitative features in optical coherence tomography angiography. Biomed Opt Express (2017) 8(9):4206. doi:10.1364/BOE.8.004206

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Alam, M, Lim, JI, Toslak, D, and Yao, X. Differential artery–vein analysis improves the performance of OCTA staging of sickle cell retinopathy. Translational Vis Sci & Technology (2019) 8(2):3. doi:10.1167/tvst.8.2.3

CrossRef Full Text | Google Scholar

28. Zahid, S, Alam, MN, Yao, X, and Lim, JI. Quantitative optical coherence tomography angiography parameters in central retinal vein occlusion. Invest Ophthalmol Vis Sci (2018) 59(9):5427.

Google Scholar

29. Chen, JL, Zahid, S, Alam, MN, Yao, X, and Lim, JI. Assessment of quantitative optical coherence tomography angiography parameters in branch retinal vein occlusion and monitoring response to treatment. Invest Ophthalmol Vis Sci (2018) 59(9):5458.

Google Scholar

30. Alam, MN, Thapa, D, Lim, JI, Cao, D, and Yao, X Automatic classification of sickle cell retinopathy using quantitative features in optical coherence tomography angiography. Invest Ophthalmol Vis Sci (2017) 58(8):1679.

Google Scholar

31. Wang, J, Hormel, TT, Tsuboi, K, Wang, X, Ding, X, Peng, X, et al. Deep learning for diagnosing and segmenting choroidal neovascularization in OCT angiography in a large real-world data set. Translational Vis Sci & Technology (2023) 12(4):15. doi:10.1167/tvst.12.4.15

CrossRef Full Text | Google Scholar

32. Lin, L, Wang, Z, Wu, J, Huang, Y, Lyu, J, Cheng, P, et al. “BSDA-Net: a boundary shape and distance aware joint learning framework for segmenting and classifying OCTA images,” in Computing and computer assisted intervention – MICCAI 2021: 24th international conference, proceedings, part VIII, Strasbourg, France, September 27–October 1, 2021 (2021). 65–75. (Boundary shape and distance, FAZ, segmentation, classification, OCTA, joint learning). doi:10.1007/978-3-030-87237-3_7

CrossRef Full Text | Google Scholar

33. Anvari, P, Ashrafkhorasani, M, Habibi, A, and Falavarjani, KG. Artifacts in optical coherence tomography angiography. J Ophthalmic & Vis Res (2021) 16:271–86. doi:10.18502/jovr.v16i2.9091

CrossRef Full Text | Google Scholar

34. Lee, CS, Tyring, AJ, Wu, Y, Xiao, S, Rokem, AS, DeRuyter, NP, et al. Generating retinal flow maps from structural optical coherence tomography with artificial intelligence. Sci Rep (2019) 9(1):5694. doi:10.1038/s41598-019-42042-y

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Zhang, Z, Ji, Z, Chen, Q, Yuan, S, and Fan, W. Texture-guided U-net for OCT-to-OCTA generation. In: H Ma, L Wang, and C Zhang, editors. Pattern Recognition and computer vision. Vol 13022. Lecture notes in computer science. Springer International Publishing (2021). 42–52. doi:10.1007/978-3-030-88013-2_4

CrossRef Full Text | Google Scholar

36. Li, S, Zhang, D, Li, X, Ou, C, An, L, Xu, Y, et al. Vessel-promoted OCT to OCTA image translation by heuristic contextual constraints. Med Image Anal (2024) 98:103311. doi:10.1016/j.media.2024

PubMed Abstract | CrossRef Full Text | Google Scholar

37. Li, PL, O’Neil, C, Saberi, S, Sinder, K, Wang, K, Tan, B, et al. Deep learning algorithm for generating optical coherence tomography angiography (OCTA) maps of the retinal vasculature. In: ME Zelinski, TM Taha, J Howe, AA Awwal, and KM Iftekharuddin, editors. Applications of machine learning 2020. SPIE (2020). 11511: 8. doi:10.1117/12.2568629

CrossRef Full Text | Google Scholar

38. Valizadegan, H, Nguyen, Q, and Hauskrecht, M. Learning classification models from multiple experts. J Biomed Inform (2013) 46(6):1125–35. doi:10.1016/j.jbi.2013.08.007

PubMed Abstract | CrossRef Full Text | Google Scholar

39. Le, D, Son, T, Kim, TH, Adejumo, T, Abtahi, M, Ahmed, S, et al. Deep learning-based optical coherence tomography angiography image construction using spatial vascular connectivity network. Commun Eng (2024) 3(1):28. doi:10.1038/s44172-024-00173-9

CrossRef Full Text | Google Scholar

40. Ronneberger, O, Fischer, P, and Brox, T. U-net: convolutional networks for biomedical image segmentation. In: N Navab, J Hornegger, WM Wells, and AF Frangi, editors. Medical image Computing and computer-assisted intervention – MICCAI 2015. Vol 9351. Lecture notes in computer science. Springer International Publishing (2015). 234–41. doi:10.1007/978-3-319-24574-4_28

CrossRef Full Text | Google Scholar

41. Isola, P, Zhu, JY, Zhou, T, and Efros, AA. Image-to-Image translation with conditional adversarial networks (2024). Available from: http://arxiv.org/abs/1611.07004 (Accessed February 20, 2024).

Google Scholar

42. Zhu, JY, Park, T, Isola, P, and Efros, AA. Unpaired image-to-image translation using cycle-consistent adversarial networks (2024). Available from: http://arxiv.org/abs/1703.10593 (Accessed February 20, 2024).

Google Scholar

43. Heusel, M, Ramsauer, H, Unterthiner, T, Nessler, B, and Hochreiter, S. GANs trained by a two time-scale update rule converge to a local nash equilibrium (2024). doi:10.48550/ARXIV.1706.08500

CrossRef Full Text | Google Scholar

44. Wang, S, Ma, K, Yeganeh, H, Wang, Z, and Lin, W. A patch-structure representation method for quality assessment of contrast changed images. IEEE Signal Process Lett (2015) 22(12):2387–90. doi:10.1109/LSP.2015.2487369

CrossRef Full Text | Google Scholar

45. Chao, SC, Yang, SJ, Chen, HC, Sun, CC, Liu, CH, and Lee, CY. Early macular angiography among patients with glaucoma, ocular hypertension, and normal subjects. J Ophthalmol (2019) 2019:1–7. doi:10.1155/2019/7419470

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Pedinielli, A, Bonnin, S, Sanharawi, ME, Mané, V, Erginay, A, Couturier, A, et al. Three different optical coherence tomography angiography measurement methods for assessing capillary density changes in diabetic retinopathy. Ophthalmic Surg Lasers Imaging Retina (2017) 48(5):378–84. doi:10.3928/23258160-20170428-03

PubMed Abstract | CrossRef Full Text | Google Scholar

47. Mastropasqua, R, Toto, L, Mastropasqua, A, Aloia, R, De Nicola, C, Mattei, PA, et al. Foveal avascular zone area and parafoveal vessel density measurements in different stages of diabetic retinopathy by optical coherence tomography angiography. Int J Ophthalmol (2017) 10(10):1545–1551. doi:10.18240/ijo.2017.10.11

PubMed Abstract | CrossRef Full Text | Google Scholar

48. Jhaj, G, Glazman, S, Shrier, EM, and Bodis-Wollner, I. Non-exudative age-related macular degeneration foveal avascular zone area, foveal vessel density, and ganglion cell complex thickness. Invest Ophthalmol Vis Sci (2017) 58(8):36.

Google Scholar

49. Al-Sheikh, M, Iafe, NA, Phasukkijwatana, N, Sadda, SR, and Sarraf, D. Biomarkers of neovascular activity in age-related macular degeneration using optical coherence tomography angiography. Retina (2018) 38(2):220–30. doi:10.1097/IAE.0000000000001628

PubMed Abstract | CrossRef Full Text | Google Scholar

50. Akagi, T, Uji, A, Huang, AS, Weinreb, RN, Yamada, T, Miyata, M, et al. Conjunctival and intrascleral vasculatures assessed using anterior segment optical coherence tomography angiography in normal eyes. Am J Ophthalmol (2018) 196:1–9. doi:10.1016/j.ajo.2018.08.009

PubMed Abstract | CrossRef Full Text | Google Scholar

51. Ye, H, Zheng, C, Lan, X, Zhao, L, Qiao, T, Li, X, et al. Evaluation of retinal vasculature before and after treatment of children with obstructive sleep apnea-hypopnea syndrome by optical coherence tomography angiography. Graefes Arch Clin Exp Ophthalmol (2019) 257(3):543–8. doi:10.1007/s00417-018-04207-9

PubMed Abstract | CrossRef Full Text | Google Scholar

52. Chu, Z, Lin, J, Gao, C, Xin, C, Zhang, Q, Chen, CL, et al. Quantitative assessment of the retinal microvasculature using optical coherence tomography angiography. J Biomed Opt (2016) 21(6):066008. doi:10.1117/1.JBO.21.6.066008

PubMed Abstract | CrossRef Full Text | Google Scholar

53. Li, M, Huang, K, Xu, Q, Yang, J, Zhang, Y, Ji, Z, et al. A retinal dataset for optical coherence tomography angiography study. Med. Image Anal. (2024) 93:103092. doi:10.1016/j.media.2024.103092

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: GAN, generative AI, OCT, OCTA, OCTA features

Citation: Badhon RH, Thompson AC, Lim JI, Leng T and Alam MN (2024) Quantitative characterization of retinal features in translated OCTA. Exp. Biol. Med. 249:10333. doi: 10.3389/ebm.2024.10333

Received: 02 August 2024; Accepted: 07 October 2024;
Published: 23 October 2024.

Copyright © 2024 Badhon, Thompson, Lim, Leng and Alam. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Minhaj Nur Alam, malam8@charlotte.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.