Arama Sonuçları

Listeleniyor 1 - 2 / 2
  • Yayın
    Retinal disease classification from bimodal OCT and OCTA using a CNN-ViT hybrid architecture
    (Institute of Electrical and Electronics Engineers Inc., 2025-09-21) Aydın, Ömer Faruk; Tek, Faik Boray; Turkan, Yasemin
    Retinal diseases are the leading cause of vision impairment and blindness worldwide. Early and accurate diagnosis is critical for effective treatment, and recent advances in imaging technologies such as Optical Coherence Tomography (OCT) and OCT Angiography (OCTA), have enabled detailed visualization of the retinal structure and vasculature. By leveraging these modalities, this study proposes an advanced deep learning architecture called MultiModalNet for automated multi-class retinal disease classification. MultiModalNet employs a dual-branch design, where OCTA projection maps are processed through a ResNet101 encoder, and cross-sectional slices from the OCT volume (B-scans) are analyzed using a Vision Transformer (ViT-Large). The extracted features from both branches were fused and passed through the fully connected layers for the final classification. Evaluated on the 3-class OCTA-500 dataset, which includes Age-related Macular Degeneration (AMD), Diabetic Retinopathy (DR), and Normal cases, the proposed model achieved state-of-the-art classification accuracy of 94.59 percent, significantly o utperforming single-modality baselines. This result highlights the effectiveness of integrating vascular and structural information to improve the diagnostic performance. The findings suggest that hybrid multi-modal deep learning approaches can play a transformative role in computer-aided ophthalmology, enhancing both clinical decision-making and screening workflows.
  • Yayın
    Self-supervised learning of 3D structure from 2D OCT slices for retinal disease diagnosis on UK biobank scans
    (Institute of Electrical and Electronics Engineers Inc., 2025-09-21) Nazlı, Muhammet Serdar; Turkan, Yasemin; Tek, Faik Boray
    This study presents a self-supervised learning framework for retinal disease classification using Optical Coherence Tomography (OCT) scans. To balance the contextual richness of 3D volumes with the computational efficiency of 2D architectures, we introduce a quasi-3D input generation strategy. Each input is constructed by stacking three OCT slices, sampled from channel-specific Gaussian distributions centered on the volume midplane, and arranged in a standard three-channel 2D format compatible with existing pre-trained models. These quasi-3D images are used to pre-train a Vision Transformer (ViT-Base) via a Masked Autoencoder (MAE) with a shared masking pattern, encouraging the model to reconstruct masked regions by encoding anatomical continuity across slices. Pre-training is conducted on 10,000 unlabeled OCT volumes from the UK Biobank. The encoder is then fine-tuned on the OCTA-500 dataset for three-class and four-class retinal disease classification tasks, including macular degeneration and diabetic retinopathy. The model achieves 92.57% accuracy on the three-class task, matching the performance of RETFound while using over 150 times less pre-training data and a smaller backbone.