4 sonuçlar
Arama Sonuçları
Listeleniyor 1 - 4 / 4
Yayın A haar classifier based call number detection and counting method for library books(IEEE, 2018-12-06) Kanburoğlu, Ali Buğra; Tek, Faik BorayCounting and organization of books in libraries is a routine and time-consuming task The task gets more complicated by misplaced books in shelves. In order to solve these problems, we propose an automated visual call number (book-id) detection and counting system in this paper. The method employs a Haar feature-based classifier from OpenCV library and cloud-based OCR system to decode characters from images. To develop and test the method, we have acquired and organized a dataset of 1000 book call numbers. The proposed method has been tested on 20 bookshelves images that contain 233 call numbers, which resulted in a true detection rate of 96% and false detection rate of 1.75 per image. For OCR step, the number of false recognized characters per call number was 0.76.Yayın Convolutional attention network for MRI-based Alzheimer's disease classification and its interpretability analysis(IEEE, 2021-09-17) Türkan, Yasemin; Tek, Faik BorayNeuroimaging techniques, such as Magnetic Resonance Imaging (MRI) and Positron Emission Tomography (PET), help to identify Alzheimer's disease (AD). These techniques generate large-scale, high-dimensional, multimodal neuroimaging data, which is time-consuming and difficult to interpret and classify. Therefore, interest in deep learning approaches for the classification of 3D structural MRI brain scans has grown rapidly. In this research study, we improved the 3D VGG model proposed by Korolev et al. [2]. We increased the filters in the 3D convolutional layers and then added an attention mechanism for better classification. We compared the performance of the proposed approaches for the classification of Alzheimer's disease versus mild cognitive impairments and normal cohorts on the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset. We observed that both the accuracy and area under curve results improved with the proposed models. However, deep neural networks are black boxes that produce predictions that require further explanation for medical usage. We compared the 3D-data interpretation capabilities of the proposed models using four different interpretability methods: Occlusion, 3D Ultrametric Contour Map, 3D Gradient-Weighted Class Activation Mapping, and SHapley Additive explanations (SHAP). We observed that explanation results differed in different network models and data classes.Yayın Adaptive convolution kernel for artificial neural networks(Academic Press Inc., 2021-02) Tek, Faik Boray; Çam, İlker; Karlı, DenizMany deep neural networks are built by using stacked convolutional layers of fixed and single size (often 3 × 3) kernels. This paper describes a method for learning the size of convolutional kernels to provide varying size kernels in a single layer. The method utilizes a differentiable, and therefore backpropagation-trainable Gaussian envelope which can grow or shrink in a base grid. Our experiments compared the proposed adaptive layers to ordinary convolution layers in a simple two-layer network, a deeper residual network, and a U-Net architecture. The results in the popular image classification datasets such as MNIST, MNIST-CLUTTERED, CIFAR-10, Fashion, and ‘‘Faces in the Wild’’ showed that the adaptive kernels can provide statistically significant improvements on ordinary convolution kernels. A segmentation experiment in the Oxford-Pets dataset demonstrated that replacing ordinary convolution layers in a U-shaped network with 7 × 7 adaptive layers can improve its learning performance and ability to generalize.Yayın Retinal disease classification using optical coherence tomography angiography images(Institute of Electrical and Electronics Engineers Inc., 2024) Aydın, Ömer Faruk; Nazlı, Muhammet Serdar; Tek, Faik Boray; Turkan, YaseminOptical Coherence Tomography Angiography (OCTA) is a non-invasive imaging modality widely used for the detailed visualization of retinal microvasculature, which is crucial for diagnosing and monitoring various retinal diseases. However, manual interpretation of OCTA images is labor-intensive and prone to variability, highlighting the need for automated classification methods. This study presents an aproach that utilizes transfer learning to classify OCTA images into different retinal disease categories, including age-related macular degeneration (AMD) and diapethic retinopathy (DR). We used the OCTA-500 dataset [1], the largest publicly available retinal dataset that contains images from 500 subjects with diverse retinal conditions. To address the class imbalance, we employed k-fold cross-validation and grouped various other conditions under the 'OTHERS' class. Additionally, we compared the performance of the ResNet50 model with OCTA inputs to that of the ResNet50 and RetFound (Vision Transformer) models with OCT inputs to assess the efficiency of OCTA in retinal condition classification. In the three-class (AMD, D R, Normal) classification, ResNet50-OCTA o utperformed ResNet50-OCT, but slightly underperformed compared to RetFound-OCT, which was pretrained on a large OCT dataset. In the four-class (AMD, DR, Normal, Others) classification, ResNet50-OCTA and RetFound-OCT achieved similar classification a ccuracies. This study establishes a baseline for retinal condition classification using the OCTA-500 dataset and provides a comparison between OCT and OCTA input modalities.












