6 sonuçlar
Arama Sonuçları
Listeleniyor 1 - 6 / 6
Yayın Convolutional neural network (CNN) algorithm based facial emotion recognition (FER) system for FER-2013 dataset(IEEE, 2022-11-18) Ezerceli, Özay; Eskil, Mustafa TanerFacial expression recognition (FER) is the key to understanding human emotions and feelings. It is an active area of research since human thoughts can be collected, processed, and used in customer satisfaction, politics, and medical domains. Automated FER systems had been developed and have been used to recognize humans’ emotions but it has been a quite challenging problem in machine learning due to the high intra-class variation. The first models were using known methods such as Support Vector Machines (SVM), Bayes classifier, Fuzzy Techniques, Feature Selection, Artificial Neural Networks (ANN) in their models but still, some limitations affect the accuracy critically such as subjectivity, occlusion, pose, low resolution, scale, illumination variation, etc. The ability of CNN boosts FER accuracy. Deep learning algorithms have emerged as the greatest way to produce the best results in FER in recent years. Various datasets were used to train, test, and validate the models. FER2013, CK+, JAFFE and FERG are some of the most popular datasets. To improve the accuracy of FER models, one dataset or a mix of datasets has been employed. Every dataset includes limitations and issues that have an impact on the model that is trained for it. As a solution to this problem, our state-of-the-art model based on deep learning architectures, particularly convolutional neural network architectures (CNN) with supportive techniques has been implemented. The proposed model achieved 93.7% accuracy with the combination of FER2013 and CK+ datasets for FER2013.Yayın Malaria parasite detection with deep transfer learning(IEEE, 2018-12-06) Var, Esra; Tek, Faik BorayThis study aims to automatically detect malaria parasites (Plasmodium sp) on images taken from Giemsa stained blood smears. Deep learning methods provide limited performance when sample size is low. In transfer learning, visual features are learned from large general data sets, and problem-specific classification problem can be solved successfully in restricted problem specific data sets. In this study, we apply transfer learning method to detect and classify malaria parasites. We use a popular pre-trained CNN model VGG19. We trained the model for 20 epoch on 1428 P Vivax, 1425 P Ovule, 1446 E Falciparum, 1450 P Malariae and 1440 non-parasite samples. The transfer learning model achieves %80, %83, %86, %75 precision and 83%, 86%, 86%, 79% f-measure on 19 test images.Yayın Animal sound classification using a convolutional neural network(IEEE, 2018-12-06) Şaşmaz, Emre; Tek, Faik BorayIn this paper, we investigate the problem of animal sound classification using deep learning and propose a system based on convolutional neural network architecture. As the input to the network, sound files were preprocessed to extract Mel Frequency Cepstral Coefficients (MFCC) using LibROSA library. To train and test the system we have collected 875 animal sound samples from an online sound source site for 10 different animal types. We report classification confusion matrices and the results obtained by different gradient descent optimizers. The best accuracy of 75% was obtained by Nesterov-accelerated Adaptive Moment Estimation (Nadam).Yayın Uzaktan algılanan görüntülerde bina yoğunluğu kestirimi için derin öğrenme(Institute of Electrical and Electronics Engineers Inc., 2019-09) Süberk, Nilay Tuğçe; Ateş, Hasan FehmiBu bildiri, derin öğrenme yöntemleri uygulayarak uzaktan algılamalı optik görüntülerde bina yoğunluğunun noktasal olarak kestirilmesi ile ilgilidir. Bu çalışma kapsamında, evrişimsel sinir ağına (ESA) dayalı derin öğrenme yöntemlerine başvurulmuştur. Önceden eğitilmiş, VGG-16 ve FCN-8s derin mimarileri bu probleme uyarlanmış ve ince ayar verilerek eğitilmiştir. Kestirilen değerler yerleşim bölgelerinde bina yoğunluk haritası oluşturmak için kullanılmıştır. Her iki mimarinin karşılaştırmalı benzetim sonuçları, güdümlü eğitim için binaları gösteren detaylı haritalara ihtiyaç duyulmadan hassas yoğunluk kestirimi yapılabileceğini göstermektedir.Yayın Analysis of single image super resolution models(IEEE, 2022-11-18) Köprülü, Mertali; Eskil, Mustafa TanerImage Super-Resolution (SR) is a set of image processing techniques which improve the resolution of images and videos. Deep learning approaches have made remarkable improvement in image super-resolution in recent years. This article aims and seeks to provide a comprehensive analysis on recent advances of models which has been used in image superresolution. This study has been investigated over other essential topics of current model problems, such as publicly accessible benchmark data-sets and performance evaluation measures. Finally, The study concluded these analysis by highlighting several weaknesses of existing base models as their feeding strategy and approved that the training technique which is Blind Feeding, which led several model to achieve state-of-the art.Yayın Retinal disease classification from bimodal OCT and OCTA using a CNN-ViT hybrid architecture(Institute of Electrical and Electronics Engineers Inc., 2025-09-21) Aydın, Ömer Faruk; Tek, Faik Boray; Turkan, YaseminRetinal diseases are the leading cause of vision impairment and blindness worldwide. Early and accurate diagnosis is critical for effective treatment, and recent advances in imaging technologies such as Optical Coherence Tomography (OCT) and OCT Angiography (OCTA), have enabled detailed visualization of the retinal structure and vasculature. By leveraging these modalities, this study proposes an advanced deep learning architecture called MultiModalNet for automated multi-class retinal disease classification. MultiModalNet employs a dual-branch design, where OCTA projection maps are processed through a ResNet101 encoder, and cross-sectional slices from the OCT volume (B-scans) are analyzed using a Vision Transformer (ViT-Large). The extracted features from both branches were fused and passed through the fully connected layers for the final classification. Evaluated on the 3-class OCTA-500 dataset, which includes Age-related Macular Degeneration (AMD), Diabetic Retinopathy (DR), and Normal cases, the proposed model achieved state-of-the-art classification accuracy of 94.59 percent, significantly o utperforming single-modality baselines. This result highlights the effectiveness of integrating vascular and structural information to improve the diagnostic performance. The findings suggest that hybrid multi-modal deep learning approaches can play a transformative role in computer-aided ophthalmology, enhancing both clinical decision-making and screening workflows.












