Automated segmentation of retinal vessels plays a pivotal role in early diagnosis of ophthalmic disorders. In this paper, a blood vessel segmentation algorithm using an enhanced fuzzy min-max neural network supervised classifier is proposed. The input to the network is an optimal 11-D feature vector which consists of spatial as well as frequency domain features extracted from each pixel of a fundus image. The essence of the method is its hyperbox classifier which performs online learning and gives binary output without any need of post-processing. The method is tested on publicly available databases DRIVE and STARE. The results are compared with the existing methods in the literature. The proposed method exhibits efficient performance and can be implemented in computer aided screening and diagnosis of retinal diseases. The method attains an average accuracy, sensitivity and specificity of 95.73%, 74.75% and 97.81% on DRIVE database and 95.51%, 74.65% and 97.11% on STARE database, respectively.
相似文献To improve the accuracy of retinal vessel segmentation, a retinal vessel segmentation algorithm for color fundus images based on back-propagation (BP) neural network is proposed according to the characteristics of retinal blood vessels. Four kinds of green channel image enhancement results of adaptive histogram equalization, morphological processing, Gaussian matched filtering, and Hessian matrix filtering are used to form feature vectors. The BP neural network is input to segment blood vessels. Experiments on the color fundus image libraries DRIVE and STARE show that this algorithm can obtain complete retinal blood vessel segmentation as well as connected vessel stems and terminals. When segmenting most small blood vessels, the average accuracy on the DRIVE library reaches 0.9477, and the average accuracy on the STARE library reaches 0.9498, which has a good segmentation effect. Through verification, the algorithm is feasible and effective for blood vessel segmentation of color fundus images and can detect more capillaries.
相似文献In medicine, diagnosis is as important as treatment. Retinal blood vessels are the most easily visible vessels in the whole body, and therefore, play a key role in the diagnosis of numerous diseases and eye disorders. Systematic and eye diseases cause morphologic variations, such as the growing, narrowing or branching of retinal blood vessels. Imaging-based screening of retinal blood vessels plays an important role in the identification and follow-up of eye diseases. Therefore, automatic retinal vessel segmentation can be used to diagnose and monitor those diseases. Computer-aided algorithms are required for the analysis of progression of eye diseases. This study proposes a hybrid method that provides a combination of pre-processing and data augmentation methods with a deep learning model. Pre-processing was used to solve the irregular clarification problems and to form a contrast between the background and retinal blood vessels. After pre-processing step, a convolutional neural network (CNN) was designed and then trained for the extraction of retinal blood vessels. In the training phase, data augmentation was performed to improve training performance. The CNN was trained and tested in the DRIVE database, which is commonly used in retinal blood vessel segmentation and publicly available for studies in this area. Results showed that the proposed system extracted vessels with a sensitivity of 77.78%, specificity of 97,84%, precision of 84.17% and accuracy of 95.27%.
This study also compared the results to those of previous studies. The comparison showed that the proposed method is an efficient and successful method for extracting retinal blood vessels. Moreover, the pre-processing phases improved the system performance. We believe that the proposed method and results will make contribution to the literature.
相似文献Single Shot MultiBox Detector (SSD) method using multi-scale feature maps for object detection, showing outstanding performance in object detection task. However, as a one-stage detection method, it’s difficult for SSD methods to quickly notice significant areas of objects in the image. In the SSD network structure, feature maps of different scales are used to independently predict object, and there is a lack of interaction between low-level feature maps and high-level feature maps. In this paper we propose an enhanced SSD method using interactive multi-scale attention features (MA-SSD). Our method uses the attention mechanism to generate attention features of multiple scales and adds it to the original detection branch of the SSD method, which effectively enhances the feature representation ability and improves the detection accuracy. At the same time, the feature of different detection scales interacts with each other, and all the detection branches in our method have a parallel structure, which ensures the detection efficiency. Our proposed method achieves competitive performance on the public dataset PascalVOC.
相似文献Logos are specially designed marks that identify goods, services, and organizations using distinguished characters, graphs, signals, and colors. Identifying logos can facilitate scene understanding, intelligent navigation, and object recognition. Although numerous logo recognition methods have been proposed for printed logos, a few methods have been specifically designed for logos in photos. Furthermore, most recognition methods use codebook-based approaches for the logos in photos. A codebook-based method is concerned with the generation of visual words for all the logo models. When new logos are added, the codebook reconstruction is required if effectiveness is a crucial factor. Moreover, logo detection in natural scenes is difficult because of perspective tilt and non-rigid deformation. Therefore, this study develops an extendable, but discriminating, model-based logo detection method. The proposed logo detection method is based on a support vector machine (SVM) using edge-based histograms of oriented gradient (HOGE) as features through multi-scale sliding window scanning. Thereafter, anti-distortion affine scale invariant feature transform (ASIFT) is used for logo verification with constraints on the ASIFT matching pairs and neighbors. The experimental results using the public Flickr-Logo database confirm that the proposed method has a higher retrieval and precision accuracy compared to existing model-based methods.
相似文献