共查询到20条相似文献,搜索用时 31 毫秒
1.
目的 针对目前印刷套准识别方法依赖于经验人工设计特征提取的问题,提出一种不需要人工提取图像特征的卷积神经网络模型,实现印刷套准状态的识别.方法 采用图像增强技术实现不均衡训练集的均衡化,增加训练集图像的数量,提高模型的识别准确率.设计基于AlexNet网络结构的印刷套准识别模型的结构参数,分析批处理样本数量和基础学习率对模型性能的影响规律.结果 文中方法获得的总印刷套准识别准确率为0.9860,召回率为1.0000,分类准确率几何平均数为0.9869.结论 文中方法能自动提取图像特征,不依赖于人工设计的特征提取方法.在构造的数据集上,文中方法的分类性能优于实验中的支持向量机方法. 相似文献
2.
Based on the theory of modal acoustic emission (AE), when the convolutional neural network (CNN) is used to identify rotor rub-impact faults, the training data has a small sample size, and the AE sound segment belongs to a single channel signal with less pixel-level information and strong local correlation. Due to the convolutional pooling operations of CNN, coarse-grained and edge information are lost, and the top-level information dimension in CNN network is low, which can easily lead to overfitting. To solve the above problems, we first propose the use of sound spectrograms and their differential features to construct multi-channel image input features suitable for CNN and fully exploit the intrinsic characteristics of the sound spectra. Then, the traditional CNN network structure is improved, and the outputs of all convolutional layers are connected as one layer constitutes a fused feature that contains information at each layer, and is input into the network’s fully connected layer for classification and identification. Experiments indicate that the improved CNN recognition algorithm has significantly improved recognition rate compared with CNN and dynamical neural network (DNN) algorithms. 相似文献
3.
Fawaz Waselallah Alsaade Theyazn H. H. Aldhyani Mosleh Hmoud Al-Adhaileh 《计算机、材料和连续体(英文)》2021,68(1):805-819
The COVID-19 pandemic poses an additional serious public health threat due to little or no pre-existing human immunity, and developing a system to identify COVID-19 in its early stages will save millions of lives. This study applied support vector machine (SVM), k-nearest neighbor (K-NN) and deep learning convolutional neural network (CNN) algorithms to classify and detect COVID-19 using chest X-ray radiographs. To test the proposed system, chest X-ray radiographs and CT images were collected from different standard databases, which contained 95 normal images, 140 COVID-19 images and 10 SARS images. Two scenarios were considered to develop a system for predicting COVID-19. In the first scenario, the Gaussian filter was applied to remove noise from the chest X-ray radiograph images, and then the adaptive region growing technique was used to segment the region of interest from the chest X-ray radiographs. After segmentation, a hybrid feature extraction composed of 2D-DWT and gray level co-occurrence matrix was utilized to extract the features significant for detecting COVID-19. These features were processed using SVM and K-NN. In the second scenario, a CNN transfer model (ResNet 50) was used to detect COVID-19. The system was examined and evaluated through multiclass statistical analysis, and the empirical results of the analysis found significant values of 97.14%, 99.34%, 99.26%, 99.26% and 99.40% for accuracy, specificity, sensitivity, recall and AUC, respectively. Thus, the CNN model showed significant success; it achieved optimal accuracy, effectiveness and robustness for detecting COVID-19. 相似文献
4.
5.
6.
With the development of deep learning and Convolutional Neural Networks
(CNNs), the accuracy of automatic food recognition based on visual data have
significantly improved. Some research studies have shown that the deeper the model is,
the higher the accuracy is. However, very deep neural networks would be affected by the
overfitting problem and also consume huge computing resources. In this paper, a new
classification scheme is proposed for automatic food-ingredient recognition based on
deep learning. We construct an up-to-date combinational convolutional neural network
(CBNet) with a subnet merging technique. Firstly, two different neural networks are
utilized for learning interested features. Then, a well-designed feature fusion component
aggregates the features from subnetworks, further extracting richer and more precise
features for image classification. In order to learn more complementary features, the
corresponding fusion strategies are also proposed, including auxiliary classifiers and
hyperparameters setting. Finally, CBNet based on the well-known VGGNet, ResNet and
DenseNet is evaluated on a dataset including 41 major categories of food ingredients and
100 images for each category. Theoretical analysis and experimental results demonstrate
that CBNet achieves promising accuracy for multi-class classification and improves the
performance of convolutional neural networks. 相似文献
7.
Aim to countermeasure the presentation attack for iris recognition system, an iris liveness detection scheme based on batch normalized convolutional neural network (BNCNN) is proposed to improve the reliability of the iris authentication system. The BNCNN architecture with eighteen layers is constructed to detect the genuine iris and fake iris, including convolutional layer, batch-normalized (BN) layer, Relu layer, pooling layer and full connected layer. The iris image is first preprocessed by iris segmentation and is normalized to 256×256 pixels, and then the iris features are extracted by BNCNN. With these features, the genuine iris and fake iris are determined by the decision-making layer. Batch normalization technique is used in BNCNN to avoid the problem of over fitting and gradient disappearing during training. Extensive experiments are conducted on three classical databases: the CASIA Iris Lamp database, the CASIA Iris Syn database and Ndcontact database. The results show that the proposed method can effectively extract micro texture features of the iris, and achieve higher detection accuracy compared with some typical iris liveness detection methods. 相似文献
8.
As a common medium in our daily life, images are important for most people
to gather information. There are also people who edit or even tamper images to
deliberately deliver false information under different purposes. Thus, in digital forensics,
it is necessary to understand the manipulating history of images. That requires to verify
all possible manipulations applied to images. Among all the image editing manipulations,
recoloring is widely used to adjust or repaint the colors in images. The color information
is an important visual information that image can deliver. Thus, it is necessary to
guarantee the correctness of color in digital forensics. On the other hand, many image
retouching or editing applications or software are equipped with recoloring function. This
enables ordinary people without expertise of image processing to apply recoloring for
images. Hence, in order to secure the color information of images, in this paper, a
recoloring detection method is proposed. The method is based on convolutional neural
network which is quite popular in recent years. Unlike the traditional linear classifier, the
proposed method can be employed for binary classification as well as multiple labels
classification. The classification performance of different structure for the proposed
architecture is also investigated in this paper. 相似文献
9.
Sagheer Abbas Yousef Alhwaiti Areej Fatima Muhammad A. Khan Muhammad Adnan Khan Taher M. Ghazal Asma Kanwal Munir Ahmad Nouh Sabri Elmitwally 《计算机、材料和连续体(英文)》2022,70(3):4563-4581
This paper presents a handwritten document recognition system based on the convolutional neural network technique. In today’s world, handwritten document recognition is rapidly attaining the attention of researchers due to its promising behavior as assisting technology for visually impaired users. This technology is also helpful for the automatic data entry system. In the proposed system prepared a dataset of English language handwritten character images. The proposed system has been trained for the large set of sample data and tested on the sample images of user-defined handwritten documents. In this research, multiple experiments get very worthy recognition results. The proposed system will first perform image pre-processing stages to prepare data for training using a convolutional neural network. After this processing, the input document is segmented using line, word and character segmentation. The proposed system get the accuracy during the character segmentation up to 86%. Then these segmented characters are sent to a convolutional neural network for their recognition. The recognition and segmentation technique proposed in this paper is providing the most acceptable accurate results on a given dataset. The proposed work approaches to the accuracy of the result during convolutional neural network training up to 93%, and for validation that accuracy slightly decreases with 90.42%. 相似文献
10.
Gulzar Ahmed Tahir Alyas Muhammad Waseem Iqbal Muhammad Usman Ashraf Ahmed Mohammed Alghamdi Adel A. Bahaddad Khalid Ali Almarhabi 《计算机、材料和连续体(英文)》2022,73(2):2967-2984
Handwritten character recognition systems are used in every field of life nowadays, including shopping malls, banks, educational institutes, etc. Urdu is the national language of Pakistan, and it is the fourth spoken language in the world. However, it is still challenging to recognize Urdu handwritten characters owing to their cursive nature. Our paper presents a Convolutional Neural Networks (CNN) model to recognize Urdu handwritten alphabet recognition (UHAR) offline and online characters. Our research contributes an Urdu handwritten dataset (aka UHDS) to empower future works in this field. For offline systems, optical readers are used for extracting the alphabets, while diagonal-based extraction methods are implemented in online systems. Moreover, our research tackled the issue concerning the lack of comprehensive and standard Urdu alphabet datasets to empower research activities in the area of Urdu text recognition. To this end, we collected 1000 handwritten samples for each alphabet and a total of 38000 samples from 12 to 25 age groups to train our CNN model using online and offline mediums. Subsequently, we carried out detailed experiments for character recognition, as detailed in the results. The proposed CNN model outperformed as compared to previously published approaches. 相似文献
11.
Jieren Cheng Yifu Liu Xiangyan Tang Victor S. Sheng Mengyang Li Junqi Li 《计算机、材料和连续体(英文)》2020,62(3):1317-1333
Distributed Denial-of-Service (DDoS) has caused great damage to the network
in the big data environment. Existing methods are characterized by low computational
efficiency, high false alarm rate and high false alarm rate. In this paper, we propose a
DDoS attack detection method based on network flow grayscale matrix feature via multiscale convolutional neural network (CNN). According to the different characteristics of
the attack flow and the normal flow in the IP protocol, the seven-tuple is defined to
describe the network flow characteristics and converted into a grayscale feature by binary.
Based on the network flow grayscale matrix feature (GMF), the convolution kernel of
different spatial scales is used to improve the accuracy of feature segmentation, global
features and local features of the network flow are extracted. A DDoS attack classifier
based on multi-scale convolution neural network is constructed. Experiments show that
compared with correlation methods, this method can improve the robustness of the
classifier, reduce the false alarm rate and the missing alarm rate. 相似文献
12.
Image retrieval for food ingredients is important work, tremendously tiring, uninteresting, and expensive. Computer vision systems have extraordinary advancements in image retrieval with CNNs skills. But it is not feasible for small-size food datasets using convolutional neural networks directly. In this study, a novel image retrieval approach is presented for small and medium-scale food datasets, which both augments images utilizing image transformation techniques to enlarge the size of datasets, and promotes the average accuracy of food recognition with state-of-the-art deep learning technologies. First, typical image transformation techniques are used to augment food images. Then transfer learning technology based on deep learning is applied to extract image features. Finally, a food recognition algorithm is leveraged on extracted deep-feature vectors. The presented image-retrieval architecture is analyzed based on a small-scale food dataset which is composed of forty-one categories of food ingredients and one hundred pictures for each category. Extensive experimental results demonstrate the advantages of image-augmentation architecture for small and medium datasets using deep learning. The novel approach combines image augmentation, ResNet feature vectors, and SMO classification, and shows its superiority for food detection of small/medium-scale datasets with comprehensive experiments. 相似文献
13.
Recently, the effectiveness of neural networks, especially convolutional neural networks, has been validated in the field of natural language processing, in which, sentiment classification for online reviews is an important and challenging task. Existing convolutional neural networks extract important features of sentences without local features or the feature sequence. Thus, these models do not perform well, especially for transition sentences. To this end, we propose a Piecewise Pooling Convolutional Neural Network (PPCNN) for sentiment classification. Firstly, with a sentence presented by word vectors, convolution operation is introduced to obtain the convolution feature map vectors. Secondly, these vectors are segmented according to the positions of transition words in sentences. Thirdly, the most significant feature of each local segment is extracted using max pooling mechanism, and then the different aspects of features can be extracted. Specifically, the relative sequence of these features is preserved. Finally, after processed by the dropout algorithm, the softmax classifier is trained for sentiment classification. Experimental results show that the proposed method PPCNN is effective and superior to other baseline methods, especially for datasets with transition sentences. 相似文献
14.
目的 为了改善传统机器检测印刷产品缺陷存在误费率高的不足。方法 提出以卷积神经网络为控制核心的印刷品缺陷检测系统。设计可在实际检测中应用的卷积神经网络,设计在线印刷质量检测系统的硬件结构。结果 对结构相同而训练次数、学习率不同的卷积神经网络进行了缺陷检测的性能对比,验证了该卷积神经网络在学习率小于0.01时,可以获得较好的识别效果;在学习率大于0.05时,网络不容易收敛。网络训练次数越多,精度越高,相应的训练时间也较长。在满足快速性和精确度的条件下,确定了适应某印刷品的缺陷检验网络训练次数为50,学习率为0.005,此时的识别率为90%。结论 经过实验证明,该检测系统具有良好的缺陷识别能力,缺陷类型的分类准确率较高。该系统具有一定的实用价值。 相似文献
15.
The Convolutional Neural Network (CNN) is a widely used deep neural network.
Compared with the shallow neural network, the CNN network has better performance and
faster computing in some image recognition tasks. It can effectively avoid the problem that
network training falls into local extremes. At present, CNN has been applied in many
different fields, including fault diagnosis, and it has improved the level and efficiency of
fault diagnosis. In this paper, a two-streams convolutional neural network (TCNN) model is
proposed. Based on the short-time Fourier transform (STFT) spectral and Mel Frequency
Cepstrum Coefficient (MFCC) input characteristics of two-streams acoustic emission (AE)
signals, an AE signal processing and classification system is constructed and compared
with the traditional recognition methods of AE signals and traditional CNN networks. The
experimental results illustrate the effectiveness of the proposed model. Compared with
single-stream convolutional neural network and a simple Long Short-Term Memory
(LSTM) network, the performance of TCNN which combines spatial and temporal features
is greatly improved, and the accuracy rate can reach 100% on the current database, which is
12% higher than that of single-stream neural network. 相似文献
16.
提出了一种基于卷积神经网络(CNN)与门控循环单元(GRU)的垂直管道气液两相流流型识别方法。该方法基于电阻层析成像(ERT)系统的重建图像,对其填充处理后进行离散余弦变换(DCT),求取最大、最小 DCT 系数的差值,选取一定帧数长度数据作为网络输入,对流型进行识别。分析了输入序列长度对CNN-GRU、CNN 及 GRU 网络分类准确的影响,确定了最佳输入向量维度分别为 60、65 及 50,使用实验数据对3种网络进行训练、测试,结果表明,CNN-GRU网络分类准确率最高,平均流型识别准确率可达 99.40%。 相似文献
17.
18.
Saraswathi Sivamani Sun Il Chon Do Yeon Choi Dong Hoon Lee Ji Hwan Park 《计算机、材料和连续体(英文)》2022,72(3):4433-4452
Existing segmentation and augmentation techniques on convolutional neural network (CNN) has produced remarkable progress in object detection. However, the nominal accuracy and performance might be downturned with the photometric variation of images that are directly ignored in the training process, along with the context of the individual CNN algorithm. In this paper, we investigate the effect of a photometric variation like brightness and sharpness on different CNN. We observe that random augmentation of images weakens the performance unless the augmentation combines the weak limits of photometric variation. Our approach has been justified by the experimental result obtained from the PASCAL VOC 2007 dataset, with object detection CNN algorithms such as YOLOv3 (You Only Look Once), Faster R-CNN (Region-based CNN), and SSD (Single Shot Multibox Detector). Each CNN model shows performance loss for varying sharpness and brightness, ranging between −80% to 80%. It was further shown that compared to random augmentation, the augmented dataset with weak photometric changes delivered high performance, but the photometric augmentation range differs for each model. Concurrently, we discuss some research questions that benefit the direction of the study. The results prove the importance of adaptive augmentation for individual CNN model, subjecting towards the robustness of object detection. 相似文献
19.
20.
Juan Pablo Villa-Pulgarin erson Alberto Ruales-Torres Daniel Arias-Garzón Mario Alejandro Bravo-Ortiz Harold Brayan Arteaga-Arteaga Alejandro Mora-Rubio Jesus Alejandro Alzate-Grisales Esteban Mercado-Ruiz M. Hassaballah Simon Orozco-Arias Oscar Cardona-Morales Reinel Tabares-Soto 《计算机、材料和连续体(英文)》2022,70(2):2131-2148
Skin cancer is one of the most severe diseases, and medical imaging is among the main tools for cancer diagnosis. The images provide information on the evolutionary stage, size, and location of tumor lesions. This paper focuses on the classification of skin lesion images considering a framework of four experiments to analyze the classification performance of Convolutional Neural Networks (CNNs) in distinguishing different skin lesions. The CNNs are based on transfer learning, taking advantage of ImageNet weights. Accordingly, in each experiment, different workflow stages are tested, including data augmentation and fine-tuning optimization. Three CNN models based on DenseNet-201, Inception-ResNet-V2, and Inception-V3 are proposed and compared using the HAM10000 dataset. The results obtained by the three models demonstrate accuracies of 98%, 97%, and 96%, respectively. Finally, the best model is tested on the ISIC 2019 dataset showing an accuracy of 93%. The proposed methodology using CNN represents a helpful tool to accurately diagnose skin cancer disease. 相似文献