首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 687 毫秒
1.
ABSTRACT

In this paper, we propose a robust subspace learning method, based on RPCA, named Robust Principal Component Analysis with Projection Learning (RPCAPL), which further improves the performance of feature extraction by projecting data samples into a suitable subspace. For Subspace Learning (SL) methods in clustering and classification tasks, it is also critical to construct an appropriate graph for discovering the intrinsic structure of the data. For this reason, we add a graph Laplacian matrix to the RPCAPL model for preserving the local geometric relationships between data samples and name the improved model as RPCAGPL, which takes all samples as nodes in the graph and treats affinity between pairs of connected samples as weighted edges. The RPCAGPL can not only globally capture the low-rank subspace structure of the data in the original space, but also locally preserve the neighbor relationship between the data samples.  相似文献   

2.
Software defect prediction plays a very important role in software quality assurance, which aims to inspect as many potentially defect-prone software modules as possible. However, the performance of the prediction model is susceptible to high dimensionality of the dataset that contains irrelevant and redundant features. In addition, software metrics for software defect prediction are almost entirely traditional features compared to the deep semantic feature representation from deep learning techniques. To address these two issues, we propose the following two solutions in this paper: (1) We leverage a novel non-linear manifold learning method - SOINN Landmark Isomap (SLIsomap) to extract the representative features by selecting automatically the reasonable number and position of landmarks, which can reveal the complex intrinsic structure hidden behind the defect data. (2) We propose a novel defect prediction model named DLDD based on hybrid deep learning techniques, which leverages denoising autoencoder to learn true input features that are not contaminated by noise, and utilizes deep neural network to learn the abstract deep semantic features. We combine the squared error loss function of denoising autoencoder with the cross entropy loss function of deep neural network to achieve the best prediction performance by adjusting a hyperparameter. We compare the SL-Isomap with seven state-of-the-art feature extraction methods and compare the DLDD model with six baseline models across 20 open source software projects. The experimental results verify that the superiority of SL-Isomap and DLDD on four evaluation indicators.  相似文献   

3.
Vehicle re-identification (ReID) aims to retrieve the target vehicle in an extensive image gallery through its appearances from various views in the cross-camera scenario. It has gradually become a core technology of intelligent transportation system. Most existing vehicle re-identification models adopt the joint learning of global and local features. However, they directly use the extracted global features, resulting in insufficient feature expression. Moreover, local features are primarily obtained through advanced annotation and complex attention mechanisms, which require additional costs. To solve this issue, a multi-feature learning model with enhanced local attention for vehicle re-identification (MFELA) is proposed in this paper. The model consists of global and local branches. The global branch utilizes both middle and high-level semantic features of ResNet50 to enhance the global representation capability. In addition, multi-scale pooling operations are used to obtain multi-scale information. While the local branch utilizes the proposed Region Batch Dropblock (RBD), which encourages the model to learn discriminative features for different local regions and simultaneously drops corresponding same areas randomly in a batch during training to enhance the attention to local regions. Then features from both branches are combined to provide a more comprehensive and distinctive feature representation. Extensive experiments on VeRi-776 and VehicleID datasets prove that our method has excellent performance.  相似文献   

4.
In the era of Big data, learning discriminant feature representation from network traffic is identified has as an invariably essential task for improving the detection ability of an intrusion detection system (IDS). Owing to the lack of accurately labeled network traffic data, many unsupervised feature representation learning models have been proposed with state-of-the-art performance. Yet, these models fail to consider the classification error while learning the feature representation. Intuitively, the learnt feature representation may degrade the performance of the classification task. For the first time in the field of intrusion detection, this paper proposes an unsupervised IDS model leveraging the benefits of deep autoencoder (DAE) for learning the robust feature representation and one-class support vector machine (OCSVM) for finding the more compact decision hyperplane for intrusion detection. Specially, the proposed model defines a new unified objective function to minimize the reconstruction and classification error simultaneously. This unique contribution not only enables the model to support joint learning for feature representation and classifier training but also guides to learn the robust feature representation which can improve the discrimination ability of the classifier for intrusion detection. Three set of evaluation experiments are conducted to demonstrate the potential of the proposed model. First, the ablation evaluation on benchmark dataset, NSL-KDD validates the design decision of the proposed model. Next, the performance evaluation on recent intrusion dataset, UNSW-NB15 signifies the stable performance of the proposed model. Finally, the comparative evaluation verifies the efficacy of the proposed model against recently published state-of-the-art methods.  相似文献   

5.
金梅  李媛媛  郝兴军  杨曼  张立国 《计量学报》2022,43(12):1573-1580
针对现有的行人重识别方法提取到的特征信息充分性与辨识性不足导致检索精度低的问题,提出一种基于非对称增强注意力与特征交叉融合的行人重识别方法。首先,构建非对称增强注意力模块,通过多重池化聚合的跨邻域通道交互注意力增强显著特征表示,使网络聚焦于图像中的行人区域;其次,考虑到网络各层特征间的差异性与关联性,构建特征交叉融合模块,利用交叉融合方式实现同层不同级特征的跨层级融合,进而实现多尺度融合;最后,水平切分输出特征以获取局部特征,从而实现在特定区域上描述行人。在Market1501、DukeMTMC-reID与CUHK03这3个公开数据集上对提出的方法进行了验证,首位命中率(Rank-1)分别达到了93.5%、85.1%和64.3%,证明了该方法在提升行人重识别性能上具有优越性。  相似文献   

6.
Recently, the effectiveness of neural networks, especially convolutional neural networks, has been validated in the field of natural language processing, in which, sentiment classification for online reviews is an important and challenging task. Existing convolutional neural networks extract important features of sentences without local features or the feature sequence. Thus, these models do not perform well, especially for transition sentences. To this end, we propose a Piecewise Pooling Convolutional Neural Network (PPCNN) for sentiment classification. Firstly, with a sentence presented by word vectors, convolution operation is introduced to obtain the convolution feature map vectors. Secondly, these vectors are segmented according to the positions of transition words in sentences. Thirdly, the most significant feature of each local segment is extracted using max pooling mechanism, and then the different aspects of features can be extracted. Specifically, the relative sequence of these features is preserved. Finally, after processed by the dropout algorithm, the softmax classifier is trained for sentiment classification. Experimental results show that the proposed method PPCNN is effective and superior to other baseline methods, especially for datasets with transition sentences.  相似文献   

7.
ABSTRACT

To detect oral tongue squamous cell carcinoma (OTSCC) using fibre optic Raman spectroscopy, we present a classification model based on convolutional neural networks (CNN) and support vector machines (SVM). 24 samples Raman spectra of OTSCC and para-carcinoma tissues from 12 patients were collected and analysed. In our proposed model, CNN is used as a feature extractor for forming a representative vector. Then the derived features are fed into an SVM classifier, which is used for OTSCC classification. Experimental results demonstrated that the area under the receiver operating characteristic curve was 99.96% and the classification error was zero (sensitivity: 99.54%, specificity: 99.54%). To show the superiority of this model, comparison results with the state-of-the-art methods showed it can obtain a competitive accuracy. These findings may pay a way to apply the proposed model in the fibre optic Raman instruments for intra-operative evaluation of OTSCC resection margins.  相似文献   

8.
倪俊帅  赵梅  胡长青 《声学技术》2020,39(3):366-371
为了改善分类系统的性能,进一步提高舰船辐射噪声分类的正确率,该文提出了一种基于深度神经网络的多特征融合分类方法。该方法首先提取舰船辐射噪声几种不同的特征,将提取的特征同时用于训练具有多个输入分支的深度神经网络,使网络直接在多种特征参数上进行联合学习,通过神经网络的输入分支和连接层实现特征融合,再对舰船辐射噪声进行分类。为了特征深度学习提取了舰船辐射噪声的频谱特征、梅尔倒谱系数和功率谱特征,并将多特征融合分类方法与在一种特征上进行深度学习分类方法的正确率进行对比。实验结果表明,基于深度学习的多特征融合分类方法可以有效地提高舰船辐射噪声分类的正确率,是一种可行的分类方法。  相似文献   

9.
针对语音情感识别中无法对关键的时空依赖关系进行建模,导致识别率低的问题,提出一种基于自身注意力(self-attention)时空特征的语音情感识别算法,利用双线性卷积神经网络、长短期记忆网络和多组注意力(multi-headattention)机制去自动学习语音信号的最佳时空表征。首先提取语音信号的对数梅尔(log-Mel)特征、一阶差分和二阶差分特征合成3D log-Mel特征集作为卷积神经网络的输入;然后综合考虑空间特征和时间依赖性关系,将双线性池化和双向长短期记忆网络的输出融合得到空间-时间特征表征,利用多组注意力机制捕获判别性强的特征;最后利用softmax函数进行分类。在IEMOCAP和EMO-DB数据库上进行实验,结果表明两种数据库的识别率分别为63.12%和87.09%,证明了此方法的有效性。  相似文献   

10.
针对传统旋转机械智能识别方法需要人为提取特征及诊断精度低的问题,基于深度学习的强大学习能力,提出一种深度卷积神经网络故障诊断模型(Deep Convolutional Neural Network Fault Diagnosis Model,DCNN-FDM)用于轴心轨迹识别。该模型包括输入模块、特征提取模块及分类模块三部分。原始图像输入模型后,经过输入模块的二值化处理及最近邻插值,统一变为尺寸大小为32×32的单通道图像;经特征提取模块中两组交替的卷积层和池化层作用,得到图形特征;最后,这些特征经全连接层的扁平化处理而张成一维向量,输入到softmax分类器中进行分类。利用奇异值差分谱方法,对实测轴心轨迹进行提纯,得到4类轴心轨迹样本集用于DCNN-FDM的训练与预测。结果表明:所提模型较传统的浅层学习模型的识别效果好,可实现转子故障的精确诊断,识别率达到97.09%。最后通过全连接层的主成分可视化分析,验证了模型具备自适应特征学习能力。  相似文献   

11.
The appearance of pedestrians can vary greatly from image to image, and different pedestrians may look similar in a given image. Such similarities and variabilities in the appearance and clothing of individuals make the task of pedestrian re-identification very challenging. Here, a pedestrian re-identification method based on the fusion of local features and gait energy image (GEI) features is proposed. In this method, the human body is divided into four regions according to joint points. The color and texture of each region of the human body are extracted as local features, and GEI features of the pedestrian gait are also obtained. These features are then fused with the local and GEI features of the person. Independent distance measure learning using the cross-view quadratic discriminant analysis (XQDA) method is used to obtain the similarity of the metric function of the image pairs, and the final similarity is acquired by weight matching. Evaluation of experimental results by cumulative matching characteristic (CMC) curves reveals that, after fusion of local and GEI features, the pedestrian reidentification effect is improved compared with existing methods and is notably better than the recognition rate of pedestrian re-identification with a single feature.  相似文献   

12.
Vehicle type classification is considered a central part of an intelligent traffic system. In recent years, deep learning had a vital role in object detection in many computer vision tasks. To learn high-level deep features and semantics, deep learning offers powerful tools to address problems in traditional architectures of handcrafted feature-extraction techniques. Unlike other algorithms using handcrated visual features, convolutional neural network is able to automatically learn good features of vehicle type classification. This study develops an optimized automatic surveillance and auditing system to detect and classify vehicles of different categories. Transfer learning is used to quickly learn the features by recording a small number of training images from vehicle frontal view images. The proposed system employs extensive data-augmentation techniques for effective training while avoiding the problem of data shortage. In order to capture rich and discriminative information of vehicles, the convolutional neural network is fine-tuned for the classification of vehicle types using the augmented data. The network extracts the feature maps from the entire dataset and generates a label for each object (vehicle) in an image, which can help in vehicle-type detection and classification. Experimental results on a public dataset and our own dataset demonstrated that the proposed method is quite effective in detection and classification of different types of vehicles. The experimental results show that the proposed model achieves 96.04% accuracy on vehicle type classification.  相似文献   

13.
刘国庆  方成刚  黄德军  龙超 《包装工程》2023,44(17):197-205
目的 针对试剂卡生产企业采用人工分选印刷缺陷的试剂卡存在效率低、成本高、易漏检的问题,提出一种基于深度神经网络YOLOv5s的改进试剂卡印刷缺陷检测算法YOLOv5s-EF。方法 通过图像预处理算法获得高质量的缺陷图像数据集,在YOLOv5s的主干特征提取网络中添加高效通道注意力(Efficient Channel Attention, ECA)机制,增强特征图中重要特征的表示能力;引入焦点损失函数(Focal Loss)来缓解正负样本不均衡的影响;结合印刷区域的定位结果,二次精确定位并构建方位特征向量,提出一种特征向量相似度匹配方法。结果 实验结果表明,本文提出的试剂卡印刷缺陷检测算法在测试集上的检测平均准确度可以达到97.3%,速度为22.6帧/s。结论 相较于其他网络模型,本文提出的方法可以实现对多种印刷缺陷的识别与定位,模型具有较好的检测速度和鲁棒性,有利于提高企业生产的智能化水平。  相似文献   

14.
Recently, the Erebus attack has proved to be a security threat to the blockchain network layer, and the existing research has faced challenges in detecting the Erebus attack on the blockchain network layer. The cloud-based active defense and one-sidedness detection strategies are the hindrances in detecting Erebus attacks. This study designs a detection approach by establishing a ReliefF_WMRmR-based two-stage feature selection algorithm and a deep learning-based multimodal classification detection model for Erebus attacks and responding to security threats to the blockchain network layer. The goal is to improve the performance of Erebus attack detection methods, by combining the traffic behavior with the routing status based on multimodal deep feature learning. The traffic behavior and routing status were first defined and used to describe the attack characteristics at diverse stages of s leak monitoring, hidden traffic overlay, and transaction identity forgery. The goal is to clarify how an Erebus attack affects the routing transfer and traffic state on the blockchain network layer. Consequently, detecting objects is expected to become more relevant and sensitive. A two-stage feature selection algorithm was designed based on ReliefF and weighted maximum relevance minimum redundancy (ReliefF_WMRmR) to alleviate the overfitting of the training model caused by redundant information and noise in multiple source features of the routing status and traffic behavior. The ReliefF algorithm was introduced to select strong correlations and highly informative features of the labeled data. According to WMRmR, a feature selection framework was defined to eliminate weakly correlated features, eliminate redundant information, and reduce the detection overhead of the model. A multimodal deep learning model was constructed based on the multilayer perceptron (MLP) to settle the high false alarm rates incurred by multisource data. Using this model, isolated inputs and deep learning were conducted on the selected routing status and traffic behavior. Redundant intermodal information was removed because of the complementarity of the multimodal network, which was followed by feature fusion and output feature representation to boost classification detection precision. The experimental results demonstrate that the proposed method can detect features, such as traffic data, at key link nodes and route messages in a real blockchain network environment. Additionally, the model can detect Erebus attacks effectively. This study provides novelty to the existing Erebus attack detection by increasing the accuracy detection by 1.05%, the recall rate by 2.01%, and the F1-score by 2.43%.  相似文献   

15.
With the development of deep learning and Convolutional Neural Networks (CNNs), the accuracy of automatic food recognition based on visual data have significantly improved. Some research studies have shown that the deeper the model is, the higher the accuracy is. However, very deep neural networks would be affected by the overfitting problem and also consume huge computing resources. In this paper, a new classification scheme is proposed for automatic food-ingredient recognition based on deep learning. We construct an up-to-date combinational convolutional neural network (CBNet) with a subnet merging technique. Firstly, two different neural networks are utilized for learning interested features. Then, a well-designed feature fusion component aggregates the features from subnetworks, further extracting richer and more precise features for image classification. In order to learn more complementary features, the corresponding fusion strategies are also proposed, including auxiliary classifiers and hyperparameters setting. Finally, CBNet based on the well-known VGGNet, ResNet and DenseNet is evaluated on a dataset including 41 major categories of food ingredients and 100 images for each category. Theoretical analysis and experimental results demonstrate that CBNet achieves promising accuracy for multi-class classification and improves the performance of convolutional neural networks.  相似文献   

16.
17.
In the area of medical image processing, stomach cancer is one of the most important cancers which need to be diagnose at the early stage. In this paper, an optimized deep learning method is presented for multiple stomach disease classification. The proposed method work in few important steps—preprocessing using the fusion of filtering images along with Ant Colony Optimization (ACO), deep transfer learning-based features extraction, optimization of deep extracted features using nature-inspired algorithms, and finally fusion of optimal vectors and classification using Multi-Layered Perceptron Neural Network (MLNN). In the feature extraction step, pre-trained Inception V3 is utilized and retrained on selected stomach infection classes using the deep transfer learning step. Later on, the activation function is applied to Global Average Pool (GAP) for feature extraction. However, the extracted features are optimized through two different nature-inspired algorithms—Particle Swarm Optimization (PSO) with dynamic fitness function and Crow Search Algorithm (CSA). Hence, both methods’ output is fused by a maximal value approach and classified the fused feature vector by MLNN. Two datasets are used to evaluate the proposed method—CUI WahStomach Diseases and Combined dataset and achieved an average accuracy of 99.5%. The comparison with existing techniques, it is shown that the proposed method shows significant performance.  相似文献   

18.
行人检测系统涉及交通安全问题,需要很高的鲁棒性,基于单特征结合单核支持向量机的方法效果有限,为解决这一问题,提出采用多特征和多核学习的方法来提升系统的鲁棒性,通过将积分信道特征、多层次导向边缘能量特征和CENTRIST特征分别与直方图交叉核、高斯核和多项式核进行线性组合,采用简单多核学习(Simple MKL)来分别计算核函数的权重系数,将多核学习方法与经典的梯度直方图特征/支持向量机、多尺度梯度直方图特征/直方图交叉核支持向量机和特征融合/直方图交叉核支持向量机的行人检测方法进行比较,实验表明所提出的行人检测算法的鲁棒性有明显提升。  相似文献   

19.
The text classification process has been extensively investigated in various languages, especially English. Text classification models are vital in several Natural Language Processing (NLP) applications. The Arabic language has a lot of significance. For instance, it is the fourth mostly-used language on the internet and the sixth official language of the United Nations. However, there are few studies on the text classification process in Arabic. A few text classification studies have been published earlier in the Arabic language. In general, researchers face two challenges in the Arabic text classification process: low accuracy and high dimensionality of the features. In this study, an Automated Arabic Text Classification using Hyperparameter Tuned Hybrid Deep Learning (AATC-HTHDL) model is proposed. The major goal of the proposed AATC-HTHDL method is to identify different class labels for the Arabic text. The first step in the proposed model is to pre-process the input data to transform it into a useful format. The Term Frequency-Inverse Document Frequency (TF-IDF) model is applied to extract the feature vectors. Next, the Convolutional Neural Network with Recurrent Neural Network (CRNN) model is utilized to classify the Arabic text. In the final stage, the Crow Search Algorithm (CSA) is applied to fine-tune the CRNN model’s hyperparameters, showing the work’s novelty. The proposed AATC-HTHDL model was experimentally validated under different parameters and the outcomes established the supremacy of the proposed AATC-HTHDL model over other approaches.  相似文献   

20.
Human action recognition under complex environment is a challenging work. Recently, sparse representation has achieved excellent results of dealing with human action recognition problem under different conditions. The main idea of sparse representation classification is to construct a general classification scheme where the training samples of each class can be considered as the dictionary to express the query class, and the minimal reconstruction error indicates its corresponding class. However, how to learn a discriminative dictionary is still a difficult work. In this work, we make two contributions. First, we build a new and robust human action recognition framework by combining one modified sparse classification model and deep convolutional neural network (CNN) features. Secondly, we construct a novel classification model which consists of the representation-constrained term and the coefficients incoherence term. Experimental results on benchmark datasets show that our modified model can obtain competitive results in comparison to other state-of-the-art models.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号