共查询到18条相似文献,搜索用时 93 毫秒
1.
2.
随着人工智能发展,语音情感识别技术的应用范围越来越广。该文以语音情感识别为出发点,介绍了语音信号特征提取方法和语音情感分类模型训练过程中中权值和参数更新的算法,并在tensorflow框架中进行试验设计和试验,通过试验分析激活函数、中间层层数、训练轮次对模型训练结果的影响。试验结果表明,当训练轮次为1 000轮、中间层层数为6个且激活函数选择elu时判定准确率较高。 相似文献
3.
目的针对传统无纺布缺陷分类检测中人工依赖性强、效率低等问题,提出一种能够满足工厂要求的卷积神经网络分类检测方法。方法首先建立包括脏点、褶皱、断裂、缺纱和无缺陷等5种共计7万张无纺布图像样本库,其次构造一个具有不同神经元个数的卷积层和池化层的神经网络,然后采用反向传播算法逐层更新权值,通过梯度下降法最小化损失函数,最后利用Softmax分类器实现无纺布的缺陷分类检测。结果构建了12层的卷积神经网络,通过2万张样本进行测试实验,无缺陷样本准确率可以达到100%,缺陷样本分类准确率均在95%以上,检测时间在35 ms以内。结论该方法能够满足工业生产线中对于无纺布缺陷实时分类检测的要求。 相似文献
4.
5.
6.
目的 为了改善传统机器检测印刷产品缺陷存在误费率高的不足。方法 提出以卷积神经网络为控制核心的印刷品缺陷检测系统。设计可在实际检测中应用的卷积神经网络,设计在线印刷质量检测系统的硬件结构。结果 对结构相同而训练次数、学习率不同的卷积神经网络进行了缺陷检测的性能对比,验证了该卷积神经网络在学习率小于0.01时,可以获得较好的识别效果;在学习率大于0.05时,网络不容易收敛。网络训练次数越多,精度越高,相应的训练时间也较长。在满足快速性和精确度的条件下,确定了适应某印刷品的缺陷检验网络训练次数为50,学习率为0.005,此时的识别率为90%。结论 经过实验证明,该检测系统具有良好的缺陷识别能力,缺陷类型的分类准确率较高。该系统具有一定的实用价值。 相似文献
7.
8.
9.
为了满足疫情形势下公共场所的智能化检测需求,该文提出了一种基于深度学习的人脸佩戴口罩识别方法,并对该方法进行识别准确率测试。该方法在可移动设备中应用,能够实现图像采集、图像处理、图像分析和结果输出的功能,其中图像分析方法是基于DarkNet-35卷积神经网络构建模型的,在主干网络中该模型采用多尺度特征学习及融合的构建模式。试验证明该模型可以检测侧拍、光线暗和佩戴不规范等问题,在测试口罩佩戴数据集上的检测精度为96.5%,检测速度为67 F/s。该口罩检测方案具有可移动、低成本以及易于部署等特点,可以满足多场所的检测需求。 相似文献
10.
目的为实现饮料易拉罐拉环背部激光打码的自动化,提出一种基于遗传算法的易拉罐罐盖图像识别新方法。方法首先搭建一套易拉罐盖激光自动打码机,基于所搭建的实验系统,利用CCD相机实时采集罐盖图像。对所采集到的图像进行中值滤波和灰度增强处理,在此基础上,研究基于遗传算法的罐盖图像阈值分割新方法,分析、确定算法的关键参数(个体数目、交叉率、变异率等),由此得到罐盖的二值化图像,并对算法处理结果进行误差分析。结果遗传算法经过约15代的迭代计算,能够收敛,获取到最优的图像阈值,整个算法的运行时间约30 ms,最终的图像精度约为7.9 pixel。结论基于遗传算法的图像阈值分割实时性好,分割后的图像精度高,与传统的Ostu阈值分割法相比,得到的信息更加丰厚,能抑制光线不均所造成的图像干扰。同时对遗传算法阈值分割后的图像进行了sobel边缘检测,得到了清晰的罐盖边缘,为激光打码的准确定位奠定了基础。 相似文献
11.
目的 针对目前印刷套准识别方法依赖于经验人工设计特征提取的问题,提出一种不需要人工提取图像特征的卷积神经网络模型,实现印刷套准状态的识别.方法 采用图像增强技术实现不均衡训练集的均衡化,增加训练集图像的数量,提高模型的识别准确率.设计基于AlexNet网络结构的印刷套准识别模型的结构参数,分析批处理样本数量和基础学习率对模型性能的影响规律.结果 文中方法获得的总印刷套准识别准确率为0.9860,召回率为1.0000,分类准确率几何平均数为0.9869.结论 文中方法能自动提取图像特征,不依赖于人工设计的特征提取方法.在构造的数据集上,文中方法的分类性能优于实验中的支持向量机方法. 相似文献
12.
Nianbin Wang Ming He Jianguo Sun Hongbin Wang Lianke Zhou Ci Chu Lei Chen 《计算机、材料和连续体(英文)》2019,58(1):169-181
Underwater target recognition is a key technology for underwater acoustic countermeasure. How to classify and recognize underwater targets according to the noise information of underwater targets has been a hot topic in the field of underwater acoustic signals. In this paper, the deep learning model is applied to underwater target recognition. Improved anti-noise Power-Normalized Cepstral Coefficients (ia-PNCC) is proposed, based on PNCC applied to underwater noises. Multitaper and normalized Gammatone filter banks are applied to improve the anti-noise capacity. The method is combined with a convolutional neural network in order to recognize the underwater target. Experiment results show that the acoustic feature presented by ia-PNCC has lower noise and are well-suited to underwater target recognition using a convolutional neural network. Compared with the combination of convolutional neural network with single acoustic feature, such as MFCC (Mel-scale Frequency Cepstral Coefficients) or LPCC (Linear Prediction Cepstral Coefficients), the combination of the ia-PNCC with a convolutional neural network offers better accuracy for underwater target recognition. 相似文献
13.
Based on the theory of modal acoustic emission (AE), when the convolutional neural network (CNN) is used to identify rotor rub-impact faults, the training data has a small sample size, and the AE sound segment belongs to a single channel signal with less pixel-level information and strong local correlation. Due to the convolutional pooling operations of CNN, coarse-grained and edge information are lost, and the top-level information dimension in CNN network is low, which can easily lead to overfitting. To solve the above problems, we first propose the use of sound spectrograms and their differential features to construct multi-channel image input features suitable for CNN and fully exploit the intrinsic characteristics of the sound spectra. Then, the traditional CNN network structure is improved, and the outputs of all convolutional layers are connected as one layer constitutes a fused feature that contains information at each layer, and is input into the network’s fully connected layer for classification and identification. Experiments indicate that the improved CNN recognition algorithm has significantly improved recognition rate compared with CNN and dynamical neural network (DNN) algorithms. 相似文献
14.
Fawaz Waselallah Alsaade Theyazn H. H. Aldhyani Mosleh Hmoud Al-Adhaileh 《计算机、材料和连续体(英文)》2021,68(1):805-819
The COVID-19 pandemic poses an additional serious public health threat due to little or no pre-existing human immunity, and developing a system to identify COVID-19 in its early stages will save millions of lives. This study applied support vector machine (SVM), k-nearest neighbor (K-NN) and deep learning convolutional neural network (CNN) algorithms to classify and detect COVID-19 using chest X-ray radiographs. To test the proposed system, chest X-ray radiographs and CT images were collected from different standard databases, which contained 95 normal images, 140 COVID-19 images and 10 SARS images. Two scenarios were considered to develop a system for predicting COVID-19. In the first scenario, the Gaussian filter was applied to remove noise from the chest X-ray radiograph images, and then the adaptive region growing technique was used to segment the region of interest from the chest X-ray radiographs. After segmentation, a hybrid feature extraction composed of 2D-DWT and gray level co-occurrence matrix was utilized to extract the features significant for detecting COVID-19. These features were processed using SVM and K-NN. In the second scenario, a CNN transfer model (ResNet 50) was used to detect COVID-19. The system was examined and evaluated through multiclass statistical analysis, and the empirical results of the analysis found significant values of 97.14%, 99.34%, 99.26%, 99.26% and 99.40% for accuracy, specificity, sensitivity, recall and AUC, respectively. Thus, the CNN model showed significant success; it achieved optimal accuracy, effectiveness and robustness for detecting COVID-19. 相似文献
15.
Jieren Cheng Yifu Liu Xiangyan Tang Victor S. Sheng Mengyang Li Junqi Li 《计算机、材料和连续体(英文)》2020,62(3):1317-1333
Distributed Denial-of-Service (DDoS) has caused great damage to the network
in the big data environment. Existing methods are characterized by low computational
efficiency, high false alarm rate and high false alarm rate. In this paper, we propose a
DDoS attack detection method based on network flow grayscale matrix feature via multiscale convolutional neural network (CNN). According to the different characteristics of
the attack flow and the normal flow in the IP protocol, the seven-tuple is defined to
describe the network flow characteristics and converted into a grayscale feature by binary.
Based on the network flow grayscale matrix feature (GMF), the convolution kernel of
different spatial scales is used to improve the accuracy of feature segmentation, global
features and local features of the network flow are extracted. A DDoS attack classifier
based on multi-scale convolution neural network is constructed. Experiments show that
compared with correlation methods, this method can improve the robustness of the
classifier, reduce the false alarm rate and the missing alarm rate. 相似文献
16.
Gulzar Ahmed Tahir Alyas Muhammad Waseem Iqbal Muhammad Usman Ashraf Ahmed Mohammed Alghamdi Adel A. Bahaddad Khalid Ali Almarhabi 《计算机、材料和连续体(英文)》2022,73(2):2967-2984
Handwritten character recognition systems are used in every field of life nowadays, including shopping malls, banks, educational institutes, etc. Urdu is the national language of Pakistan, and it is the fourth spoken language in the world. However, it is still challenging to recognize Urdu handwritten characters owing to their cursive nature. Our paper presents a Convolutional Neural Networks (CNN) model to recognize Urdu handwritten alphabet recognition (UHAR) offline and online characters. Our research contributes an Urdu handwritten dataset (aka UHDS) to empower future works in this field. For offline systems, optical readers are used for extracting the alphabets, while diagonal-based extraction methods are implemented in online systems. Moreover, our research tackled the issue concerning the lack of comprehensive and standard Urdu alphabet datasets to empower research activities in the area of Urdu text recognition. To this end, we collected 1000 handwritten samples for each alphabet and a total of 38000 samples from 12 to 25 age groups to train our CNN model using online and offline mediums. Subsequently, we carried out detailed experiments for character recognition, as detailed in the results. The proposed CNN model outperformed as compared to previously published approaches. 相似文献
17.
As a common medium in our daily life, images are important for most people
to gather information. There are also people who edit or even tamper images to
deliberately deliver false information under different purposes. Thus, in digital forensics,
it is necessary to understand the manipulating history of images. That requires to verify
all possible manipulations applied to images. Among all the image editing manipulations,
recoloring is widely used to adjust or repaint the colors in images. The color information
is an important visual information that image can deliver. Thus, it is necessary to
guarantee the correctness of color in digital forensics. On the other hand, many image
retouching or editing applications or software are equipped with recoloring function. This
enables ordinary people without expertise of image processing to apply recoloring for
images. Hence, in order to secure the color information of images, in this paper, a
recoloring detection method is proposed. The method is based on convolutional neural
network which is quite popular in recent years. Unlike the traditional linear classifier, the
proposed method can be employed for binary classification as well as multiple labels
classification. The classification performance of different structure for the proposed
architecture is also investigated in this paper. 相似文献
18.
Image retrieval for food ingredients is important work, tremendously tiring, uninteresting, and expensive. Computer vision systems have extraordinary advancements in image retrieval with CNNs skills. But it is not feasible for small-size food datasets using convolutional neural networks directly. In this study, a novel image retrieval approach is presented for small and medium-scale food datasets, which both augments images utilizing image transformation techniques to enlarge the size of datasets, and promotes the average accuracy of food recognition with state-of-the-art deep learning technologies. First, typical image transformation techniques are used to augment food images. Then transfer learning technology based on deep learning is applied to extract image features. Finally, a food recognition algorithm is leveraged on extracted deep-feature vectors. The presented image-retrieval architecture is analyzed based on a small-scale food dataset which is composed of forty-one categories of food ingredients and one hundred pictures for each category. Extensive experimental results demonstrate the advantages of image-augmentation architecture for small and medium datasets using deep learning. The novel approach combines image augmentation, ResNet feature vectors, and SMO classification, and shows its superiority for food detection of small/medium-scale datasets with comprehensive experiments. 相似文献