共查询到20条相似文献,搜索用时 31 毫秒
1.
Kazuko Takahashi Hiroya Takamura Manabu Okumura 《Knowledge and Information Systems》2009,19(2):185-210
Accurate estimation of class membership probability is needed for many applications in data mining and decision-making, to
which multiclass classification is often applied. Since existing methods for estimation of class membership probability are
designed for binary classification, in which only a single score outputted from a classifier can be used, an approach for
multiclass classification requires both a decomposition of a multiclass classifier into binary classifiers and a combination
of estimates obtained from each binary classifier to a target estimate. We propose a simple and general method for directly
estimating class membership probability for any class in multiclass classification without decomposition and combination,
using multiple scores not only for a predicted class but also for other proper classes. To make it possible to use multiple
scores, we propose to modify or extend representative existing methods. As a non-parametric method, which refers to the idea
of a binning method as proposed by Zadrozny et al., we create an “accuracy table” by a different method. Moreover we smooth
accuracies on the table with methods such as the moving average to yield reliable probabilities (accuracies). As a parametric
method, we extend Platt’s method to apply a multiple logistic regression. On two different datasets (open-ended data from
Japanese social surveys and the 20 Newsgroups) both with Support Vector Machines and naive Bayes classifiers, we empirically
show that the use of multiple scores is effective in the estimation of class membership probabilities in multiclass classification
in terms of cross entropy, the reliability diagram, the ROC curve and AUC (area under the ROC curve), and that the proposed
smoothing method for the accuracy table works quite well. Finally, we show empirically that in terms of MSE (mean squared
error), our best proposed method is superior to an expansion for multiclass classification of a PAV method proposed by Zadrozny
et al., in both the 20 Newsgroups dataset and the Pendigits dataset, but is slightly worse than the state-of-the-art method,
which is an expansion for multiclass classification of a combination of boosting and a PAV method, on the Pendigits dataset.
相似文献
Manabu OkumuraEmail: |
2.
Tatt Hee Oong Author VitaeNor Ashidi Mat IsaAuthor Vitae 《Applied Soft Computing》2012,12(4):1303-1308
This paper presents a new method called one-against-all ensemble for solving multiclass pattern classification problems. The proposed method incorporates a neural network ensemble into the one-against-all method to improve the generalization performance of the classifier. The experimental results show that the proposed method can reduce the uncertainty of the decision and it is comparable to the other widely used methods. 相似文献
3.
Adaptive binary tree for fast SVM multiclass classification 总被引:1,自引:0,他引:1
This paper presents an adaptive binary tree (ABT) to reduce the test computational complexity of multiclass support vector machine (SVM). It achieves a fast classification by: (1) reducing the number of binary SVMs for one classification by using separating planes of some binary SVMs to discriminate other binary problems; (2) selecting the binary SVMs with the fewest average number of support vectors (SVs). The average number of SVs is proposed to denote the computational complexity to exclude one class. Compared with five well-known methods, experiments on many benchmark data sets demonstrate our method can speed up the test phase while remain the high accuracy of SVMs. 相似文献
4.
Fault detection and diagnosis (FDD) in chemical process systems is an important tool for effective process monitoring to ensure the safety of a process. Multi-scale classification offers various advantages for monitoring chemical processes generally driven by events in different time and frequency domains. However, there are issues when dealing with highly interrelated, complex, and noisy databases with large dimensionality. Therefore, a new method for the FDD framework is proposed based on wavelet analysis, kernel Fisher discriminant analysis (KFDA), and support vector machine (SVM) classifiers. The main objective of this work was to combine the advantages of these tools to enhance the performance of the diagnosis on a chemical process system. Initially, a discrete wavelet transform (DWT) was applied to extract the dynamics of the process at different scales. The wavelet coefficients obtained during the analysis were reconstructed using the inverse discrete wavelet transform (IDWT) method, which were then fed into the KFDA to produce discriminant vectors. Finally, the discriminant vectors were used as inputs for the SVM classification task. The SVM classifiers were utilized to classify the feature sets extracted by the proposed method. The performance of the proposed multi-scale KFDA-SVM method for fault classification and diagnosis was analysed and compared using a simulated Tennessee Eastman process as a benchmark. The results showed the improvements of the proposed multiscale KFDA-SVM framework with an average 96.79% of classification accuracy over the multi-scale KFDA-GMM (84.94%), and the established independent component analysis-SVM method (95.78%) of the faults in the Tennessee Eastman process. 相似文献
5.
We extend extreme learning machine (ELM) classifiers to complex Reproducing Kernel Hilbert Spaces (RKHS) where the input/output variables as well as the optimization variables are complex-valued. A new family of classifiers, called complex-valued ELM (CELM) suitable for complex-valued multiple-input–multiple-output processing is introduced. In the proposed method, the associated Lagrangian is computed using induced RKHS kernels, adopting a Wirtinger calculus approach formulated as a constrained optimization problem similarly to the conventional ELM classifier formulation. When training the CELM, the Karush–Khun–Tuker (KKT) theorem is used to solve the dual optimization problem that consists of satisfying simultaneously smallest training error as well as smallest norm of output weights criteria. The proposed formulation also addresses aspects of quaternary classification within a Clifford algebra context. For 2D complex-valued inputs, user-defined complex-coupled hyper-planes divide the classifier input space into four partitions. For 3D complex-valued inputs, the formulation generates three pairs of complex-coupled hyper-planes through orthogonal projections. The six hyper-planes then divide the 3D space into eight partitions. It is shown that the CELM problem formulation is equivalent to solving six real-valued ELM tasks, which are induced by projecting the chosen complex kernel across the different user-defined coordinate planes. A classification example of powdered samples on the basis of their terahertz spectral signatures is used to demonstrate the advantages of the CELM classifiers compared to their SVM counterparts. The proposed classifiers retain the advantages of their ELM counterparts, in that they can perform multiclass classification with lower computational complexity than SVM classifiers. Furthermore, because of their ability to perform classification tasks fast, the proposed formulations are of interest to real-time applications. 相似文献
6.
The high dimensionality of microarray datasets endows the task of multiclass tissue classification with various difficulties—the
main challenge being the selection of features deemed relevant and non-redundant to form the predictor set for classifier
training. The necessity of varying the emphases on relevance and redundancy, through the use of the degree of differential
prioritization (DDP) during the search for the predictor set is also of no small importance. Furthermore, there are several
types of decomposition technique for the feature selection (FS) problem—all-classes-at-once, one-vs.-all (OVA) or pairwise
(PW). Also, in multiclass problems, there is the need to consider the type of classifier aggregation used—whether non-aggregated
(a single machine), or aggregated (OVA or PW). From here, first we propose a systematic approach to combining the distinct
problems of FS and classification. Then, using eight well-known multiclass microarray datasets, we empirically demonstrate
the effectiveness of the DDP in various combinations of FS decomposition types and classifier aggregation methods. Aided by
the variable DDP, feature selection leads to classification performance which is better than that of rank-based or equal-priorities
scoring methods and accuracies higher than previously reported for benchmark datasets with large number of classes. Finally,
based on several criteria, we make general recommendations on the optimal choice of the combination of FS decomposition type
and classifier aggregation method for multiclass microarray datasets. 相似文献
7.
8.
Simulated Annealing has been a very successful general algorithm for the solution of large, complex combinatorial optimization problems. Since its introduction, several applications in different fields of engineering, such as integrated circuit placement, optimal encoding, resource allocation, logic synthesis, have been developed. In parallel, theoretical studies have been focusing on the reasons for the excellent behavior of the algorithm. This paper reviews most of the important results on the theory of Simulated Annealing, placing them in a unified framework. New results are reported as well.This research was sponsored by SRC and DARPA monitored by SNWSC under contract numbers N00039-87-C-012 and N00039-88-C-0292. 相似文献
9.
This paper introduces a novel framework for 3D head model recognition based on the recently proposed 2D subspace analysis method. Two main contributions have been made. First, a 2D version of clustering-based discriminant analysis (CDA) is proposed, which combines the capability to model the multiple cluster structure embedded within a single class with the computational advantage that is characteristic of 2D subspace analysis methods. Second, we extend the applications of 2D subspace methods to the field of 3D head model classification by characterizing these models with 2D feature sets. 相似文献
10.
图分割问题是一种典型的NP-hard 问题, 如何对其进行高效求解一直都是学界和工业界的一个难题. 本文构建了一种新型的确定性退火控制算法, 提供了图分割问题的一种高质量近似解法. 算法主要由两部分构成: 全局收敛的迭代过程以及屏障函数最小点组成的收敛路径. 我们证明了,当屏障因子从足够大的实数降为0, 沿着一系列由屏障问题最小点组成的收敛路径可以得到图分割问题的一种高质量的近似解. 仿真计算结果表明本文所构建算法相比已有方法的优越性 相似文献
11.
Jigang Xie Author Vitae Zhengding Qiu Author Vitae Author Vitae Yanqiang Zhang Author Vitae 《Pattern recognition》2007,40(11):3292-3298
Many real-world classification tasks involve discriminations between two unbalanced classes in imprecise environments, in which either the training data do not represent a random sample of the target population or the class distribution may shift over time in the target population. In such situations, in order to minimize the misclassification costs, the class distribution in target population must be known for selecting the optimal threshold. Forman has presented a method, based on the distribution generated on training data and the distribution on unlabeled test data, for estimating the number of positives in target population. However, when the data size is small, it is difficult to reliably generate these distributions for estimating the number of positives. This paper presents a novel algorithm to generate these distributions based on the bootstrap and Fisher discriminant analysis. Experiment results on five UCI data sets demonstrate its effectiveness. 相似文献
12.
A novel model for Fisher discriminant analysis is developed in this paper. In the new model, maximal Fisher criterion values of discriminant vectors and minimal statistical correlation between feature components extracted by discriminant vectors are simultaneously required. Then the model is transformed into an extreme value problem, in the form of an evaluation function. Based on the evaluation function, optimal discriminant vectors are worked out. Experiments show that the method presented in this paper is comparative to the winner between FSLDA and ULDA. 相似文献
13.
José Carlos Castillo Davide Carneiro Juan Serrano-Cuerda Paulo Novais Antonio Fernández-Caballero José Neves 《International journal of systems science》2014,45(4):810-824
The society is changing towards a new paradigm in which an increasing number of old adults live alone. In parallel, the incidence of conditions that affect mobility and independence is also rising as a consequence of a longer life expectancy. In this paper, the specific problem of falls of old adults is addressed by devising a technological solution for monitoring these users. Video cameras, accelerometers and GPS sensors are combined in a multi-modal approach to monitor humans inside and outside the domestic environment. Machine learning techniques are used to detect falls and classify activities from accelerometer data. Video feeds and GPS are used to provide location inside and outside the domestic environment. It results in a monitoring solution that does not imply the confinement of the users to a closed environment. 相似文献
14.
15.
16.
Arthur Tenenhaus Alain Giron Michel Béra Bernard Fertil 《Computational statistics & data analysis》2007,51(9):4083-4100
“Kernel logistic PLS” (KL-PLS) is a new tool for supervised nonlinear dimensionality reduction and binary classification. The principles of KL-PLS are based on both PLS latent variables construction and learning with kernels. The KL-PLS algorithm can be seen as a supervised dimensionality reduction (complexity control step) followed by a classification based on logistic regression. The algorithm is applied to 11 benchmark data sets for binary classification and to three medical problems. In all cases, KL-PLS proved its competitiveness with other state-of-the-art classification methods such as support vector machines. Moreover, due to successions of regressions and logistic regressions carried out on only a small number of uncorrelated variables, KL-PLS allows handling high-dimensional data. The proposed approach is simple and easy to implement. It provides an efficient complexity control by dimensionality reduction and allows the visual inspection of data segmentation. 相似文献
17.
Abstract
The bag-of-words approach to text document representation
typically results in vectors of the order of 5000–20,000
components as the representation of documents. To make effective
use of various statistical classifiers, it may be necessary to
reduce the dimensionality of this representation. We point out
deficiencies in class discrimination of two popular such
methods, Latent Semantic Indexing (LSI), and sequential feature
selection according to some relevant criterion. As a remedy, we
suggest feature transforms based on Linear Discriminant Analysis
(LDA). Since LDA requires operating both with large and dense
matrices, we propose an efficient intermediate dimension
reduction step using either a random transform or LSI. We report
good classification results with the combined feature transform
on a subset of the Reuters-21578 database. Drastic reduction of
the feature vector dimensionality from 5000 to 12 actually
improves the classification performance.An erratum to this article can be found at 相似文献
18.
Daniela G. Calò 《Computational statistics & data analysis》2007,52(1):471-482
Gaussian mixture models (GMM) are commonly employed in nonparametric supervised classification. In high-dimensional problems it is often the case that information relevant to the separation of the classes is contained in a few directions. A GMM fitting procedure oriented to supervised classification is proposed, with the aim of reducing the number of free parameters. It resorts to projection pursuit as a dimension reduction method and combines it with GM modelling of class-conditional densities. In its derivation, issues regarding the forward and backward projection pursuit algorithms are discussed. The proposed procedure avoids the “curse of dimensionality”, is able to model structure in subspaces and regularizes the classification model. Its performance is illustrated on a simulation experiment and on a real data set, in comparison with other GMM-based classification methods. 相似文献
19.
针对文本情感分类准确率不高的问题,提出基于CCA-VSM分类器和KFD的多级文本情感分类方法。采用典型相关性分析对文档的权重特征向量和词性特征向量进行降维,在约简向量集上构建向量空间模型,根据模型之间的差异度设计VSM分类器,筛选出与测试文档差异度较小的R个模型作为核Fisher判别的输入,最终判别出文档的情感观点。实验结果表明:该方法比传统支持向量机有较高的分类准确率和较快的分类速度,权重特征和词性特征对分类准确率的影响较大。 相似文献
20.
A Naive Bayes approach for URL classification with supervised feature selection and rejection framework 下载免费PDF全文
Web page classification has become a challenging task due to the exponential growth of the World Wide Web. Uniform Resource Locator (URL)‐based web page classification systems play an important role, but high accuracy may not be achievable as URL contains minimal information. Nevertheless, URL‐based classifiers along with rejection framework can be used as a first‐level filter in a multistage classifier, and a costlier feature extraction from contents may be done in later stages. However, noisy and irrelevant features present in URL demand feature selection methods for URL classification. Therefore, we propose a supervised feature selection method by which relevant URL features are identified using statistical methods. We propose a new feature weighting method for a Naive Bayes classifier by embedding the term goodness obtained from the feature selection method. We also propose a rejection framework to the Naive Bayes classifier by using posterior probability for determining the confidence score. The proposed method is evaluated on the Open Directory Project and WebKB data sets. Experimental results show that our method can be an effective first‐level filter. McNemar tests confirm that our approach significantly improves the performance. 相似文献