?ukasiewicz presented two different analyses of modal notions by means of many-valued logics: (1) the linearly ordered systems ?3,..., Open image in new window ,..., \(\hbox {L}_{\omega }\); (2) the 4-valued logic ? he defined in the last years of his career. Unfortunately, all these systems contain “?ukasiewicz type (modal) paradoxes”. On the other hand, Brady’s 4-valued logic BN4 is the basic 4-valued bilattice logic. The aim of this paper is to show that BN4 can be strengthened with modal operators following ?ukasiewicz’s strategy for defining truth-functional modal logics. The systems we define lack “?ukasiewicz type paradoxes”. Following Brady, we endow them with Belnap–Dunn type bivalent semantics. 相似文献
In this paper, we propose the problem of online cost-sensitive classifier adaptation and the first algorithm to solve it. We assume that we have a base classifier for a cost-sensitive classification problem, but it is trained with respect to a cost setting different to the desired one. Moreover, we also have some training data samples streaming to the algorithm one by one. The problem is to adapt the given base classifier to the desired cost setting using the steaming training samples online. To solve this problem, we propose to learn a new classifier by adding an adaptation function to the base classifier, and update the adaptation function parameter according to the streaming data samples. Given an input data sample and the cost of misclassifying it, we update the adaptation function parameter by minimizing cost-weighted hinge loss and respecting previous learned parameter simultaneously. The proposed algorithm is compared to both online and off-line cost-sensitive algorithms on two cost-sensitive classification problems, and the experiments show that it not only outperforms them on classification performances, but also requires significantly less running time.
Requirements analysis is the software engineering stage that is closest to the users’ world. It also involves tasks that are knowledge intensive. Thus, the use of Bayesian networks (BNs) to model this knowledge would be a valuable aid. These probabilistic models could manage the imprecision and ambiguities usually present in requirements engineering (RE). In this work, we conduct a literature review focusing on where and how BNs are applied on subareas of RE in order to identify which gaps remain uncovered and which methods might engineers employ to incorporate this intelligent technique into their own requirements processes. The scarcity of identified studies (there are only 20) suggests that not all RE areas have been properly investigated in the literature. The evidence available for adopting BNs into RE is sufficiently mature yet the methods applied are not easily translatable to other topics. Nonetheless, there are enough studies supporting the applicability of synergistic cooperation between RE and BNs. This work provides a background for understanding the current state of research encompassing RE and BNs. Functional, non-functional and -ilities requirements artifacts are enhanced by the use of BNs. These models were obtained by interacting with experts or by learning from databases. The most common criticism from the point of view of BN experts is that the models lack validation, whereas requirements engineers point to the lack of a clear application method for BNs and the lack of tools for incorporating them as built-in help functions. 相似文献
The problem of district design for the implementation of arc routing activities is addressed. The aim is to partition a road network into a given number of sectors to facilitate the organization of the operations to be implemented within the region. This problem arises in numerous applications such as postal delivery, meter readings, winter gritting, road maintenance, and municipal solid waste collection. An integer linear programming model is proposed where a novel set of node parity constraints to favor Eulerian districts is introduced. Series of instances were solved to assess the impact of these parity constraints on the objective function and deadhead distance. Networks with up to 401 nodes and 764 edges were successfully solved. The model is useful at a tactical level as it can be used to promote workload balance, compactness, deadhead distance reduction and parity in districts. 相似文献
The performance of state-of-the-art speaker verification in uncontrolled environment is affected by different variabilities. Short duration variability is very common in these scenarios and causes the speaker verification performance to decrease quickly while the duration of verification utterances decreases. Linear discriminant analysis (LDA) is the most common session variability compensation algorithm, nevertheless it presents some shortcomings when trained with insufficient data. In this paper we introduce two methods for session variability compensation to deal with short-length utterances on i-vector space. The first method proposes to incorporate the short duration variability information in the within-class variance estimation process. The second proposes to compensate the session and short duration variabilities in two different spaces with LDA algorithms (2S-LDA). First, we analyzed the behavior of the within and between class scatters in the first proposed method. Then, both proposed methods are evaluated on telephone session from NIST SRE-08 for different duration of the evaluation utterances: full (average 2.5 min), 20, 15, 10 and 5 s. The 2S-LDA method obtains good results on different short-length utterances conditions in the evaluations, with a EER relative average improvement of 1.58%, compared to the best baseline (WCCN[LDA]). Finally, we applied the 2S-LDA method in speaker verification under reverberant environment, using different reverberant conditions from Reverb challenge 2013, obtaining an improvement of 8.96 and 23% under matched and mismatched reverberant conditions, respectively. 相似文献
We introduce mobile agents for mobile crowdsensing. Crowdsensing campaigns are designed through different roles that are implemented as mobile agents. The role-based tasks of mobile agents include collecting data, analyzing data and sharing data in the campaign. Mobile agents execute and control the campaign autonomously as a multi-agent system and migrate in the opportunistic network of participants’ devices. Mobile agents take into account the available resources in the devices and match participants’ privacy requirements to the campaign requirements. Sharing of task results in real-time facilitates cooperation towards the campaign goal while maintaining a selected global measure, such as energy efficiency. We discuss current challenges in crowdsensing and propose mobile agent based solutions for campaign execution and monitoring, addressing data collection and participant-related issues. We present a software framework for mobile agents-based crowdsensing that is seamlessly integrated into the Web. A set of simulations are conducted to compare mobile agent-based campaigns with existing crowdsensing approaches. We implemented and evaluated a small-scale real-world mobile agent based campaign for pedestrian flock detection. The simulation and evaluation results show that mobile agent based campaigns produce comparable results with less energy consumption when the number of agents is relatively small and enables in-network data processing with sharing of data and task results with insignificant overhead. 相似文献
An important question for the upcoming Semantic Web is how to best combine open world ontology languages, such as the OWL-based ones, with closed world rule-based languages. One of the most mature proposals for this combination is known as hybrid MKNF knowledge bases (Motik and Rosati, 2010 [52]), and it is based on an adaptation of the Stable Model Semantics to knowledge bases consisting of ontology axioms and rules. In this paper we propose a well-founded semantics for nondisjunctive hybrid MKNF knowledge bases that promises to provide better efficiency of reasoning, and that is compatible with both the OWL-based semantics and the traditional Well-Founded Semantics for logic programs. Moreover, our proposal allows for the detection of inconsistencies, possibly occurring in tightly integrated ontology axioms and rules, with only little additional effort. We also identify tractable fragments of the resulting language. 相似文献
Higman??s lemma is an important result in infinitary combinatorics, which has been formalized in several theorem provers. In this paper we present a formalization and proof of Higman??s Lemma in the ACL2 theorem prover. Our formalization is based on a proof by Murthy and Russell, where the key termination argument is justified by the multiset relation induced by a well-founded relation. To our knowledge, this is the first mechanization of this proof. 相似文献
In this work we develop a procedure to deform a given surface triangulation to obtain its alignment with interior curves.
These curves are defined by splines in a parametric space and, subsequently, mapped to the surface triangulation. We have
restricted our study to orthogonal mapping, so we require the curves to be included in a patch of the surface that can be
orthogonally projected onto a plane (our parametric space). For example, the curves can represent interfaces between different
materials or boundary conditions, internal boundaries or feature lines. Another setting in which this procedure can be used
is the adaption of a reference mesh to changing curves in the course of an evolutionary process. Specifically, we propose
a new method that moves the nodes of the mesh, maintaining its topology, in order to achieve two objectives simultaneously:
the piecewise approximation of the curves by edges of the surface triangulation and the optimization of the resulting mesh.
We will designate this procedure as projecting/smoothing method and it is based on the smoothing technique that we have introduced for surface triangulations in previous works. The
mesh quality improvement is obtained by an iterative process where each free node is moved to a new position that minimizes a certain objective function. The minimization process is done on the parametric
plane attending to the surface piece-wise approximation and to an algebraic quality measure (mean ratio) of the set of triangles that are connected to the free node. So, the 3-D local projecting/smoothing problem is reduced to a 2-D optimization problem. Several applications of this method
are presented. 相似文献