The ever-increasing size of data emanating from mobile devices and sensors, dictates the use of distributed systems for storing and querying these data. Typically, such data sources provide some spatio-temporal information, alongside other useful data. The RDF data model can be used to interlink and exchange data originating from heterogeneous sources in a uniform manner. For example, consider the case where vessels report their spatio-temporal position, on a regular basis, by using various surveillance systems. In this scenario, a user might be interested to know which vessels were moving in a specific area for a given temporal range. In this paper, we address the problem of efficiently storing and querying spatio-temporal RDF data in parallel. We specifically study the case of SPARQL queries with spatio-temporal constraints, by proposing the DiStRDF system, which is comprised of a Storage and a Processing Layer. The DiStRDF Storage Layer is responsible for efficiently storing large amount of historical spatio-temporal RDF data of moving objects. On top of it, we devise our DiStRDF Processing Layer, which parses a SPARQL query and produces corresponding logical and physical execution plans. We use Spark, a well-known distributed in-memory processing framework, as the underlying processing engine. Our experimental evaluation, on real data from both aviation and maritime domains, demonstrates the efficiency of our DiStRDF system, when using various spatio-temporal range constraints.
Through the Preservation of Complex Objects Symposia (POCOS), leading researchers and practitioners have managed to propose strategies for preserving digital art and computer games. The same is not true for the preservation of archaeological visualizations. This article therefore discusses the following question: “Can emulation be employed to effectively preserve obsolete computer visualizations from the Archaeology domain?” Guidelines and test results coming from this work would be of great benefit to the archaeological community, and would contribute knowledge to other research communities, specifically those interested in similar data types/3D visualizations. 相似文献
In this paper, pair programming is empirically investigated from the perspective of developer personalities and temperaments
and how they affect pair effectiveness. A controlled experiment was conducted to investigate the impact of developer personalities
and temperaments on communication, pair performance and pair viability-collaboration. The experiment involved 70 undergraduate
students and the objective was to compare pairs of heterogeneous developer personalities and temperaments with pairs of homogeneous
personalities and temperaments, in terms of pair effectiveness. Pair effectiveness is expressed in terms of pair performance, measured by communication, velocity, design correctness and passed acceptance tests, and pair collaboration-viability measured by developers’ satisfaction, knowledge acquisition and participation. The results have shown that there is important
difference between the two groups, indicating better communication, pair performance and pair collaboration-viability for
the pairs with heterogeneous personalities and temperaments. In order to provide an objective assessment of the differences
between the two groups of pairs, a number of statistical tests and stepwise Discriminant Analysis were used.
Ignatios DeligiannisEmail:
Panagiotis Sfetsos
is an Assistant Professor at the Department of Informatics at the Alexander Technological Educational Institute of Thessaloniki,
Greece. He received his B.Sc. in Computer Science and Statistics from the University of Uppsala, Sweden (1981), and the Ph.D.
degree in Computer Science from the Aristotle University of Thessaloniki (2007). His Ph.D. Thesis was on “Experimentation
in Object Oriented Technology and Agile Methods”. His research interests include empirical software evaluation, measurement,
testing, quality, agile methods and especially extreme programming.
Ioannis G. Stamelos
is an Associate Professor of Computer Science at the Aristotle University of Thessaloniki, Dept. of Informatics. He received
a degree in Electrical Engineering from the Polytechnic School of Thessaloniki (1983) and the Ph. D. degree in Computer Science
from the Aristotle University of Thessaloniki (1988). He teaches object-oriented programming, software engineering, software
project management and enterprise information systems at the graduate and postgraduate level. His research interests include
empirical software evaluation and management, software education and open source software engineering. He is author of 90
scientific papers and member of the IEEE Computer Society.
Lefteris Angelis
received his B.Sc. and Ph.D. degree in Mathematics from Aristotle University of Thessaloniki (A.U.Th.). He works currently
as an Assistant Professor at the Department of Informatics of A.U.Th. His research interests involve statistical methods with
applications in software engineering and information systems, computational methods in mathematics and statistics, planning
of experiments and simulation techniques.
Ignatios Deligiannis
is an Associate Professor at Alexander Technological Education Institute of Thessaloniki, Greece. His main interests are Object-Oriented
software methods, and in particular design assessment and measurement. He received his B.Sc. in Computer Science from Lund
University, Sweden, in 1979, and then worked for several years in software development at Siemens Telecommunications industry.
He was member of ESERG (Empirical Software Engineering Research Group at Bournemouth University, UK). Currently, he is a research
partner of Software Engineering Group::Plase laboratory, Aristotle University of Thessaloniki, Greece.
相似文献
Today, the publication of microdata poses a privacy threat: anonymous personal records can be re-identified using third data sources. Past research has tried to develop a concept of privacy guarantee that an anonymized data set should satisfy before publication, culminating in the notion of t-closeness. To satisfy t-closeness, the records in a data set need to be grouped into Equivalence Classes (ECs), such that each EC contains records of indistinguishable quasi-identifier values, and its local distribution of sensitive attribute (SA) values conforms to the global table distribution of SA values. However, despite this progress, previous research has not offered an anonymization algorithm tailored for t-closeness. In this paper, we cover this gap with SABRE, a SA Bucketization and REdistribution framework for t-closeness. SABRE first greedily partitions a table into buckets of similar SA values and then redistributes the tuples of each bucket into dynamically determined ECs. This approach is facilitated by a property of the Earth Mover’s Distance (EMD) that we employ as a measure of distribution closeness: If the tuples in an EC are picked proportionally to the sizes of the buckets they hail from, then the EMD of that EC is tightly upper-bounded using localized upper bounds derived for each bucket. We prove that if the t-closeness constraint is properly obeyed during partitioning, then it is obeyed by the derived ECs too. We develop two instantiations of SABRE and extend it to a streaming environment. Our extensive experimental evaluation demonstrates that SABRE achieves information quality superior to schemes that merely applied algorithms tailored for other models to t-closeness, and can be much faster as well. 相似文献
The Beyond 3G (B3G) radio landscape will consist of cognitive heterogeneous wireless networks, operating in the framework of diverse co-operative associations among different classes of operators and providers, for the accommodation of the demands of users with multimode and/or multihoming enabled terminals. In this context, the optimized spectrum and radio resource utilization will be key factor for accomplishment of the purposes of both users and operators/providers, namely the satisfaction of user’s needs and the augmentation of profit, respectively. In this paper, we focus on an architecture for the management and optimization of spectrum and radio resource utilization in such composite wireless environments, and we analytically present the respective information flow among and from/to the functional entities involved in this architecture. The proposed management architecture can operate in the framework of different business scenarios and is based on related work that has been conducted within the IEEE 1900.4 standard. 相似文献
We present a general two-player quantum game simulator that can simulate any two-player quantum game described by a 2×2 payoff matrix (two strategy games).The user can determine the payoff matrices for both players, their strategies and the amount of entanglement between their initial strategies. The outputs of the simulator are the expected payoffs of each player as a function of the other player's strategy parameters and the amount of entanglement. The simulator also produces contour plots that divide the strategy spaces of the game in regions in which players can get larger payoffs if they choose to use a quantum strategy against any classical one. We also apply the simulator to two well-known quantum games, the Battle of Sexes and the Chicken game.
Program summary
Program title: Quantum Game Simulator (QGS)Catalogue identifier: AEED_v1_0Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEED_v1_0.htmlProgram obtainable from: CPC Program Library, Queen's University, Belfast, N. IrelandLicensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.htmlNo. of lines in distributed program, including test data, etc.: 3416No. of bytes in distributed program, including test data, etc.: 583 553Distribution format: tar.gzProgramming language: Matlab R2008a (C)Computer: Any computer that can sufficiently run Matlab R2008aOperating system: Any system that can sufficiently run Matlab R2008aClassification: 4.15Nature of problem: Simulation of two player quantum games described by a payoff matrix.Solution method: The program calculates the matrices that comprise the Eisert setup for quantum games based on the quantum circuit model. There are 5 parameters that can be altered. We define 3 of them as constant. We play the quantum game for all possible values for the other 2 parameters and store the results in a matrix.Unusual features: The software provides an easy way of simulating any two-player quantum games.Running time: Approximately 0.4 sec (Region Feature) and 0.3 sec (Payoff Feature) on a Intel Core 2 Duo GHz with 2 GB of memory under Windows XP. 相似文献
Secure authentication features of communication and electronic commerce protocols involve computationally expensive and memory intensive cryptographic operations that have the potential to be turned into denial-of-service (DoS) exploits. Recent proposals attempt to improve DoS resistance by implementing a trade-off between the resources required for the potential victim(s) with the resources used by a prospective attacker. Such improvements have been proposed for the Internet Key Exchange (IKE), the Just Fast Keying (JFK) key agreement protocol and the Secure Sockets Layer (SSL/TLS) protocol. In present article, we introduce probabilistic model checking as an efficient tool-assisted approach for systematically quantifying DoS security threats. We model a security protocol with a fixed network topology using probabilistic specifications for the protocol participants. We attach into the protocol model, a probabilistic attacker model which performs DoS related actions with assigned cost values. The costs for the protocol participants and the attacker reflect the level of some resource expenditure (memory, processing capacity or communication bandwidth) for the associated actions. From the developed model we obtain a Discrete Time Markov Chain (DTMC) via property preserving discrete-time semantics. The DTMC model is verified using the PRISM model checker that produces probabilistic estimates for the analyzed DoS threat. In this way, it is possible to evaluate the level of resource expenditure for the attacker, beyond which the likelihood of widespread attack is reduced and subsequently to compare alternative design considerations for optimal resistance to the analyzed DoS threat. Our approach is validated through the analysis of the Host Identity Protocol (HIP). The HIP base-exchange is seen as a cryptographic key-exchange protocol with special features related to DoS protection. We analyze a serious DoS threat, for which we provide probabilistic estimates, as well as results for the associated attacker and participants' costs. 相似文献
Enterprises increasingly recognize the compelling economic and operational benefits from virtualizing and pooling IT resources in the cloud. Nevertheless, the significant and valuable transformation of organizations that adopt cloud computing is accompanied by a number of security threats that should be considered. In this paper, we outline significant security challenges presented when migrating to a cloud environment and propose PaaSword – a novel holistic framework that aspires to alleviate these challenges. Specifically, the proposed framework involves a context-aware security model, the necessary policies enforcement mechanism along with a physical distribution, encryption and query middleware. 相似文献
In this paper, the strictly convex quadratic program (QP) arising in model predictive control (MPC) for constrained linear systems is reformulated as a system of piecewise affine equations. A regularized piecewise smooth Newton method with exact line search on a convex, differentiable, piecewise-quadratic merit function is proposed for the solution of the reformulated problem. The algorithm has considerable merits when applied to MPC over standard active set or interior point algorithms. Its performance is tested and compared against state-of-the-art QP solvers on a series of benchmark problems. The proposed algorithm is orders of magnitudes faster, especially for large-scale problems and long horizons. For example, for the challenging crude distillation unit model of Pannocchia, Rawlings, and Wright (2007) with 252 states, 32 inputs, and 90 outputs, the average running time of the proposed approach is 1.57 ms. 相似文献