In this paper we investigate how formal software verification systems can be improved by utilising parallel assignment in weakest precondition computations.We begin with an introduction to modern software verification systems. Specifically, we review the method in which software abstractions are built using counterexample-guided abstraction refinement (CEGAR). The classical NP-complete parallel assignment problem is first posed, and then an additional restriction is added to create a special case in which the problem is tractable with an O(n2) algorithm. The parallel assignment problem is then discussed in the context of weakest precondition computations. In this special situation where statements can be assumed to execute truly concurrently, we show that any sequence of simple assignment statements without function calls can be transformed into an equivalent parallel assignment block.Results of compressing assignment statements into a parallel form with this algorithm are presented for a wide variety of software applications. The proposed algorithms were implemented in the ComFoRT reasoning framework [J. Ivers and N. Sharygina. Overview of ComFoRT: A model checking reasoning framework. Technical Report CMU/SEI-2004-TN-018, Carnegie Mellon Software Engineering Institute, 2004] and used to measure the improvement in the verification of real software systems. This improvement in time proved to be significant for many classes of software. 相似文献
In this paper, we will present a technique for measuring visibility distances under foggy weather conditions using a camera mounted onboard a moving vehicle. Our research has focused in particular on the problem of detecting daytime fog and estimating visibility distances; thanks to these efforts, an original method has been developed, tested and patented. The approach consists of dynamically implementing Koschmieder's law. Our method enables computing the meteorological visibility distance, a measure defined by the International Commission on Illumination (CIE) as the distance beyond which a black object of an appropriate dimension is perceived with a contrast of less than 5%. Our proposed solution is an original one, featuring the advantage of utilizing a single camera and necessitating the presence of just the road and sky in the scene. As opposed to other methods that require the explicit extraction of the road, this method offers fewer constraints by virtue of being applicable with no more than the extraction of a homogeneous surface containing a portion of the road and sky within the image. This image preprocessing also serves to identify the level of compatibility of the processed image with the set of Koschmieder's model hypotheses.
Nicolas Hautiére graduated from the École Nationale des Travaux Publics de l'État, France (2002). He received his M.S. and Ph.D. degree in computer vision, respectively, in 2002 and 2005 from Saint-Étienne University (France). From 2002, he is a researcher in the Laboratoire Central des Ponts et Chaussées (LCPC), Paris, France. His research interests include trafic engineering, computer vision, and pattern recognition.
Jean-Philippe Tarel graduated from the École Nationale des Ponts et Chaussées, Paris, France (1991). He received his Ph.D. degree in Applied Mathematics from Paris IX-Dauphine University in 1996 and he was with the Institut National de Recherche en Informatique et Automatique (INRIA) from 1991 to 1996. From 1997 to 1998, he was a research associate at Brown University, USA. From 1999, he is a researcher in the Laboratoire Central des Ponts et Chaussées (LCPC), Paris, France, and from 2001 to 2003 in the INRIA. His research interests include computer vision, pattern recognition, and shape modeling.
Jean Lavenant graduated from the École Nationale des Travaux Publics de l'État, Lyon, France (2001). He received the M.S. degree in Computer Vision from Jean Monnet university of Saint-Étienne in 2001. In 2001, he was a researcher in the Laboratoire Central des Ponts et Chaussées (LCPC). In 2002, he was a system engineer in Chicago (USA). He is currently an engineer for the french ministry of transports.
Didier Aubert received the M.S. and Ph.D. degree, respectively, in 1985 and 1989 from the National Polytechnical Institut of Grenoble (INPG). From 1989--1990, he worked as a research scientist on the development of an automatic road following system for the NAVLAB at Carnegie Mellon University. From 1990–1994, he worked in the research department of a private company (ITMI). During this period he was the project leader of several projects dealing with computer vision. He is currently a researcher at INRETS since 1995 and works on Road traffic measurements, crowd monitoring, automated highway systems, and driving assistance systems for vehicles. He is an image processing expert for several companies, teaches at Universities (Paris VI, Paris XI, ENPC, ENST) and is at the editorial board of RTS (Research - Transport - Safety). 相似文献
Assessing the probability of rare and extreme events is an important issue in the risk management of financial portfolios.
Extreme value theory provides the solid fundamentals needed for the statistical modelling of such events and the computation
of extreme risk measures. The focus of the paper is on the use of extreme value theory to compute tail risk measures and the
related confidence intervals, applying it to several major stock market indices. 相似文献
Free binary decision diagrams (FBDDs) are graph-based data structures representing Boolean functions with the constraint (additional
to binary decision diagram) that each variable is tested at most once during the computation. The function EARn is the following Boolean function defined for n × n Boolean matrices: EARn(M) = 1 iff the matrix M contains two equal adjacent rows. We prove that each FBDD computing EARn has size at least
and we present a construction of such diagrams of size approximately
. 相似文献
Conditions for the emergence of cooperation in a spatial common-pool resource game are studied. This combines in a unique
way local and global interactions. A fixed number of harvesters are located on a spatial grid. Harvesters choose among three
strategies: defection, cooperation, and enforcement. Individual payoffs are affected by both global factors, namely, aggregate
harvest and resource stock level, and local factors, such as the imposition of sanctions on neighbors by enforcers. The evolution
of strategies in the population is driven by social learning through imitation, based on local interaction or locally available
information. Numerous types of equilibria exist in these settings. An important new finding is that clusters of cooperators
and enforcers can survive among large groups of defectors. We discuss how the results contrast with the non-spatial, but otherwise
similar, game of Sethi and Somanathan (American Economic Review 86(4):766–789, 1996).
相似文献
This paper first presents a novel approach for modelling facial features, Local Directional Texture (LDT), which exploits the unique directional information in image textures for the problem of face recognition. A variant of LDT with privacy-preserving temporal strips (TS) is then considered to achieve faceless recognition with a higher degree of privacy while maintaining high accuracy. The TS uses two strips of pixel blocks from the temporal planes, XT and YT, for face recognition. By removing the reliance on spatial context (i.e., XY plane) for this task, the proposed method withholds facial appearance information from public view, where only one-dimensional temporal information that varies across time are extracted for recognition. Thus, privacy is assured, yet without impeding the facial recognition task which is vital for many security applications such as street surveillance and perimeter access control. To validate the reliability of the proposed method, experiments were carried out using the Honda/UCSD, CK+, CAS(ME)2 and CASME II databases. The proposed method achieved a recognition rate of 98.26% in the standard video-based face recognition database, Honda/UCSD. It also offers a 81.92% reduction in the dimension length required for storing the extracted features, in contrast to the conventional LBP-TOP.
GPU Shape Grammars provide a solution for interactive procedural generation, tuning and visualization of massive environment elements for both video games and production rendering. Our technique generates detailed models without explicit geometry storage. To this end we reformulate the grammar expansion for generation of detailed models at the tesselation control and geometry shader stages. Using the geometry generation capabilities of modern graphics hardware, our technique generated massive, highly detailed models. GPU Shape Grammars integrate within a scalable framework by introducing automatic generation of levels of detail at reduced cost. We apply our solution for interactive generation and rendering of scenes containing thousands of buildings and trees. 相似文献
Some conditions relating to the automata involved in the W-testing method are discussed. It is also shown how to use the
method for reduced automata instead of minimal automata. New design test conditions (weak output distinguishable, strong test-complete and output delimited type) are considered for the generalised stream X-machines (stream X-machines with basic functions replaced by relations and having as output strings of symbols rather than single
symbols). It is proved that testing methods similar to those already developed for ordinary deterministic stream X-machines
may be applied for generalised stream X-machines with output delimited types. A particular case of generalised stream X-machine
with output delimited type is the X-machine with output delimiter, which produces outputs having a distinct right end character.
Received October 2000 / Accepted in revised form January 2001 相似文献