Dense stereo algorithms are able to estimate disparities at all pixels including untextured regions. Typically these disparities are evaluated at integer disparity steps. A subsequent sub-pixel interpolation often fails to propagate smoothness constraints on a sub-pixel level.We propose to increase the sub-pixel accuracy in low-textured regions in four possible ways: First, we present an analysis that shows the benefit of evaluating the disparity space at fractional disparities. Second, we introduce a new disparity smoothing algorithm that preserves depth discontinuities and enforces smoothness on a sub-pixel level. Third, we present a novel stereo constraint (gravitational constraint) that assumes sorted disparity values in vertical direction and guides global algorithms to reduce false matches, especially in low-textured regions. Finally, we show how image sequence analysis improves stereo accuracy without explicitly performing tracking. Our goal in this work is to obtain an accurate 3D reconstruction. Large-scale 3D reconstruction will benefit heavily from these sub-pixel refinements.Results based on semi-global matching, obtained with the above mentioned algorithmic extensions are shown for the Middlebury stereo ground truth data sets. The presented improvements, called ImproveSubPix, turn out to be one of the top-performing algorithms when evaluating the set on a sub-pixel level while being computationally efficient. Additional results are presented for urban scenes. The four improvements are independent of the underlying type of stereo algorithm. 相似文献
The accuracy of optical flow estimation algorithms has been improving steadily as evidenced by results on the Middlebury optical flow benchmark. The typical formulation, however, has changed little since the work of Horn and Schunck. We attempt to uncover what has made recent advances possible through a thorough analysis of how the objective function, the optimization method, and modern implementation practices influence accuracy. We discover that “classical” flow formulations perform surprisingly well when combined with modern optimization and implementation techniques. One key implementation detail is the median filtering of intermediate flow fields during optimization. While this improves the robustness of classical methods it actually leads to higher energy solutions, meaning that these methods are not optimizing the original objective function. To understand the principles behind this phenomenon, we derive a new objective function that formalizes the median filtering heuristic. This objective function includes a non-local smoothness term that robustly integrates flow estimates over large spatial neighborhoods. By modifying this new term to include information about flow and image boundaries we develop a method that can better preserve motion details. To take advantage of the trend towards video in wide-screen format, we further introduce an asymmetric pyramid downsampling scheme that enables the estimation of longer range horizontal motions. The methods are evaluated on the Middlebury, MPI Sintel, and KITTI datasets using the same parameter settings. 相似文献
In this article we present the results of a comparison of six globally available land-cover products for the Mekong Basin – an area that spans 795,000 km2 and comprises parts of six riparian countries: China, Myanmar, Thailand, Laos, Cambodia, and Vietnam. The basin covers most climatic zones: from high-altitude, snow-covered mountainous regions in the north, to subtropical and tropical rainforest areas and agricultural land further south. The geopolitically important region not only is home to over 72,000,000 inhabitants, but also is a centre of attention of several environmental modelling experts, trying to assess future hydrologic dynamics, climate variability, as well probable land-use developments in the area.We compare land-cover products of the University of Maryland, UMD 1992–1993, the GLC 2000 product, the GlobCover products of 2004–2006 and 2009, as well as the MODIS-derived land-cover products of 2001 and 2009. For harmonization of individual legends, the Land Cover Classification System, LCCS, has been employed. However, even after harmonization, cross-tabulation among the products reveals extreme differences, where the impact of differing classification algorithms weighs higher than the impact of temporal coincidence of products. Especially, differences within mixed-vegetation classes are large, strongly impacting the overall assessment of forested land, other vegetated land, and even cultivated land in the Mekong Basin. The findings presented here are of high relevance for the modelling community as well as Mekong-related environmental studies, which should consider global remote-sensing-derived products with caution and solid background knowledge. 相似文献
Molecular visualization is often challenged with rendering of large molecular structures in real time. We introduce a novel approach that enables us to show even large protein complexes. Our method is based on the level‐of‐detail concept, where we exploit three different abstractions combined in one visualization. Firstly, molecular surface abstraction exploits three different surfaces, solvent‐excluded surface (SES), Gaussian kernels and van der Waals spheres, combined as one surface by linear interpolation. Secondly, we introduce three shading abstraction levels and a method for creating seamless transitions between these representations. The SES representation with full shading and added contours stands in focus while on the other side a sphere representation of a cluster of atoms with constant shading and without contours provide the context. Thirdly, we propose a hierarchical abstraction based on a set of clusters formed on molecular atoms. All three abstraction models are driven by one importance function classifying the scene into the near‐, mid‐ and far‐field. Moreover, we introduce a methodology to render the entire molecule directly using the A‐buffer technique, which further improves the performance. The rendering performance is evaluated on series of molecules of varying atom counts. 相似文献
This paper proposes a simplicity-oriented approach and framework for language-to-language transformation of, in particular, graphical languages. Key to simplicity is the decomposition of the transformation specification into sub-rule systems that separately specify purpose-specific aspects. We illustrate this approach by employing a variation of Plotkin’s Structural Operational Semantics (SOS) for pattern-based transformations of typed graphs in order to address the aspect ‘computation’ in a graph rewriting fashion. Key to our approach are two generalizations of Plotkin’s structural rules: the use of graph patterns as the matching concept in the rules, and the introduction of node and edge types. Types do not only allow one to easily distinguish between different kinds of dependencies, like control, data, and priority, but may also be used to define a hierarchical layering structure. The resulting Type-based Structural Operational Semantics (TSOS) supports a well-structured and intuitive specification and realization of semantically involved language-to-language transformations adequate for the generation of purpose-specific views or input formats for certain tools, like, e.g., model checkers. A comparison with the general-purpose transformation frameworks ATL and Groove, illustrates along the educational setting of our graphical WebStory language that TSOS provides quite a flexible format for the definition of a family of purpose-specific transformation languages that are easy to use and come with clear guarantees.
In the last years, a variety of processes respectively process steps have been investigated for the production of niobium powder. This is due to the fact that niobium capacitors could be a viable alternative to tantalum capacitors from a performance, availability, and price point of view. The reduction of niobium pentoxide by magnesium results in fine powders with high specific surface area but has the disadvantages of a very exothermic nature and the formation of magnesium niobate. It is shown in this work that the application of a continuously operating cyclone reactor and the use of niobium(IV) oxide as raw material solve these problems. A good control of the highly exothermic reaction within the cyclone reactor was achieved in the cyclone reactor by the ratio between gas flow rate and powder flow rate as well as by a proper preheating of the gas. 相似文献
The introduction of moving loads in the Floating Frame of Reference Formulation is presented. We derive the kinematics and governing equations of motion of a general flexible multibody system and their extension to moving loads. The equivalence of convective effects with Coriolis and centripetal forces is shown. These effects are measured numerically and their significance in moving loads traveling at high speed is confirmed. A method is presented to handle discontinuities when moving loads separate from the flexible structure. The method is extended from beam models to general flexible structures obtained by means of the Finite Element Method. An interpolation method for the deformation field of the modal representation of these bodies is introduced.The work is concluded by application of the method to modern mechanical problems in numerical simulations. 相似文献
Rapid advances in image acquisition and storage technology underline the need for real-time algorithms that are capable of solving large-scale image processing and computer-vision problems. The minimums–tcut problem, which is a classical combinatorial optimization problem, is a prominent building block in many vision and imaging algorithms such as video segmentation, co-segmentation, stereo vision, multi-view reconstruction, and surface fitting to name a few. That is why finding a real-time algorithm which optimally solves this problem is of great importance. In this paper, we introduce to computer vision the Hochbaum’s pseudoflow (HPF) algorithm, which optimally solves the minimum s–t cut problem. We compare the performance of HPF, in terms of execution times and memory utilization, with three leading published algorithms: (1) Goldberg’s and Tarjan’s Push-Relabel; (2) Boykov’s and Kolmogorov’s augmenting paths; and (3) Goldberg’s partial augment-relabel. While the common practice in computer-vision is to use either BK or PRF algorithms for solving the problem, our results demonstrate that, in general, HPF algorithm is more efficient and utilizes less memory than these three algorithms. This strongly suggests that HPF is a great option for many real-time computer-vision problems that require solving the minimum s–t cut problem. 相似文献