首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
Model-based recognition of 3D objects from single images   总被引:1,自引:0,他引:1  
In this work, we treat major problems of object recognition which have received relatively little attention lately. Among them are the loss of depth information in the projection from a 3D object to a single 2D image, and the complexity of finding feature correspondences between images. We use geometric invariants to reduce the complexity of these problems. There are no geometric invariants of a projection from 3D to 2D. However, given certain modeling assumptions about the 3D object, such invariants can be found. The modeling assumptions can be either a particular model or a generic assumption about a class of models. Here, we use such assumptions for single-view recognition. We find algebraic relations between the invariants of a 3D model and those of its 2D image under general projective projection. These relations can be described geometrically as invariant models in a 3D invariant space, illuminated by invariant “light rays,” and projected onto an invariant version of the given image. We apply the method to real images  相似文献   

2.
3.
In this paper, we derive new geometric invariants for structured 3D points and lines from single image under projective transform, and we propose a novel model-based 3D object recognition algorithm using them. Based on the matrix representation of the transformation between space features (points and lines) and the corresponding projected image features, new geometric invariants are derived via the determinant ratio technique. First, an invariant for six points on two adjacent planes is derived, which is shown to be equivalent to Zhu's result [1], but in simpler formulation. Then, two new geometric invariants for structured lines are investigated: one for five lines on two adjacent planes and the other for six lines on four planes. By using the derived invariants, a novel 3D object recognition algorithm is developed, in which a hashing technique with thresholds and multiple invariants for a model are employed to overcome the over-invariant and false alarm problems. Simulation results on real images show that the derived invariants remain stable even in a noisy environment, and the proposed 3D object recognition algorithm is quite robust and accurate.  相似文献   

4.
Local invariants for recognition   总被引:2,自引:0,他引:2  
  相似文献   

5.
There are three projective invariants of a set of six points in general position in space. It is well known that these invariants cannot be recovered from one image, however an invariant relationship does exist between space invariants and image invariants. This invariant relationship is first derived for a single image. Then this invariant relationship is used to derive the space invariants, when multiple images are available. This paper establishes that the minimum number of images for computing these invariants is three, and the computation of invariants of six points from three images can have as many as three solutions. Algorithms are presented for computing these invariants in closed form. The accuracy and stability with respect to image noise, selection of the triplets of images and distance between viewing positions are studied both through real and simulated images. Applications of these invariants are also presented. Both the results of Faugeras (1992) and Hartley et al. (1992) for projective reconstruction and Sturm's method (1869) for epipolar geometry determination from two uncalibrated images with at least seven points are extended to the case of three uncalibrated images with only six points  相似文献   

6.
A central task of computer vision is to automatically recognize objects in real-world scenes. The parameters defining image and object spaces can vary due to lighting conditions, camera calibration and viewing position. It is therefore desirable to look for geometric properties of the object which remain invariant under such changes in the observation parameters. The study of such geometric invariance is a field of active research. This paper presents the theory and computation of projective invariants formed from points and lines using the geometric algebra framework. This work shows that geometric algebra is a very elegant language for expressing projective invariants using n views. The paper compares projective invariants involving two and three cameras using simulated and real images. Illustrations of the application of such projective invariants in visual guided grasping, camera self-localization and reconstruction of shape and motion complement the experimental part.  相似文献   

7.

Orthogonal moments and their invariants to geometric transformations for gray-scale images are widely used in many pattern recognition and image processing applications. In this paper, we propose a new set of orthogonal polynomials called adapted Gegenbauer–Chebyshev polynomials (AGC). This new set is used as a basic function to define the orthogonal adapted Gegenbauer–Chebyshev moments (AGCMs). The rotation, scaling, and translation invariant property of (AGCMs) is derived and analyzed. We provide a novel series of feature vectors of images based on the adapted Gegenbauer–Chebyshev orthogonal moments invariants (AGCMIs). We practice a novel image classification system using the proposed feature vectors and the fuzzy k-means classifier. A series of experiments is performed to validate this new set of orthogonal moments and compare its performance with the existing orthogonal moments as Legendre invariants moments, the Gegenbauer and Tchebichef invariant moments using three different image databases: the MPEG7-CE Shape database, the Columbia Object Image Library (COIL-20) database and the ORL-faces database. The obtained results ensure the superiority of the proposed AGCMs over all existing moments in representation and recognition of the images.

  相似文献   

8.
9.
Moment invariants for recognition under changing viewpoint and illumination   总被引:1,自引:0,他引:1  
Generalised color moments combine shape and color information and put them on an equal footing. Rational expressions of such moments can be designed, that are invariant under both geometric deformations and photometric changes. These generalised color moment invariants are effective features for recognition under changing viewpoint and illumination. The paper gives a systematic overview of such moment invariants for several combinations of deformations and photometric changes. Their validity and potential is corroborated through a series of experiments. Both the cases of indoor and outdoor images are considered, as illumination changes tend to differ between these circumstances. Although the generalised color moment invariants are extracted from planar surface patches, it is argued that invariant neighbourhoods offer a concept through which they can also be used to deal with 3D objects and scenes.  相似文献   

10.
熊艳  彭嘉雄 《计算机学报》1997,20(2):179-184
本文提出了一种由未定标图象估计三维射不变量的新算法。实验结果表明了本算法的有效性。  相似文献   

11.
12.
几何不变量,特别是射影不变量,是基于单视点灰度图像识别三维物体的一条有效途径.但理论研究表明,只有特定的几何约束结构,才具有射影不变量.所以,研究并发现这种几何约束结构就具有十分重要的意义.该文提出了一种新的由相邻3平面上5条直线组成的几何约束结构及其所具有的射影不变量.该结构较Sugimoto提出的几何约束结构简单,可从结构同样复杂的物体中获得更多的几何不变量,有利于提高物体识别的稳定性;同时,由于该结构大量存在于由多面体组合而构成的人造物体及地面建筑物中,因此它非常适合这类物体的识别.实验验证了文中提出的几何约束结构具有不随物体成像视点改变的射影不变量.  相似文献   

13.
Image distortion induced by the relative motion between an observer and the scene is an important cue for recovering the motion and the structure of the scene. It is known that the distortion in images can be described by transformation groups, such as Euclidean, affine, and projective groups. In this paper, we investigate how the moments of image curves are changed by group transformations, and we derive a relationship between the change in image moments and the invariant vector fields of the transformation groups. The results are used to formalize a method for extracting invariant vector fields of affine transformations from changes in the moments of orientation of curve segments in images. The method is applied to a realtime robot visual navigation task.  相似文献   

14.
Model-based invariants for 3-D vision   总被引:5,自引:4,他引:1  
Invariance under a group of 3-D transformations seems a desirable component of an efficient 3-D shape representation. We propose representations which are invariant under weak perspective to either rigid or linear 3-D transformations, and we show how they can be computed efficiently from a sequence of images with a linear and incremental algorithm. We show simulated results with perspective projection and noise, and the results of model acquisition from a real sequence of images. The use of linear computation, together with the integration through time of invariant representations, offers improved robustness and stability. Using these invariant representations, we derive model-based projective invariant functions of general 3-D objects. We discuss the use of the model-based invariants with existing recognition strategies: alignment without transformation, and constant time indexing from 2-D images of general 3-D objects.  相似文献   

15.
16.
Orthogonal variant moments features in image analysis   总被引:1,自引:0,他引:1  
Moments are statistical measures used to obtain relevant information about a certain object under study (e.g., signals, images or waveforms), e.g., to describe the shape of an object to be recognized by a pattern recognition system. Invariant moments (e.g., the Hu invariant set) are a special kind of these statistical measures designed to remain constant after some transformations, such as object rotation, scaling, translation, or image illumination changes, in order to, e.g., improve the reliability of a pattern recognition system. The classical moment invariants methodology is based on the determination of a set of transformations (or perturbations) for which the system must remain unaltered. Although very well established, the classical moment invariants theory has been mainly used for processing single static images (i.e. snapshots) and the use of image moments to analyze images sequences or video, from a dynamic point of view, has not been sufficiently explored and is a subject of much interest nowadays. In this paper, we propose the use of variant moments as an alternative to the classical approach. This approach presents clear differences compared to the classical moment invariants approach, that in specific domains have important advantages. The difference between the classical invariant and the proposed variant approach is mainly (but not solely) conceptual: invariants are sensitive to any image change or perturbation for which they are not invariant, so any unexpected perturbation will affect the measurements (i.e. is subject to uncertainty); on the contrary, a variant moment is designed to be sensitive to a specific perturbation, i.e., to measure a transformation, not to be invariant to it, and thus if the specific perturbation occurs it will be measured; hence any unexpected disturbance will not affect the objective of the measurement confronting thus uncertainty. Furthermore, given the fact that the proposed variant moments are orthogonal (i.e. uncorrelated) it is possible to considerably reduce the total inherent uncertainty. The presented approach has been applied to interesting open problems in computer vision such as shape analysis, image segmentation, tracking object deformations and object motion tracking, obtaining encouraging results and proving the effectiveness of the proposed approach.  相似文献   

17.
Geometric invariants and object recognition   总被引:10,自引:4,他引:6  
  相似文献   

18.
Studies the computation of projective invariants in pairs of images from uncalibrated cameras and presents a detailed study of the projective and permutation invariants for configurations of points and/or lines. Two basic computational approaches are given, one algebraic and one geometric. In each case, invariants are computed in projective space or directly from image measurements. Finally, we develop combinations of those projective invariants which are insensitive to permutations of the geometric primitives of each of the basic configurations  相似文献   

19.
Moment functions defined using a polar coordinate representation of the image space, such as radial moments and Zernike moments, are used in several recognition tasks requiring rotation invariance. However, this coordinate representation does not easily yield translation invariant functions, which are also widely sought after in pattern recognition applications. This paper presents a mathematical framework for the derivation of translation invariants of radial moments defined in polar form. Using a direct application of this framework, translation invariant functions of Zernike moments are derived algebraically from the corresponding central moments. Both derived functions are developed for non-symmetrical as well as symmetrical images. They mitigate the zero-value obtained for odd-order moments of the symmetrical images. Vision applications generally resort to image normalization to achieve translation invariance. The proposed method eliminates this requirement by providing a translation invariance property in a Zernike feature set. The performance of the derived invariant sets is experimentally confirmed using a set of binary Latin and English characters.  相似文献   

20.
Recently, orthogonal moments have become efficient tools for two-dimensional and three-dimensional (2D and 3D) image not only in pattern recognition, image vision, but also in image processing and applications engineering. Yet, there is still a major difficulty in 3D rotation invariants. In this paper, we propose new sets of invariants for 2D and 3D rotation, scaling and translation based on orthogonal radial Hahn moments. We also present theoretical mathematics to derive them. Thus, this paper introduces in the first case new 2D radial Hahn moments based on polar representation of an object by one-dimensional orthogonal discrete Hahn polynomials, and a circular function. In the second case, we present new 3D radial Hahn moments using a spherical representation of volumetric image by one-dimensional orthogonal discrete Hahn polynomials and a spherical function. Further 2D and 3D invariants are derived from the proposed 2D and 3D radial Hahn moments respectively, which appear as the third case. In order to test the proposed approach, we have resolved three issues: the image reconstruction, the invariance of rotation, scaling and translation, and the pattern recognition. The result of experiments show that the Hahn moments have done better than the Krawtchouk moments, with and without noise. Simultaneously, the mentioned reconstruction converges quickly to the original image using 2D and 3D radial Hahn moments, and the test images are clearly recognized from a set of images that are available in COIL-20 database for 2D image, and Princeton shape benchmark (PSB) database for 3D image.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号