首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 20 毫秒
1.
A prerequisite for higher-level visual tasks such as object recognition is a segmentation of the image into distinct two-dimensional regions. While it has long been assumed that the human visual system jointly exploits region and boundary cues for image segmentation, we report the results of psychophysical experiments which suggest that the visual system relies on geometric properties of bounding contours such as closure and not on the texture of the two-dimensional regions they partition. These findings suggest that the visual system may code and links contours into coherent shapes before surface properties are conjoined.  相似文献   

2.
Image segmentation is an important early stage in visual processing in which the visual system groups together parts of the image that belong together, prior to or in conjunction with object recognition. Two principal processes may be involved in image segmentation: an edge-based process that uses feature contrasts to mark boundaries of coherent regions, and a region-based process that groups similar features over a larger scale. Earlier, we have shown that motion and colour interact strongly in image segmentation by the human visual system. Here we explore the nature of this interaction in terms of edge- and region-based processes. We measure performance on a region-based colour segmentation task in the presence of distinct types of motion information, in the form of edges and regions which in themselves do not reveal the location of the colour target. The results show that both motion edges and regions may guide the integrative process required for this colour segmentation task. Motion edges appear to act by delimiting areas over which to integrate colour information, whereas motion similarities define primitive surfaces within which colour grouping and segmentation processes are deployed.  相似文献   

3.
The visual system groups image elements that belong to an object and segregates them from other objects and the background. Important cues for this grouping process are the Gestalt criteria, and most theories propose that these are applied in parallel across the visual scene. Here, we find that Gestalt grouping can indeed occur in parallel in some situations, but we demonstrate that there are also situations where Gestalt grouping becomes serial. We observe substantial time delays when image elements have to be grouped indirectly through a chain of local groupings. We call this chaining process incremental grouping and demonstrate that it can occur for only a single object at a time. We suggest that incremental grouping requires the gradual spread of object-based attention so that eventually all the object's parts become grouped explicitly by an attentional labeling process. Our findings inspire a new incremental grouping theory that relates the parallel, local grouping process to feedforward processing and the serial, incremental grouping process to recurrent processing in the visual cortex. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

4.
5.
The visual system analyses information by decomposing complex objects into simple components (visual features) that are widely distributed across the cortex. When several objects are present simultaneously in the visual field, a mechanism is required to group (bind) together visual features that belong to each object and to separate (segment) them from features of other objects. An attractive scheme for binding visual features into a coherent percept consists of synchronizing the activity of their neural representations. If synchrony is important in binding, one would expect that binding and segmentation are facilitated by visual displays that are temporally manipulated to induce stimulus-dependent synchrony. Here we show that visual grouping is indeed facilitated when elements of one percept are presented at the same time as each other and are temporally separated (on a scale below the integration time of the visual system) from elements of another percept or from background elements. Our results indicate that binding is due to a global mechanism of grouping caused by synchronous neural activation, and not to a local mechanism of motion computation.  相似文献   

6.
This article examines how the human visual system represents the shapes of 3-dimensional (3D) objects. One long-standing hypothesis is that object shapes are represented in terms of volumetric component parts and their spatial configuration. This hypothesis is examined in 3 experiments using a whole-part matching paradigm in which participants match object parts to whole novel 3D object shapes. Experiments 1 and 2, consistent with volumetric image segmentation, show that whole-part matching is faster for volumetric component parts than for either open or closed nonvolumetric regions of edge contour. However, the results of Experiment 3 show that an equivalent advantage is found for bounded regions of edge contour that correspond to object surfaces. The results are interpreted in terms of a surface-based model of 3D shape representation, which proposes edge-bounded 2-dimensional polygons as basic primitives of surface shape. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

7.
We review and discuss different classes of image segmentation methods. The usefulness of these methods is illustrated by a number of clinical cases. Segmentation is the process of assigning labels to pixels in 2D images or voxels in 3D images. Typically the effect is that the image is split up into segments, also called regions or areas. In medical imaging it is essential for quantification of outlined structures and for 3D visualization of relevant image data. Based on the level of implemented model knowledge we have classified these methods into (1) manual delineation, (2) low-level segmentation, and (3) model-based segmentation. Pure manual delineation of structures in a series of images is time-consuming and user-dependent and should therefore be restricted to quick experiments. Low-level segmentation analyzes the image locally at each pixel in the image and is practically limited to high-contrast images. Model-based segmentation uses knowledge of object structure such as global shape or semantic context. It typically requires an initialization, for example in the form of a rough approximation of the contour to be found. In practice it turns out that the use of high-level knowledge, e.g. anatomical knowledge, in the segmentation algorithm is quite complicated. Generally, the number of clinical applications decreases with the level and extent of prior knowledge needed by the segmentation algorithm. Most problems of segmentation inaccuracies can be overcome by human interaction. Promising segmentation methods for complex images are therefore user-guided and thus semi-automatic. They require manual intervention and guidance and consist of fast and accurate refinement techniques to assist the human operator.  相似文献   

8.
闫东阳  明冬萍 《工程科学学报》2017,39(11):1735-1742
在遥感影像分割分类中,种子区域生长算法是一种常见的分割算法.传统的种子区域生长算法只能提取单一连续的、纹理简单的目标地物,而对具有复杂纹理和多光谱特征的遥感影像,分割时存在分割效果差、不能同时有效地提取多个地物的问题.针对以上问题,本文提出了一种改进的面向对象的自动多种子区域生长算法.该方法适用于同时提取多个目标地物,且分割效果好.该方法首先使用一种改进的中值滤波对影像进行平滑处理,使目标内部一致性更高,同时保留纹理信息.然后通过一定的准则进行自动种子选取并进行生长,最后对生长后的区域进行碎斑合并处理,最终得到多种对象的分割结果.本文采用三组不同大小的1 m空间分辨率的航空影像进行实验,通过与分水岭以及传统单种子区域生长算法的多组实验对比,发现该方法可以面向全局对象,自动选取覆盖各种地物类型的种子,同时对多种地物目标进行分割处理,可为后续面向对象影像分析和应用提供可靠的数据基础.  相似文献   

9.
A model of recognition is described based on cell properties in the ventral cortical stream of visual processing in the primate brain. At a critical intermediate stage in this system, 'Elaborate' feature sensitive cells respond selectively to visual features in a way that depends on size (+/- 1 octave), orientation (+/- 45 degrees) but does not depend on position within central vision (+/- 5 degrees). These features are simple conjunctions of 2-D elements (e.g. a horizontal dark area above a dark smoothly convex area). They can arise either as elements of an object's surface pattern or as a 3-D component bounded by an object's external contour. By requiring a combination of several such features without regard to their position within the central region of the visual image, 'Pattern' sensitive cells at higher levels can exhibit selectivity for complex configurations that typify objects seen under particular viewing conditions. Given that input features to such Pattern sensitive cells are specified in approximate size and orientation, initial cellular 'representations' of the visual appearance of object type (or object example) are also selective for orientation and size. At this level, sensitivity to object view (+/- 60 degrees) arises because visual features disappear as objects are rotated in perspective. Processing is thus viewer-centred and the neurones only respond to objects seen from particular viewing conditions or 'object instances'. Combined sensitivity to multiple features (conjunctions of elements) independent of their position, establishes selectivity for the configurations of object parts (from one view) because rearranged configurations of the same parts yield images lacking some of the 2-D visual features present in the normal configuration. Different neural populations appear to be selectively tuned to particular components of the same biological object (e.g. face, eyes, hands, legs), perhaps because the independent articulation of these components gives rise to correlated activity in different sets of input visual features. Generalisation over viewing conditions for a given object can be established by hierarchically pooling outputs of view-condition specific cells with pooling operations dependent on the continuity in experience across viewing conditions. Different object parts are seen together and different views are seen in succession when the observer walks around the object. The view specific coding that characterises the selectivity of cells in the temporal lobe can be seen as a natural consequence of selective experience of objects from particular vantage points. View specific coding for the face and body also has great utility in understanding complex social signals, a property that may not be feasible with object-centred processing.  相似文献   

10.
This article examines the possibility that the visual system treats dynamic cues as instances of new perceptual objects under some circumstances. Using the contingent capture paradigm (C. L. Folk, R. W. Remington, & J. C. Johnston, 1992), the author compared luminance change cues of different magnitude for their ability to capture attention when participants were set for new objects. Whereas small luminance changes failed to produce attentional capture, large luminance changes indeed captured attention, suggesting that they were treated as compatible with the participants' attentional set for new objects. It is argued that sufficiently large luminance transients led to a disruption of spatiotemporal object continuity and precipitated the emergence of a new perceptual object. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

11.
Nine experiments examined the means by which visual memory for individual objects is structured into a larger representation of a scene. Participants viewed images of natural scenes or object arrays in a change detection task requiring memory for the visual form of a single target object. In the test image, 2 properties of the stimulus were independently manipulated: the position of the target object and the spatial properties of the larger scene or array context. Memory performance was higher when the target object position remained the same from study to test. This same-position advantage was reduced or eliminated following contextual changes that disrupted the relative spatial relationships among contextual objects (context deletion, scrambling, and binding change) but was preserved following contextual change that did not disrupt relative spatial relationships (translation). Thus, episodic scene representations are formed through the binding of objects to scene locations, and object position is defined relative to a larger spatial representation coding the relative locations of contextual objects. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

12.
Top-down learning of low-level vision tasks   总被引:1,自引:0,他引:1  
Perceptual tasks such as edge detection, image segmentation, lightness computation and estimation of three-dimensional structure are considered to be low-level or mid-level vision problems and are traditionally approached in a bottom-up, generic and hard-wired way. An alternative to this would be to take a top-down, object-class-specific and example-based approach. In this paper, we present a simple computational model implementing the latter approach. The results generated by our model when tested on edge-detection and view-prediction tasks for three-dimensional objects are consistent with human perceptual expectations. The model's performance is highly tolerant to the problems of sensor noise and incomplete input image information. Results obtained with conventional bottom-up strategies show much less immunity to these problems. We interpret the encouraging performance of our computational model as evidence in support of the hypothesis that the human visual system may learn to perform supposedly low-level perceptual tasks in a top-down fashion.  相似文献   

13.
To locate objects in the environment, animals and humans use visual and nonvisual information. We were interested in children's ability to relocate an object on the basis of self-motion and local and distal color cues for orientation. Five- to 9-year-old children were tested on an object location memory task in which, between presentation and test, the availability of local and distal cues was manipulated. Additionally, participants' viewpoint could be changed. We used a Bayesian model selection approach to compare our hypotheses. We found that, to remain oriented in space, 5-year-olds benefit from visual information in general, 7-year-olds benefit from visual cues when a viewpoint change takes place, and 9-year-olds do not benefit from the availability of visual cues for orientation but rely on self-movement cues instead. Results are discussed in terms of the adaptive combination model (Newcombe & Huttenlocher, 2006). (PsycINFO Database Record (c) 2011 APA, all rights reserved)  相似文献   

14.
In manual search tasks designed to assess infants' knowledge of the object concept, why does search for objects hidden by darkness precede search for objects hidden by visible occluders by several months? A graded representations account explains this décalage by proposing that the conflicting visual input from occluders directly competes with object representations, whereas darkness merely weakens representations. This study tests the prediction that representations of objects hidden by darkness are strong enough for infants to bind auditory cues to them and support search, whereas representations of objects hidden by occluders are not. Six-and-half-month-olds were presented with audible or silent objects that remained visible, became hidden by darkness, or became hidden by a visible occluder. Search required engaging in the same means-end action in all conditions. As predicted, auditory cues increased search when objects were hidden by darkness but not when they were hidden by a visible occluder. Results are discussed in the context of different facets of object concept development highlighted by graded representations perspectives and core knowledge perspectives and in relation to other work on multimodal object representations. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

15.
Typical natural visual scenes contain many objects, which need to be segregated from each other and from the background. Present theories subdivide the processes responsible for this segregation into a pre-attentive and attentive system. The pre-attentive system segregates image regions that 'pop out' rapidly and in parallel across the visual field. In the primary visual cortex, responses to pre-attentively selected image regions are enhanced. When objects do not segregate automatically from the rest of the image, the time-consuming attentive system is recruited. Here we investigate whether attentive selection is also associated with a modulation of firing rates in area V1 of the brain in monkeys trained to perform a curve-tracing task. Neuronal responses to the various segments of a target curve were simultaneously enhanced relative to responses evoked by a distractor curve, even if the two curves crossed each other. This indicates that object-based attention is associated with a response enhancement at the earliest level of the visual cortical processing hierarchy.  相似文献   

16.
Change blindness, or the failure to detect (often large) changes to visual scenes, has been demonstrated in a variety of different situations. Failures to detect auditory changes are far less studied, and thus little is known about the nature of change deafness. Five experiments were conducted to explore the processes involved in change deafness by measuring explicit change detection as well as auditory object encoding. The experiments revealed that considerable change deafness occurs, even though auditory objects are encoded quite well. Familiarity with the objects did not affect detection or recognition performance. Whereas spatial location was not an effective cue, fundamental frequency and the periodicity/aperiodicity of the sounds provided important cues for the change-detection task. Implications for the mechanisms responsible for change deafness and auditory sound organization are discussed. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

17.
Humans see whole objects from input fragmented in space and time, yet spatiotemporal object perception is poorly understood. The authors propose the theory of spatiotemporal relatability (STR), which describes the visual information and processes that allow visible fragments revealed at different times and places, due to motion and occlusion, to be assembled into unitary perceived objects. They present a formalization of STR that specifies spatial and temporal relations for object formation. Predictions from the theory regarding conditions that lead to unit formation were tested and confirmed in experiments with dynamic and static, occluded and illusory objects. Moreover, the results support the identity hypothesis of a common process for amodal and modal contour interpolation and provide new evidence regarding the relative efficiency of static and dynamic object formation. STR postulates a mental representation, the dynamic visual icon, that briefly maintains shapes and updates positions of occluded fragments to connect them with visible regions. The theory offers a unified account of interpolation processes for static, dynamic, occluded, and illusory objects. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

18.
In laboratory experiments, infants are sensitive to patterns of visual features that co-occur (e.g., Fiser & Aslin, 2002). Once infants learn the statistical regularities, however, what do they do with that knowledge? Moreover, which patterns do infants learn in the cluttered world outside of the laboratory? Across 4 experiments, we show that 9-month-olds use this sensitivity to make inferences about object properties. In Experiment 1, 9-month-old infants expected co-occurring visual features to remain fused (i.e., infants looked longer when co-occurring features split apart than when they stayed together). Forming such expectations can help identify integral object parts for object individuation, recognition, and categorization. In Experiment 2, we increased the task difficulty by presenting the test stimuli simultaneously with a different spatial layout from the familiarization trials to provide a more ecologically valid condition. Infants did not make similar inferences in this more distracting test condition. However, Experiment 3 showed that a social cue did allow inferences in this more difficult test condition, and Experiment 4 showed that social cues helped infants choose patterns among distractor patterns during learning as well as during test. These findings suggest that infants can use feature co-occurrence to learn about objects and that social cues shape such foundational learning in distraction-filled environments. (PsycINFO Database Record (c) 2011 APA, all rights reserved)  相似文献   

19.
The mechanisms underlying segmentation and selection of visual stimuli over time were investigated in patients with posterior parietal damage. In a modified visual search task, a preview of old objects preceded search of a new set for a target while the old items remained. In Experiment 1, control participants ignored old and prioritized new items, but patients had severe difficulties finding the target (especially on the contralesional side). In Experiment 2, simplified displays yielded analogous results, ruling out search ease as a crucial factor in poor preview search. In Experiment 3, outlines around distractor groups (to aid segmentation) improved conjunction but not preview search, suggesting a specific deficit in spatiotemporal segmentation. Experiment 4 ruled out spatial disengagement problems as a factor. The data emphasize the role of spatiotemporal segmentation cues in preview search and the parietal lobe in the role of these cues to prioritize search of new stimuli. (PsycINFO Database Record (c) 2010 APA, all rights reserved)  相似文献   

20.
Theories of image segmentation suggest that the human visual system may use two distinct processes to segregate figure from background: a local process that uses local feature contrasts to mark borders of coherent regions and a global process that groups similar features over a larger spatial scale. We performed psychophysical experiments to determine whether and to what extent the global similarity process contributes to image segmentation by motion and color. Our results show that for color, as well as for motion, segmentation occurs first by an integrative process on a coarse spatial scale, demonstrating that for both modalities the global process is faster than one based on local feature contrasts. Segmentation by motion builds up over time, whereas segmentation by color does not, indicating a fundamental difference between the modalities. Our data suggest that segmentation by motion proceeds first via a cooperative linking over space of local motion signals, generating almost immediate perceptual coherence even of physically incoherent signals. This global segmentation process occurs faster than the detection of absolute motion, providing further evidence for the existence of two motion processes with distinct dynamic properties.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号