共查询到20条相似文献,搜索用时 15 毫秒
1.
How to make robot vision work robustly under varying lighting conditions and without the constraint of the current color-coded environment are two of the most challenging issues in the RoboCup community. In this paper, we present a robust omnidirectional vision sensor to deal with these issues for the RoboCup Middle Size League soccer robots, in which two novel algorithms are applied. The first one is a camera parameters auto-adjusting algorithm based on image entropy. The relationship between image entropy and camera parameters is verified by experiments, and camera parameters are optimized by maximizing image entropy to adapt the output of the omnidirectional vision to the varying illumination. The second one is a ball recognition method based on the omnidirectional vision without color classification. The conclusion is derived that the ball on the field can be imaged to be an ellipse approximately in our omnidirectional vision, and the arbitrary FIFA ball can be recognized by detecting the ellipse imaged by the ball. The experimental results show that a robust omnidirectional vision sensor can be realized by using the two algorithms mentioned above. 相似文献
2.
3.
4.
In this work, a comparison of the performances of a stereo and a monocular vision system for the 3D pose estimation of a planar target in very challenging conditions is presented. In particular, the systems have been designed to detect in real time a target moving with a maximum speed of 1 m/s, in a range of distances from 0.5 to 4 m from the cameras, with an accuracy of less than 1 cm (referred to the estimation of the real world coordinates) and with a field of view of 80 deg. A theoretical evaluation and experimental results to assess the performance of the proposed systems are presented. Our analysis demonstrates the good accuracy in terms of target position estimation of the presented approaches not only for close range applications, but also for mid-to-long range ones. 相似文献
5.
6.
7.
8.
针对铝管生产过程中对准确、可量化的自动缺陷检测系统的迫切需要,本文引入一种由图像采集、缺陷检测、缺陷处理等模块组成的铝管缺陷检测系统。平板探测器获取由X光高压电源产生,穿过铝管的X射线并把所形成的数字图像通过USB端口发送至检测服务器。检测服务器使用机器视觉算法检测图像中的缺陷。当服务器检测到缺陷时,会向PCI板上指定位输出信号,报警装置接到信号后报警提醒工作人员。实验表明该系统能够自动、准确的标记出铝管中存在的缺陷,达到了系统的设计目标。 相似文献
9.
Interoperable vision component for object detection and 3D pose estimation for modularized robot control 总被引:1,自引:0,他引:1
Yasushi Mae Jaeil Choi Hideyasu Takahashi Kenichi Ohara Tomohito Takubo Tatsuo Arai 《Mechatronics》2011,21(6):983-992
Finding objects and tracking their poses are essential functions for service robots, in order to manipulate objects and interact with humans. We present novel algorithms for local feature matching for object detection, and 3D pose estimation. Our feature matching algorithm takes advantage of local geometric consistency for better performance, and the new 3D pose estimation algorithm solves the pose in a closed-form using homography, followed by a non-linear optimization step for stability. Advantages of our approach include better performance, minimal prior knowledge for the target pattern, and easy implementation and portability as a modularized software component. We have implemented our approach along with both CPU and GPU-based feature extraction, and built an interoperable component that can be used in any Robot Technology (RT)-based control system. Experiment shows that our approach produces very robust results for the estimated 3D pose, and maintain very low false positive rate. It is also fast enough to be used in on-line applications. We integrated our vision component in an autonomous robot system with a search-and-grasp task, and tested it with several objects that are found in ordinary domestic environment. We present the details of our approach, the design of our modular component design, and the results of the experiments in this paper. 相似文献
10.
传统的凸轮轴表面缺陷多为产后抽检,相比在线全检存在漏检与缺陷率上升等现象,且只能事后发现;而人工在线全检不但会使成本上升、也对人力资源提出了考验。为此实现自动实时在线全检就成为急需解决的课题。设计了基于机器视觉的凸轮轴表面缺陷在线自动检测系统。系统安装在凸轮轴生产流水线两侧,搭建特定光源,在凸轮轴移动、停止、旋转过程中通过高速相机对其表面进行图像捕获,并由工控机进行缺陷判定与定位。根据轴类表面缺陷的特征,设计了缺陷分割算法和缺陷区域标记算法,对凸轮轴表面的外伤、砂眼、研磨不良等典型缺陷进行分辨。算法可以准确提取目标缺陷区域,标记缺陷位置并统计缺陷特征对缺陷进行判定。该系统可在0.44 s每根轴的速度下,检测出凸轮轴表面直径大于1 mm 的缺陷,并通过人机交互界面显示缺陷所在位置。完全可以取代产后抽检及人工在线全检,同时还可以提高检测效率与检测精度。 相似文献
11.
As a popular image manipulation technique, object removal can be achieved by image-inpainting without any noticeable traces, which poses huge challenges to passive image forensics. The existing detection approach utilizes full search for block matching, resulting in high computational complexity. This paper presents an efficient forgery detection algorithm for object removal by exemplar-based inpainting, which integrates central pixel mapping (CPM), greatest zero-connectivity component labeling (GZCL) and fragment splicing detection (FSD). CPM speeds up suspicious block search by efficiently matching those blocks with similar hash values and then finding the suspicious pairs. To improve the detection precision, GZCL is used to mark the tampered pixels in suspected block pairs. FSD is adopted to distinguish and locate tampered regions from its best-match regions. Experimental results show that the proposed algorithm can reduce up to 90% of the processing time and maintain a detection precision above 85% under different kinds of object-removed images. 相似文献
12.
Keypoint-based object detection achieves better performance without positioning calculations and extensive prediction. However, they have heavy backbone, and high-resolution is restored using upsampling that obtain unreliable features. We propose a self-constrained parallelism keypoint-based lightweight object detection network (SCPNet), which speeds inference, drops parameters, widens receptive fields, and makes prediction accurate. Specifically, the parallel multi-scale fusion module (PMFM) with parallel shuffle blocks (PSB) adopts parallel structure to obtain reliable features and reduce depth, adopts repeated multi-scale fusion to avoid too many parallel branches. The self-constrained detection module (SCDM) has a two-branch structure, with one branch predicting corners, and employing entad offset to match high-quality corner pairs, and the other branch predicting center keypoints. The distances between the paired corners’ geometric centers and the center keypoints are used for self-constrained detection. On MS-COCO 2017 and PASCAL VOC, SCPNet’s results are competitive with the state-of-the-art lightweight object detection. https://github.com/mengdie-wang/SCPNet.git. 相似文献
13.
针对红外数据集规模小,标记样本少的特点,提出了一种红外目标检测网络的半监督迁移学习方法,主要用于提高目标检测网络在小样本红外数据集上的训练效率和泛化能力,提高深度学习模型在训练样本较少的红外目标检测等场景当中的适应性。文中首先阐述了在标注样本较少时无标注样本对提高模型泛化能力、抑制过拟合方面的作用。然后提出了红外目标检测网络的半监督迁移学习流程:在大量的RGB图像数据集中训练预训练模型,后使用少量的有标注红外图像和无标注红外图像对网络进行半监督学习调优。另外,文中提出了一种特征相似度加权的伪监督损失函数,使用同一批次样本的预测结果相互作为标注,以充分利用无标注图像内相似目标的特征分布信息;为降低半监督训练的计算量,在伪监督损失函数的计算中,各目标仅将其特征向量邻域范围内的预测目标作为伪标注。实验结果表明,文中方法所训练的目标检测网络的测试准确率高于监督迁移学习所获得的网络,其在Faster R-CNN上实现了1.1%的提升,而在YOLO-v3上实现了4.8%的显著提升,验证了所提出方法的有效性。 相似文献
14.
条纹投影轮廓术能较好地兼顾系统灵活性与测量精度,是光学三维表面成像与测量的主流技术。利用条纹投影轮廓术进行三维成像,首先需要建立合适的系统模型,然后通过系统标定来确定描述模型的系统参数,最后利用标定的系统模型进行三维重建,获得物体的三维表面形貌。由此可见,系统标定与系统模型密不可分,对三维成像的性能有直接影响。根据相位-三维映射和双目立体视觉两类不同的工作原理,对条纹投影轮廓术的系统模型和系统标定方法进行了综述,并简要总结了评估系统精度的方法和依据。 相似文献
15.
16.
17.
Representing contextual features at multiple scales is important for RGB-D SOD. Recently, due to advances in backbone convolutional neural networks (CNNs) revealing stronger multi-scale representation ability, many methods achieved comprising performance. However, most of them represent multi-scale features in a layer-wise manner, which ignores the fine-grained global contextual cues in a single layer. In this paper, we propose a novel global contextual exploration network (GCENet) to explore the performance gain of multi-scale contextual features in a fine-grained manner. Concretely, a cross-modal contextual feature module (CCFM) is proposed to represent the multi-scale contextual features at a single fine-grained level, which can enlarge the range of receptive fields for each network layer. Furthermore, we design a multi-scale feature decoder (MFD) that integrates fused features from CCFM in a top-down way. Extensive experiments on five benchmark datasets demonstrate that the proposed GCENet outperforms the other state-of-the-art (SOTA) RGB-D SOD methods. 相似文献
18.
19.
简要分析了交通流检测技术的发展现状,结合当前智能交通系统的应用需求,利用连续三帧差分的运动估计方法来构建初始背景,并采用统计打分的策略实时地对背景进行更新;同时提出了一种简单而有效的阴影消除算法以提高交通流参数检测的准确度。另外,针对现有交通流检测系统无车辆跟踪这一环节,可能导致流量多计数的问题,本文提出同时利用车辆的位置信息、颜色信息和分形维信息对车辆进行匹配跟踪的策略。大量实验证明该检测算法能快速、有效地检测出各种交通流参数,为实现交通管理的自动化奠定基础。 相似文献
20.
The cutting-edge RGB saliency models are prone to fail for some complex scenes, while RGB-D saliency models are often affected by inaccurate depth maps. Fortunately, light field images can provide a sufficient spatial layout depiction of 3D scenes. Therefore, this paper focuses on salient object detection of light field images, where a Similarity Retrieval-based Inference Network (SRI-Net) is proposed. Due to various focus points, not all focal slices extracted from light field images are beneficial for salient object detection, thus, the key point of our model lies in that we attempt to select the most valuable focal slice, which can contribute more complementary information for the RGB image. Specifically, firstly, we design a focal slice retrieval module (FSRM) to choose an appropriate focal slice by measuring the foreground similarity between the focal slice and RGB image. Secondly, in order to combine the original RGB image and the selected focal slice, we design a U-shaped saliency inference module (SIM), where the two-stream encoder is used to extract multi-level features, and the decoder is employed to aggregate multi-level deep features. Extensive experiments are conducted on two widely used light field datasets, and the results firmly demonstrate the superiority and effectiveness of the proposed SRI-Net. 相似文献