首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到19条相似文献,搜索用时 187 毫秒
1.
针对UWB定位性能易受障碍物遮挡、非视距干扰的问题,提出了一种新的UWB指纹匹配定位算法。该算法利用基站与各定位标签之间的距离信息建立指纹库,并在KNN定位算法的基础上,引入了模糊推理方法,通过模糊规则处理得到待定位节点与k个参考节点的匹配度,把该匹配度作为权值对KNN算法进行加权,获得初始定位,同时,创新性地提出了位置优化阈T,根据阈值T和初始定位结果与k参考节点的欧式距离大小,判断是否进行二次模糊加权处理。测试显示,该算法定位误差保持在10cm左右,并且和一次模糊推理的加权KNN算法比较,优化算法定位精度提高了17.8%,提高了UWB室内定位的精确度和稳健性  相似文献   

2.
郭伟斌  王洪光  姜勇  孙鹏 《机器人》2012,34(5):620-627
为了保证轮臂复合式巡检机器人自动越障时行走轮能够可靠抓线,提出一种基于图像的视觉伺服抓线控制方法.在图像空间定义输电线的偏距和偏角来表征图像特征的变化,并设计出位姿解耦伺服控制律.由于视觉雅可比矩阵与机器人状态有关且可能发生奇异,采用模糊方法进行机器人控制.通过实验确定模糊集合的论域,设计了两层模糊控制器并编制了自动抓线程序.进行了上百次实验,其抓线控制的偏距和偏角误差分别在25像素和2以内,能够可靠抓线,实际越障中只需小于40s的时间即可完成抓线.结果表明该方法具有准确、可靠、效率高的特点,能够满足准确、快速、可靠抓线的任务要求.  相似文献   

3.
基于节点生长k-均值聚类算法的强化学习方法   总被引:3,自引:0,他引:3  
处理连续状态强化学习问题,主要方法有两类:参数化的函数逼近和自适应离散划分.在分析了现有对连续状态空间进行自适应划分方法的优缺点的基础上,提出了一种基于节点生长k均值聚类算法的划分方法,分别给出了在离散动作和连续动作两种情况下该强化学习方法的算法步骤.在离散动作的MountainCar问题和连续动作的双积分问题上进行仿真实验.实验结果表明,该方法能够根据状态在连续空间的分布,自动调整划分的精度,实现对于连续状态空间的自适应划分,并学习到最佳策略.  相似文献   

4.
本文提出了一种用于机器人的大系统递阶协调自适应控制方法.一方面补偿机器人因建模的不准确以及负载的未知带来的误差;另一方面引入大系统的递阶协调控制实现系统控制算法的并行性,为机器人的在线控制研究提供了一条途径.本文最后给出的例子证明了该方法是可实现的.  相似文献   

5.
本文提出了一种新的、有效的机器人自适应控制方式,克服了其他方法由于模型不准或计算量大等所带来的一系列问题。本文首先将 Lagrange 运动方程转化为 ARMA 模型,并用虚拟噪声补偿模型误差(即由于线性化、解耦、观测不准和干扰等误差).然后利用改进的 Kalman 自适应滤波算法在线进行参数辨识和状态估计,将获得的参数用于机器人控制系统自适应控制器的设计.最后给出了该算法的仿真结果并对此进行了讨论。  相似文献   

6.
基于增强学习的多机器人系统优化控制是近年来机器人学与分布式人工智能的前沿研究领域.多机器人系统具有分布、异构和高维连续空间等特性,使得面向多机器人系统的增强学习的研究面临着一系列挑战,为此,对其相关理论和算法的研究进展进行了系统综述.首先,阐述了多机器人增强学习的基本理论模型和优化目标;然后,在对已有学习算法进行对比分析的基础上,重点探讨了多机器人增强学习理论与应用研究中的困难和求解思路,给出了若干典型问题和应用实例;最后,对相关研究进行了总结和展望.  相似文献   

7.
基于FNN的覆冰机器人越障机械臂轨迹跟踪控制   总被引:1,自引:1,他引:0       下载免费PDF全文
覆冰机器人除冰时要跨越各种障碍物。采用卡尔曼滤波学习算法,将自适应模糊神经网络控制器用于覆冰机器人越障时的机械臂轨迹跟踪控制,解决了BP算法实时性差的问题。经过仿真实验论证,该方法对覆冰机器人越障时的机械臂轨迹跟踪控制具有很好的效果,表明控制策略和理论分析的可行性。  相似文献   

8.
由于除冰机器人多在天气恶劣,覆冰较厚的输电线路上工作,现有的基于视觉伺服越障策略存在图像质量差,冰、线区分难等不足。根据模糊逻辑和粒子群优化原理,提出了一种除冰机器人在线越障和路径规划方法。该方法通过模糊规划器实现除冰机器人机械臂的无障跟踪和平稳越障。在此基础上,以机械臂末端经过路径长度和与目标点距离的综合最小为目标,利用粒子群算法对模糊规划器输出角度进行在线优化。仿真结果表明:与传统的模糊越障规划相比,该方法不仅满足除冰机器人实时规划和自主越障的要求,缩短了机械臂经过轨迹的长度,提高了除冰机器人的工作效率和续航能力,为实际工程应用中除冰机器人的能源短缺问题,提供了一种节约使用方案,具有一定的工程应用价值。  相似文献   

9.
传统KNN算法是在基于距离的离群检测算法的基础上提出的一种在大数据集下进行离群点挖掘的算法,然而KNN算法只以最近的第k个邻居的距离作为判断是否是离群点的标准有时也失准确性.给出了一种在大数据集下基于KNN的离群点检测算法,即在传统KNN方法的基础上为每个数据点增加了权重,权重值为与最近的k个邻居的平均距离,离群点为那些与第k个邻居的距离最大且相同条件下权重最大的点.算法能提高离群点检测的准确性,通过实验验证了算法的可行性,并与传统KNN算法的性能进行了对比.  相似文献   

10.
通过观察可以发现连续七近邻查询中KNN发生改变的必要条件是第k个邻居发生变化,因此不需要监测所有k近邻,只需要监测第k个邻居即可.该方法采用边界线来监测第k个邻居的变化,不过这需要将原始空间转变为时间-距离(TD)空间后进行操作.在TD空间中每一个对象用一个时间函数来表示,通过监测当前第七个邻居的前视矩形区域来构造边界线.实验结果表明,边界线算法在七非常大的时候是最有效的.  相似文献   

11.
Robust motion control is fundamental to autonomous mobile robots. In the past few years, reinforcement learning (RL) has attracted considerable attention in the feedback control of wheeled mobile robot. However, it is still difficult for RL to solve problems with large or continuous state spaces, which is common in robotics. To improve the generalization ability of RL, this paper presents a novel hierarchical RL approach for optimal path tracking of wheeled mobile robots. In the proposed approach, a graph Laplacian-based hierarchical approximate policy iteration (GHAPI) algorithm is developed, in which the basis functions are constructed automatically using the graph Laplacian operator. In GHAPI, the state space of an Markov decision process is divided into several subspaces and approximate policy iteration is carried out on each subspace. Then, a near-optimal path-tracking control strategy can be obtained by GHAPI combined with proportional-derivative (PD) control. The performance of the proposed approach is evaluated by using a P3-AT wheeled mobile robot. It is demonstrated that the GHAPI-based PD control can obtain better near-optimal control policies than previous approaches.  相似文献   

12.
Reinforcement learning (RL) is a popular method for solving the path planning problem of autonomous mobile robots in unknown environments. However, the primary difficulty faced by learning robots using the RL method is that they learn too slowly in obstacle-dense environments. To more efficiently solve the path planning problem of autonomous mobile robots in such environments, this paper presents a novel approach in which the robot’s learning process is divided into two phases. The first one is to accelerate the learning process for obtaining an optimal policy by developing the well-known Dyna-Q algorithm that trains the robot in learning actions for avoiding obstacles when following the vector direction. In this phase, the robot’s position is represented as a uniform grid. At each time step, the robot performs an action to move to one of its eight adjacent cells, so the path obtained from the optimal policy may be longer than the true shortest path. The second one is to train the robot in learning a collision-free smooth path for decreasing the number of the heading changes of the robot. The simulation results show that the proposed approach is efficient for the path planning problem of autonomous mobile robots in unknown environments with dense obstacles.  相似文献   

13.
A new formation navigation approach derived from multi-robots cooperative online FastSLAM is proposed. In this approach,the leader and follower robots are defined.The posteriori estimation of the leader robot state is treated as a relative reference for all follower robots to correct their state priori estimations.The control volume of individual follower will be achieved from the results of the corrected estimation.All robots are observed as landmarks with known associations by the others and are considered in their landmarks updating.By the method,the errors of the robot posterior estimations are reduced and the formation is well kept.The simulation and physical experiment results show that the multi-robots relative localization accuracy is improved and the formation navigation control is more stable and efficient than normal leader-following strategy.The algorithm is easy in implementation.  相似文献   

14.
《Advanced Robotics》2013,27(10):1125-1142
This paper presents a novel approach for acquiring dynamic whole-body movements on humanoid robots focused on learning a control policy for the center of mass (CoM). In our approach, we combine both a model-based CoM controller and a model-free reinforcement learning (RL) method to acquire dynamic whole-body movements in humanoid robots. (i) To cope with high dimensionality, we use a model-based CoM controller as a basic controller that derives joint angular velocities from the desired CoM velocity. The balancing issue can also be considered in the controller. (ii) The RL method is used to acquire a controller that generates the desired CoM velocity based on the current state. To demonstrate the effectiveness of our approach, we apply it to a ball-punching task on a simulated humanoid robot model. The acquired whole-body punching movement was also demonstrated on Fujitsu's Hoap-2 humanoid robot.  相似文献   

15.
Reinforcement learning (RL) for robot control is an important technology for future robots since it enables us to design a robot’s behavior using the reward function. However, RL for high degree-of-freedom robot control is still an open issue. This paper proposes a discrete action space DCOB which is generated from the basis functions (BFs) given to approximate a value function. The remarkable feature is that, by reducing the number of BFs to enable the robot to learn quickly the value function, the size of DCOB is also reduced, which improves the learning speed. In addition, a method WF-DCOB is proposed to enhance the performance, where wire-fitting is utilized to search for continuous actions around each discrete action of DCOB. We apply the proposed methods to motion learning tasks of a simulated humanoid robot and a real spider robot. The experimental results demonstrate outstanding performance.  相似文献   

16.
足球机器人系统仿真中的碰撞研究   总被引:2,自引:0,他引:2  
薛方正  冯挺  徐心和 《机器人》2005,27(1):78-81
总结了近年来足球机器人碰撞的研究成果,指出当前一些碰撞模型的不足,进而指出了数字仿真系统的离散特性对足球机器人的碰撞性质的影响. 详细地分析了足球机器人的运动碰撞情况. 最后提出了足球机器人的碰撞模型并给出了具体的算法. 实践证明该算法是有效的.  相似文献   

17.
基于模糊神经网络的强化学习及其在机器人导航中的应用   总被引:5,自引:0,他引:5  
段勇  徐心和 《控制与决策》2007,22(5):525-529
研究基于行为的移动机器人控制方法.将模糊神经网络与强化学习理论相结合,构成模糊强化系统.它既可获取模糊规则的结论部分和模糊隶属度函数参数,也可解决连续状态空间和动作空间的强化学习问题.将残差算法用于神经网络的学习,保证了函数逼近的快速性和收敛性.将该系统的学习结果作为反应式自主机器人的行为控制器,有效地解决了复杂环境中的机器人导航问题.  相似文献   

18.
Industrial robots are versatile mechanical systems that require accurate tracking of continuous end-effector trajectories. However, a variety of control problems are encountered due to the deviation between the desired and actual paths.In this study, a new continuous path planning method based on an interpolation of orientation scheme is applied for precise path generation in robot welding. This method guarantees minimum deviation of positioning and orientation errors. Also, a new trajectory error evaluation strategy is developed to describe the trajectory errors at the effect points, which are very important in some robot jobs such as arc-welding operations.The simulation study of circular motions in arc-welding operations shows the effectiveness of the proposed approach.  相似文献   

19.
双轮驱动移动机器人的学习控制器设计方法*   总被引:1,自引:0,他引:1  
提出一种基于增强学习的双轮驱动移动机器人路径跟随控制方法,通过将机器人运动控制器的优化设计问题建模为Markov决策过程,采用基于核的最小二乘策略迭代算法(KLSPI)实现控制器参数的自学习优化。与传统表格型和基于神经网络的增强学习方法不同,KLSPI算法在策略评价中应用核方法进行特征选择和值函数逼近,从而提高了泛化性能和学习效率。仿真结果表明,该方法通过较少次数的迭代就可以获得优化的路径跟随控制策略,有利于在实际应用中的推广。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号