[1]WU Ming,LI Guangyu,WEI Zhenhua,et al.Bearing only target tracking and observability control of a mobile robot[J].CAAI Transactions on Intelligent Systems,2022,17(5):919-930.[doi:10.11992/tis.202107066]
Copy

Bearing only target tracking and observability control of a mobile robot

References:
[1] 蔡自兴, 邹小兵. 移动机器人环境认知理论与技术的研究[J]. 机器人, 2004, 26(1): 87–91
CAI Zixing, ZOU Xiaobing. Research on environmental cognition theory and methodology for mobile robots[J]. Robot, 2004, 26(1): 87–91
[2] YANG Nan, WANG Rui, GAO Xiang, et al. Challenges in monocular visual odometry: photometric calibration, motion bias, and rolling shutter effect[J]. IEEE robotics and automation letters, 2018, 3(4): 2878–2885.
[3] WANG C C, THORPE C, THRUN S, et al. Simultaneous localization, mapping and moving object tracking[J]. The international journal of robotics research, 2007, 26(9): 889–916.
[4] 伍明, 李琳琳, 孙继银. 基于概率数据关联交互多模滤波的移动机器人未知环境下动态目标跟踪[J]. 机器人, 2012, 34(6): 668–679
WU Ming, LI Linlin, SUN Jiyin. PDA-IMM based moving object tracking with mobile robots in unknown environments[J]. Robot, 2012, 34(6): 668–679
[5] MUNARO M, LEWIS C, CHAMBERS D, et al. RGB-D human detection and tracking for industrial environments[M]//Intelligent Autonomous Systems 13. Cham: Springer International Publishing, 2015: 1655?1668.
[6] ZHANG Rui, WANG Zhaokui, ZHANG Yulin. Astronaut visual tracking of flying assistant robot in space station based on deep learning and probabilistic model[J]. International journal of aerospace engineering, 2018, 2018: 1–17.
[7] CASTLER O, MURRAYDW. Object recognition and localization while tracking and mapping[C]// Proceedings of 8th IEEE International Symposium on Mixed andAugmented Reality. Orlando, IEEE, 2009: 179?180.
[8] KLEIN G, MURRAY D. Parallel tracking and mapping for small AR workspaces[C]// Proceedings of6th IEEE and ACM International Symposium on Mixed and Augmented Reality. Nara, IEEE, 2007: 225?234.
[9] DE OLIVEIRA D CB, DE SOUZA DA SILVA R L. Moving object tracking for SLAM-based augmented reality[J]. Journalof mobile multimedia, 2021, 17(4): 577–602.
[10] DAI W, ZHANG Y, ZHENG Y, et al. RGB-D SLAM with moving object tracking in dynamic environments[J]. IET cyber-systems and robotics, 2021, 3(4): 11.
[11] LI Guihai, CHEN Songlin. Visual slam in dynamic scenes based on object tracking and static points detection[J]. Journal of intelligent & robotic systems, 2022, 104(2): 1–10.
[12] CIVERA J, DAVISON A J, MONTIEL J M M. Inverse depth parametrization for monocular SLAM[J]. IEEE transactions on robotics, 2008, 24(5): 932–945.
[13] DISSANAYAKE M W M G, NEWMAN P, CLARK S, et al. A solution to the simultaneous localization and map building (SLAM) problem[J]. IEEE transactions on robotics and automation, 2001, 17(3): 229–241.
[14] WANGSIRIPITAK S, MURRAY D W. Avoiding moving outliers in visual SLAM by tracking moving objects[C]// Proceedings of IEEE International Conference on Robotics and Automation. Kobe, IEEE, 2009: 375?380.
[15] MIGLIORE D, RIGAMONTI R, MARZORATI D, et al. Use a single camera for simultaneous localization and mapping with mobile object tracking in dynamic environments[C]//Proceedings of IEEE International Conference on Robotics and Automation, Kobe, Japan. IEEE, 2009: 889?895
[16] BESCOS B, CAMPOS C, TARDóS J D, et al. DynaSLAM II: tightly-coupled multi-object tracking and SLAM[J]. IEEE robotics and automation letters, 2021, 6(3): 5191–5198.
[17] LIU Yuzhen, LIU Jiacheng, HAO Yun, et al. A switching-coupled backend for simultaneous localization and dynamic object tracking[J]. IEEE robotics and automation letters, 2021, 6(2): 1296–1303.
[18] SUALEH M, KIM G W. Semantics aware dynamic SLAM based on 3D MODT[J]. Sensors (Basel, Switzerland), 2021, 21(19): 6355.
[19] LIU Jiacheng, MENG Ziyang, YOU Zheng. A robust visual SLAM system in dynamic man-made environments[J]. Science China technological sciences, 2020, 63(9): 1628–1636.
[20] OH R, SHIYifang, CHOI J W. A hybrid Newton-raphson and particle swarm optimization method for target motion analysis by batch processing[J]. Sensors (Basel, Switzerland), 2021, 21(6): 2033.
[21] ZHUK R S. Automatic detection and tracking the moving objects observed by an unmanned aerial vehicles video camera[J]. Informatics, 2021, 18(2): 83–97.
[22] WATANABE Y. Stochastically optimized monocular vision-based navigation and guidance[D]. USA: Georgia Institute of Technology, 2008: 57.
[23] KIM J, ROCK S. Stochastic feedback controller design considering the dual effect[J]//AIAA Guidance, Navigation, and Control Conference and Exhibit. Keystone, Colorado. Reston, Virginia: AIAA, 2006: 6090.
[24] SCH?LLERC, ARAVANTINOS V, LAY F, et al. What the constant velocity model can teach us about pedestrian motion prediction[J]. IEEE robotics and automation letters, 2020, 5(2): 1696–1703.
[25] ZHANG Z. A flexible new technique for camera calibration[J]. IEEE transactionson pattern analysis and machine intelligence, 2000, 22(11): 1330–1334.
[26] DAI Y, LEE S G. The leader-follower formation control of non-holonomic mobile robots[J]. International journal of control, automation and systems, 2012, 10(2): 350–361.
Similar References:

Memo

-

Last Update: 1900-01-01

Copyright © CAAI Transactions on Intelligent Systems