Abstract:Fusion of vision-based and inertial pose estimation has many high-potential applications in navigation, robotics, and augmented reality. Our research aims at the development of a fully mobile, completely self-contained tracking system, that is able to estimate sensor motion from known 3D scene structure. This requires a highly modular and scalable software architecture for algorithm design and testing. As the main contribution of this paper, we discuss the design of our hybrid tracker and emphasize important f… Show more
“…Also in Eino et al [15], where inertial data from an IMU is fused with the velocity estimation from a vision algorithm, no details about the scale problem are reported. Ribo et al [16] proposed an EKF to fuse vision and inertial data to estimate the 6DoF attitude. There and in [10,[17][18][19] the authors use a priori knowledge to overcome the scale problem.…”
The fusion of inertial and visual data is widely used to improve an object's pose estimation. However, this type of fusion is rarely used to estimate further unknowns in the visual framework. In this paper we present and compare two different approaches to estimate the unknown scale parameter in a monocular SLAM framework. Directly linked to the scale is the estimation of the object's absolute velocity and position in 3D. The first approach is a spline fitting task adapted from Jung and Taylor and the second is an extended Kalman filter. Both methods have been simulated offline on arbitrary camera paths to analyze their behavior and the quality of the resulting scale estimation. We then embedded an online multi rate extended Kalman filter in the Parallel Tracking and Mapping (PTAM) algorithm of Klein and Murray together with an inertial sensor. In this inertial/monocular SLAM framework, we show a real time, robust and fast converging scale estimation. Our
“…Also in Eino et al [15], where inertial data from an IMU is fused with the velocity estimation from a vision algorithm, no details about the scale problem are reported. Ribo et al [16] proposed an EKF to fuse vision and inertial data to estimate the 6DoF attitude. There and in [10,[17][18][19] the authors use a priori knowledge to overcome the scale problem.…”
The fusion of inertial and visual data is widely used to improve an object's pose estimation. However, this type of fusion is rarely used to estimate further unknowns in the visual framework. In this paper we present and compare two different approaches to estimate the unknown scale parameter in a monocular SLAM framework. Directly linked to the scale is the estimation of the object's absolute velocity and position in 3D. The first approach is a spline fitting task adapted from Jung and Taylor and the second is an extended Kalman filter. Both methods have been simulated offline on arbitrary camera paths to analyze their behavior and the quality of the resulting scale estimation. We then embedded an online multi rate extended Kalman filter in the Parallel Tracking and Mapping (PTAM) algorithm of Klein and Murray together with an inertial sensor. In this inertial/monocular SLAM framework, we show a real time, robust and fast converging scale estimation. Our
“…Ribo et al [8,9] present a wearable AR system that is mounted on a helmet. It consists of a real-time 3D visualization subsystem (composed by a stereo see-through HMD) and a real-time tracking subsystem (composed by a camera and an IMU).…”
Section: Previous Work On Hybrid Trackingmentioning
The precise localization of human operators in robotic workplaces is an important requirement to be satisfied in order to develop human-robot interaction tasks. Human tracking provides not only safety for human operators, but also context information for intelligent human-robot collaboration. This paper evaluates an inertial motion capture system which registers full-body movements of an user in a robotic manipulator workplace. However, the presence of errors in the global translational measurements returned by this system has led to the need of using another localization system, based on Ultra-WideBand (UWB) technology. A Kalman filter fusion algorithm which combines the measurements of these systems is developed. This algorithm unifies the advantages of both technologies: high data rates from the motion capture system and global translational precision from the UWB localization system. The developed hybrid system not only tracks the movements of all limbs of the user as previous motion capture systems, but is also able to position precisely the user in the environment.
“…The 3D pose is computed from artificial landmarks [7] as depicted in figure 3(a). To avoid deficits in visual tracking of the landmark, an inertial tracker located at the top of the AR gear aids the tracking process [23]. By means of this hybrid tracking approach, the precise position and orientation of the user T pose can be computed yielding only a very small relative mean distance error of 0.5%.…”
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.