Hyungpil Moon, Jose Martinez-Carranza, Titus Cieslewski, Matthias Faessler, Davide Falanga, Alessandro Simovic, Davide Scaramuzza, Shuo Li, Michael Ozo, Christophe De Wagter, Guido de Croon, Sunyou Hwang, Sunggoo Jung, Hyunchul Shim, Haeryang Kim, Minhyuk Park, Tsz-Chiu Au, Si Jung Kim, Challenges and implemented technologies used in autonomous drone racing, Intelligent Service Robotics, Vol. 12 (2), 2019. (Journal Article)
Autonomous Drone Racing (ADR) is a challenge for autonomous drones to navigate a cluttered indoor environment without relying on any external sensing in which all the sensing and computing must be done on board. Although no team could complete the whole racing track so far, most successful teams implemented waypoint tracking methods and robust visual recognition of the gates of distinct colors because the complete environmental information
was given to participants before the events. In this paper, we introduce the purpose of ADR as a benchmark testing ground for autonomous drone
technologies and analyze the challenges and technologies used in the two previous ADRs held in IROS 2016 and IROS 2017. Six teams that participated in these events present their implemented technologies that cover modifyed ORBSLAM, robust alignment method for waypoints deployment, sensor fusion for motion estimation, deep learning for gate detection and motion control, and stereo-vision for gate detection. |
|
Davide Falanga, Suseong Kim, Davide Scaramuzza, How Fast Is Too Fast? The Role of Perception Latency in High-Speed Sense and Avoid, IEEE Robotics and Automation Letters, Vol. 4 (2), 2019. (Journal Article)
In this letter, we study the effects that perception latency has on the maximum speed a robot can reach to safely navigate through an unknown cluttered environment. We provide a general analysis that can serve as a baseline for future quantitative reasoning for design tradeoffs in autonomous robot navigation. We consider the case where the robot is modeled as a linear secondorder system with bounded input and navigates through static obstacles. Also, we focus on a scenario where the robot wants to reach a target destination in as little time as possible, and therefore cannot change its longitudinal velocity to avoid obstacles. We show how the maximum latency that the robot can tolerate to guarantee safety is related to the desired speed, the range of its sensing pipeline, and the actuation limitations of the platform (i.e., the maximum acceleration it can produce). As a particular case study, we compare monocular and stereo frame-based cameras against novel, low-latency sensors, such as event cameras, in the case of quadrotor flight. To validate our analysis, we conduct experiments on a quadrotor platform equipped with an event camera to detect and avoid obstacles thrown towards the robot. To the best of our knowledge, this is the first theoretical work in which perception and actuation limitations are jointly considered to study the performance of a robotic platform in high-speed navigation. |
|
Davide Falanga, Kevin Kleber, Stefano Mintchev, Dario Floreano, Davide Scaramuzza, The Foldable Drone: A Morphing Quadrotor That Can Squeeze and Fly, IEEE Robotics and Automation Letters, Vol. 4 (2), 2019. (Journal Article)
The recent advances in state estimation, perception, and navigation algorithms have significantly contributed to the ubiquitous use of quadrotors for inspection, mapping, and aerial imaging. To further increase the versatility of quadrotors, recent works investigated the use of an adaptive morphology, which consists of modifying the shape of the vehicle during flight to suit a specific task or environment. However, these works either increase the complexity of the platform or decrease its controllability. In this letter, we propose a novel, simpler, yet effective morphing design for quadrotors consisting of a frame with four independently rotating arms that fold around the main frame. To guarantee stable flight at all times, we exploit an optimal control strategy that adapts on the fly to the drone morphology. We demonstrate the versatility of the proposed adaptive morphology in different tasks, such as negotiation of narrow gaps, close inspection of vertical surfaces, and object grasping and transportation. The experiments are performed on an actual, fully autonomous quadrotor relying solely on onboard visual-inertial sensors and compute. No external motion tracking systems and computers are used. This is the first work showing stable flight without requiring any symmetry of the morphology. |
|
Suseong Kim, Davide Falanga, Davide Scaramuzza, Computing the Forward Reachable Set for a Multirotor Under First-Order Aerodynamic Effects, IEEE Robotics and Automation Letters, Vol. 3 (4), 2019. (Journal Article)
Collision avoidance plays a crucial role in safe multirotor flight in cluttered environments. Even though a given reference trajectory is designed to be collision free, it might lead to collision due to imperfect tracking caused by external disturbances. In this work, we tackle this problem by computing the Forward Reachable Set (FRS), which is the set of positions and velocities that a multirotor can reach while following a reference trajectory due to tracking errors. Hence, if the FRS is computed before flight, we can utilize it to check the safety of a given trajectory in terms of collision avoidance. To compute a realistic FRS that covers an agile flight envelope, we consider first-order aerodynamic effects, which have the most salient influence on the vehicle. For computing FRS, we conduct a thorought stability analysis including these aerodynamic effects. Then, we present a FRS computation method which can easily be adapted to newly given reference trajectories. The presented method is validated by comparing the FRS with real flight data collected during agile and high-speed flight. In addition, we compare the FRS computed with and without compensating for firstorder aerodynamic effect to highlight their significance on the trajectory tracking performance. To the best of our knowledge, this is the first attempt to compute FRSs by incorporating firstorder aerodynamic effects for multirotors. |
|
Elias Mueggler, Guillermo Gallego, Henri Rebecq, Davide Scaramuzza, Continuous-Time Visual-Inertial Odometry for Event Cameras, IEEE Transactions on Robotics, Vol. 34 (6), 2019. (Journal Article)
Event cameras are bio-inspired vision sensors that output pixel-level brightness changes instead of standard intensity frames. They offer significant advantages over standard cameras, namely a very high dynamic range, no motion blur, and a latency in the order of microseconds. However, due to the fundamentally different structure of the sensor’s output, new algorithms that exploit the high temporal resolution and the asynchronous nature of the sensor are required. Recent work has shown that a continuous-time representation of the event camera pose can deal with the high temporal resolution and asynchronous nature of this sensor in a principled way. In this paper, we leverage such a continuous-time representation to perform visual-inertial odometry with an event camera. This representation allows direct integration of the asynchronous events with micro-second accuracy and the inertial measurements at high frequency. The event camera trajectory is approximated by a smooth curve in the space of rigid-body motions using cubic splines. This formulation significantly reduces the number of variables in trajectory estimation problems. We evaluate our method on real data from several scenes and compare the results against ground truth from a motion-capture system. We show that our method provides improved accuracy over the result of a state-of-the-art visual odometry method for event cameras. We also show that both the map orientation and scale can be recovered accurately by fusing events and inertial data. To the best of our knowledge, this is the first work on visual-inertial fusion with event cameras using a continuous-time framework. |
|
Zichao Zhang, Guillermo Gallego, Davide Scaramuzza, On the Comparison of Gauge Freedom Handling in Optimization-Based Visual-Inertial State Estimation, IEEE Robotics and Automation Letters, Vol. 3 (3), 2019. (Journal Article)
It is well known that visual-inertial state estimation is possible up to a four degrees-of-freedom (DoF) transformation (rotation around gravity and translation), and the extra DoFs (“gauge freedom”) have to be handled properly. While different approaches for handling the gauge freedom have been used in practice, no previous study has been carried out to systematically analyze their differences. In this paper, we present the first comparative analysis of different methods for handling the gauge freedom in optimization-based visual-inertial state estimation.We experimentally compare three commonly used approaches: fixing the unobservable states to some given values, setting a prior on such states, or letting the states evolve freely during optimization. Specifically, we show that (i) the accuracy and computational time of the three methods are similar, with the free gauge approach being slightly faster; (ii) the covariance estimation from the free gauge approach appears dramatically different, but is actually tightly related to the other approaches. Our findings are validated both in simulation and on real-world datasets and can be useful for designing optimization-based visual-inertial state estimation algorithms. |
|
Henri Rebecq, Guillermo Gallego, Elias Müggler, Davide Scaramuzza, EMVS: Event-Based Multi-View Stereo - 3D Reconstruction with an Event Camera in Real-Time, International Journal of Computer Vision, Vol. 126 (12), 2018. (Journal Article)
Event cameras are bio-inspired vision sensors that output pixel-level brightness changes instead of standard intensity frames. They offer significant advantages over standard cameras, namely a very high dynamic range, no motion blur, and a latency in the order of microseconds. However, because the output is composed of a sequence of asynchronous events rather than actual intensity images, traditional vision algorithms cannot be applied, so that a paradigm shift is needed. We introduce the problem of event-based multi-view stereo (EMVS) for event cameras and propose a solution to it. Unlike traditional MVS methods, which address the problem of estimating dense 3D structure from a set of known viewpoints, EMVS estimates semi-dense 3D structure from an event camera with known trajectory. Our EMVS solution elegantly exploits two inherent properties of an event camera: (1) its ability to respond to scene edges—which naturally provide semi-dense geometric information without any pre-processing operation—and (2) the fact that it provides continuous measurements as the sensor moves. Despite its simplicity (it can be implemented in a few lines of code), our algorithm is able to produce accurate, semi-dense depth maps, without requiring any explicit data association or intensity estimation. We successfully validate our method on both synthetic and real data. Our method is computationally very efficient and runs in real-time on a CPU. |
|
Zichao Zhang, Davide Scaramuzza, A Tutorial on Quantitative Trajectory Evaluation for Visual(-Inertial) Odometry, In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, 2018-11-01. (Conference or Workshop Paper published in Proceedings)
In this tutorial, we provide principled methods to quantitatively evaluate the quality of an estimated trajectory from visual(-inertial) odometry (VO/VIO), which is the foundation of benchmarking the accuracy of different algorithms. First, we show how to determine the transformation type to use in trajectory alignment based on the specific sensing modality (i.e., monocular, stereo and visual-inertial). Second, we describe commonly used error metrics (i.e., the absolute trajectory error and the relative error) and their strengths and weaknesses. To make the methodology presented for VO/VIO applicable to other setups, we also generalize our formulation to any given sensing modality. To facilitate the reproducibility of related research, we publicly release our implementation of the methods described in this tutorial. |
|
Davide Falanga, Philipp Foehn, Peng Lu, Davide Scaramuzza, PAMPC: Perception-Aware Model Predictive Control for Quadrotors, In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, 2018-11-01. (Conference or Workshop Paper published in Proceedings)
We present the first perception-aware model predictive control framework for quadrotors that unifies control and planning with respect to action and perception objectives. Our framework leverages numerical optimization to compute trajectories that satisfy the system dynamics and require control inputs within the limits of the platform. Simultaneously, it optimizes perception objectives for robust and reliable sensing by maximizing the visibility of a point of interest and minimizing its velocity in the image plane. Considering both perception and action objectives for motion planning and control is challenging due to the possible conflicts arising from their respective requirements. For example, for a quadrotor to track a reference trajectory, it needs to rotate to align its thrust with the direction of the desired acceleration. However, the perception objective might require to minimize such rotation to maximize the visibility of a point of interest. A model-based optimization framework, able to consider both perception and action objectives and couple them through the system dynamics, is therefore necessary. Our perception-aware model predictive control framework works in a receding-horizon fashion by iteratively solving a non-linear optimization problem. It is capable of running in real-time, fully onboard our lightweight, small-scale quadrotor using a low-power ARM computer, together with a visual-inertial odometry pipeline. We validate our approach in experiments demonstrating (I) the conflict between perception and action objectives, and (II) improved behavior in extremely challenging lighting conditions. |
|
Henri Rebecq, Daniel Gehrig, Davide Scaramuzza, ESIM: an Open Event Camera Simulator, In: Conference on Robot Learning (CoRL), Zurich, 2018, CoRL, Conference on Robot Learning (CoRL), Zurich, 2018., 2018-10-01. (Conference or Workshop Paper published in Proceedings)
Event cameras are revolutionary sensors that work radically differently from standard cameras. Instead of capturing intensity images at a fixed rate, event cameras measure changes of intensity asynchronously, in the form of a stream of events, which encode per-pixel brightness changes. In the last few years, their outstanding properties (asynchronous sensing, no motion blur, high dynamic range) have led to exciting vision applications, with very low-latency and high robustness. However, these sensors are still scarce and expensive to get, slowing down progress of the research community. To address these issues, there is a huge demand for cheap, high-quality synthetic, labeled event for algorithm prototyping, deep learning and algorithm benchmarking. The development of such a simulator, however, is not trivial since event cameras work fundamentally differently from framebased cameras. We present the first event camera simulator that can generate a large amount of reliable event data. The key component of our simulator is a theoretically sound, adaptive rendering scheme that only samples frames when necessary, through a tight coupling between the rendering engine and the event simulator. We release an open source implementation of our simulator. |
|
Elia Kaufmann, Antonio Loquercio, Rene Ranftl, Alexey Dosovitskiy, Vladlen Koltun, Davide Scaramuzza, Deep Drone Racing: Learning Agile Flight in Dynamic Environments, In: Conference on Robotic Learning (CoRL), Zurich, 2018, CoRL, Conference on Robotic Learning (CoRL), Zurich, 2018., 2018-10-01. (Conference or Workshop Paper published in Proceedings)
Autonomous agile flight brings up fundamental challenges in robotics, such as coping with unreliable state estimation, reacting optimally to dynamically changing environments, and coupling perception and action in real time under severe resource constraints. In this paper, we consider these challenges in the context of autonomous, vision-based drone racing in dynamic environments. Our approach combines a convolutional neural network (CNN) with a state-of-the-art path-planning and control system. The CNN directly maps raw images into a robust representation in the form of a waypoint and desired speed. This information is then used by the planner to generate a short, minimum-jerk trajectory segment and corresponding motor commands to reach the desired goal. We demonstrate our method in autonomous agile flight scenarios, in which a vision-based quadrotor traverses drone-racing tracks with possibly moving gates. Our method does not require any explicit map of the environment and runs fully onboard. We extensively test the precision and robustness of the approach in simulation and in the physical world. We also evaluate our method against state-of-the-art navigation approaches and professional human drone pilots. |
|
Alex Scheitlin, Improving Build Failure Resolution Through In-IDE Assistance, University of Zurich, Faculty of Business, Economics and Informatics, 2018. (Bachelor's Thesis)
Continuous integration is an essential part of modern software engineering and helps programmers to automate their build process (e.g., compiling source code, packaging binary code, or running tests). While developers benefit from earlier caught bugs, more stable code bases, and shorter software release cycles, they still face several problems when working with continuous integration. For example, if a software build fails, additional time is needed to resolve detected problems by build servers (i.e., build failures). Developers first need to locate the root of failure and then fix the error. For that purpose, they leave their Integrated Development Environment (IDE) and scan through long log files on the build server.
To simplify and reduce the effort required for this process, the vision is to bridge the gap between the build-related information available on the build server and the local development environment where the source code needs to be fixed. Therefore, CAESAR (Ci Assistant for (Build Failure) Resolution and Summarization) was developed, which is an IDE plugin that leverages detailed build-related information to support developers in debugging build failures directly within their IDE. The tool summarizes build logs, classifies errors, shows the files in which the errors occurred, gives hints or error descriptions, and enables to directly debug the error. This is all possible without leaving the IDE.
To evaluate the usefulness of CAESAR, a controlled experiment was conducted, showing that on average, developers could resolve build failures 48.4% faster when working with CAESAR. Especially, the evaluation of the experiment showed that developers working with CAESAR were able to reduce the time needed to identify the error in the build log and locate it within the source code. Lastly, participants stated that, thanks to CAESARís assistance, context switches between their own IDE and the build server were no longer required. |
|
Guillermo Gallego, Henri Rebecq, Davide Scaramuzza, A Unifying Contrast Maximization Framework for Event Cameras, with Applications to Motion, Depth, and Optical Flow Estimation, In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2018-07-18. (Conference or Workshop Paper published in Proceedings)
We present a unifying framework to solve several computer vision problems with event cameras: motion, depth and optical flow estimation. The main idea of our framework is to find the point trajectories on the image plane that are best aligned with the event data by maximizing an objective function: the contrast of an image of warped events. Our method implicitly handles data association between the events, and therefore, does not rely on additional appearance information about the scene. In addition to accurately recovering the motion parameters of the problem, our framework produces motion-corrected edge-like images with high dynamic range that can be used for further scene analysis. The proposed method is not only simple, but more importantly, it is, to the best of our knowledge, the first method that can be successfully applied to such a diverse set of important vision tasks with event cameras. |
|
Ana I. Maqueda, Antonio Loquercio, Guillermo Gallego, Narciso Garcia, Davide Scaramuzza, Event-Based Vision Meets Deep Learning on Steering Prediction for Self-Driving Cars, In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2018-07-18. (Conference or Workshop Paper published in Proceedings)
Event cameras are bio-inspired vision sensors that naturally
capture the dynamics of a scene, filtering out redundant
information. This paper presents a deep neural
network approach that unlocks the potential of event cameras
on a challenging motion-estimation task: prediction
of a vehicle’s steering angle. To make the best out of this
sensor–algorithm combination, we adapt state-of-the-art
convolutional architectures to the output of event sensors
and extensively evaluate the performance of our approach
on a publicly available large scale event-camera dataset
(1000 km). We present qualitative and quantitative explanations
of why event cameras allow robust steering prediction
even in cases where traditional cameras fail, e.g. challenging
illumination conditions and fast motion. Finally, we
demonstrate the advantages of leveraging transfer learning
from traditional to event-based vision, and show that our
approach outperforms state-of-the-art algorithms based on
standard cameras. |
|
Riccardp Spica, Davide Falanga, Erik Cristofalo, Eduardo Montijano, Davide Scaramuzza, Mac Schwager, A Real-Time Game Theoretic Planner for Autonomous Two-Player Drone Racing, In: Robotics: Science and Systems, s.n., Robotics: Science and Systems, 2018-06-01. (Conference or Workshop Paper published in Proceedings)
To be successful in multi-player drone racing, a player must not only follow the race track in an optimal way, but also compete with other drones through strategic blocking, faking, and opportunistic passing while avoiding collisions. Since unveiling one’s own strategy to the adversaries is not desirable, this requires each player to independently predict the other players’ future actions. Nash equilibria are a powerful tool to model this and similar multi-agent coordination problems in which the absence of communication impedes full coordination between the agents. In this paper, we propose a novel receding horizon planning algorithm that, exploiting sensitivity analysis within an iterated best response computational scheme, can approximate Nash equilibria in real time. We demonstrate that our solution effectively competes against alternative strategies in a large number of drone racing simulations. |
|
Jeffrey Delmerico, Davide Scaramuzza, A Benchmark Comparison of Monocular Visual-Inertial Odometry Algorithms for Flying Robots, In: IEEE International Conference on Robotics and Automation (ICRA), 2018., IEEE, IEEE International Conference on Robotics and Automation (ICRA), 2018., 2018-05-21. (Conference or Workshop Paper published in Proceedings)
Flying robots require a combination of accuracy and low latency in their state estimation in order to achieve stable and robust flight. However, due to the power and payload constraints of aerial platforms, state estimation algorithms must provide these qualities under the computational constraints of embedded hardware. Cameras and inertial measurement units (IMUs) satisfy these power and payload constraints, so visualinertial odometry (VIO) algorithms are popular choices for state estimation in these scenarios, in addition to their ability to operate without external localization from motion capture or global positioning systems. It is not clear from existing results in the literature, however, which VIO algorithms perform well under the accuracy, latency, and computational constraints of a flying robot with onboard state estimation. This paper evaluates an array of publicly-available VIO pipelines (MSCKF, OKVIS, ROVIO, VINS-Mono, SVO+MSF, and SVO+GTSAM) on different hardware configurations, including several singleboard computer systems that are typically found on flying robots. The evaluation considers the pose estimation accuracy, per-frame processing time, and CPU and memory load while processing the EuRoC datasets, which contain six degree of freedom (6DoF) trajectories typical of flying robots. We present our complete results as a benchmark for the research community. Narrated video presentation: https://youtu.be/ymI3FmwU9AY |
|
Philipp Föhn, Davide Scaramuzza, Onboard State Dependent LQR for Agile Quadrotors, In: IEEE International Conference on Robotics and Automation (ICRA), 2018., IEEE, IEEE International Conference on Robotics and Automation (ICRA), 2018., 2018-05-21. (Conference or Workshop Paper published in Proceedings)
State-of-the-art approaches in quadrotor control split the problem into multiple cascaded subproblems, exploiting the different time scales of the rotational and translational dynamics. They calculate a desired acceleration as input for a cascaded attitude controller but omit the attitude dynamics. These approaches use limits on the desired acceleration to maintain feasibility and robustness through the control cascade. We propose an implementation of an LQR controller, which: (I) is linearized depending on the quadrotor’s state; (II) unifies the control of rotational and translational states; (III) handles time-varying system dynamics and control parameters. Our implementation is efficient enough to compute the full linearization and solution of the LQR at a minimum of 10Hz on the vehicle using a common ARM processor. We show four successful experiments: (I) controlling at hover state with large disturbances; (II) tracking along a trajectory; (III) tracking along an infeasible trajectory; (IV) tracking along a trajectory with disturbances. All the experiments were done using only onboard visual inertial state estimation and LQR computation. To the best of our knowledge, this is the first implementation and evaluation of a state-dependent LQR capable of onboard computation while providing this amount of versatility and performance. Video of the experiments: https://youtu.be/8OVsJNgNfa0 Narrated video presentation: https://youtu.be/c7gHF-NJjPo |
|
Ruben Gomez-Ojeda, Zichao Zhang, Javier Gonzalez-Jimenez, Davide Scaramuzza, Learning-based Image Enhancement for Visual Odometry in Challenging HDR Environments, In: IEEE International Conference on Robotics and Automation (ICRA), 2018., IEEE, IEEE International Conference on Robotics and Automation (ICRA), 2018., 2018-05-21. (Conference or Workshop Paper published in Proceedings)
One of the main open challenges in visual odometry (VO) is the robustness to difficult illumination conditions or high dynamic range (HDR) environments. The main difficulties in these situations come from both the limitations of the sensors and the inability to perform a successful tracking of interest points because of the bold assumptions in VO, such as brightness constancy. We address this problem from a deep learning perspective, for which we first fine-tune a deep neural network with the purpose of obtaining enhanced representations of the sequences for VO. Then, we demonstrate how the insertion of long short term memory allows us to obtain temporally consistent sequences, as the estimation depends on previous states. However, the use of very deep networks enlarges the computational burden of the VO framework; therefore, we also propose a convolutional neural network of reduced size capable of performing faster. Finally, we validate the enhanced representations by evaluating the sequences produced by the two architectures in several state-of-art VO algorithms, such as ORB-SLAM and DSO. |
|
Titus Cieslewski, Siddharth Choudhary, Davide Scaramuzza, Data-Efficient Decentralized Visual SLAM, In: IEEE International Conference on Robotics and Automation (ICRA), 2018., IEEE, IEEE International Conference on Robotics and Automation (ICRA), 2018., 2018-05-01. (Conference or Workshop Paper published in Proceedings)
Decentralized visual simultaneous localization and mapping (SLAM) is a powerful tool for multi-robot applications in environments where absolute positioning is not available. Being visual, it relies on cheap, lightweight and versatile cameras, and, being decentralized, it does not rely on communication to a central entity. In this work, we integrate state-of-theart decentralized SLAM components into a new, complete decentralized visual SLAM system. To allow for data association and optimization, existing decentralized visual SLAM systems exchange the full map data among all robots, incurring large data transfers at a complexity that scales quadratically with the robot count. In contrast, our method performs efficient data association in two stages: first, a compact full-image descriptor is deterministically sent to only one robot. Then, only if the first stage succeeded, the data required for relative pose estimation is sent, again to only one robot. Thus, data association scales linearly with the robot count and uses highly compact place representations. For optimization, a state-of-theart decentralized pose-graph optimization method is used. It exchanges a minimum amount of data which is linear with trajectory overlap. We characterize the resulting system and identify bottlenecks in its components. The system is evaluated on publicly available datasets and we provide open access to the code. A narrated video presentation can be found at: https://youtu.be/zEBfCA5tVOk |
|
Zichao Zhang, Davide Scaramuzza, Perception-aware Receding Horizon Navigation for MAVs, In: IEEE International Conference on Robotics and Automation (ICRA), 2018., IEEE, IEEE International Conference on Robotics and Automation (ICRA), 2018., 2018-05-01. (Conference or Workshop Paper published in Proceedings)
To reach a given destination safely and accurately, a micro aerial vehicle needs to be able to avoid obstacles and minimize its state estimation uncertainty at the same time. To achieve this goal, we propose a perception-aware receding horizon approach. In our method, a single forwardlooking camera is used for state estimation and mapping. Using the information from the monocular state estimation and mapping system, we generate a library of candidate trajectories and evaluate them in terms of perception quality, collision probability, and distance to the goal. The best trajectory to execute is then selected as the one that maximizes a reward function based on these three metrics. To the best of our knowledge, this is the first work that integrates active vision within a receding horizon navigation framework for a goal reaching task. We demonstrate by simulation and real-world experiments on an actual quadrotor that our active approach leads to improved state estimation accuracy in a goal-reaching task when compared to a purely-reactive navigation system, especially in difficult scenes (e.g., weak texture). A video showing the experiments can be found at https://youtu.be/761zxZMeQNo A narrated video presentation can be found here: https://www.youtube.com/watch?v=FK6S_CRXiuI |
|