Davide Scaramuzza, Omnidirectional Camera, In: Computer Vision: A Reference Guide, Springer US, New York, p. 552 - 560, 2014. (Book Chapter)
|
|
Davide Scaramuzza, Amazon-Drohne hebt noch lange nicht ab, In: 20 Minuten, p. 12, 3 December 2013. (Newspaper Article)
Schon 2015 sollen Amazon-Drohnen Pakete zustellen. Experten zweifeln am Zeitplan. Ist das Ganze nur ein PR-Gag? |
|
Christian Forster, Matia Pizzoli, Davide Scaramuzza, Air-ground localization and map augmentation using monocular dense reconstruction, In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Institute of Electrical and Electronics Engineers, Tokyo, Japan, 2013-11-03. (Conference or Workshop Paper published in Proceedings)
We propose a new method for the localization of a Micro Aerial Vehicle (MAV) with respect to a ground robot. We solve the problem of registering the 3D maps computed by the robots using different sensors: a dense 3D reconstruction from the MAV monocular camera is aligned with the map computed from the depth sensor on the ground robot. Once aligned, the dense reconstruction from the MAV is used to augment the map computed by the ground robot, by extending it with the information conveyed by the aerial views. The overall approach is novel, as it builds on recent developments in live dense reconstruction from moving cameras to address the problem of air-ground localization. The core of our contribution is constituted by a novel algorithm integrating dense reconstructions from monocular views, Monte Carlo localization, and an iterative pose refinement. In spite of the radically different vantage points from which the maps are acquired, the proposed method achieves high accuracy whereas appearance-based, state-of-the-art approaches fail. Experimental validation in indoor and outdoor scenarios reported an accuracy in position estimation of 0.08 meters and real time performance. This demonstrates that our new approach effectively overcomes the limitations imposed by the difference in sensors and vantage points that negatively affect previous techniques relying on matching visual features. |
|
Davide Scaramuzza, Die fliegenden Kuriere, In: Die Zeit, p. 1, 8 September 2013. (Newspaper Article)
|
|
Davide Scaramuzza, Der Spion von deinem Fenster, In: WeltWoche, p. 1, 28 March 2013. (Newspaper Article)
|
|
Andrea Censi, Davide Scaramuzza, Calibration by correlation using metric embedding from non-metric similarities, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 35 (10), 2013. (Journal Article)
This paper presents a new intrinsic calibration method that allows us to calibrate a generic single-view point camera. From the video sequence obtained while the camera undergoes random motion, we compute the pairwise time correlation of the luminance signal for the pixels. We show that the pairwise correlation of any pixels pair is a function of the distance between the pixel directions on the visual sphere. This leads to formalizing calibration as a problem of metric embedding from non-metric measurements: we want to find the disposition of pixels on the visual sphere, from similarities that are an unknown function of the distances. This problem is a generalization of multidimensional scaling (MDS) that has so far resisted a comprehensive observability analysis and a generic solution. We show that the observability depends both on the local geometric properties as well as on the global topological properties of the target manifold. It follows that, in contrast to the Euclidean case, on the sphere we can recover the scale of the points distribution. We describe an algorithm that is robust across manifolds and can recover a metrically accurate solution when the metric information is observable. We demonstrate the performance of the algorithm for several cameras (pin-hole, fish-eye, omnidirectional). |
|
Special issue: Micro-UAV Perception and Control, Edited by: Nathan Michael, Davide Scaramuzza, Vijay Kumar, Springer, New York, 2012-08-01. (Edited Scientific Work)
|
|
Roy Anati, Davide Scaramuzza, Konstantinos G Derpanis, Kostas Daniilidis, Robot localization using soft object detection, In: IEEE International Conference on Robotics and Automation, Institute of Electrical and Electronics Engineers, 2012-05-14. (Conference or Workshop Paper published in Proceedings)
In this paper, we give a new double twist to the robot localization problem. We solve the problem for the case of prior maps which are semantically annotated perhaps even sketched by hand. Data association is achieved not through the detection of visual features but the detection of object classes used in the annotation of the prior maps. To avoid the caveats of general object recognition, we propose a new representation of the query images that consists of a vector of the detection scores for each object class. Given such soft object detections we are able to create hypotheses about pose and to refine them through particle filtering. As opposed to small confined office and kitchen spaces, our experiment takes place in a large open urban rail station with multiple semantically ambiguous places. The success of our approach shows that our new representation is a robust way to exploit the plethora of existing prior maps for GPS-denied environments avoiding the data association problems when matching point clouds or visual features. |
|
Lefteris Doitsidis, Stephan Weiss, Alessandro Renzaglia, Markus W Achtelik, Elias Kosmatopoulos, Roland Siegwart, Davide Scaramuzza, Optimal surveillance coverage for teams of micro aerial vehicles in GPS-Denied environments using onboard vision, Autonomous Robots, Vol. 33 (1-2), 2012. (Journal Article)
This paper deals with the problem of deploying a team of flying robots to perform surveillance-coverage missions over a terrain of arbitrary morphology. In such missions, a key factor for the successful completion of the mission is the knowledge of the terrain’s morphology. The focus of this paper is on the implementation of a two-step procedure that allows us to optimally align a team of flying vehicles for the aforementioned task. Initially, a single robot constructs a map of the area using a novel monocular-vision-based approach. A state-of-the-art visual-SLAM algorithm tracks the pose of the camera while, simultaneously, autonomously, building an incremental map of the environment. The map generated is processed and serves as an input to an optimization procedure using the cognitive, adaptive methodology initially introduced in Renzaglia et al. (Proceedings of the IEEE international conference on robotics and intelligent system (IROS), Taipei, Taiwan, pp. 3314–3320, 2010). The output of this procedure is the optimal arrangement of the robots team, which maximizes the monitored area. The efficiency of our approach is demonstrated using real data collected from aerial robots in different outdoor areas. |
|
Friedrich Fraundorfer, Davide Scaramuzza, Visual Odometry: Part II - Matching, Robustness, and Applications, IEEE Robotics and Automation Magazine, Vol. 19 (2), 2012. (Journal Article)
Part II of the tutorial has summarized the remaining building blocks of the VO pipeline: specifically, how to detect and match salient and repeatable features across frames and robust estimation in the presence of outliers and bundle adjustment. In addition, error propagation, applications, and links to publicly available code are included. VO is a well understood and established part of robotics. VO has reached a maturity that has allowed us to successfully use it for certain classes of applications: space, ground, aerial, and underwater. In the presence of loop closures, VO can be used as a building block for a complete SLAM algorithm to reduce motion drift. Challenges that still remain are to develop and demonstrate large-scale and long-term implementations, such as driving autonomous cars for hundreds of miles. Such systems have recently been demonstrated using Lidar and Radar sensors [86]. However, for VO to be used in such systems, technical issues regarding robustness and, especially, long-term stability have to be resolved. Eventually, VO has the potential to replace Lidar-based systems for egomotion estimation, which are currently leading the state of the art in accuracy, robustness, and reliability. VO offers a cheaper and mechanically easier-to-manufacture solution for egomotion estimation, while, additionally, being fully passive. Furthermore, the ongoing miniaturization of digital cameras offers the possibility to develop smaller and smaller robotic systems capable of ego-motion estimation. |
|