Stephan Ihrke’s research while affiliated with Fraunhofer Institute for Transportation and Infrastructure Systems and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (5)


Precise Edge Tracking of Vehicles Using a Static Camera Setup
  • Conference Paper

November 2018

·

14 Reads

Stephan Ihrke

BOP: Benchmark for 6D Object Pose Estimation: 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part X
  • Chapter
  • Full-text available

September 2018

·

591 Reads

·

164 Citations

Lecture Notes in Computer Science

·

·

·

[...]

·

We propose a benchmark for 6D pose estimation of a rigid object from a single RGB-D input image. The training data consists of a texture-mapped 3D object model or images of the object in known 6D poses. The benchmark comprises of: (i) eight datasets in a unified format that cover different practical scenarios, including two new datasets focusing on varying lighting conditions, (ii) an evaluation methodology with a pose-error function that deals with pose ambiguities, (iii) a comprehensive evaluation of 15 diverse recent methods that captures the status quo of the field, and (iv) an online evaluation system that is open for continuous submission of new results. The evaluation shows that methods based on point-pair features currently perform best, outperforming template matching methods, learning-based methods and methods based on 3D local features. The project website is available at bop.felk.cvut.cz.

Download

BOP: Benchmark for 6D Object Pose Estimation

August 2018

·

523 Reads

·

1 Citation

We propose a benchmark for 6D pose estimation of a rigid object from a single RGB-D input image. The training data consists of a texture-mapped 3D object model or images of the object in known 6D poses. The benchmark comprises of: i) eight datasets in a unified format that cover different practical scenarios, including two new datasets focusing on varying lighting conditions, ii) an evaluation methodology with a pose-error function that deals with pose ambiguities, iii) a comprehensive evaluation of 15 diverse recent methods that captures the status quo of the field, and iv) an online evaluation system that is open for continuous submission of new results. The evaluation shows that methods based on point-pair features currently perform best, outperforming template matching methods, learning-based methods and methods based on 3D local features. The project website is available at bop.felk.cvut.cz.


Smart Ubiquitous Projection: Discovering Surfaces for the Projection of Adaptive Content

May 2016

·

122 Reads

·

9 Citations

Ubiquitous projection or "display everywhere" is a popular paradigm, according to which regular rooms are augmented with projected digital content in order to create immersive interactive environments. In this work, we revisit this concept, where instead of considering every physical surface and object as a display, we seek to determine areas that are suitable for the projection and interaction with digital information. After determining a set of requirements that such surfaces need to fulfil, we describe a novel computer vision-based technique to automatically detect rectangular surface regions that are deemed adequate for projection and mark those areas as available placeholders for users to use as "clean" displays. As a proof of concept, we show how content can be adaptively laid out in those placeholders using a simple tablet UI.


Fig. 1. Our tracking pipeline. For each frame t the RGB-D image(a) is processed by the forest to predict object probabilities and local object coordinates(b). We use the observed depth from the original image, the forest predictions and the particles from the last frame together with our motion model(d) to construct our proposal distribution(c). Particles are sampled(e) according to the proposal distribution, then weighted(f) and resampled(g). Our final pose estimate is calculated as mean of the resampled particles(h).
Fig. 2. To construct our proposal distribution we first calculate a continuous representation of the prior distribution for the pose at the current frame (a-d). Next we determine two pose estimates H local t+1 (light gray) and H global t+1
Fig. 3. Example images of the dataset provided by Choi and Christensen [8]. 
Fig. 4. Averaged translation and rotation Fig. 5. Reconstructed motion trajectory 
Fig. 4. Averaged translation and rotation RMSE on the dataset of [8].

+2

6-DOF Model Based Tracking via Object Coordinate Regression

November 2014

·

572 Reads

·

79 Citations

Lecture Notes in Computer Science

This work investigates the problem of 6-Degrees-Of-Freedom (6-DOF) object tracking from RGB-D images, where the object is rigid and a 3D model of the object is known. As in many previous works, we utilize a Particle Filter (PF) framework. In order to have a fast tracker, the key aspect is to design a clever proposal distribution which works reliably even with a small number of particles. To achieve this we build on a recently developed state-of-the-art system for single image 6D pose estimation of known 3D objects, using the concept of so-called 3D object coordinates. The idea is to train a random forest that regresses the 3D object coordinates from the RGB-D image. Our key technical contribution is a two-way procedure to integrate the random forest predictions in the proposal distribution generation. This has many practical advantages, in particular better generalization ability with respect to occlusions, changes in lighting and fast-moving objects. We demonstrate experimentally that we exceed state-of-the-art on a given, public dataset. To raise the bar in terms of fast-moving objects and object occlusions, we also create a new dataset, which will be made publicly available.

Citations (4)


... This PPF-based approach effectively addresses issues such as a huge number of missing points and a high proportion of outliers, making it inherently suitable for multi-instance scenarios. In the BOP challenge [23][24][25], up until 2019, PPF-based methods consistently outperformed machine learning-based methods. PPF-based methods are considered the most suitable for multi-instance pose estimation when dealing with non-RGB data or data without training. ...

Reference:

Robust multi-view PPF-based method for multi-instance pose estimation
BOP: Benchmark for 6D Object Pose Estimation: 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part X

Lecture Notes in Computer Science

... AR applications can support content arrangement by suggesting and refining object placement through methods such as surface detection [87,92,121], object auto-clustering [117], or object relocation [80,91]. Full automation that minimizes manual effort has also been explored, like adaptation to the physical environments [19,38,97], user context [34,60,80], or original layouts [20]. ...

Smart Ubiquitous Projection: Discovering Surfaces for the Projection of Adaptive Content
  • Citing Conference Paper
  • May 2016

... Zhong et al. [56] used the Rigid Pose dataset for their evaluation. Furthermore, the ACCV14 dataset [57], an RGB-D dataset, was used for their evaluation. The Princeton [41] dataset is an RGB-D dataset used by Rasoulidanesh et al. [40] for evaluating their method for tracking the object along with depth. ...

6-DOF Model Based Tracking via Object Coordinate Regression

Lecture Notes in Computer Science