liu.seSearch for publications in DiVA
Change search
Link to record
Permanent link

Direct link
BETA
Mester, Rudolf
Publications (10 of 15) Show all publications
Ochs, M., Bradler, H. & Mester, R. (2016). Enhanced Phase Correlation for Reliable and Robust Estimation of Multiple Motion Distributions. In: IMAGE AND VIDEO TECHNOLOGY, PSIVT 2015: . Paper presented at 7th Pacific-Rim Symposium on Image and Video Technology (PSIVT) (pp. 368-379). Springer Publishing Company, 9431
Open this publication in new window or tab >>Enhanced Phase Correlation for Reliable and Robust Estimation of Multiple Motion Distributions
2016 (English)In: IMAGE AND VIDEO TECHNOLOGY, PSIVT 2015, Springer Publishing Company, 2016, Vol. 9431, p. 368-379Conference paper, Published paper (Refereed)
Abstract [en]

Phase correlation is one of the classic methods for sparse motion or displacement estimation. It is renowned in the literature for high precision and insensitivity against illumination variations. We propose several important enhancements to the phase correlation (PhC) method which render it more robust against those situations where a motion measurement is not possible (low structure, too much noise, too different image content in the corresponding measurement windows). This allows the method to perform self-diagnosis in adverse situations. Furthermore, we extend the PhC method by a robust scheme for detecting and classifying the presence of multiple motions and estimating their uncertainties. Experimental results on the Middlebury Stereo Dataset and on the KITTI Optical Flow Dataset show the potential offered by the enhanced method in contrast to the PhC implementation of OpenCV.

Place, publisher, year, edition, pages
Springer Publishing Company, 2016
Series
Lecture Notes in Computer Science, ISSN 0302-9743
Keywords
Optical flow; Motion estimation; Phase correlation
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:liu:diva-127789 (URN)10.1007/978-3-319-29451-3_30 (DOI)000374173000030 ()978-3-319-29451-3; 978-3-319-29450-6 (ISBN)
Conference
7th Pacific-Rim Symposium on Image and Video Technology (PSIVT)
Available from: 2016-05-12 Created: 2016-05-12 Last updated: 2016-05-12
Piccini, T., Persson, M., Nordberg, K., Felsberg, M. & Mester, R. (2015). Good Edgels to Track: Beating the Aperture Problem with Epipolar Geometry. In: Agapito, Lourdes and Bronstein, Michael M. and Rother, Carsten (Ed.), COMPUTER VISION - ECCV 2014 WORKSHOPS, PT II: . Paper presented at 13th European Conference on Computer Vision (ECCV) (pp. 652-664). Elsevier
Open this publication in new window or tab >>Good Edgels to Track: Beating the Aperture Problem with Epipolar Geometry
Show others...
2015 (English)In: COMPUTER VISION - ECCV 2014 WORKSHOPS, PT II / [ed] Agapito, Lourdes and Bronstein, Michael M. and Rother, Carsten, Elsevier, 2015, p. 652-664Conference paper, Published paper (Refereed)
Abstract [en]

An open issue in multiple view geometry and structure from motion, applied to real life scenarios, is the sparsity of the matched key-points and of the reconstructed point cloud. We present an approach that can significantly improve the density of measured displacement vectors in a sparse matching or tracking setting, exploiting the partial information of the motion field provided by linear oriented image patches (edgels). Our approach assumes that the epipolar geometry of an image pair already has been computed, either in an earlier feature-based matching step, or by a robustified differential tracker. We exploit key-points of a lower order, edgels, which cannot provide a unique 2D matching, but can be employed if a constraint on the motion is already given. We present a method to extract edgels, which can be effectively tracked given a known camera motion scenario, and show how a constrained version of the Lucas-Kanade tracking procedure can efficiently exploit epipolar geometry to reduce the classical KLT optimization to a 1D search problem. The potential of the proposed methods is shown by experiments performed on real driving sequences.

Place, publisher, year, edition, pages
Elsevier, 2015
Series
Lecture Notes in Computer Science, ISSN 0302-9743, E-ISSN 1611-3349 ; 8926
Keywords
Densification; Tracking; Epipolar geometry; Lucas-Kanade; Feature extraction; Edgels; Edges
National Category
Other Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:liu:diva-121565 (URN)10.1007/978-3-319-16181-5_50 (DOI)000362495500050 ()978-3-319-16180-8 (ISBN)
Conference
13th European Conference on Computer Vision (ECCV)
Available from: 2015-09-25 Created: 2015-09-25 Last updated: 2018-01-23Bibliographically approved
Pinggera, P., Franke, U. & Mester, R. (2015). High-Performance Long Range Obstacle Detection Using Stereo Vision. In: 2015 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS): . Paper presented at IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 1308-1313). IEEE
Open this publication in new window or tab >>High-Performance Long Range Obstacle Detection Using Stereo Vision
2015 (English)In: 2015 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), IEEE , 2015, p. 1308-1313Conference paper, Published paper (Refereed)
Abstract [en]

Reliable detection of obstacles at long range is crucial for the timely response to hazards by fast-moving safety-critical platforms like autonomous cars. We present a novel method for the joint detection and localization of distant obstacles using a stereo vision system on a moving platform. The approach is applicable to both static and moving obstacles and pushes the limits of detection performance as well as localization accuracy. The proposed detection algorithm is based on sound statistical tests using local geometric criteria which implicitly consider non-flat ground surfaces. To achieve maximum performance, it operates directly on image data instead of precomputed stereo disparity maps. A careful experimental evaluation on several datasets shows excellent detection performance and localization accuracy up to very large distances, even for small obstacles. We demonstrate a parallel implementation of the proposed system on a GPU that executes at real-time speeds.

Place, publisher, year, edition, pages
IEEE, 2015
Series
IEEE International Conference on Intelligent Robots and Systems, ISSN 2153-0858
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:liu:diva-127068 (URN)10.1109/IROS.2015.7353537 (DOI)000371885401068 ()978-1-4799-9994-1 (ISBN)
Conference
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
Available from: 2016-04-13 Created: 2016-04-13 Last updated: 2016-04-13
Fanani, N., Barnada, M. & Mester, R. (2015). Motion Priors Estimation for Robust Matching Initialization in Automotive Applications. In: Advances in Visual Computing: 11th International Symposium, ISVC 2015, Las Vegas, NV, USA, December 14-16, 2015, Proceedings, Part I. Paper presented at 11th International Symposium, ISVC 2015, Las Vegas, NV, USA, December 14-16, 2015 (pp. 115-126). SPRINGER INT PUBLISHING AG, 9474
Open this publication in new window or tab >>Motion Priors Estimation for Robust Matching Initialization in Automotive Applications
2015 (English)In: Advances in Visual Computing: 11th International Symposium, ISVC 2015, Las Vegas, NV, USA, December 14-16, 2015, Proceedings, Part I, SPRINGER INT PUBLISHING AG , 2015, Vol. 9474, p. 115-126Conference paper, Published paper (Refereed)
Abstract [en]

Tracking keypoints through a video sequence is a crucial first step in the processing chain of many visual SLAM approaches. This paper presents a robust initialization method to provide the initial match for a keypoint tracker, from the 1st frame where a keypoint is detected to the 2nd frame, that is: when no depth information is available. We deal explicitly with the case of long displacements. The starting position is obtained through an optimization that employs a distribution of motion priors based on pyramidal phase correlation, and epipolar geometry constraints. Experiments on the KITTI dataset demonstrate the significant impact of applying a motion prior to the matching. We provide detailed comparisons to the state-of-the-art methods.

Place, publisher, year, edition, pages
SPRINGER INT PUBLISHING AG, 2015
Series
Lecture Notes in Computer Science, ISSN 0302-9743 ; 9474
National Category
Computer Vision and Robotics (Autonomous Systems)
Identifiers
urn:nbn:se:liu:diva-129182 (URN)10.1007/978-3-319-27857-5_11 (DOI)000376400300011 ()9783319278575 (ISBN)9783319278568 (ISBN)
Conference
11th International Symposium, ISVC 2015, Las Vegas, NV, USA, December 14-16, 2015
Available from: 2016-06-13 Created: 2016-06-13 Last updated: 2018-01-10Bibliographically approved
Persson, M., Piccini, T., Felsberg, M. & Mester, R. (2015). Robust Stereo Visual Odometry from Monocular Techniques. In: 2015 IEEE Intelligent Vehicles Symposium (IV): . Paper presented at 2015 IEEE Intelligent Vehicles Symposium (IV), Seoul, South Korea, June 28 2015-July 1 2015 (pp. 686-691). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Robust Stereo Visual Odometry from Monocular Techniques
2015 (English)In: 2015 IEEE Intelligent Vehicles Symposium (IV), Institute of Electrical and Electronics Engineers (IEEE), 2015, p. 686-691Conference paper, Published paper (Refereed)
Abstract [en]

Visual odometry is one of the most active topics in computer vision. The automotive industry is particularly interested in this field due to the appeal of achieving a high degree of accuracy with inexpensive sensors such as cameras. The best results on this task are currently achieved by systems based on a calibrated stereo camera rig, whereas monocular systems are generally lagging behind in terms of performance. We hypothesise that this is due to stereo visual odometry being an inherently easier problem, rather than than due to higher quality of the state of the art stereo based algorithms. Under this hypothesis, techniques developed for monocular visual odometry systems would be, in general, more refined and robust since they have to deal with an intrinsically more difficult problem. In this work we present a novel stereo visual odometry system for automotive applications based on advanced monocular techniques. We show that the generalization of these techniques to the stereo case result in a significant improvement of the robustness and accuracy of stereo based visual odometry. We support our claims by the system results on the well known KITTI benchmark, achieving the top rank for visual only systems∗ .

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2015
Series
Intelligent Vehicle, IEEE Symposium, ISSN 1931-0587
Keywords
Visual odometry, VSLAM, structure from motion
National Category
Computer Vision and Robotics (Autonomous Systems)
Identifiers
urn:nbn:se:liu:diva-121829 (URN)10.1109/IVS.2015.7225764 (DOI)000380565800112 ()978-146737266-4 (ISBN)
Conference
2015 IEEE Intelligent Vehicles Symposium (IV), Seoul, South Korea, June 28 2015-July 1 2015
Available from: 2015-10-08 Created: 2015-10-08 Last updated: 2018-01-11Bibliographically approved
Bradler, H., Anne Wiegand, B. & Mester, R. (2015). The Statistics of Driving Sequences - and what we can learn from them. In: 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOP (ICCVW): . Paper presented at IEEE International Conference on Computer Vision Workshops (pp. 106-114). IEEE
Open this publication in new window or tab >>The Statistics of Driving Sequences - and what we can learn from them
2015 (English)In: 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOP (ICCVW), IEEE , 2015, p. 106-114Conference paper, Published paper (Refereed)
Abstract [en]

The motion of a driving car is highly constrained and we claim that powerful predictors can be built that learn the typical egomotion statistics, and support the typical tasks of feature matching, tracking, and egomotion estimation. We analyze the statistics of the ground truth data given in the KITTI odometry benchmark sequences and confirm that a coordinated turn motion model, overlaid by moderate vibrations, is a very realistic model. We develop a predictor that is able to significantly reduce the uncertainty about the relative motion when a new image frame comes in. Such predictors can be used to steer the matching process from frame n to frame n + 1. We show that they can also be employed to detect outliers in the temporal sequence of egomotion parameters.

Place, publisher, year, edition, pages
IEEE, 2015
National Category
Computer Vision and Robotics (Autonomous Systems)
Identifiers
urn:nbn:se:liu:diva-130844 (URN)10.1109/ICCVW.2015.24 (DOI)000380434700015 ()978-0-7695-5720-5 (ISBN)
External cooperation:
Conference
IEEE International Conference on Computer Vision Workshops
Available from: 2016-08-26 Created: 2016-08-26 Last updated: 2018-01-10
Mester, R. (2014). Motion Estimation Revisited: an Estimation-Theoretic Approach. In: 2014 IEEE SOUTHWEST SYMPOSIUM ON IMAGE ANALYSIS AND INTERPRETATION (SSIAI 2014): . Paper presented at IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI) (pp. 113-116). IEEE
Open this publication in new window or tab >>Motion Estimation Revisited: an Estimation-Theoretic Approach
2014 (English)In: 2014 IEEE SOUTHWEST SYMPOSIUM ON IMAGE ANALYSIS AND INTERPRETATION (SSIAI 2014), IEEE , 2014, p. 113-116Conference paper, Published paper (Refereed)
Abstract [en]

The present paper analyzes some previously unexplored aspects of motion estimation that are fundamental both for discrete block matching as well as for differential optical flow approaches a la Lucas-Kanade. It aims at providing a complete estimation-theoretic approach that makes the assumptions about noisy observations of samples from a continuous signal of a certain class explicit. It turns out that motion estimation is a combination of simultaneously estimating the true underlying continuous signal and optimizing the displacement between two hypothetical copies of this unknown signal. Practical schemes such as the current variants of Lucas-Kanade are just approximations to the fundamental estimation problem identified in the present paper. Derivatives appear as derivatives to the continuous signal representation kernels, not as ad hoc discrete derivative masks. The formulation via an explicit signal space defined by kernels is a precondition for analyzing e.g. the convergence range of iterative displacement estimation procedures, and for systematically chosing preconditioning filters. The paper sets the stage for further in-depth analysis of some fundamental issues that have so far been overlooked or ignored in motion analysis.

Place, publisher, year, edition, pages
IEEE, 2014
Keywords
optical flow; differential motion estimation; brightness constancy constraint equation; block matching; signal model; image derivatives
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:liu:diva-119602 (URN)10.1109/SSIAI.2014.6806042 (DOI)000355255900029 ()978-1-4799-4053-0 (ISBN)
Conference
IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI)
Available from: 2015-06-22 Created: 2015-06-22 Last updated: 2015-06-22
Mester, R. & Conrad, C. (2014). When patches match - a statistical view on matching under illumination variation. In: 2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR): . Paper presented at 22nd International Conference on Pattern Recognition (ICPR) (pp. 4364-4369). IEEE COMPUTER SOC
Open this publication in new window or tab >>When patches match - a statistical view on matching under illumination variation
2014 (English)In: 2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), IEEE COMPUTER SOC , 2014, p. 4364-4369Conference paper, Published paper (Refereed)
Abstract [en]

We discuss matching measures (scores and residuals) for comparing image patches under unknown affine photometric (=intensity) transformations. In contrast to existing methods, we derive a fully symmetric matching measure which reflects the fact that both copies of the signal are affected by measurement errors (noise), not only one. As it turns out, this evolves into an eigensystem problem; however a simple direct solution for all entities of interest can be given. We strongly advocate for constraining the estimated gain ratio and the estimated mean value offset to realistic ranges, thus preventing the matching scheme from locking into unrealistic correspondences.

Place, publisher, year, edition, pages
IEEE COMPUTER SOC, 2014
Series
International Conference on Pattern Recognition, ISSN 1051-4651
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:liu:diva-121335 (URN)10.1109/ICPR.2014.747 (DOI)000359818004084 ()978-1-4799-5208-3 (ISBN)
Conference
22nd International Conference on Pattern Recognition (ICPR)
Available from: 2015-09-14 Created: 2015-09-14 Last updated: 2015-09-14
Koschorrek, P., Piccini, T., Öberg, P., Felsberg, M., Nielsen, L. & Mester, R. (2013). A multi-sensor traffic scene dataset with omnidirectional video. In: 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW): . Paper presented at IEEE Conference on Computer Vision and Patter Recognition : Workshop on Ground Truth, June 28, 2013, Portland, U.S.A. (pp. 727-734). IEEE conference proceedings
Open this publication in new window or tab >>A multi-sensor traffic scene dataset with omnidirectional video
Show others...
2013 (English)In: 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), IEEE conference proceedings, 2013, p. 727-734Conference paper, Published paper (Refereed)
Abstract [en]

The development of vehicles that perceive their environment, in particular those using computer vision, indispensably requires large databases of sensor recordings obtained from real cars driven in realistic traffic situations. These datasets should be time shaped for enabling synchronization of sensor data from different sources. Furthermore, full surround environment perception requires high frame rates of synchronized omnidirectional video data to prevent information loss at any speeds.

This paper describes an experimental setup and software environment for recording such synchronized multi-sensor data streams and storing them in a new open source format. The dataset consists of sequences recorded in various environments from a car equipped with an omnidirectional multi-camera, height sensors, an IMU, a velocity sensor, and a GPS. The software environment for reading these data sets will be provided to the public, together with a collection of long multi-sensor and multi-camera data streams stored in the developed format.

Place, publisher, year, edition, pages
IEEE conference proceedings, 2013
Series
IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, ISSN 2160-7508
National Category
Computer Vision and Robotics (Autonomous Systems)
Identifiers
urn:nbn:se:liu:diva-93277 (URN)10.1109/CVPRW.2013.110 (DOI)000331116100112 ()978-0-7695-4990-3 (ISBN)
Conference
IEEE Conference on Computer Vision and Patter Recognition : Workshop on Ground Truth, June 28, 2013, Portland, U.S.A.
Available from: 2013-05-29 Created: 2013-05-29 Last updated: 2018-01-11
Conrad, C., Mertz, M. & Mester, R. (2013). Contour-relaxed Superpixels. In: Heyden, A., Kahl, F., Olsson, C., Oskarsson, M., Tai, X.-C. (Ed.), : . Paper presented at EMMCVPR 2013. 9th International Conference Energy Minimization Methods in Computer Vision and Pattern Recognition, August 19-21, Lund, Sweden (pp. 280-293). Springer Berlin/Heidelberg
Open this publication in new window or tab >>Contour-relaxed Superpixels
2013 (English)In: / [ed] Heyden, A., Kahl, F., Olsson, C., Oskarsson, M., Tai, X.-C., Springer Berlin/Heidelberg, 2013, p. 280-293Conference paper, Published paper (Refereed)
Abstract [en]

We propose and evaluate a versatile scheme for image pre-segmentation that generates a partition of the image into a selectable number of patches (’superpixels’), under the constraint of obtaining maximum homogeneity of the ’texture’ inside of each patch, and maximum accordance of the contours with both the image content as well as a Gibbs-Markov random field model. In contrast to current state-of-the art approaches to superpixel segmentation, ’homogeneity’ does not limit itself to smooth region-internal signals and high feature value similarity between neighboring pixels, but is applicable also to highly textured scenes. The energy functional that is to be maximized for this purpose has only a very small number of design parameters, depending on the particular statistical model used for the images.

The capability of the resulting partitions to deform according to the image content can be controlled by a single parameter. We show by means of an extensive comparative experimental evaluation that the compactness-controlled contour-relaxed superpixels method outperforms the state-of-the art superpixel algorithms with respect to boundary recall and undersegmentation error while being faster or on a par with respect to runtime.

Place, publisher, year, edition, pages
Springer Berlin/Heidelberg, 2013
Series
Lecture Notes in Computer Science, ISSN 0302-9743, E-ISSN 1611-3349 ; 8081
National Category
Signal Processing
Identifiers
urn:nbn:se:liu:diva-93239 (URN)10.1007/978-3-642-40395-8_21 (DOI)978-3-642-40394-1 (ISBN)978-3-642-40395-8 (ISBN)
Conference
EMMCVPR 2013. 9th International Conference Energy Minimization Methods in Computer Vision and Pattern Recognition, August 19-21, Lund, Sweden
Projects
ELLIIT
Funder
ELLIIT - The Linköping‐Lund Initiative on IT and Mobile Communications
Available from: 2013-05-28 Created: 2013-05-28 Last updated: 2018-02-20Bibliographically approved
Organisations

Search in DiVA

Show all publications