Lidé

Ing. Denys Rozumnyi

Všechny publikace

Motion-from-Blur: 3D Shape and Motion Estimation of Motion-blurred Objects in Videos

  • Autoři: Ing. Denys Rozumnyi, Oswald, M.R., Ferrari, V., Pollefeys, M.
  • Publikace: Proceeding 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Piscataway: IEEE, 2022. p. 15969-15978. ISSN 2575-7075. ISBN 978-1-6654-6946-3.
  • Rok: 2022
  • DOI: 10.1109/CVPR52688.2022.01552
  • Odkaz: https://doi.org/10.1109/CVPR52688.2022.01552
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    We propose a method for jointly estimating the 3D motion, 3D shape, and appearance of highly motion-blurred objects from a video. To this end, we model the blurred appearance of a fast moving object in a generative fashion by parametrizing its 3D position, rotation, velocity, acceleration, bounces, shape, and texture over the duration of a predefined time window spanning multiple frames. Using differentiable rendering, we are able to estimate all parameters by minimizing the pixel-wise reprojection error to the input video via backpropagating through a rendering pipeline that accounts for motion blur by averaging the graphics output over short time intervals. For that purpose, we also estimate the camera exposure gap time within the same optimization. To account for abrupt motion changes like bounces, we model the motion trajectory as a piece-wise polynomial, and we are able to estimate the specific time of the bounce at sub-frame accuracy. Experiments on established benchmark datasets demonstrate that our method outperforms previous methods for fast moving object deblurring and 3D reconstruction.

DeFMO: Deblurring and Shape Recovery of Fast Moving Objects

  • Autoři: Ing. Denys Rozumnyi, Oswald, M.R., Ferrari, V., prof. Ing. Jiří Matas, Ph.D., Pollefeys, M.
  • Publikace: Proceedings of 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). USA: IEEE Computer Society, 2021. p. 3455-3464. ISSN 2575-7075. ISBN 978-1-6654-4509-2.
  • Rok: 2021
  • DOI: 10.1109/CVPR46437.2021.00346
  • Odkaz: https://doi.org/10.1109/CVPR46437.2021.00346
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    Objects moving at high speed appear significantly blurred when captured with cameras. The blurry appearance is especially ambiguous when the object has complex shape or texture. In such cases, classical methods, or even humans, are unable to recover the object's appearance and motion. We propose a method that, given a single image with its estimated background, outputs the object's appearance and position in a series of sub-frames as if captured by a high-speed camera (i.e. temporal super-resolution). The proposed generative model embeds an image of the blurred object into a latent space representation, disentangles the background, and renders the sharp appearance. Inspired by the image formation model, we design novel self-supervised loss function terms that boost performance and show good generalization capabilities. The proposed DeFMO method is trained on a complex synthetic dataset, yet it performs well on real-world data from several datasets. DeFMO outperforms the state of the art and generates high-quality temporal super-resolution frames.

FMODetect: Robust Detection of Fast Moving Objects

  • Autoři: Ing. Denys Rozumnyi, prof. Ing. Jiří Matas, Ph.D., Šroubek, F., Pollefeys, M., Oswald, M.R.
  • Publikace: ICCV2021: Proceedings of the International Conference on Computer Vision. Piscataway: IEEE, 2021. p. 3521-3529. ISSN 2380-7504. ISBN 978-1-6654-2812-5.
  • Rok: 2021
  • DOI: 10.1109/ICCV48922.2021.00352
  • Odkaz: https://doi.org/10.1109/ICCV48922.2021.00352
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    We propose the first learning-based approach for fast moving objects detection. Such objects are highly blurred and move over large distances within one video frame. Fast moving objects are associated with a deblurring and matting problem, also called deblatting. We show that the separation of deblatting into consecutive matting and deblurring allows achieving real-time performance, i.e. an order of magnitude speed-up, and thus enabling new classes of application. The proposed method detects fast moving objects as a truncated distance function to the trajectory by learning from synthetic data. For the sharp appearance estimation and accurate trajectory estimation, we propose a matting and fitting network that estimates the blurred appearance without background, followed by an energy minimization based deblurring. The state-of-the-art methods are outperformed in terms of recall, precision, trajectory estimation, and sharp appearance reconstruction. Compared to other methods, such as deblatting, the inference is of several orders of magnitude faster and allows applications such as real-time fast moving object detection and retrieval in large video collections.

Tracking by Deblatting

  • DOI: 10.1007/s11263-021-01480-w
  • Odkaz: https://doi.org/10.1007/s11263-021-01480-w
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    Objects moving at high speed along complex trajectories often appear in videos, especially videos of sports. Such objects travel a considerable distance during exposure time of a single frame, and therefore, their position in the frame is not well defined. They appear as semi-transparent streaks due to the motion blur and cannot be reliably tracked by general trackers. We propose a novel approach called Tracking by Deblatting based on the observation that motion blur is directly related to the intra-frame trajectory of an object. Blur is estimated by solving two intertwined inverse problems, blind deblurring and image matting, which we call deblatting. By postprocessing, non-causal Tracking by Deblatting estimates continuous, complete, and accurate object trajectories for the whole sequence. Tracked objects are precisely localized with higher temporal resolution than by conventional trackers. Energy minimization by dynamic programming is used to detect abrupt changes of motion, called bounces. High-order polynomials are then fitted to smooth trajectory segments between bounces. The output is a continuous trajectory function that assigns location for every real-valued time stamp from zero to the number of frames. The proposed algorithm was evaluated on a newly created dataset of videos from a high-speed camera using a novel Trajectory-IoU metric that generalizes the traditional Intersection over Union and measures the accuracy of the intra-frame trajectory. The proposed method outperforms the baselines both in recall and trajectory accuracy. Additionally, we show that from the trajectory function precise physical calculations are possible, such as radius, gravity, and sub-frame object velocity. Velocity estimation is compared to the high-speed camera measurements and radars. Results show high performance of the proposed method in terms of Trajectory-IoU, recall, and velocity estimation.

Sub-Frame Appearance and 6D Pose Estimation of Fast Moving Objects

  • Autoři: Ing. Denys Rozumnyi, Kotera, J., Šroubek, F., prof. Ing. Jiří Matas, Ph.D.,
  • Publikace: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). USA: IEEE Computer Society, 2020. p. 6777-6785. ISSN 2575-7075. ISBN 978-1-7281-7168-5.
  • Rok: 2020
  • DOI: 10.1109/CVPR42600.2020.00681
  • Odkaz: https://doi.org/10.1109/CVPR42600.2020.00681
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    We propose a novel method that tracks fast moving objects, mainly non-uniform spherical, in full 6 degrees of freedom, estimating simultaneously their 3D motion trajectory, 3D pose and object appearance changes with a time step that is a fraction of the video frame exposure time. The sub-frame object localization and appearance estimation allows realistic temporal super-resolution and precise shape estimation. The method, called TbD-3D (Tracking by Deblatting in 3D) relies on a novel reconstruction algorithm which solves a piece-wise deblurring and matting problem. The 3D rotation is estimated by minimizing the reprojection error. As a second contribution, we present a new challenging dataset with fast moving objects that change their appearance and distance to the camera. High-speed camera recordings with zero lag between frame exposures were used to generate videos with different frame rates annotated with ground-truth trajectory and pose.

Intra-frame Object Tracking by Deblatting

  • Autoři: Kotera, J., Ing. Denys Rozumnyi, Šroubek, F., prof. Ing. Jiří Matas, Ph.D.,
  • Publikace: 2019 IEEE International Conference on Computer Vision Workshops (ICCVW 2019). Los Alamitos: IEEE Computer Society, 2019. p. 2300-2309. ISSN 2473-9944. ISBN 978-1-7281-5023-9.
  • Rok: 2019
  • DOI: 10.1109/ICCVW.2019.00283
  • Odkaz: https://doi.org/10.1109/ICCVW.2019.00283
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    Objects moving at high speed along complex trajectories often appear in videos, especially videos of sports. Such objects elapse non-negligible distance during exposure time of a single frame and therefore their position in the frame is not well defined. They appear as semi-transparent streaks due to the motion blur and cannot be reliably tracked by standard trackers. We propose a novel approach called Tracking by Deblatting based on the observation that motion blur is directly related to the intra-frame trajectory of an object. Blur is estimated by solving two intertwined inverse problems, blind deblurring and image matting, which we call deblatting. The trajectory is then estimated by fitting a piecewise quadratic curve, which models physically justifiable trajectories. As a result, tracked objects are precisely localized with higher temporal resolution than by conventional trackers. The proposed TbD tracker was evaluated on a newly created dataset of videos with ground truth obtained by a high-speed camera using a novel Trajectory-IoU metric that generalizes the traditional Intersection over Union and measures the accuracy of the intra-frame trajectory. The proposed method outperforms baseline both in recall and trajectory accuracy.

Learned Semantic Multi-Sensor Depth Map Fusion

  • Autoři: Ing. Denys Rozumnyi, Cherabier, I., Pollefeys, M., Oswald, M.
  • Publikace: 2019 IEEE International Conference on Computer Vision Workshops (ICCVW 2019). Los Alamitos: IEEE Computer Society, 2019. p. 2089-2099. ISSN 2473-9944. ISBN 978-1-7281-5023-9.
  • Rok: 2019
  • DOI: 10.1109/ICCVW.2019.00264
  • Odkaz: https://doi.org/10.1109/ICCVW.2019.00264
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    Volumetric depth map fusion based on truncated signed distance functions has become a standard method and is used in many 3D reconstruction pipelines. In this paper, we are generalizing this classic method in multiple ways: 1) Semantics: Semantic information enriches the scene representation and is incorporated into the fusion process. 2) Multi-Sensor: Depth information can originate from different sensors or algorithms with very different noise and outlier statistics which are considered during data fusion. 3) Scene denoising and completion: Sensors can fail to recover depth for certain materials and light conditions, or data is missing due to occlusions. Our method denoises the geometry, closes holes and computes a watertight surface for every semantic class. 4) Learning: We propose a neural network reconstruction method that unifies all these properties within a single powerful framework. Our method learns sensor or algorithm properties jointly with semantic depth fusion and scene completion and can also be used as an expert system, e.g. to unify the strengths of various photometric stereo algorithms. Our approach is the first to unify all these properties. Experimental evaluations on both synthetic and real data sets demonstrate clear improvements.

Non-causal Tracking by Deblatting

  • DOI: 10.1007/978-3-030-33676-9_9
  • Odkaz: https://doi.org/10.1007/978-3-030-33676-9_9
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    Tracking by Deblatting (Deblatting = deblurring and matting) stands for solving an inverse problem of deblurring and image matting for tracking motion-blurred objects. We propose non-causal Tracking by Deblatting which estimates continuous, complete and accurate object trajectories. Energy minimization by dynamic programming is used to detect abrupt changes of motion, called bounces. High-order polynomials are fitted to segments, which are parts of the trajectory separated by bounces. The output is a continuous trajectory function which assigns location for every real-valued time stamp from zero to the number of frames. Additionally, we show that from the trajectory function precise physical calculations are possible, such as radius, gravity or sub-frame object velocity. Velocity estimation is compared to the high-speed camera measurements and radars. Results show high performance of the proposed method in terms of Trajectory-IoU, recall and velocity estimation.

The World of Fast Moving Objects

  • Autoři: Ing. Denys Rozumnyi, Kotěra, J., Šroubek, F., Novotný, L., prof. Ing. Jiří Matas, Ph.D.,
  • Publikace: CVPR 2017: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition. IEEE Computer Society Press, 2017. p. 4838-4846. ISSN 1063-6919. ISBN 978-1-5386-0457-1.
  • Rok: 2017
  • DOI: 10.1109/CVPR.2017.514
  • Odkaz: https://doi.org/10.1109/CVPR.2017.514
  • Pracoviště: Katedra kybernetiky, Skupina vizuálního rozpoznávání
  • Anotace:
    The notion of a Fast Moving Object (FMO), i.e. an object that moves over a distance exceeding its size within the exposure time, is introduced. FMOs may, and typically do, rotate with high angular speed. FMOs are very common in sports videos, but are not rare elsewhere. In a single frame, such objects are often barely visible and appear as semitransparent streaks. A method for the detection and tracking of FMOs is proposed. The method consists of three distinct algorithms, which form an efficient localization pipeline that operates successfully in a broad range of conditions. We show that it is possible to recover the appearance of the object and its axis of rotation, despite its blurred appearance. The proposed method is evaluated on a new annotated dataset. The results show that existing trackers are inadequate for the problem of FMO localization and a new approach is required. Two applications of localization, temporal superresolution and highlighting, are presented.

Coplanar Repeats by Energy Minimization

  • DOI: 10.5244/C.30.107
  • Odkaz: https://doi.org/10.5244/C.30.107
  • Pracoviště: Skupina vizuálního rozpoznávání
  • Anotace:
    This paper proposes an automated method to detect, group and rectify arbitrarily arranged coplanar repeated elements via energy minimization. The proposed energy functional combines several features that model how planes with coplanar repeats are projected into images and captures global interactions between different coplanar repeat groups and scene planes. An inference framework based on a recent variant of α-expansion is described and fast convergence is demonstrated. We compare the proposed method to two widely-used geometric multi-model fitting methods using a new dataset of annotated images containing multiple scene planes with coplanar repeats in varied arrangements. The evaluation shows a significant improvement in the accuracy of rectifications computed from coplanar repeats detected with the proposed method versus those detected with the baseline methods.

Za stránku zodpovídá: Ing. Mgr. Radovan Suk