Lidé

prof. Ing. Tomáš Svoboda, Ph.D.

Proděkan pro rozvoj, garant programu Kybernetika a robotika - doktorský

Všechny publikace

Learning to Predict Lidar Intensities

  • DOI: 10.1109/TITS.2020.3037980
  • Odkaz: https://doi.org/10.1109/TITS.2020.3037980
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    We propose a data-driven method for simulating lidar sensors. The method reads computer-generated data, and (i) extracts geometrically simulated lidar point clouds and (ii) predicts the strength of the lidar response – lidar intensities. Qualitative valuation of the proposed pipeline demonstrates the ability to predict systematic ailures such as no/low responses on polished parts of car bodyworks and windows, for strong responses on reflective surfaces such as traffic signs and license/registration plates. We also experimentally show that enhancing the training set by such simulated data improves the segmentation accuracy on the real dataset with limited access to real data. Implementation of the resulting lidar simulator for the GTA V game, as well as the accompanying large dataset, is made publicly available.

Pose consistency KKT-loss for weakly supervised learning of robot-terrain interaction model

  • DOI: 10.1109/LRA.2021.3076957
  • Odkaz: https://doi.org/10.1109/LRA.2021.3076957
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    We address the problem of self-supervised learning for predicting the shape of supporting terrain (i.e. the terrain which will provide rigid support for the robot during its traversal) from sparse input measurements. The learning method exploits two types of ground-truth labels: dense 2.5D maps and robot poses, both estimated by a usual SLAM procedure from offline recorded measurements. We show that robot poses are required because straightforward supervised learning from the 3D maps only suffers from: (i) exaggerated height of the supporting terrain caused by terrain flexibility (vegetation, shallow water, snow or sand) and (ii) missing or noisy measurements caused by high spectral absorbance or non-Lambertian reflectance of the measured surface. We address the learning from robot poses by introducing a novel KKT-loss, which emerges as the distance from necessary Karush-Kuhn-Tucker conditions for constrained local optima of a simplified first-principle model of the robot-terrain interaction. We experimentally verify that the proposed weakly supervised learning from ground-truth robot poses boosts the accuracy of predicted support heightmaps and increases the accuracy of estimated robot poses. All experiments are conducted on a dataset captured by a real platform. Both the dataset and codes which replicates experiments in the paper are made publicly available as a part of the submission.

DARPA Subterranean Challenge: Multi-robotic exploration of underground environments

  • DOI: 10.1007/978-3-030-43890-6_22
  • Odkaz: https://doi.org/10.1007/978-3-030-43890-6_22
  • Pracoviště: Centrum umělé inteligence, Vidění pro roboty a autonomní systémy, Multirobotické systémy
  • Anotace:
    The Subterranean Challenge (SubT) is a contest organised by the Defense Advanced Research Projects Agency (DARPA). The contest reflects the requirement of increasing safety and efficiency of underground search-and-rescue missions. In the SubT challenge, teams of mobile robots have to detect, localise and report positions of specific objects in an underground environment. This paper provides a description of the multi-robot heterogeneous exploration system of our CTU-CRAS team, which scored third place in the Tunnel Circuit round, surpassing the performance of all other non-DARPA-funded competitors. In addition to the description of the platforms, algorithms and strategies used, we also discuss the lessons-learned by participating at such contest.

Simultaneous exploration and segmentation for search and rescue

  • DOI: 10.1002/rob.21847
  • Odkaz: https://doi.org/10.1002/rob.21847
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    We consider the problem of active victim segmentation during a search‐and‐rescue (SAR) exploration mission. The robot is equipped with a multimodal sensor suite consisting of a camera, lidar, and pan‐tilt thermal sensor. The robot enters an unknown scene, builds a 3D model incrementally, and the proposed method simultaneously (a) segments the victims from incomplete multimodal measurements and (b) controls the motion of the thermal camera. Both of these tasks are difficult due to the lack of natural training data and the limited number of real‐world trials. In particular, we overcome the absence of training data for the segmentation task by employing a manually designed generative model, which provides a semisynthetic training data set. The limited number of real‐world trials is tackled by self‐supervised initialization and optimization‐based guiding of the motion control learning. In addition to that, we provide a quantitative evaluation of the proposed method on several real testing scenarios using the real SAR robot. Finally, we also provide a data set which will allow for further development of algorithms on the real data.

Tracked Robot Odometry for Obstacle Traversal in Sensory Deprived Environment

  • Autoři: Kubelka, V., Reinštein, M., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: IEEE-ASME TRANSACTIONS ON MECHATRONICS. 2019, 24(6), 2745-2755. ISSN 1083-4435.
  • Rok: 2019
  • DOI: 10.1109/TMECH.2019.2945031
  • Odkaz: https://doi.org/10.1109/TMECH.2019.2945031
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    Mobile-tracked robots are suitable for traversing rough terrain. However, standard exteroceptive localization methods (visual or laser SLAM) may be unreliable due to smoke, dust, fog, or insufficient lighting in harsh conditions of urban search and rescue missions. During extensive end-user evaluations in real-world conditions of such scenarios, we have observed that the accuracy of dead-reckoning localization suffers while traversing vertical obstacles. We propose to combine explicit modeling of robot kinematics and data-driven approach based on machine learning. The proposed method is experimentally verified indoors and outdoors traversing various obstacles. Indoors, a reference position has been recorded as well to assess the accuracy of our solution. The experimental dataset is released to the public to help the robotics community.

Data-driven Policy Transfer with Imprecise Perception Simulation

  • DOI: 10.1109/LRA.2018.2857927
  • Odkaz: https://doi.org/10.1109/LRA.2018.2857927
  • Pracoviště: Katedra kybernetiky, Vidění pro roboty a autonomní systémy
  • Anotace:
    This paper presents a complete pipeline for learning continuous motion control policies for a mobile robot when only a non-differentiable physics simulator of robot-terrain interactions is available. The multi-modal state estimation of the robot is also complex and difficult to simulate, so we simultaneously learn a generative model which refines simulator outputs. We propose a coarse-to-fine learning paradigm, where the coarse motion planning is alternated with guided learning and policy transfer to the real robot. The policy is jointly optimized with the generative model. We evaluate the method on a real-world platform.

Controlling Robot Morphology From Incomplete Measurements

  • DOI: 10.1109/TIE.2016.2580125
  • Odkaz: https://doi.org/10.1109/TIE.2016.2580125
  • Pracoviště: Katedra kybernetiky, Vidění pro roboty a autonomní systémy
  • Anotace:
    Mobile robots with complex morphology are essential for traversing rough terrains in Urban Search & Rescue missions. Since teleoperation of the complex morphology causes high cognitive load of the operator, the morphology is controlled autonomously. The autonomous control measures the robot state and surrounding terrain which is usually only partially observable, and thus the data are often incomplete. We marginalize the control over the missing measurements and evaluate an explicit safety condition. If the safety condition is violated, tactile terrain exploration by the body-mounted robotic arm gathers the missing data.

Fast Simulation of Vehicles with Non-deformable Tracks

  • DOI: 10.1109/IROS.2017.8206546
  • Odkaz: https://doi.org/10.1109/IROS.2017.8206546
  • Pracoviště: Katedra kybernetiky, Vidění pro roboty a autonomní systémy
  • Anotace:
    This paper presents a novel technique that allows for both computationally fast and sufficiently plausible simulation of vehicles with non-deformable tracks. The method is based on an effect we have called Contact Surface Motion. A comparison with several other methods for simulation of tracked vehicle dynamics is presented with the aim to evaluate methods that are available off-the-shelf or with minimum effort in general-purpose robotics simulators. The proposed method is implemented as a plugin for the open-source physics-based simulator Gazebo using the Open Dynamics Engine.

Learning for Active 3D Mapping

  • DOI: 10.1109/ICCV.2017.171
  • Odkaz: https://doi.org/10.1109/ICCV.2017.171
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    We propose an active 3D mapping method for depth sen- sors, which allow individual control of depth-measuring rays, such as the newly emerging solid-state lidars. The method simultaneously (i) learns to reconstruct a dense 3D occupancy map from sparse depth measurements, and (ii) optimizes the reactive control of depth-measuring rays. To make the first step towards the online control optimization, we propose a fast prioritized greedy algorithm, which needs to update its cost function in only a small fraction of possible rays. The approximation ratio of the greedy algorithm is derived. An experimental evaluation on the subset of the KITTI dataset demonstrates significant improvement in the 3D map accuracy when learning-to-reconstruct from sparse measurements is coupled with the optimization of depth measuring rays.

Point cloud registration from local feature correspondences—Evaluation on challenging datasets

  • DOI: 10.1371/journal.pone.0187943
  • Odkaz: https://doi.org/10.1371/journal.pone.0187943
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    Registration of laser scans, or point clouds in general, is a crucial step of localization and mapping with mobile robots or in object modeling pipelines. A coarse alignment of the point clouds is generally needed before applying local methods such as the Iterative Closest Point (ICP) algorithm. We propose a feature-based approach to point cloud registration and evaluate the proposed method and its individual components on challenging real-world datasets. For a moderate overlap between the laser scans, the method provides a superior registration accuracy compared to state-of-the-art methods including Generalized ICP, 3D Normal-Distribution Transform, Fast Point-Feature Histograms, and 4-Points Congruent Sets. Compared to the surface normals, the points as the underlying features yield higher performance in both keypoint detection and establishing local reference frames. Moreover, sign disambiguation of the basis vectors proves to be an important aspect in creating repeatable local reference frames. A novel method for sign disambiguation is proposed which yields highly repeatable reference frames.

Autonomous Flipper Control with Safety Constraints

  • DOI: 10.1109/IROS.2016.7759447
  • Odkaz: https://doi.org/10.1109/IROS.2016.7759447
  • Pracoviště: Katedra kybernetiky, Vidění pro roboty a autonomní systémy
  • Anotace:
    Policy Gradient methods require many real-world trials. Some of the trials may endanger the robot system and cause its rapid wear. Therefore, a safe or at least gentle-to-wear exploration is a desired property. We incorporate bounds on the probability of unwanted trials into the recent Contextual Relative Entropy Policy Search method. The proposed algorithm is evaluated on the task of autonomous flipper control for a real Search and Rescue rover platform.

Improving multimodal data fusion for mobile robots by trajectory smoothing

  • DOI: 10.1016/j.robot.2016.07.006
  • Odkaz: https://doi.org/10.1016/j.robot.2016.07.006
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    Localization of mobile robots is still an important topic, especially in case of dynamically changing, complex environments such as in Urban Search & Rescue (USAR). In this paper we aim for improving the reliability and precision of localization of our multimodal data fusion algorithm. Multimodal data fusion requires resolving several issues such as significantly different sampling frequencies of the individual modalities. We compare our proposed solution with the well-proven and popular Rauch–Tung–Striebel smoother for the Extended Kalman filter. Furthermore, we improve the precision of our data fusion by incorporating scale estimation for the visual modality.

Touching without vision: terrain perception in sensory deprived environments

  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    In this paper we demonstrate a combined hardware and software solution that enhances sensor suite and perception capabilities of a mobile robot intended for real Urban Search & Rescue missions. A common fail-case, when exploring unknown environment of a disaster site, is the outage o r deterioration of exteroceptive sensory measurements that the robot heavily relies on especially for localization and navigation purposes. Deprivation of visual and laser modalities caused by dense smoke motivated us to develop a novel solution comprised of force sensor arrays embedded into tracks of our platform. Furthermore, we also exploit a robotic arm for active perception in cases when the prediction based on force sensors is too uncertain. Beside the integration of hardware, we also propose a framework exploiting Gaussian proces ses followed by Gibb's sampling to process raw sensor measurements and provide probabilistic interpretation of the underlying terrain pro file. In the final, the profile is perceived by proprioceptive means only and successfully substitutes for the lack of exteroceptive measurements in the close vicinity of the robot, when traversing unknown and unseen obstacles. We evaluated our solution on real world terrains.

Robust Data Fusion of Multimodal Sensory Information for Mobile Robots

  • Autoři: Kubelka, V., Oswald, L., Pomerleau, F., Colas, F., prof. Ing. Tomáš Svoboda, Ph.D., Reinštein, M.
  • Publikace: Journal of Field Robotics. 2015, 32(4), 447-473. ISSN 1556-4959.
  • Rok: 2015
  • DOI: 10.1002/rob.21535
  • Odkaz: https://doi.org/10.1002/rob.21535
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Urban search and rescue (USAR) missions for mobile robots require reliable state estimation systems resilient to conditions given by the dynamically changing environment. We design and evaluate a data fusion system for localization of a mobile skid-steer robot intended for USAR missions. We exploit a rich sensor suite including both proprioceptive (inertial measurement unit and tracks odometry) and exteroceptive sensors (omnidirectional camera and rotating laser rangefinder). To cope with the specificities of each sensing modality (such as significantly differing sampling frequencies), we introduce a novel fusion scheme based on an extended Kalman filter for six degree of freedom orientation and position estimation. We demonstrate the performance on field tests of more than 4.4 km driven under standard USAR conditions. Part of our datasets include ground truth positioning, indoor with a Vicon motion capture system and outdoor with a Leica theodolite tracker. The overall median accuracy of localization-achieved by combining all four modalities-was 1.2% and 1.4% of the total distance traveled for indoor and outdoor environments, respectively. To identify the true limits of the proposed data fusion, we propose and employ a novel experimental evaluation procedure based on failure case scenarios. In this way, we address the common issues such as slippage, reduced camera field of view, and limited laser rangefinder range, together with moving obstacles spoiling the metric map. We believe such a characterization of the failure cases is a first step toward identifying the behavior of state estimation under such conditions. We release all our datasets to the robotics community for possible benchmarking.

Safe Exploration for Reinforcement Learning in Real Unstructured Environments

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    In USAR (Urban Search and Rescue) missions, robots are often required to operate in an unknown environment and with imprecise data coming from their sensors. However, it is highly desired that the robots only act in a safe manner and do not perform actions that could probably make damage to them. To train some tasks with the robot, we utilize reinforcement learning (RL). This machine learning method however requires the robot to perform actions leading to unknown states, which may be dangerous. We develop a framework for training a safety function which constrains possible actions to a subset of really safe actions. Our approach utilizes two basic concepts. First, a "core" of the safety function is given by a cautious simulator and possibly also by manually given examples. Second, a classifier training phase is performed (using Neyman-Pearson SVMs), which extends the safety function to the states where the simulator fails to recognize safe states.

TRADR Project: Long-Term Human-Robot Teaming for Robot Assisted Disaster Response

  • Autoři: Kruijff-Korbayová, I., Colas, F., Gianni, M., Pirri, F., de Greeff, J., Hindriks, K., Neerincx, M., Ogren, P., prof. Ing. Tomáš Svoboda, Ph.D., Worst, R.
  • Publikace: KI - Künstliche Intelligenz, German Journal on Artificial Intelligence. 2015, 29(2), 193-201. ISSN 0933-1875.
  • Rok: 2015
  • DOI: 10.1007/s13218-015-0352-5
  • Odkaz: https://doi.org/10.1007/s13218-015-0352-5
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Abstract This paper describes the project TRADR: Long- Term Human-Robot Teaming for Robot Assisted Disaster Response. Experience shows that any incident serious enough to require robot involvement will most likely involve a sequence of sorties over several hours, days and even months. TRADR focuses on the challenges that thus arise for the persistence of environment models, multi- robot action models, and human-robot teaming, in order to allow incremental capability improvement over the dura- tion of a mission. TRADR applies a user centric design approach to disaster response robotics, with use cases involving the response to a medium to large scale industrial accident by teams consisting of human rescuers and several robots (both ground and airborne). This paper describes the fundamentals of the project: the motivation, objectives and approach in contrast to related work.

Designing, developing, and deploying systems to support human-robot teams in disaster response

  • Autoři: Kruijff, G.J.M., Kruijff-Korbayova, I., Keshavdas, S., Larochelle, B., Janíček, M., Colas, F., Liu, M., Pomerleau, F., Siegwart, R., Neerincx, M.A., Looije, R., Smets, N.J.J.M, Mioch, T., van Diggelen, J., Pirri, F., Gianni, M., Ferri, F., Menna, M., Worst, R., Linder, T., Tretyakov, V., Surmann, H., prof. Ing. Tomáš Svoboda, Ph.D., Reinštein, M., doc. Ing. Karel Zimmermann, Ph.D., Petříček, T., Hlaváč, V.
  • Publikace: Advanced Robotics. 2014, 28(23), 1547-1570. ISSN 0169-1864.
  • Rok: 2014
  • DOI: 10.1080/01691864.2014.985335
  • Odkaz: https://doi.org/10.1080/01691864.2014.985335
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    This paper describes our experience in designing, developing and deploying systems for supporting human-robot teams during disaster response. It is based on R&D performed in the EU-funded project NIFTi. NIFTi aimed at building intelligent, collaborative robots that could work together with humans in exploring a disaster site, to make a situational assessment. To achieve this aim, NIFTi addressed key scientific design aspects in building up situation awareness in a human-robot team, developing systems using a user-centric methodology involving end users throughout the entire R&D cycle, and regularly deploying implemented systems under real-life circumstances for experimentation and testing. This has yielded substantial scientific advances in the state-of-the-art in robot mapping, robot autonomy for operating in harsh terrain, collaborative planning, and human-robot interaction. NIFTi deployed its system in actual disaster response activities in Northern Italy, in July 2012, aiding in structure damage assessment.

Experience in System Design for Human-Robot Teaming in Urban Search & Rescue

  • Autoři: Kruijff, G.J.M., Janíček, M., Keshavdas, S., Larochelle, B., Zender, H., Smets, N.J.J.M., Mioch, T., Neerincx, M.A., Diggelen, J.V., Colas, F., Liu, M., Pomerleau, F., Siegwart, R., Hlaváč, V., prof. Ing. Tomáš Svoboda, Ph.D., Petříček, T., Reinštein, M., doc. Ing. Karel Zimmermann, Ph.D., Pirri, F., Gianni, M., Papadakis, P., Sinha, A., Balmer, P., Tomatis, N., Worst, R., Linder, T., Surmann, H., Tretyakov, V., Corrao, S., Pratzler-Wanczura, S., Sulk, M.
  • Publikace: Field and Service Robotics. Heidelberg: Springer, 2014. p. 111-125. Springer Tracts in Advanced Robotics. ISSN 1610-7438. ISBN 978-3-642-40685-0.
  • Rok: 2014
  • DOI: 10.1007/978-3-642-40686-7_8
  • Odkaz: https://doi.org/10.1007/978-3-642-40686-7_8
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    The paper describes experience with applying a user-centric design methodology in developing systems for human-robot teaming in Urban Search & Rescue. A human-robot team consists of several semi-autonomous robots (rovers/UGVs, microcopter/ UAVs), several humans at an off-site command post (mission commander, UGV operators) and one on-site human (UAV operator). This system has been developed in close cooperation with several rescue organizations, and has been deployed in a real-life tunnel accident use case. The human-robot team jointly explores an accident site, communicating using a multi-modal team interface, and spoken dialogue. The paper describes the development of this complex socio-technical system per se, as well as recent experience in evaluating the performance of this system.

Non-Rigid Object Detection with Local Interleaved Sequential Alignment (LISA)

  • DOI: 10.1109/TPAMI.2013.171
  • Odkaz: https://doi.org/10.1109/TPAMI.2013.171
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    This paper shows that the successively evaluated features used in a sliding window detection process to decide about object presence/absence also contain knowledge about object deformation. We exploit these detection features to estimate the object deformation. Estimated deformation is then immediately applied to not yet evaluated features to align them with the observed image data. In our approach, the alignment estimators are jointly learned with the detector. The joint process allows for the learning of each detection stage from less deformed training samples than in the previous stage. For the alignment estimation we propose regressors that approximate non-linear regression functions and compute the alignment parameters extremely fast.

Safe Exploration Techniques for Reinforcement Learning - An Overview

  • DOI: 10.1007/978-3-319-13823-7_31
  • Odkaz: https://doi.org/10.1007/978-3-319-13823-7_31
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We overview different approaches to safety in (semi)autonomous robotics. Part icularly, we focus on how to achieve safe behavior of a robot if it is requested to perform ex ploration of unknown states. Presented methods are studied from the viewpoint of reinforcement learning, a partially-supervised machine learning method. To collect training data for this a lgorithm, the robot is required to freely explore the state space - which can lead to possibly dangerous situations. The role of safe exploration is to provide a framework allowing explora tion while preserving safety. The examined methods range from simple algorithms to sophisticat ed methods based on previous experience or state prediction. Our overview also addresses the i ssues of how to define safety in the real-world applications (apparently absolute safety is un achievable in the continuous and random real world). In the conclusion we also suggest several ways that are worth researching more thoroughly.

Exploiting Features - Locally Interleaved Sequential Alignment for Object Detection

  • Autoři: doc. Ing. Karel Zimmermann, Ph.D., Hurych, D., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: Computer Vision - ACCV 2012, 11th Asian Conference on Computer Vision, Part 1. Heidelberg: Springer, 2013, pp. 446-459. Lecture Notes in Computer Science. ISSN 0302-9743. ISBN 978-3-642-37330-5. Available from: ftp://cmp.felk.cvut.cz/pub/cmp/articles/hurycd1/hurych-accv2012.pdf
  • Rok: 2013
  • DOI: 10.1007/978-3-642-37331-2_34
  • Odkaz: https://doi.org/10.1007/978-3-642-37331-2_34
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We exploit image features multiple times in order to make sequential decision process faster and better performing. In the decision process features providing knowledge about the object presence or absence in a given detection window are successively evaluated. We show that these features also provide information about object position within the evaluated window. The classification process is sequentially interleaved with estimating the correct position. The position estimate is used for steering the features yet to be evaluated. This locally interleaved sequential alignment (LISA) allows to run an object detector on sparser grid which speeds up the process. The position alignment is jointly learned with the detector. We achieve a better detection rate since the method allows for training the detector on perfectly aligned image samples. For estimation of the alignment we propose a learnable regressor that approximates a non-linear regression function and runs in ne2076-1465gligible time.

Mutual On-Line Learning for Detection and Tracking in High-Resolution Images

  • Autoři: Hurych, D., doc. Ing. Karel Zimmermann, Ph.D., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: Proceedings of the International Joint Conference on Computer Vision, Imaging and Computer Graphics. Theory and Applications (VISIGRAPP2011). Heidelberg: Springer, 2013, pp. 240-256. Communications in Computer and Information Science. ISSN 1865-0929. ISBN 978-3-642-32349-2.
  • Rok: 2013
  • DOI: 10.1007/978-3-642-32350-8_15
  • Odkaz: https://doi.org/10.1007/978-3-642-32350-8_15
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    This paper addresses object detection and tracking in high-resolution omnidirectional images. The foreseen application is a visual subsystem of a rescue robot equipped with an omnidirectional camera, which demands real-time efficiency and robustness against changing viewpoint. Object detectors typically do not guarantee specific frame rate. The detection time may vastly depend on a scene complexity and image resolution. The adapted tracker can often help to overcome the situation, where the appearance of the object is far from the training set. On the other hand, once a tracker is lost, it almost never finds the object again. We propose a combined solution where a very efficient tracker (based on sequential linear predictors) incrementally accommodates varying appearance and speeds up the whole process. Next we propose to incrementally update the detector with examples collected by the tracker. We experimentally show that the performance of the combined algorithm, measured by a ratio between false positives and false negatives, outperforms both individual algorithms. The tracker allows to run the expensive detector only sparsely enabling the combined solution to run in real-time on 12 MPx images from a high resolution omnidirectional camera (Ladybug3).

Area-weighted Surface Normals for 3D Object Recognition

  • Autoři: Petříček, T., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: ICPR 2012: Proceedings of 21st International Conference on Pattern Recognition. New York: IEEE, 2012, pp. 1492-1496. ISSN 1051-4651. ISBN 978-4-9906441-0-9.
  • Rok: 2012
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    This paper presents a method for feature-based 3D object recognition in cluttered scenes. It deals with the problem of non-uniform sampling density which is inherent in typical range sensing methods. We suggest a method operating on polygonal meshes which overcomes the problem by exploiting surface area in both establishing local frames and creating feature descriptors. The method is able to recognize even highly occluded objects and outperforms state of the art in terms of recognition rate on a standard publicly available dataset.

A Unified Framework for Planning and Execution-Monitoring of Mobile Robots

  • Autoři: Gianni, M., Papadakis, P., Pirri, F., Liu, M., Pomerleau, F., Colas, F., doc. Ing. Karel Zimmermann, Ph.D., prof. Ing. Tomáš Svoboda, Ph.D., Petříček, T., Kruijff, G., Khambhaita, H., Zender, H.
  • Publikace: Automated Action Planning for Autonomous Mobile Robots: Papers from the AAAI Workshop (WS-11-09). Menlo Park: AAAI Press, 2011, pp. 39-44. ISBN 978-1-57735-525-0.
  • Rok: 2011
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We present an original integration of high level planning and execution with incoming perceptual information from vision, SLAM, topological map segmentation and dialogue. The task of the robot system, implementing the integrated model, is to explore unknown areas and report detected objects to an operator, by speaking loudly. The knowledge base of the planner maintains a graph-based representation of the metric map that is dynamically constructed via an unsupervised topological segmentation method, and augmented with information about the type and position of detected objects, within the map, such as cars or containers. According to this knowledge the cognitive robot can infer strategies in so generating parametric plans that are instantiated from the perceptual processes. Finally, a model-based approach for the execution and control of the robot system is proposed.

Detection of unseen patches trackable by linear predictors

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Linear predictors (LPs) are being used for tracking because of their computational efficiency which is better than steepest descent methods (e.g. Lucas-Kanade). The only disadvantage of LPs is the necessary learning phase which hinders the predictors applicability as a general patch tracker. We address this limitation and propose to learn a bank of LPs off-line and develop an on-line detector which selects image regions that could be tracked by some predictor from the bank. The proposed detector differs significantly from the usual solutions that attempt to find the closest match between a candidate patch and a database of exemplars. We construct the detector directly from the learned linear predictor. The detector positively detects the learned patches, but also many other image patches, which were not used in LP learning phase. This means, that the LP is able to track also previously unseen image patches, the appearances of which are often significantly diverse from the patches used.

Fast Learnable Object Tracking and Detection in High-resolution Omnidirectional Images

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    This paper addresses object detection and tracking in high-resolution omnidirectional images. The foreseen application is a visual subsystem of a rescue robot equipped with an omnidirectional camera, which demands real time efficiency and robustness against changing viewpoint. Object detectors typically do not guarantee specific frame rate. The detection time may vastly depend on a scene complexity and image resolution. The adapted tracker can often help to overcome the situation, where the appearance of the object is far from the training set. On the other hand, once a tracker is lost, it almost never finds the object again. We propose a combined solution where a very efficient tracker (based on sequential linear predictors) incrementally accommodates varying appearance and speeds up the whole process. We experimentally show that the performance of the combined algorithm, measured by a ratio between false positives and false negatives, outperforms both individual algorithms.

Improving Cascade of Classifiers by Sliding Window Alignment in Between

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We improve an object detector based on cascade of classifiers by a local alignment of the sliding window. The detector needs to operate on a relatively sparse grid in order to achieve a real time performance on high-resolution images. The proposed local alignment in the middle of the cascade improves its recognition performance whilst retaining the necessary speed. We show that the moment of the alignment matters and discuss the performance in terms of false negatives and false positives. The proposed method is tested on a car detection problem.

Incremental learning and validation of sequential predictors in video browsing application

  • Autoři: Hurych, D., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: VISIGRAPP 2010: International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications. Setúbal: Institute for Systems and Technologies of Information, Control and Communication, 2010. pp. 467-474. ISBN 978-989-674-028-3.
  • Rok: 2010
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Loss-of-track detection (tracking validation) and automatic tracker adaptation to new object appearances are attractive topics in computer vision. We apply very efficient learnable sequential predictors in order to address both issues. Validation is done by clustering of the sequential predictor responses. No aditional object model for validation is needed. The paper also proposes an incremental learning procedure that accommodates changing object appearance, which mainly improves the recall of the tracker/detector. Exemplars for the incremental learning are collected automatically, no user interaction is required. The aditional training examples are selected automatically using the tracker stability computed for each potential aditional training example. Coupled with a sparsely applied SIFT or SURF based detector the method is employed for object localization in videos. Our Matlab implementation scans videosequences up to eight times faster than the actual frame rate. A standard-lengt

Active Shape Model and Linear Predictors for Face Association Refinement

  • Autoři: Hurych, D., prof. Ing. Tomáš Svoboda, Ph.D., Trojanová, J., US, Y.
  • Publikace: The Ninth IEEE International Workshop on Visual Surveillance 2009. Los Alamitos: IEEE Computer Society Press, 2009, pp. 1193-1200. ISBN 978-1-4244-4441-0.
  • Rok: 2009
  • DOI: 10.1109/ICCVW.2009.5457473
  • Odkaz: https://doi.org/10.1109/ICCVW.2009.5457473
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    This paper summarizes results of face association experiments on real low resolution data from airport and the Labeled faces in the Wild (LFW) database. The objective of experiments is to evaluate different face alignment methods and their contribution to face association as such. The first alignment method used is Sequential Learnable Linear Predictor (SLLiP), originally developed for object tracking. The second method is well known face alignment method Active Shape Model (ASM). Both methods are compared versus face association without alignment. In case of high resolution LFW database the ASM rapidly increases the association results, on the other hand for real low resolution airport data the SLLiP method brought more improvement than ASM.

Anytime learning for the NoSLLiP tracker

  • DOI: 10.1016/j.physletb.2003.10.07
  • Odkaz: https://doi.org/10.1016/j.physletb.2003.10.07
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Anytime learning for the Sequence of Learned Linear Predictors (SLLiP) tracker is proposed. Learning might be time consuming for large problems, we present an anytime learning algorithm which, after a very short initialization period, provides a solution with defined precision. As SLLiP tracking requires only a fraction of the processing power of an ordinary PC, the learning can continue in a parallel background thread continuously delivering improved SLLiPs, ie. faster, with lower computational complexity, with the same pre-defined precision. The proposed approach is verified on publicly-available sequences with approximately 12000 ground truthed frames. The learning time is shown to be twenty times smaller than learning based on linear programming proposed in the paper that introduced the SLLiP tracker [TR]. Its robustness and accuracy is similar. Superiority in frame-rate and robustness with respect to the SIFT detector, Lucas-Kanade tracker and Jurie's tracker is also demonstrated.

Tracking by an Optimal Sequence of Linear Predictors

  • DOI: 10.1109/TPAMI.2008.119
  • Odkaz: https://doi.org/10.1109/TPAMI.2008.119
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We propose a learning approach to tracking explicitly minimizing the computational complexity of the tracking process subject to user-defined probability of failure (loss-of-lock) and precision. The tracker is formed by a Number of Sequences of Learned Linear Predictors (NoSLLiP). Robustness of NoSLLiP is achieved by modeling the object as a collection of local motion predictors --- object motion is estimated by the outlier-tolerant Ransac algorithm from local predictions. Efficiency of the NoSLLiP tracker stems from (i) the simplicity of the local predictors and (ii) from the fact that all design decisions - the number of local predictors used by the tracker, their computational complexity (ie the number of observations the prediction is based on), locations as well as the number of Ransac iterations are all subject to the optimization (learning) process. All time-consuming operations are performed during the learning stage - t.

Combination of Stochastic and AdaBoost Approach for Object Tracking and Recognition in Video

  • Autoři: Vlček, P., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: Proceedings of Workshop 2008. Praha: Czech Technical University in Prague, 2008, pp. 122-123. ISBN 978-80-01-04016-4.
  • Rok: 2008
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    The digital video processing becomes more and more important area of the computer vision. Between the quite developed methods for static images processing and video processing there are many clear differences, for example the lower overall image quality of the video, the higher volume of the video data and the real-time processing requirement. In this work we focus on the task of 3D tracking of the human head for the application in automated indexing of the feature-length movies. One of the most successful real-time tracking algorithms is the CONDENSATION algorithm and a well known approach to face detection is the Viola-Jones detector, based on the AdaBoost learning algorithm. We combine the two approaches and design a 3D head tracking algorithm, which is able to automatically learn the head appearance and track the full-angle head turnaround.

Simultaneous learning of motion and appearance

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    A new learning method for motion estimation of objects with significantly varying appearance is proposed. Varying object appearance is represented by a low dimensional space of appearance parameters. The appearance mapping and motion estimation method are optimized simultaneously. Appearance parameters are estimated by unsupervised learning. The method is experimentally verified by a tracking application on sequences which exhibit strong variable illumination, non-rigid deformations and self-occlusions.

Adaptive parameter optimization for real-time tracking

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Adaptation of a tracking procedure combined in a common way with a Kalman filter is formulated as an constrained optimization problem, where a trade-off between precision and loss-of-lock probability is explicitly taken into account. While the tracker is learned in order to minimize computational complexity during a learning stage, in a tracking stage the precision is maximized online under a constraint imposed by the loss-of-lock probability resulting in an optimal setting of the tracking procedure. We experimentally show that the proposed method converges to a steady solution in all variables. In contrast to a common Kalman filter based tracking, we achieve a significantly lower state covariance matrix. We also show, that if the covariance matrix is continuously updated, the method is able to adapt to a different situations. If a dynamic model is precise enough the tracker is allowed to spend a longer time with a fine motion estimation, however, if the motion gets saccadic, i.e. unpr

Adaptive Single-view 3D Tracking of the Human Head by Incremental Texture Wrapping

  • Autoři: Vlček, P., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: ISCAM 2007: International Conference in Applied Mathematics for Undergraduate and Graduate Students. Bratislava: FEI, Slovak University of Technology, 2007. p. 76-79.
  • Rok: 2007
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We consider a single-camera model-based 3D head tracking problem. The existing methods usually estimate only limited out of plane rotations, because the back of the head contains no good features to track. To solve this problem, we propose an adaptive tracking algorithm with a simple 3D ellipsoidal textured head model. The rotational and translational parts of the motion are estimated separately. The translation is estimated by a particle filter contour tracking. The remaining relative rotation is then estimated by a combination of feature tracker with \sc RANSAC-based outlier detection.

Image Processing, Analysis and Machine Vision - A MATLAB Companion

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Comprehensive textbook about Matlab programming in image processing and computer vision

Learning Efficient Linear Predictors for Motion Estimation

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    A novel object representation for tracking is proposed. The tracked object is represented as a constellation of spatially localised linear predictors which are learned on a single training image. In the learning stage, sets of pixels whose intensities allow for optimal least square predictions of the transformations are selected as a support of the linear predictor. The approach comprises three contributions: learning object specific linear predictors, explicitly dealing with the predictor precision - computational complexity trade-off and selecting a view-specific set of predictors suitable for global object motion estimate. Robustness to occlusion is achieved by RANSAC procedure. The learned tracker is very efficient, achieving frame rate generally higher than 30 frames per second despite the Matlab implementation.

Multiview 3D Tracking with an Incrementally Constructed 3D Model

  • DOI: 10.1109/3DPVT.2006.101
  • Odkaz: https://doi.org/10.1109/3DPVT.2006.101
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    A novel object representation for tracking is proposed. The tracked object is represented as a constellation of spatially localised linear predictors which are learned on a single training image. In the learning stage, sets of pixels whose intensities allow for optimal least square predictions of the transformations are selected as a support of the linear predictor. The approach comprises three contributions: learning object specific linear predictors, explicitly dealing with the predictor precision - computational complexity trade-off and selecting a view-specific set of predictors suitable for global object motion estimate. Robustness to occlusion is achieved by RANSAC procedure. The learned tracker is very efficient, achieving frame rate generally higher than 30 frames per second despite the Matlab implementation.

A Convenient Multi-Camera Self-Calibration for Virtual Environments

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Martinec, D., Pajdla, T.
  • Publikace: PRESENCE: Teleoperators and Virtual Environments. 2005, 14(4), 407-422. ISSN 1054-7460.
  • Rok: 2005
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Virtual immersive environments or telepresence setups often consist of multiple cameras which have to be calibrated. We present a convenient method for doing this. The minimum is three cameras, but there is no upper limit. The method is fully automatic and a freely moving bright spot is the only calibration object. A set of virtual 3D points is made by waving the bright spot through the working volume. Its projections are found with sub-pixel precision and verified by a robust RANSAC analysis. The cameras do not have to see all points, only reasonable overlap between camera subgroups is necessary. Projective structures are computed via rank-4 factorization and the Euclidean stratification is done by imposing geometric constraints. This linear estimate initializes a post-processing computation of non-linear distortion which is also fully automatic. We suggest a trick on how to use a very ordinary laser pointer as the calibration object. We show that it is possible to calibrate an immers

A Software for Complete Calibration of Multicamera Systems

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: Image and Video Communications and Processing, Proceedings od SPIE-IS&T Electronic Imaging. Bellingham: SPIE, 2005. pp. 115-128. ISBN 0-8194-5658-6.
  • Rok: 2005
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We present a software for complete metric calibration of synchronized multicamera setups. The minimum is three cameras, but there is no upper limit. The method is fully automatic and a freely moving bright spot is the only calibration object. No camera pre-calibration is required. The software computes complete set of intrinsic camera parameters including the parameters of non-linear distortion as well as external orientation of the cameras in one common coordinate system. The software is written in Matlab, runs in non-interactive mode and produces both textual and intuitive graphical output. It is very robust and may be run with the same set of parameters in different setups. The software is free.

Probabilistic Estimation of Articulated Body Model from Multiview Data

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    An optimization algorithm and statistical description of articulated body model estimation is proposed. The optimization algorithm fits the model into segmented multiview images. The input of our algorithm is a sequence of segmented images captured by several cameras and a structure of the articulated model. The output of the optimization procedure is shape and motion of the articulated model. The optimization runs over all cameras and all images in the sequence. We focus on description and optimization of probability distribution of the model parameters given segmented multiview sequence. We demonstrate the performance of the algorithm on real sequences of walking human.

Cinematographic Rules Applied to a Camera Network

  • Autoři: Doubek, P., Geys, I., prof. Ing. Tomáš Svoboda, Ph.D., Van Gool, L.
  • Publikace: Omnivis 2004: The Fifth Workshop on Omnidirectional Vision, Camera Networks and Non-Classical Cameras. Praha: Czech Technical University in Prague, 2004. p. 17-29.
  • Rok: 2004
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    We present a camera network system consisting of several modules of 2-3 low end cameras attached to one computer. It is not possible for a human to observe all the information coming from such a network simultaneously. Our system is designed to select the best viewpoint for each part of the video sequence, thus automatically creating one real-time video stream that contains the most important data. It acts as a combination of a director and a cameraman. Cinematography developed its own terminology, techniques and rules, how to make a good movie. We illustrate here some of these techniques and how they can be applied to a camera network, to solve the best viewpoint selection problem. Our system consists of only fixed cameras, but the output is not constrained to already existing views. A virtual zoom can be applied to select only a part of the view. We propose a view interpolation algorithm which makes it possible to create new intermediate views from the existing camera i

Epipolar Geometry for Central Catadioptric Cameras

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Central catadioptric cameras are cameras which combine lenses and mirrors to capture a very wide field of view with a central projection. In this paper we extend the classical epipolar geometry of perspective cameras to all central catadioptric cameras. The epipolar geometry is formulated as the geometry of corresponding rays in a three-dimensional space. Then, using the model of image formation of central catadioptric cameras, the constraint on corresponding image points is derived. It is shown that the corresponding points lie on epipolar conics. In addition, the shape of the conics for all types of central catadioptric cameras is classified. Finally, the theory is verified by experiments with real and simulated central catadioptric cameras.

Reliable 3D reconstruction from a few catadioptric images

  • Autoři: Doubek, P., prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: Proceedings of the IEEE Workshop on Omnidirectional Vision. Los Alamitos: IEEE Computer Society Press, 2002. pp. 71-78. ISBN 0-7695-1629-7.
  • Rok: 2002
  • Pracoviště: Katedra kybernetiky
  • Anotace:
    This paper proposes a scheme for the reliable reconstruction of indoor scenes from few catadioptric images. A set of hand-detected correspondences are established across (not necessarily all) images. Our improved method is used for the estimation of the essential matrix from points which have appropriately normalized coordinates. Hartley's decomposition is used for estimation of motion parameters. A heuristic is suggested for selecting the point pairs which are {\em most reliable\/} for 3D reconstruction. The known mid-point method is applied for computing the 3D model of a real scene. The parameters of the catradioptric sensor are approximately known but no precise self-calibration method is performed. The experiments show that a reliable 3D reconstruction is possible even without complicated non-linear self-calibration and/or recontruction methods.

Matching in Catadioptric Images with Appropriate Windows and Outliers Removal

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T.
  • Publikace: Computer Analysis of Images and Patterns: Proceedings of the 9th International Conference. Berlin: Springer, 2001. pp. 733-740. ISBN 3-540-42513-6.
  • Rok: 2001

Epipolar Geometry of Central Panoramic Cameras

  • Autoři: Pajdla, T., prof. Ing. Tomáš Svoboda, Ph.D., Hlaváč, V.
  • Publikace: Panoramic Vision: Sensors, Theory, and Applications. New York: Springer, 2000. p. 85-114. ISBN 0-387-95111-3.
  • Rok: 2000

Panoramic Cameras for 3D Computation

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T.
  • Publikace: Proceedings of the Czech Pattern Recognition Workshop. Prague: Czech Pattern Recognition Society, 2000. pp. 63-70. ISBN 80-238-5215-9.
  • Rok: 2000

Central Panoramic Cameras: Design, Epipolar Geometry, Egomotion

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T., Hlaváč, V.
  • Publikace: Proceedings of Workshop 99. Praha: České vysoké učení technické v Praze, 1999, pp. 94.
  • Rok: 1999

Central Panoramic Cameras: Design and Geometry

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T., Hlaváč, V.
  • Publikace: Proceedings of the Computer Vision Winter Workshop. Ljubljana: IEEE Slovenia Section, 1998, pp. 120-133. ISBN 961-6062-13-1.
  • Rok: 1998

Epipolar Geometry for Panoramic Cameras

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T., Hlaváč, V.
  • Publikace: Proceedings of the fift European Conference on Computer Vision. Berlin: Springer, 1998. pp. 218-232. ISBN 3-540-64569-1.
  • Rok: 1998

Epipolar Geometry for Panoramic Cameras

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D.,
  • Publikace: Poster 1998. Praha: České vysoké učení technické v Praze, Fakulta elektrotechnická, 1998, pp. 40-41.
  • Rok: 1998

Epipolar Geometry for Panoramic Cameras

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T., Hlaváč, V.
  • Publikace: Workshop 98. Praha: České vysoké učení technické v Praze, 1998. pp. 177-178.
  • Rok: 1998

Motion Estimation Using Central Panoramic Cameras

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T., Hlaváč, V.
  • Publikace: IEEE International Conference on Intelligent Vehicles. Stuttgart: Causal Productions, 1998, pp. 335-340. ISBN 0-876346-16-7.
  • Rok: 1998

A Badly Calibrated Camera in Ego-Motion Analysis, an Error Propagation

  • Pracoviště: Katedra kybernetiky
  • Anotace:
    Nepřesně kalibrovaná kamera při odhadování vlastního pohybu, propagace chyby

A Badly Calibrated Camera in Ego-Motion Estimation and Propagation of Uncertainly

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Sturm, P.
  • Publikace: Czech Pattern Recognition Workshop '97. Prague: Czech Pattern Recognition Society, 1997, pp. 59-63.
  • Rok: 1997

A Badly Calibrated Camera in Egomotion, Propagation of Uncertainty

From the Uncertainty in Camera Calibration to the Uncertainty in Ego-Motion

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Sturm, P.
  • Publikace: Poster 1997. Praha: České vysoké učení technické v Praze, Fakulta elektrotechnická, 1997, pp. 29.
  • Rok: 1997

What Precision with a Badly Calibrated Camera in Ego-Motion Estimation, an Error Analysis

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Sturm, P.
  • Publikace: Proceedings of 21st Workshop of the Austrian Association for Pattern Recognition. Graz: Österreichische Computer Gesselschaft, 1997, pp. 59-68. ISBN 3-85403-103-3.
  • Rok: 1997

Efficient Estimation of Essential Matrix in Motion Analysis

Fast Method for Computing Essential Matrix in Motion Analysis

  • Autoři: prof. Ing. Tomáš Svoboda, Ph.D., Pajdla, T.
  • Publikace: Poster 1995. Praha: České vysoké učení technické v Praze, Fakulta elektrotechnická, 1995, pp. 157.
  • Rok: 1995

Za stránku zodpovídá: Ing. Mgr. Radovan Suk