Lidé

Ing. Filipe Gama

Všechny publikace

Automatic infant 2D pose estimation from videos: Comparing seven deep neural network methods

  • DOI: 10.3758/s13428-025-02816-x
  • Odkaz: https://doi.org/10.3758/s13428-025-02816-x
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    Automatic markerless estimation of infant posture and motion from ordinary videos carries great potential for movement studies "in the wild", facilitating understanding of motor development and massively increasing the chances of early diagnosis of disorders. There is rapid development of human pose estimation methods in computer vision thanks to advances in deep learning and machine learning. However, these methods are trained on datasets that feature adults in different contexts. This work tests and compares seven popular methods (AlphaPose, DeepLabCut/DeeperCut, Detectron2, HRNet, MediaPipe/BlazePose, OpenPose, and ViTPose) on videos of infants in supine position and in more complex settings. Surprisingly, all methods except DeepLabCut and MediaPipe have competitive performance without additional finetuning, with ViTPose performing best. Next to standard performance metrics (average precision and recall), we introduce errors expressed in the neck-mid-hip (torso length) ratio and additionally study missing and redundant detections, and the reliability of the internal confidence ratings of the different methods, which are relevant for downstream tasks. Among the networks with competitive performance, only AlphaPose could run close to real time (27 fps) on our machine. We provide documented Docker containers or instructions for all the methods we used, our analysis scripts, and the processed data at https://hub.docker.com/u/humanoidsctu and https://osf.io/x465b/

Goal-directed tactile exploration for body model learning through self-touch on a humanoid robot

  • DOI: 10.1109/TCDS.2021.3104881
  • Odkaz: https://doi.org/10.1109/TCDS.2021.3104881
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    An early integration of tactile sensing into motor coordination is the norm in animals, but still a challenge for robots. Tactile exploration through touches on the body gives rise to first body models and bootstraps further development such as reaching competence. Reaching to one’s own body requires connections of the tactile and motor space only. Still, the problems of high dimensionality and motor redundancy persist. Through an embodied computational model for the learning of self-touch on a simulated humanoid robot with artificial sensitive skin, we demonstrate that this task can be achieved (i) effectively and (ii) efficiently at scale by employing the computational frameworks for the learning of internal models for reaching: intrinsic motivation and goal babbling. We relate our results to infant studies on spontaneous body exploration as well as reaching to vibrotactile targets on the body. We analyze the reaching configurations of one infant followed weekly between 4 and 18 months of age and derive further requirements for the computational model: accounting for (iii) continuous rather than sporadic touch and (iv) consistent redundancy resolution. Results show the general success of the learning models in the touch domain, but also point out limitations in achieving fully continuous touch.

Self-touch and other spontaneous behavior patterns in early infancy

  • DOI: 10.1109/ICDL53763.2022.9962203
  • Odkaz: https://doi.org/10.1109/ICDL53763.2022.9962203
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    Children are not born tabula rasa. However, interacting with the environment through their body movements in the first months after birth is critical to building the models or representations that are the foundation for everything that follows. We present longitudinal data on spontaneous behavior of three infants observed between about 8 and 25 weeks of age in supine position. We combined manual scoring of video recordings with an automatic extraction of motion data in order to study infants’ behavioral patterns and developmental progression such as: (i) spatial distribution of self-touches on the body, (ii) spatial patterns and regularities of hand movements, (iii) midline crossing, (iv) preferential use of one arm, and (v) dynamic patterns of movements indicative of goal-directedness. From the patterns observed in this pilot data set, we can speculate on the development of first body and peripersonal space representations. Several methods of extracting 3D kinematics from videos have recently been made available by the computer vision community. We applied one of these methods on infant videos and provide guidelines on its possibilities and limitations—a methodological contribution to automating the analysis of infant videos. In the future, we plan to use the patterns we extracted from the recordings as inputs to embodied computational models of learning of body representations in infancy.

Active exploration for body model learning through self-touch on a humanoid robot with artificial skin

  • Autoři: Ing. Filipe Gama, Shcherban, M., Rolf, M., doc. Mgr. Matěj Hoffmann, Ph.D.,
  • Publikace: Development and Learning and Epigenetic Robotics (ICDL-EpiRob), 2020 Joint IEEE 10th International Conference on. Piscataway: IEEE, 2020. ISSN 2161-9484. ISBN 978-1-7281-7306-1.
  • Rok: 2020
  • DOI: 10.1109/ICDL-EpiRob48136.2020.9278035
  • Odkaz: https://doi.org/10.1109/ICDL-EpiRob48136.2020.9278035
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    The mechanisms of infant development are far from understood. Learning about one's own body is likely a foundation for subsequent development. Here we look specifically at the problem of how spontaneous touches to the body in early infancy may give rise to first body models and bootstrap further development such as reaching competence. Unlike visually elicited reaching, reaching to own body requires connections of the tactile and motor space only, bypassing vision. Still, the problems of high dimensionality and redundancy of the motor system persist. In this work, we present an embodied computational model on a simulated humanoid robot with artificial sensitive skin on large areas of its body. The robot should autonomously develop the capacity to reach for every tactile sensor on its body. To do this efficiently, we employ the computational framework of intrinsic motivations and variants of goal babbling-as opposed to motor babbling-that prove to make the exploration process faster and alleviate the ill-posedness of learning inverse kinematics. Based on our results, we discuss the next steps in relation to infant studies: what information will be necessary to further ground this computational model in behavioral data.

Za stránku zodpovídá: Ing. Mgr. Radovan Suk