Lidé

Ing. Filipe Gama

Všechny publikace

Goal-directed tactile exploration for body model learning through self-touch on a humanoid robot

  • DOI: 10.1109/TCDS.2021.3104881
  • Odkaz: https://doi.org/10.1109/TCDS.2021.3104881
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    An early integration of tactile sensing into motor coordination is the norm in animals, but still a challenge for robots. Tactile exploration through touches on the body gives rise to first body models and bootstraps further development such as reaching competence. Reaching to one’s own body requires connections of the tactile and motor space only. Still, the problems of high dimensionality and motor redundancy persist. Through an embodied computational model for the learning of self-touch on a simulated humanoid robot with artificial sensitive skin, we demonstrate that this task can be achieved (i) effectively and (ii) efficiently at scale by employing the computational frameworks for the learning of internal models for reaching: intrinsic motivation and goal babbling. We relate our results to infant studies on spontaneous body exploration as well as reaching to vibrotactile targets on the body. We analyze the reaching configurations of one infant followed weekly between 4 and 18 months of age and derive further requirements for the computational model: accounting for (iii) continuous rather than sporadic touch and (iv) consistent redundancy resolution. Results show the general success of the learning models in the touch domain, but also point out limitations in achieving fully continuous touch.

Self-touch and other spontaneous behavior patterns in early infancy

  • DOI: 10.1109/ICDL53763.2022.9962203
  • Odkaz: https://doi.org/10.1109/ICDL53763.2022.9962203
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    Children are not born tabula rasa. However, interacting with the environment through their body movements in the first months after birth is critical to building the models or representations that are the foundation for everything that follows. We present longitudinal data on spontaneous behavior of three infants observed between about 8 and 25 weeks of age in supine position. We combined manual scoring of video recordings with an automatic extraction of motion data in order to study infants’ behavioral patterns and developmental progression such as: (i) spatial distribution of self-touches on the body, (ii) spatial patterns and regularities of hand movements, (iii) midline crossing, (iv) preferential use of one arm, and (v) dynamic patterns of movements indicative of goal-directedness. From the patterns observed in this pilot data set, we can speculate on the development of first body and peripersonal space representations. Several methods of extracting 3D kinematics from videos have recently been made available by the computer vision community. We applied one of these methods on infant videos and provide guidelines on its possibilities and limitations—a methodological contribution to automating the analysis of infant videos. In the future, we plan to use the patterns we extracted from the recordings as inputs to embodied computational models of learning of body representations in infancy.

Active exploration for body model learning through self-touch on a humanoid robot with artificial skin

  • Autoři: Ing. Filipe Gama, Shcherban, M., Rolf, M., doc. Mgr. Matěj Hoffmann, Ph.D.,
  • Publikace: Development and Learning and Epigenetic Robotics (ICDL-EpiRob), 2020 Joint IEEE 10th International Conference on. Piscataway: IEEE, 2020. ISSN 2161-9484. ISBN 978-1-7281-7306-1.
  • Rok: 2020
  • DOI: 10.1109/ICDL-EpiRob48136.2020.9278035
  • Odkaz: https://doi.org/10.1109/ICDL-EpiRob48136.2020.9278035
  • Pracoviště: Vidění pro roboty a autonomní systémy
  • Anotace:
    The mechanisms of infant development are far from understood. Learning about one's own body is likely a foundation for subsequent development. Here we look specifically at the problem of how spontaneous touches to the body in early infancy may give rise to first body models and bootstrap further development such as reaching competence. Unlike visually elicited reaching, reaching to own body requires connections of the tactile and motor space only, bypassing vision. Still, the problems of high dimensionality and redundancy of the motor system persist. In this work, we present an embodied computational model on a simulated humanoid robot with artificial sensitive skin on large areas of its body. The robot should autonomously develop the capacity to reach for every tactile sensor on its body. To do this efficiently, we employ the computational framework of intrinsic motivations and variants of goal babbling-as opposed to motor babbling-that prove to make the exploration process faster and alleviate the ill-posedness of learning inverse kinematics. Based on our results, we discuss the next steps in relation to infant studies: what information will be necessary to further ground this computational model in behavioral data.

Za stránku zodpovídá: Ing. Mgr. Radovan Suk