Infants' Developing Environment: Integration of Computer Vision and Human Annotation to Quantify where Infants Go, what They Touch, and what They See

Danyang Han, Nicolas Aziere, Tieqiao Wang, Ori Ossmy, Ajay Krishna, Hanzhi Wang, Ruiting Shen, Sinisa Todorovic, Karen Adolph

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Infants learn through interactions with the environment. Thus, to understand infants' early learning experiences, it is critical to quantify their natural learning input - where infants go, what they touch, and what they see. Wearable sensors can record locomotor and hand movements, but cannot recover the context that prompted the behaviors. Egocentric views from head cameras and eye trackers require annotation to process the videos and miss much of the surrounding context. Third-person video captures infant behavior in the entire scene but may misrepresent the egocentric view. Moreover, third-person video requires machine or human annotation to make sense of the behaviors, and either method alone is sorely lacking. Computer-vision is not sufficiently reliable to quantify much of infants' complex, variable behavior, and human annotation cannot reliably quantify 3D coordinates of behavior without laborious hand digitization. Thus, we pioneered a new system of behavior detection from third-person video that capitalizes on the integrated power of computer vision and human annotation to quantify infants' locomotor, manual, and egocentric visual interactions with the environment. Our system estimates a real infant's interaction with a physical environment during free play by projecting a "virtual"infant in a "virtual"3D environment with known coordinates of all furniture, objects, and surfaces. Our methods for using human-in-the-loop computer vision have broad applications for reliable quantification of locomotor, manual, and visual behaviors outside the purview of standard algorithms or human annotation alone.

Original languageEnglish (US)
Title of host publication2024 IEEE International Conference on Development and Learning, ICDL 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350348552
DOIs
StatePublished - 2024
Event2024 IEEE International Conference on Development and Learning, ICDL 2024 - Austin, United States
Duration: May 20 2024May 23 2024

Publication series

Name2024 IEEE International Conference on Development and Learning, ICDL 2024

Conference

Conference2024 IEEE International Conference on Development and Learning, ICDL 2024
Country/TerritoryUnited States
CityAustin
Period5/20/245/23/24

Keywords

  • behavior
  • computer vision
  • infant
  • locomotion
  • manual exploration
  • visual

ASJC Scopus subject areas

  • Artificial Intelligence
  • Computer Vision and Pattern Recognition
  • Human-Computer Interaction
  • Cognitive Neuroscience
  • Sensory Systems

Fingerprint

Dive into the research topics of 'Infants' Developing Environment: Integration of Computer Vision and Human Annotation to Quantify where Infants Go, what They Touch, and what They See'. Together they form a unique fingerprint.

Cite this