Vlad Olaru’s research while affiliated with Institute of Mathematics of the Romanian Academy and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (9)


Reconstructing Three-Dimensional Models of Interacting Humans
  • Preprint

August 2023

·

30 Reads

·

·

Elisabeta Oneata

·

[...]

·

Cristian Sminchisescu

Understanding 3d human interactions is fundamental for fine-grained scene analysis and behavioural modeling. However, most of the existing models predict incorrect, lifeless 3d estimates, that miss the subtle human contact aspects--the essence of the event--and are of little use for detailed behavioral understanding. This paper addresses such issues with several contributions: (1) we introduce models for interaction signature estimation (ISP) encompassing contact detection, segmentation, and 3d contact signature prediction; (2) we show how such components can be leveraged to ensure contact consistency during 3d reconstruction; (3) we construct several large datasets for learning and evaluating 3d contact prediction and reconstruction methods; specifically, we introduce CHI3D, a lab-based accurate 3d motion capture dataset with 631 sequences containing 2,525 contact events, 728,664 ground truth 3d poses, as well as FlickrCI3D, a dataset of 11,216 images, with 14,081 processed pairs of people, and 81,233 facet-level surface correspondences. Finally, (4) we propose methodology for recovering the ground-truth pose and shape of interacting people in a controlled setup and (5) annotate all 3d interaction motions in CHI3D with textual descriptions. Motion data in multiple formats (GHUM and SMPLX parameters, Human3.6m 3d joints) is made available for research purposes at \url{https://ci3d.imar.ro}, together with an evaluation server and a public benchmark.


Predictable Robots for Autistic Children --- Variance in Robot Behaviour, Idiosyncrasies in Autistic Children's Characteristics, and Child-Robot Engagement
  • Article
  • Full-text available

August 2021

·

306 Reads

·

13 Citations

ACM Transactions on Computer-Human Interaction

Predictability is important to autistic individuals, and robots have been suggested to meet this need as they can be programmed to be predictable, as well as elicit social interaction. The effectiveness of robot-assisted interventions designed for social skill learning presumably depends on the interplay between robot predictability, engagement in learning, and the individual differences between different autistic children. To better understand this interplay, we report on a study where 24 autistic children participated in a robot-assisted intervention. We manipulated the variance in the robot’s behaviour as a way to vary predictability, and measured the children’s behavioural engagement, visual attention, as well as their individual factors. We found that the children will continue engaging in the activity behaviourally, but may start to pay less visual attention over time to activity-relevant locations when the robot is less predictable. Instead, they increasingly start to look away from the activity. Ultimately, this could negatively influence learning, in particular for tasks with a visual component. Furthermore, severity of autistic features and expressive language ability had a significant impact on behavioural engagement. We consider our results as preliminary evidence that robot predictability is an important factor for keeping children in a state where learning can occur.

Download


Learning Complex 3D Human Self-Contact

May 2021

·

7 Reads

·

30 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

Monocular estimation of three dimensional human self-contact is fundamental for detailed scene analysis including body language understanding and behaviour modeling. Existing 3d reconstruction methods do not focus on body regions in self-contact and consequently recover configurations that are either far from each other or self-intersecting, when they should just touch. This leads to perceptually incorrect estimates and limits impact in those very fine-grained analysis domains where detailed 3d models are expected to play an important role. To address such challenges we detect self-contact and design 3d losses to explicitly enforce it. Specifically, we develop a model for Self-Contact Prediction (SCP), that estimates the body surface signature of self-contact, leveraging the localization of self-contact in the image, during both training and inference. We collect two large datasets to support learning and evaluation: (1) HumanSC3D, an accurate 3d motion capture repository containing 1,032 sequences with 5,058 contact events and 1,246,487 ground truth 3d poses synchronized with images collected from multiple views, and (2) FlickrSC3D, a repository of 3,969 images, containing 25,297 surface-to-surface correspondences with annotated image spatial support. We also illustrate how more expressive 3d reconstructions can be recovered under self-contact signature constraints and present monocular detection of face-touch as one of the multiple applications made possible by more accurate self-contact models.


Learning Complex 3D Human Self-Contact

December 2020

·

20 Reads

Monocular estimation of three dimensional human self-contact is fundamental for detailed scene analysis including body language understanding and behaviour modeling. Existing 3d reconstruction methods do not focus on body regions in self-contact and consequently recover configurations that are either far from each other or self-intersecting, when they should just touch. This leads to perceptually incorrect estimates and limits impact in those very fine-grained analysis domains where detailed 3d models are expected to play an important role. To address such challenges we detect self-contact and design 3d losses to explicitly enforce it. Specifically, we develop a model for Self-Contact Prediction (SCP), that estimates the body surface signature of self-contact, leveraging the localization of self-contact in the image, during both training and inference. We collect two large datasets to support learning and evaluation: (1) HumanSC3D, an accurate 3d motion capture repository containing 1,032 sequences with 5,058 contact events and 1,246,487 ground truth 3d poses synchronized with images collected from multiple views, and (2) FlickrSC3D, a repository of 3,969 images, containing 25,297 surface-to-surface correspondences with annotated image spatial support. We also illustrate how more expressive 3d reconstructions can be recovered under self-contact signature constraints and present monocular detection of face-touch as one of the multiple applications made possible by more accurate self-contact models.




A Parallel Framework for Parametric Maximum Flow Problems in Image Segmentation

September 2015

·

19 Reads

This paper presents a framework that supports the implementation of parallel solutions for the widespread parametric maximum flow computational routines used in image segmentation algorithms. The framework is based on supergraphs, a special construction combining several image graphs into a larger one, and works on various architectures (multi-core or GPU), either locally or remotely in a cluster of computing nodes. The framework can also be used for performance evaluation of parallel implementations of maximum flow algorithms. We present the case study of a state-of-the-art image segmentation algorithm based on graph cuts, Constrained Parametric Min-Cut (CPMC), that uses the parallel framework to solve parametric maximum flow problems, based on a GPU implementation of the well-known push-relabel algorithm. Our results indicate that real-time implementations based on the proposed techniques are possible.


Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments

December 2013

·

2,226 Reads

·

3,811 Citations

IEEE Transactions on Pattern Analysis and Machine Intelligence

We introduce a new dataset, Human3.6M, of 3.6 Million 3D Human poses, acquired by recording the performance of 11 subjects, under 4 different viewpoints, for training realistic human sensing systems and for evaluating the next generation of human pose estimation models. Besides increasing the size the current state of the art datasets by several orders of magnitude, we aim to complement such datasets with a diverse set of poses encountered in typical human activities (taking photos, posing, greeting, eating, etc.), with synchronized image, motion capture and depth data, and with accurate 3D body scans of all subjects involved. We also provide mixed reality videos where 3D human models are animated using motion capture data and inserted using correct 3D geometry, in complex real environments, viewed with moving cameras, and under occlusion. Finally, we provide large scale statistical models and detailed evaluation baselines for the dataset illustrating its diversity and the scope for improvement by future work in the research community. The dataset and code for the associated large-scale learning models, features, visualization tools, as well as the evaluation server, are available online at http://vision.imar.ro/human3.6m.

Citations (6)


... A consistent comparative analysis in terms of Mean-Per-Joint Position Error, Mean-Per-Single-Joint Position Error (MPSJPE) and runtime is made possible by the fact that all these works make reference to the public Human3.6M (H3.6M) dataset [39], [40], taken into consideration and exploited in many research studies [19]. The results demonstrate that the proposed method surpasses state-of-the-art systems in both MPJPE and runtime. ...

Reference:

A Real-Time Human Pose Measurement System for Human-In-The-Loop Dynamic Simulators
Learning Complex 3D Human Self-Contact
  • Citing Article
  • May 2021

Proceedings of the AAAI Conference on Artificial Intelligence

... We evaluate our model variants on the fit3D dataset [10], as it is the only publicly available sports dataset with SMPL- X annotations. The fit3D dataset comprises videos of human subjects performing various fitness exercises, specifically designed for studying repetitive human motion in a fitness context. ...

AIFit: Automatic 3D Human-Interpretable Feedback Models for Fitness Training
  • Citing Conference Paper
  • June 2021

... It has also been suggested that there may be an optimal sequence of interaction objectives, as indicated by Baraka et al. (2022). Ensuring the robot's behaviours are predictable can benefit attention, as high variability in speech, motion, and responses might lead to reduced attention levels over time (Schadenberg et al., 2021). However, it's worth noting that a previous study found no significant differences between contingent and non-contingent robot actions (Peca et al., 2015). ...

Predictable Robots for Autistic Children --- Variance in Robot Behaviour, Idiosyncrasies in Autistic Children's Characteristics, and Child-Robot Engagement

ACM Transactions on Computer-Human Interaction

... Beyond 3D skeleton data, several datasets incorporate 3D body mesh information, enabling high-fidelity modeling of human-human interactions through more comprehensive body structure representation. Datasets like 3DPW [141], Chi3D [54], MultiHuman [279], and Hi4D [256] utilize parametric body models, such as SMPL [163], SMPL-X [134], and GHUM [243], to capture subtle variations in fullbody shape and pose. In contrast, some datasets, like MHHI [133] and ExPI [70], adopt a different approach by constructing custom 3D human meshes from captured motion actors, achieving more realistic motion representation and better alignment between 3D meshes and visual data. ...

Three-Dimensional Reconstruction of Human Interactions
  • Citing Conference Paper
  • June 2020

... used by scientists to quantify behavior for further research [14]. Human pose estimation can be used for psychological therapy to treat certain mental disorders, such as autism in children [15]. • Virtual Try-on for clothing: In recent years, online shopping has become increasingly popular, especially for fashion clothing. ...

3D Human Sensing, Action and Emotion Recognition in Robot Assisted Therapy of Children with Autism
  • Citing Conference Paper
  • June 2018

... An example of the latter is the Human3.6M database [54], which labels no less than 3.6 million singular human poses, yet all of them are based on video data from 11 subjects instructed to simulate performing less than 20 everyday tasks. ...

Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments
  • Citing Article
  • December 2013

IEEE Transactions on Pattern Analysis and Machine Intelligence