loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Authors: Noura Joudieh 1 ; Djadja Jean Delest Djadja 2 ; Ludovic Hamon 2 and Sébastien George 2

Affiliations: 1 Faculty of Sciences, Lebanese University, Hadat, Lebanon ; 2 LIUM, Le Mans University, Le Mans, France

Keyword(s): Virtual Learning Environment, Gesture Learning- Evaluation Set up, 3D Checkpoints, Random Forest.

Abstract: Nowadays, Virtual Learning Environments (VLE) dedicated to learning gestures are more and more used in sports, surgery, and in every domain where accurate and complex technical skills are required. Indeed, one can learn from the observation and imitation of a recorded task, performed by the teacher, through a 3D virtual avatar. In addition, the student’s performance can be automatically compared to that of the teacher by considering kinematic, dynamic, or geometric properties. The motions of the body parts or the manipulated objects can be considered as a whole, or temporally and spatially decomposed into a set of ordered steps, to make the learning process easier. In this context, CheckPoints (CPs) i.e. simple 3D shapes acting as “visible landmarks”, with which a body part or an object must go through, can help in the definition of those steps. However, manually setting CPs can be a tedious task especially when they are numerous. In this paper, we propose a machine learning-based sy stem that predicts the number and the 3D position of CPs, given some demonstrations of the task to learn in the VLE. The underlying pipeline used two models: (a) the “window model” predicts the temporal parts of the demonstrated motion that may hold a CP and (b) the “position model” predicts the 3D position of the CP for each predicted part from (a). The pipeline is applied to three learning activities: (i) glass manipulation (ii), geometric shapes drawing and (iii), a dilution process in biology. For each activity, the F1-score is equal to or higher than 70% for the “window model”, while the Normalized Root Mean Squared Error (NRMSE) is below 0.07 for the “position model”. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 44.200.86.95

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Joudieh, N.; Jean Delest Djadja, D.; Hamon, L. and George, S. (2023). Automatic Prediction of 3D Checkpoints for Technical Gesture Learning in Virtual Environments. In Proceedings of the 18th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2023) - GRAPP; ISBN 978-989-758-634-7; ISSN 2184-4321, SciTePress, pages 39-51. DOI: 10.5220/0011618100003417

@conference{grapp23,
author={Noura Joudieh. and Djadja {Jean Delest Djadja}. and Ludovic Hamon. and Sébastien George.},
title={Automatic Prediction of 3D Checkpoints for Technical Gesture Learning in Virtual Environments},
booktitle={Proceedings of the 18th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2023) - GRAPP},
year={2023},
pages={39-51},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0011618100003417},
isbn={978-989-758-634-7},
issn={2184-4321},
}

TY - CONF

JO - Proceedings of the 18th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2023) - GRAPP
TI - Automatic Prediction of 3D Checkpoints for Technical Gesture Learning in Virtual Environments
SN - 978-989-758-634-7
IS - 2184-4321
AU - Joudieh, N.
AU - Jean Delest Djadja, D.
AU - Hamon, L.
AU - George, S.
PY - 2023
SP - 39
EP - 51
DO - 10.5220/0011618100003417
PB - SciTePress