Next Article in Journal
Epileptic Seizures Detection in EEG Signals Using Fusion Handcrafted and Deep Learning Features
Next Article in Special Issue
Utility and Usability of Two Forms of Supplemental Vibrotactile Kinesthetic Feedback for Enhancing Movement Accuracy and Efficiency in Goal-Directed Reaching
Previous Article in Journal
Enabling the Industrial Internet of Things to Cloud Continuum in a Real City Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning-Based Estimation of Ground Reaction Forces and Knee Joint Kinetics from Inertial Sensors While Performing a Vertical Drop Jump

1
Dipartimento di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, 20133 Milano, Italy
2
E4Sport Laboratory, Politecnico di Milano, 23900 Lecco, Italy
3
Dipartimento di Meccanica, Politecnico di Milano, 20133 Milano, Italy
4
Department of Biomedical Sciences for Health, Università degli Studi di Milano, 20133 Milano, Italy
*
Author to whom correspondence should be addressed.
Submission received: 18 October 2021 / Revised: 8 November 2021 / Accepted: 16 November 2021 / Published: 19 November 2021
(This article belongs to the Special Issue Applications, Wearables and Sensors for Sports Performance Assessment)

Abstract

:
Nowadays, the use of wearable inertial-based systems together with machine learning methods opens new pathways to assess athletes’ performance. In this paper, we developed a neural network-based approach for the estimation of the Ground Reaction Forces (GRFs) and the three-dimensional knee joint moments during the first landing phase of the Vertical Drop Jump. Data were simultaneously recorded from three commercial inertial units and an optoelectronic system during the execution of 112 jumps performed by 11 healthy participants. Data were processed and sorted to obtain a time-matched dataset, and a non-linear autoregressive with external input neural network was implemented in Matlab. The network was trained through a train-test split technique, and performance was evaluated in terms of Root Mean Square Error (RMSE). The network was able to estimate the time course of GRFs and joint moments with a mean RMSE of 0.02 N/kg and 0.04 N·m/kg, respectively. Despite the comparatively restricted data set and slight boundary errors, the results supported the use of the developed method to estimate joint kinetics, opening a new perspective for the development of an in-field analysis method.

1. Introduction

The Vertical Drop Jump (VDJ) is a plyometric training exercise commonly used by coaches to develop a wide variety of athletic biomotor qualities, such as speed strength, sprinting speed, and explosive power of the lower limb [1]. The VDJ requires the athlete to step from a measured drop height and, after landing on the ground with a limited leg flexion, to perform a maximal vertical jump, with a short ground-contact period [1,2].
The VDJ is usually part of a training program to enhance performance and neuromuscular readiness, and it has been widely used as clinical test for injury-risk screening purpose for Anterior Cruciate Ligament (ACL) non-contact injury assessment and evaluation of training interventions [1,3,4,5], while the evidence of its capability of predicting future injuries is still scarce [6,7,8,9,10,11].
Biomechanical factors and specific at-risk movement patterns have been assessed in the VDJ with complex 3D motion analysis set-ups based on optical motion capture systems. So far, such systems are the gold standard to evaluate kinematics and kinetics of lower extremity in a controlled laboratory setting, performed in order to replicate postures consistent with real at-risk sporting situations [7,12,13,14].
The set of kinematic and kinetic variables obtained by using 3-D systems enables the evaluation of lower extremity musculoskeletal loading and of at-risk multi-planar knee movements for screening purposes [9,11,15] leading to significant advances in the understanding of landing biomechanics.
However, these methods are not easy to apply as a large-scale screening tool: they are time-consuming, expensive, and their measurements are potentially affected by methodological issues (e.g., errors while identifying anatomical landmarks) or skin movement artefacts [7,12,15].
Additionally, the athlete is required to move into to a restricted area of action (close to a force plate and/or an instrumented laboratory) while wearing the equipment for acquisition (markers), and those restrictions may influence the movement pattern.
All these considerations indicate that there is a need for a simple and accurate method to capture knee kinematics and kinetics to assess the risk of injury during jumping tasks.
The use of wearable inertial-based systems using Inertial Measurement Units (IMUs) is a promising alternative to the use of the gold standard optoelectronic systems [12,16,17,18]. IMU-based systems do not require the use of cameras or complex laboratory settings to perform the test. Inertial sensors can be easily attached with straps to the limbs of the athlete and addressed separately to define body parts, allowing for a direct determination of kinematic parameters, providing data of angular acceleration and velocity, as well as spatial orientation of the individual body segments [12,19].
However, it should be noticed that, similar to the optical markers, IMUs should be placed in specific bony landmarks and not directly on the muscle to reduce soft tissue artefacts caused by the displacement of the sensor of the underlying tissues with respect to the bone [17,18,20].
IMUs are designed to be small, lightweight, and wireless, thus permitting full movements while participating in a sport or a test, and they do not require a predefined area of action to which the movement must be restricted. As the athlete is free to move, the collected data are potentially more consistent with a real sport situation (i.e., increased ecological validity of testing) and can be used to enhance athletic performance as well as to identify at-risk motion patterns [12,21,22].
IMU-based systems have already been successfully used to detect kinematic parameters in track and field applications [21] as well as in gait analysis [20,23,24].
IMU-based systems offer the possibility to directly assess temporal, dynamic, and kinematic parameters but IMUs raw data per se cannot be used as input parameter for inverse dynamic calculations. Different approaches based on machine learning methods and Artificial Neural Networks (ANNs) are used to support human motion analysis to assess kinetic parameters such as GRFs and joint moments [25,26,27].
ANNs are inspired by the structure of the human brain and serve as universal approximator, able to realise an arbitrary mapping of one vector space onto another vector space [23,28]. Instead of explicitly programming the solution of one specific task, ANNs are able to capture and use some a priori unknown information hidden in given data (training dataset), learn from them, and apply the learned input–output relationship to new data, allowing for a prediction of the behaviour of a system with respect to previously unseen data [19,23,29].
Many studies addressed the usability of ANNs to estimate ground reaction forces or joint moments by means of kinematic data obtained from an optical motion analysis system [26] or where such methods enable the use of IMU sensors for the measurement and, simultaneously, optical data for training purpose [23]. Most of the existing research uses optical motion capture systems to apply kinetics prediction methods to kinematic data. Advancements of the combined use of IMU data and ANNs, when properly validated [30], might have the potential to overcome the actual limitations of a time-consuming traditional laboratory set-up [12,25,31].
Various studies [5,9,10,32] have estimated knee joint load during jumping tasks such as the VDJ, but no study has directly assessed knee joint load using wearable sensors so far. In particular, we have no evidence of ANN-based approaches to predict knee joint kinetics during the landing phase. Thus, the aim of this study was to develop and validate an ANN allowing for the estimation of knee joint moments and GRFs during the Vertical Drop Jump by using IMUs data.
The findings of this study could help to develop a portable, practical, and field-based method to quantify and monitor knee kinetics. IMU-instrumented assessment would allow us to easily collect large datasets of objective biomechanical data in order to monitor an athlete’s workload, to set up effective training programs [26], as well as to identify at-risk behaviors that may result in a ligamentous injury.

2. Materials and Methods

2.1. Participants and Experimental Procedures

The study involved the simultaneous collection of data with commercial IMU sensors (Physilog® by Gait Up Ltd., Lausanne, Switzerland), an optoelectronic marker-based motion capture system (SMART DX 400 system by BTS Bioengineering SPA, Milan, Italy), and a force plate (AMTI OR6-7 force platform by Advanced Mechanical Technology, Inc., Watertown, MA, USA) during the execution of the Vertical Drop Jump.
Eleven healthy subjects (five women, six men), aged between 23 and 29 years, were recruited on a voluntary basis and asked to perform up to 10 VDJs in a row. Inclusion criteria were not being under- or overweighted (body mass index between 18.5 and 24.9) and the absence of any knee-injury history. Of the considered cohort, three subjects were regular amateur sports practitioners, whilst the other were students with an active lifestyle, occasionally attending the gym and who walk for daily commuting.
Before the tests, each subject gave written informed consent to participate in the study. Subjects’ anthropometric data are reported in Table 1.
Each session began with a short warm up after which the subjects were instructed to drop off a 30 cm box and perform a maximal VDJ upon landing.
Participants were asked to stand upright on the box and instructed to start the movement by stepping out from the box with a single leg as well as to place arms on the hips throughout the execution of the jump, according to the reference technical execution model [1]. All subjects were previously acquainted with the experimental tasks and were able to cope with the provided instructions and to complete the test.

2.2. Experimental Setup

The stereophotogrammetric system used for the optical data acquisition consisted of a SMART DX 400 system (BTS Bioengineering SPA, Milan, Italy) with a sampling frequency of 100 Hz. The system was calibrated prior to each usage according to the manufacturer’s specification.
The laboratory was also equipped with two 46.5 × 51.8 cm2 AMTI OR6-7 force platforms (Advanced Mechanical Technology, Inc., Watertown, MA, USA), sampling at 200 Hz, to collect Ground Reaction Force (GRF) data. The contact phase of the first landing of each VDJ was defined as the period where the unfiltered vertical GRF exceeded 10 N. Each VDJ was considered successful if the subject landed with one foot on each of the two force plates and all the reflective markers were visible by the cameras throughout the whole jump. Optical and ground reaction forces data were automatically synchronized.

2.2.1. Marker Set

Reflective skin markers (Ø = 10–15 mm) coated with an aluminum powder film and mounted on a plastic tip were attached over anatomical landmarks on the legs and pelvis of each subject according to a lower-body model [33] using 30 markers (Figure 1).
Landmarks’ position was manually determined by palpation, by identifying regions with reduced thickness of the tissues interposed between bone and skin. According to this approach, 18 markers were bilaterally placed on the superior-anterior iliac spine, posterior iliac crest, lateral and medial femoral epicondyle, lateral and medial malleolus, calcaneus and foot (corresponding to the 1st and 5th metatarsal heads).
Furthermore, 12 markers fixed at four clusters were attached to subject’s thigh and shank on both legs. Each cluster included three markers located at a fixed distance, and it defines a technical reference system fixed to the segment to which it is attached. The anthropometric model was composed of rigid segments, representing the bony segments of the subject, and joint kinematics was defined as the relative orientation in time between two adjacent segments [34].

2.2.2. IMU Devices

Three wearable IMU devices (Physilog® by Gait Up Ltd., Lausanne, Switzerland) were used to measure and acquire the VDJs. Sensors were configured via the Research ToolKit Desktop software and organized into a multiple sensor network, including a single master sensor and two slave sensors to record synchronous signals.
The IMUs synchronized regularly during the measurement, and they were easily turned on/off wirelessly through the provided mobile application (GaitUp companion Android application. Lausanne, Switzerland). Raw data were stored on the onboard SD memory of each sensor and then transferred to a PC for offline analysis. The slave sensors were placed via Velcro® straps on the participant’s right thigh and shank, whilst the master sensor was placed with a rubber clip on the sacrum, between the posterior and superior iliac spine.
Each sensor is a six-axes stand-alone unit integrating a micro-controller, an internal memory, a battery (Lithium-Ion Polymer Accumulator, 3.7 V 140 mAh by Renata, Itingen, Switzerland), a barometric pressure sensor, and a 3D accelerometer and a 3D gyroscope with programmable ranges. Device settings were set to a sampling frequency of 128 Hz and a dynamic range of ±16 g for the accelerometer and ±2000 degrees/s for the gyroscope.
Each unit is a compact (47.5 mm × 26.5 mm × 10 mm) and lightweight (36 g) wearable device for sensing movement, allowing for a long-term motion recording and on-board processing. Physilog® does not require any calibration before the acquisition because this are estimated during the recording and its right-handed coordinate system is oriented as displayed in Figure 2, together with the locations of the sensors.
An optical maker was placed on the master sensor and used as a synchronization reference for the data recorded by the motion capture system and the IMUs.
Prior to each jumping session, an operator moved the master sensor (and the attached marker therein) up and down five or more times to obtain a quasi-periodical signal recognizable from both the IMU and optical system records.

2.3. Data Analysis and Processing

A total of 112 successful VDJs were recorded, and different software applications were used for complete data processing and analysis. The optical raw data were processed with SmartTracker (BTS Bioengineering, Milan, Italy) and then exported in c3d format for further processing in Visual3D (version 6.03.06, C-Motion, Inc. Germantown, MD, USA). Visual3D makes use of the relative positions of at least three non-aligned markers to define body segment as a rigid body, described by scaled inertial properties [32]. Such a model (Figure 3) was first fit to the static recording of each subject. Then, the calibrated model was matched to the drop jump records to calculate the joint angles (kinematics) and then the joint moments (kinetics). In particular, joint kinetics was obtained by applying the inverse dynamics analysis to the kinematics of the biomechanical model and to the location, magnitude, and direction of the ground reaction forces acting on the foot during the first landing phase of the VDJ (externally applied forces).
Joint moments were computed according to this approach for the hip, knee, and ankle. In the study, only the right knee joints moments were selected and used for further analysis. The conventions used to represent the moments are reported in Table 2.
Additionally, Visual3D allowed for event detection. In particular, the first landing phase of each jump (in the following, landing phase) was defined as the interval between the time instant in which the feet were in contact with the ground, when the vertical GRF exceeded the 10-N threshold, and the moment in which the feet lifted from the ground to perform the jump. Data were exported to Matlab (version R2020b, The MathWorks Inc, Natick, MA, USA) for further analysis with custom scripts.
The raw data simultaneously collected by the IMUs were transferred via USB to the computer and then accessed through Research ToolKit (RTK) Desktop software (Gait Up Ltd., Lausanne, Switzerland). An example of a signal displayed by the sensor placed on the sacrum is shown in Figure 4.
For all the IMU sensors, the data collected in the same measurement were automatically assembled in a cluster and synchronized. Physilog® 5 data were also exported to Matlab for further analysis.

2.4. Data Synchronization and Dataset Structure

To use the data acquired by the IMUs as input for the ANN, signals must be time-synchronized with those acquired by the marker-based motion capture system.
In this study, the synchronization trigger was the initial periodic movement of the master sensor attached to a reflective marker.
The synchronization was performed within Matlab. First, the vertical marker coordinate (y-axis) with respect to the global reference system of the laboratory, acquired at 100 Hz, was resampled to 128 Hz. The signal acquired by the master sensor along the z-axis of its own reference system was processed by subtracting the gravity component and by filtering it using a standard fourth order low-pass Butterworth filter with a 10-Hz cut frequency, according to [35]. Signals were manually cut to isolate a window containing the spikes to be matched and used as reference for the synchronization. The cross-correlation of the cut signals was calculated, and the resultant delay was used to align the full signals, as well as all the signals recorded by IMU units (Figure 5).
After synchronization, the 3-axes signals of right knee moment and right GRFs extracted from Visual3D were cut to isolate their values during the first landing phase of the VDJ using the event detection information obtained via Visual3D and resampled to 128 Hz (Figure 6).
Prior to cutting, IMU data had to be filtered to remove noise and to preserve the meaningful components of the signals by applying a standard fourth order low-pass Butterworth filter with a 32-Hz cut frequency. The aligned accelerometer and gyroscope signals of the three inertial units were cut in the same way as the signals recorded by the optoelectronic system (i.e., in correspondence with the landing phase). For each subject, a 25-column table containing the matched signals of both systems in the first landing phase of each drop jump, as well as a reference column with the indexes of the frames of each jump, was obtained (Figure 7).
The full data table was created by concatenating the tables of each subject.
Additionally, two other tables were generated in the same way using the IMU signals filtered with a cut frequency of 12 Hz and 24 Hz to evaluate the sensitivity of the outcome to the filter settings.

2.5. Neural Network Implementation

In the current research, a non-linear autoregressive network with an external input (NARX) neural network was implemented in Matlab. The NARX is a recurrent dynamic neural network with feedback connections enclosing several layers of the network. The NARX model is a nonlinear generalization of the Autoregressive Exogenous (ARX) model that is used as a standard instrument in black-box system identification [36] and is commonly used in time series modelling in an extensive variety of dynamic systems [37].
A standard feedforward NARX network can predict the output response through regression of past output and exogenous input values [38].
The NARX model output is defined by the following equation:
y t = f y t 1 , y t 2 ,   ,   y t n y ,   u t 1 , u t 2 , , u t n u
where f ( . ) is the mapping function of the y(t) output response that is regressed by the previous time series values and past values of input data (u) and n y and n u are the time delays. This implementation allows for a vector ARX model, where the input and the output can be multidimensional.
The NARX neural network model allows for two different architectures, shown in Figure 8. The first is called series–parallel architecture (SP mode, open-loop), where the output y ^ t is computed by using only the actual values of both endogenous (u(t)) and exogenous inputs (y(t)). The second one is the parallel architecture (closed loop), where the output y ^ t is computed by using the actual values of the endogenous inputs u t and the past predicted values of the estimated output y ^ t that are fed back.
In this study, the series–parallel architecture was used as the true values of the input series (the data collected by the IMU units), and the true outputs (the GRFs and the right knee joint moments) were always available during the training of the network.
The training of the neural network was accomplished through a train–test technique. The data contained in the data matrix, which contains known input and output values, was split into a training and a test set: the VDJs performed by seven of the 11 subjects were randomly chosen and used as the training set whilst the data of the remaining four subjects served as the test set.
Data recorded by the IMUs were given as input, and the optical motion capture data served as output (target). Input and output details are reported in Table 3.
The NARX was implemented via the in-built function narxnet of the Matlab neural network toolbox. The ANN topology was determined through a trial-and-error procedure, starting from a set of hyperparameters (i.e., number of hidden layers, number of neurons therein, and so on) that could be retrieved from similar existing research (MUNDT). Then, such parameters were adjusted iteratively based on the network’s output metrics.
The input and feedback delays were set as default, ranging from 1:2, while the number of hidden layers and neurons was systematically optimized via a trial-and-error process in order to adapt the capacity of the network.
A delayed version of the input time series data was prepared using the function preparets that automatically shifted the input and target series to as many steps as needed to fill the initial input and layer delay states. These input and layer states were then employed as inputs to MATLAB’s built-in neural network training function [39].
Weights and biases were randomly set whilst configuring the network on data, and regularization was introduced to prevent overfitting and to improve the network’s ability to generalize by modifying the in-built Matlab default performance function. The regularization factor was set to 0.5.
The training dataset was used to train the neural network with a back-propagation algorithm. Different back-propagation algorithms were tested (e.g., Resilient backpropagation and the Levenberg–Marquadt algorithm) together with different activation functions for each layer, belonging to the sigmoid (hyperbolic tangent sigmoid and logarithmic sigmoid transfer) and linear (positive linear and linear transfer) function families.
During the training process, the results produced for the training set by the model were compared with the real value for each output vector fed to the net. According to this comparison and the specific learning algorithms, weights and biases were tuned to optimize the network performance function, defined as the regularized mean square error (MSE) between the network output and the target output, so that the computed values could match the known real output values.
After the training phase, the network was applied to the test set.
The network was cross validated (20-fold cross validation): the full dataset was iteratively split into different training (n = 7 subjects) and test (n = 4 subjects) sets in order to obtain insight into the network’s ability to generalize to an independent dataset and to predict new data that were not used in the training phase.
The network’s performance was evaluated as the resulting average Root Mean Square Error (RMSE), a standard statistical metric used to measure the error of the model in predicting quantitative data. Formally, RMSE is defined as
R M S E =   i = 1 n y i ^ y i 2 n =   i = 1 n e i 2 n
where y 1 ^ ... y n ^ are the predicted values, y i y n are the real values (thus, e 1 e n are the errors), and n is the number of observations.
The underlying assumption when presenting the RMSE is that the errors are unbiased and follow a normal distribution to provide a complete picture of the error distribution [40] RMSE was calculated for each predicted target.
Additionally, the error histogram containing the errors between target values and predicted values after training the network was obtained. The histogram indicates how predicted values differ from the target and it was calculated for both the training and test set.
The total error range on the x-axis was divided into 20 smaller bins, and the number of samples from the dataset in a particular bin are reported on the y-axis. Positive and negative error bins, accounting for the direction of the bias, were separated by a vertical orange line, corresponding to zero error.
The closer the predicted values to the target, the better the model performance.

3. Results

The topology of the developed NARX Neural Network is shown in Figure 9.
The input layer consisted of two 1:2 Tap Delay Lines that took the multi-dimensional endogenous time series x(t) (IMUs data) and the exogenous time series y(t) (right GRFs and right knee moments) and transmitted them in their time-delayed form into the subsequent layers. The computational engine of the network was made by two fully-connected hidden layers, composed of 1000 neurons each, where a linear activation function defined how the weighted sum of the input was transformed into the output from the nodes, finally yielding to the outputs of the whole network through the output layer.
Weights and biases were randomly initialized in each periodic training phase, and different backpropagation algorithms were tested to achieve the result.
Finally, the Resilient Backpropagation algorithm was chosen to train the final NARX network. The number of epochs was optimized via a trial-and-error process, and it was finally set to 5000, and the training stopped when the last epoch was reached.
The network was able to compute target values that closely matched the known real output values of the test set.
An example of the predicted patterns of the 3D right GRFs and the right knee moments obtained for the VDJs performed by one of the subjects chosen as the test set are reported in Figure 10, compared to their real (measured) values.
The predicted curves followed nearly the same pattern as the real curves, except at the very first contact with the ground, corresponding to the first frames of each landing phase of the VDJs.
The mean RMSE values achieved for the prediction of GRFs and right knee moments by the developed neural network are reported in Table 4.
Two additional datasets with the same structure of the main dataset were obtained by using a 12-Hz and a 24-Hz cut off frequency for the fourth order Butterworth filter applied to IMU data. The network was then trained by using those two datasets to see how the noise could affect the network’s ability to predict the data. However, the cut-off frequency slightly influenced the ability of the network to learn from the data.
The error histogram between target and predicted values, indicating how the predictions differed from the targets, was also obtained for both the training and test set. Two error histograms were obtained, one displaying the errors on the GRFs and the other displaying the errors on the right knee moments (Figure 11).
Most of the samples from the dataset were within the bins that were very close to the zero-error bar, indicating that the network was able to learn from the training set and then apply the learned pattern to the test set.

4. Discussion

In this paper, we showed that it was possible to estimate articular moments and GRFs during the first landing phase of a VDJ with an average RMSE of 0.02 N/kg and 0.04 N·m/kg by exploiting the acceleration and angular velocity recorded via inertial sensors combined with a neural network-based machine learning method.

4.1. Novelty of the Study

The predicted knee joint moments and GRFs showed a strong agreement with their measured values, according to their RMSE. Generally, there is no fixed threshold limit for RMSE but besides a small error, it would be optimal if the RMSE calculated on the training set was similar to the RMSE calculated on the test set to prevent overfitting (RMSE train > RMSE test) or underfitting (RMSE train < RMSE test).
The RMSE values were comparable for all the predicted variables, and they all respected the previously mentioned constraints, accounting for the capability of the network to predict kinetic data with a narrow error with respect to the articular moments and the GRFs acquired with traditional motion analysis methods.
A strength of the developed method lies in the ability of the network to simultaneously predict six output waveforms without the need of different models to estimate the two groups of kinetic variables during the execution of the same task.
Particular relevance was given here to the estimation of the 3D GRFs directly from the data recorded by inertial sensors, similar to previous work [41] in which the authors focused on the estimation of the vertical GRF (vGRF) during running using data from three body-worn IMUs by feeding them into two concatenated ANNs, where the first one mapped the relative orientations (expressed as quaternions) of the lower limbs to joint angles whilst the second one mapped the estimated joint angles in combination with vertical sensor acceleration with respect to vGRF.
Their model returned excellent results in matching the actual force profiles measured by the force plate, estimating the vGRF with an accuracy < 0.27 BW in terms of RMSE, which is in line with the results achieved in this study (RMSE vertical GRF = 0.04 N/Kg). However, it should be noticed that Wouda’s model [41] was developed to estimate the vertical GRFs during running, a highly dynamic and cyclic action in which the GRFs vary rapidly and therefore the information of the joint angles need to be included to achieve a better prediction of the kinetics.
The Physilog® sensors used in this study allowed for the extraction of quaternions; however, as the analysis of the GRFs was conducted on a small temporal window referring to the first landing phase of the VDJ, it was decided not to include them within the model. It could be expected though that a deep network is able to incorporate sensor orientation, whose information is available as raw data.
The prediction performance of the knee joint moments was consistent with that achieved by Stetter et al. in 2020 [42], in which the authors developed an ANN able to estimate knee flexion and adduction moment during various locomotion tasks based on data obtained by two wearable inertial sensors located on the right thigh and shank. The moments estimated by the network showed a good agreement with their inverse dynamics-calculated counterpart, despite varying between tasks. In particular, the corresponding RMSE values obtained in this study for the knee FE moment and knee AA moment were 0.07 N·Kg/m and 0.03 N·Kg/m, respectively, lying almost in the same range of those reported in Table 4.
It should be noticed that task-specific modelling may lead to an increased estimation accuracy for individual tasks, and in turn to a lower RMSE with respect to a multi-tasking model, but it has the disadvantage that each task must be modelled by itself [42]. Following this observation, the network developed in this study was able to provide an accurate estimation for the VDJ kinetics, but it may not be suitable to make accurate predictions for data from other sporting tasks due to its high specificity for the current application.
The overall results of the study also confirmed the hypothesis of Mundt et al. (2019) [12] concerning the applicability of a feed-forward (FF) neural network to predict the joint moments and GRFs during sport-specific movements based on data from inertial sensors. The network developed hereby was not a standard FF network but a NARX network able to predict the output response through the regression of past outputs and exogenous input values. However, the use of the series–parallel architecture of the NARX allowed for the obtainment of a network exploiting the advantage of having two different series of input values, resulting in a more accurate prediction on the outputs, together with a pure feed-forward mechanism, as previously stated by Mundt et al. (2019) [12].
The network was developed on a dataset in which the duration in terms of frames of the first landing phase of each drop jump was not normalized and it showed a high variability according to the jumping behaviour of the subjects. As the network was able to learn the jumping pattern of each subject starting from non-time normalized series, the information about the individual jumping behaviour of each subject during the first landing phase of the drop jump was preserved, allowing for its potential use as indicator of the power output of the athlete [1].

4.2. Limitations

The network was able to learn from the data during the training phase and to then apply the learned pattern to new unseen data; however, it should be noticed that the network was not always able to make a sufficiently accurate prediction of the knee moments of the GRFs at the very first instant of the landing phase of the VDJ, corresponding to the first 2–3 frames of each jump (boundary artefacts). Prediction errors accumulated at the beginning of the curve but then the predicted joint moment followed nearly the same pattern as the measured waveform (Figure 12).
Despite this limitation, the ANN was able to give an accurate prediction of the values of the curves throughout the remaining duration of the landing phase of each jump. As the actual shape and peak of the varo/valgus moment was preserved, it potentially allowed for their use to make an evaluation of the landing behaviour for screening purposes for ACL injury.
Additional limitations should be considered. The present method focused on the estimation of the kinetic of the limb where two of the three IMUs were attached, without exploring the possibility of simultaneously predicting the same variables on the contralateral limb. Additionally, at present, the developed model did not allow for an online analysis of the kinetic variables during the execution of the drop jump and it still requires the use of data from force plates.
In fact, the detection of the events defining the beginning and the end of each landing phase was not implemented by using data directly from the IMUs, such as in Zago et al., (2021) [43], but it needed to be done via the data recorded by the force plates and processed with Visual3D (or an equivalent biomechanics software) afterwards.
The last concern was about the relatively restricted dataset used to train the network. The study was conducted during a global pandemic, and due to the limitation of the access to the laboratory and to the health protocols, only eleven subjects could take part to the trials. The performance of a machine learning method adaptively increases with the amount of data available for training, allowing the network to learn more motion patterns to improve its performance regarding new data [12,42]. Thus, the results achieved on a small dataset may not be extended to a wider dataset as the jumping behaviour is highly variable between subjects.

5. Conclusions

The current study investigated the possibility of an ANN-based method to predict knee joint kinetics and GRFs during the first landing phase of a VDJ, based on data recorded by three inertial sensors.
Despite the comparatively restricted dataset and the accumulation of boundary errors at the very beginning of the landing phase, the results supported the use of wearable inertial sensors in combination with machine learning-based methods to estimate joint kinetics in a sports application as the VDJ, highlighting the great potential of such methods. This approach may allow for the replacement of time-consuming inverse dynamics calculation, for trials with less equipment, time, and effort, as well as being feasible to test subjects in a natural environment and not only in a laboratory set-up [12,16].
Looking ahead, it should be investigated if a more complex architecture of the network may lead to an improvement in the prediction accuracy at the beginning of the landing phase, allowing for a consistent learning of the motion pattern along the curve and thus overcoming the error accumulation previously discussed.
Further improvement should be implemented to incorporate other targets into the prediction, e.g., hip adduction/abduction moment, which has been shown to have a key role in the ACL injury mechanism. The final goal would be to embed such neural networks into a mobile application controlling the sensors to display the moments at the knee joint and the GRFs in real time.
The method described in this study represents a promising starting point for further improvements leading to new pathways for in-field diagnosis methods that in turn could be applied to the early identification of at-risk landing behaviours and potentially highlight a link between injury risk and IMU-based instrumented assessment.

Author Contributions

Conceptualization, S.C., M.Z. and M.G.; methodology, S.C. and M.Z.; software, S.C. and M.Z.; validation, S.C. and M.Z.; formal analysis, S.C. and M.Z.; investigation, F.B.; resources, M.G., M.T. and C.S.; data curation, S.C., M.Z. and F.B.; writing—original draft preparation, S.C.; writing—review and editing, M.Z.; visualization, F.B.; supervision, M.G., M.T. and C.S.; project administration, M.G., M.T. and C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and was approved by the Ethics Committee of Politecnico di Milano (protocol code 06/2020, Milan 04/02/2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

A raw dataset is available online at the following repository: “VDJ_NNdataset”, Mendeley Data, V1, doi:10.17632/8hsrymnzb6.1.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pedley, J.S.; Lloyd, R.S.; Read, P.; Moore, I.S.; Oliver, J.L. Drop jump: A technical model for scientific application. Strength Cond. J. 2017, 39, 36–44. [Google Scholar] [CrossRef]
  2. Eiras, A.E.; Ladeira, R.; Silva, P.A.; André, N.; Pereira, R.; Machado, M. Drop jump and muscle damage markers. Serb. J. Sport. Sci. 2009, 3, 81–84. [Google Scholar]
  3. Augustsson, S.R.; Tranberg, R.; Zügner, R.; Augustsson, J. Vertical drop jump landing depth influences knee kinematics in female recreational athletes. Phys. Ther. Sport 2018, 33, 133–138. [Google Scholar] [CrossRef] [PubMed]
  4. Chimera, N.J.; Warren, M. Use of clinical movement screening tests to predict injury in sport. World J. Orthop. 2016, 7, 202–217. [Google Scholar] [CrossRef] [PubMed]
  5. Mok, K.M.; Petushek, E.; Krosshaug, T. Reliability of knee biomechanics during a vertical drop jump in elite female athletes. Gait Posture 2016, 46, 173–178. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Kristianslund, E.; Krosshaug, T. Comparison of drop jumps and sport-specific sidestep cutting: Implications for anterior cruciate ligament injury risk screening. Am. J. Sports Med. 2013, 41, 684–688. [Google Scholar] [CrossRef] [PubMed]
  7. Smith, H.C.; Johnson, R.J.; Shultz, S.J.; Tourville, T.; Holterman, L.A.; Slauterbeck, J.; Pamela Vacek, M.; Beynnon, B.D. A Prospective Evaluation of the Landing Error Scoring System (LESS) as a Screening Tool for Anterior Cruciate Ligament Injury Risk. Am. J. Sports Med. 2012, 40, 521–526. [Google Scholar] [CrossRef] [PubMed]
  8. Cronström, A.; Creaby, M.W.; Ageberg, E. Do knee abduction kinematics and kinetics predict future anterior cruciate ligament injury risk? A systematic review and meta-analysis of prospective studies. BMC Musculoskelet. Disord. 2020, 21, 1–11. [Google Scholar] [CrossRef]
  9. Lucarno, S.; Zago, M.; Rossi, E.; Muratore, S.; Baroncini, G.; Alberti, G.; Gualtieri, D. Influence of age and sex on drop jump landing strategies in élite youth soccer players. Int. J. Sports Sci. Coach. 2020, 16, 344–351. [Google Scholar] [CrossRef]
  10. Hewett, T.E.; Myer, G.D.; Ford, K.R.; Paterno, M.V.; Quatman, C.E. Mechanisms, Prediction, and Prevention of ACL Injuries: Cut Risk With Three Sharpened and Validated Tools. J. Orthop. Res. 2016, 34, 1843–1855. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Hewett, T.E.; Myer, G.D.; Ford, K.R.; Heidt, R.S.; Colosimo, A.J.; McLean, S.G.; Van Den Bogert, A.J.; Paterno, M.V.; Succop, P. Biomechanical measures of neuromuscular control and valgus loading of the knee predict anterior cruciate ligament injury risk in female athletes: A prospective study. Am. J. Sports Med. 2005, 33, 492–501. [Google Scholar] [CrossRef] [Green Version]
  12. Mundt, M.; David, S.; Koeppe, A.; Bamer, F.; Markert, B.; Potthast, W. Intelligent prediction of kinetic parameters during cutting manoeuvres. Med. Biol. Eng. Comput. 2019, 57, 1833–1841. [Google Scholar] [CrossRef] [PubMed]
  13. Ekegren, C.L.; Miller, W.C.; Celebrin, R.G.; Eng, J.J.; MacIntyre, D.L. Reliability and validity of observational risk screening in evaluating dynamic knee valgus. J. Orthop. Sports Phys. Ther. 2009, 39, 665–674. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Fox, A.S.; Bonacci, J.; McLean, S.G.; Spittle, M.; Saunders, N. What is normal? Female lower limb kinematic profiles during athletic tasks used to examine anterior cruciate ligament injury risk: A systematic review. Sports Med. 2014, 44, 815–832. [Google Scholar] [CrossRef] [PubMed]
  15. Nilstad, A.; Krosshaug, T.; Mok, K.M.; Bahr, R.; Andersen, T.E. Association between anatomical characteristics, knee laxity, muscle strength, and peak knee valgus during vertical drop-jump landings. J. Orthop. Sports Phys. Ther. 2015, 45, 998–1005. [Google Scholar] [CrossRef] [PubMed]
  16. Dowling, A.V.; Favre, J.; Andriacchi, T.P. A wearable system to assess risk for anterior cruciate ligament injury during jump landing: Measurements of temporal events, jump height, and sagittal plane kinematics. J. Biomech. Eng. 2011, 133, 1–7. [Google Scholar] [CrossRef]
  17. Robert-Lachaine, X.; Mecheri, H.; Larue, C.; Plamondon, A. Validation of inertial measurement units with an optoelectronic system for whole-body motion analysis. Med. Biol. Eng. Comput. 2017, 55, 609–619. [Google Scholar] [CrossRef] [PubMed]
  18. Teufl, W.; Miezal, M.; Taetz, B.; Frohlichi, M.; Bleser, G. Validity of inertial sensor based 3D joint kinematics of static and dynamic sport and physiotherapy specific movements. PLoS ONE 2019, 14, e0213064. [Google Scholar] [CrossRef] [Green Version]
  19. Wang, L.; Cheng, L.; Zhao, G. Machine Learning for Human Motion Analysis: Theory and Practice; IGI Global: Hershey, PA, USA, 2009. [Google Scholar] [CrossRef] [Green Version]
  20. Seel, T.; Raisch, J.; Schauer, T. IMU-based joint angle measurement for gait analysis. Sensors 2014, 14, 6891–6909. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Schmidt, M.; Jaitner, T.; Nolte, K.; Rheinländer, C.; Wille, S.; Wehn, N. A wearable inertial sensor unit for jump diagnosis in multiple athletes. In Proceedings of the icSPORTS 2014—2nd International Congress on Sports Sciences Research and Technology Support, Rome, Italy, 24–26 October 2014; pp. 216–220. [Google Scholar] [CrossRef] [Green Version]
  22. Adesida, Y.; Papi, E.; McGregor, A.H. Exploring the role of wearable technology in sport kinematics and kinetics: A systematic review. Sensors 2019, 19, 1597. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Mundt, M.; Koeppe, A.; David, S.; Witter, T.; Bamer, F.; Potthast, W.; Markert, B. Estimation of Gait Mechanics Based on Simulated and Measured IMU Data Using an Artificial Neural Network. Front. Bioeng. Biotechnol. 2020, 8, 1–16. [Google Scholar] [CrossRef]
  24. Teufl, W.; Miezal, M.; Taetz, B.; Fröhlich, M.; Bleser, G. Validity, test-retest reliability and long-term stability of magnetometer free inertial sensor based 3D joint kinematics. Sensors 2018, 18, 1980. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Mundt, M.; Koeppe, A.; Bamer, F.; Potthast, W.; Pforzheim, A.C. Prediction of joint kinetics based on joint kinematics using neural networks. In Proceedings of the 36th Conference of the International Society of Biomechanics in Sports, Auckland, New Zealand, 10–14 September 2018; pp. 794–797. [Google Scholar]
  26. Stetter, B.J.; Ringhof, S.; Krafft, F.C.; Sell, S.; Stein, T. Estimation of knee joint forces in sport movements using wearable sensors and machine learning. Sensors 2019, 19, 3690. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Lim, H.; Kim, B.; Park, S. Prediction of lower limb kinetics and kinematics during walking by a single IMU on the lower back using machine learning. Sensors 2019, 20, 130. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Maggiora, G.M.; Elrod, D.W.; Laboratories, U.; Street, H. Model-Free Mapping Devices. J. Chem. Inf. Comput. Sci. 1992, 32, 732–741. [Google Scholar] [CrossRef]
  29. Svozil, D.; Kvasnieka, V.; Pospichal, J. Introduction to multi-layer feed-forward neural networks. Comput. Sci. 1997, 39, 43–62. [Google Scholar] [CrossRef]
  30. Argent, R.; Bevilacqua, A.; Keogh, A.; Daly, A.; Caulfield, B. The importance of real-world validation of machine learning systems in wearable exercise biofeedback platforms: A case study. Sensors 2021, 21, 2346. [Google Scholar] [CrossRef]
  31. Karatsidis, A.; Bellusci, G.; Schepers, H.M.; de Zee, M.; Andersen, M.S.; Veltink, P.H. Estimation of ground reaction forces and moments during gait using only inertial motion capture. Sensors 2017, 17, 75. [Google Scholar] [CrossRef] [Green Version]
  32. Bates, N.A.; Ford, K.R.; Myer, G.D.; Hewett, T.E. Impact Differences in Ground Reaction Force and Center of Mass Between the First and Second Landing Phases of a Drop Vertical Jump and their Implications for Injury Risk Assessment. J. Biomech. 2013, 46, 1237–1241. [Google Scholar] [CrossRef] [Green Version]
  33. Collins, T.D.; Ghoussayni, S.N.; Ewins, D.J.; Kent, J.A. A six degrees-of-freedom marker set for gait analysis: Repeatability and comparison with a modified Helen Hayes set. Gait Posture 2009, 30, 173–180. [Google Scholar] [CrossRef]
  34. Winter, D.A. Biomechanics and Motor Control of Human Movement, 4th ed.; Wiley: Hoboken, NJ, USA, 2009; ISBN 9780470398180. [Google Scholar]
  35. Crenna, F.; Rossi, G.B.; Berardengo, M. Filtering biomechanical signals in movement analysis. Sensors 2021, 21, 4580. [Google Scholar] [CrossRef]
  36. Chen, S.; Billings, S.A.; Grant, P.M. Non-linear system identification using neural networks. Int. J. Control 1990, 51, 1191–1214. [Google Scholar] [CrossRef]
  37. Boussaada, Z.; Curea, O.; Remaci, A.; Camblong, H.; Bellaaj, N.M. A nonlinear autoregressive exogenous (NARX) neural network model for the prediction of the daily direct solar radiation. Energies 2018, 11, 620. [Google Scholar] [CrossRef] [Green Version]
  38. Hosen, M.S.; Jaguemont, J.; Van Mierlo, J.; Berecibar, M. Battery lifetime prediction and performance assessment of different modeling approaches. iScience 2021, 24, 102060. [Google Scholar] [CrossRef] [PubMed]
  39. Khan, S.A.; Thakore, V.; Behal, A.; Bölöni, L.; Hickman, J.J. Comparative analysis of system identification techniques for nonlinear modeling of the neuron-microelectrode junction. J. Comput. Theor. Nanosci. 2013, 10, 573–580. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE)?—Arguments against avoiding RMSE in the literature. Geosci. Model Dev. 2014, 7, 1247–1250. [Google Scholar] [CrossRef] [Green Version]
  41. Wouda, F.J.; Giuberti, M.; Bellusci, G.; Maartens, E.; Reenalda, J.; van Beijnum, B.J.F.; Veltink, P.H. Estimation of vertical ground reaction forces and sagittal knee kinematics during running using three inertial sensors. Front. Physiol. 2018, 9, 1–14. [Google Scholar] [CrossRef]
  42. Stetter, B.J.; Krafft, F.C.; Ringhof, S.; Stein, T.; Sell, S. A Machine Learning and Wearable Sensor Based Approach to Estimate External Knee Flexion and Adduction Moments During Various Locomotion Tasks. Front. Bioeng. Biotechnol. 2020, 8, 9. [Google Scholar] [CrossRef]
  43. Zago, M.; Tarabini, M.; Spiga, M.D.; Ferrario, C.; Bertozzi, F.; Sforza, C.; Galli, M. Machine-learning based determination of gait events from foot-mounted inertial units. Sensors 2021, 21, 839. [Google Scholar] [CrossRef]
Figure 1. Subject equipped with the marker set used in the study: (a) front and (b) back view.
Figure 1. Subject equipped with the marker set used in the study: (a) front and (b) back view.
Sensors 21 07709 g001
Figure 2. Visual representation of the Physilog sensor’s reference frame (left) and representation of the locations of the sensors with their reference frame (right).
Figure 2. Visual representation of the Physilog sensor’s reference frame (left) and representation of the locations of the sensors with their reference frame (right).
Sensors 21 07709 g002
Figure 3. First landing phase of the Vertical Drop Jump in Visual3D, frontal (a) and lateral (b) views. The green spheres represent the markers on the anatomical landmarks whilst the red arrows are the vertical GRFs.
Figure 3. First landing phase of the Vertical Drop Jump in Visual3D, frontal (a) and lateral (b) views. The green spheres represent the markers on the anatomical landmarks whilst the red arrows are the vertical GRFs.
Sensors 21 07709 g003
Figure 4. Signals recorded by the IMU placed on the sacrum, displayed in the RKS. The signals are the 3D accelerations (top) and the 3D angular velocities (bottom).
Figure 4. Signals recorded by the IMU placed on the sacrum, displayed in the RKS. The signals are the 3D accelerations (top) and the 3D angular velocities (bottom).
Sensors 21 07709 g004
Figure 5. (a) Vertical position of the reference marker along the y-axis of the laboratory (vertical, positive upwards) and acceleration data along the z-axis of the master sensor, attached to the reference marker (vertical, positive downwards). The spikes used as synchronization trigger are visible at the very beginning of each signal (red square); (b) Aligned signals of reference marker and acceleration of the master by using the delay information coming from the cross-correlation calculation.
Figure 5. (a) Vertical position of the reference marker along the y-axis of the laboratory (vertical, positive upwards) and acceleration data along the z-axis of the master sensor, attached to the reference marker (vertical, positive downwards). The spikes used as synchronization trigger are visible at the very beginning of each signal (red square); (b) Aligned signals of reference marker and acceleration of the master by using the delay information coming from the cross-correlation calculation.
Sensors 21 07709 g005
Figure 6. Right knee joint moments (ac) and right GFRs (df) during the 112 VDJs. The number of frames that composes each VDJ is highly variable, depending on the jumping behavior of the subjects. Each color corresponds to the VDJs performed by a subject.
Figure 6. Right knee joint moments (ac) and right GFRs (df) during the 112 VDJs. The number of frames that composes each VDJ is highly variable, depending on the jumping behavior of the subjects. Each color corresponds to the VDJs performed by a subject.
Sensors 21 07709 g006
Figure 7. Visual structure of the dataset. The first column contains the reference index of the frame of each jump, columns 2–19 contain the data recorded by the three IMUs, placed on the sacrum (SA), right leg (RL), and right thigh (RT), whilst columns 20–25 contain the 3D right GRFs and the right knee moments. Numbers in parentheses indicate how many features (columns) belong to each source.
Figure 7. Visual structure of the dataset. The first column contains the reference index of the frame of each jump, columns 2–19 contain the data recorded by the three IMUs, placed on the sacrum (SA), right leg (RL), and right thigh (RT), whilst columns 20–25 contain the 3D right GRFs and the right knee moments. Numbers in parentheses indicate how many features (columns) belong to each source.
Sensors 21 07709 g007
Figure 8. Architectures of the NARX neural network. (a) Series–Parallel. (b) Parallel.
Figure 8. Architectures of the NARX neural network. (a) Series–Parallel. (b) Parallel.
Sensors 21 07709 g008
Figure 9. NARX Neural Network topology.
Figure 9. NARX Neural Network topology.
Sensors 21 07709 g009
Figure 10. Real and predicted kinetic curves of subject 8. The predicted variables are the the Flexion/Extension (FE) moment, the Adduction/Abduction (AA) moment, and the Rotation (R) moment at the right knee joint, and the 3D GRFs.
Figure 10. Real and predicted kinetic curves of subject 8. The predicted variables are the the Flexion/Extension (FE) moment, the Adduction/Abduction (AA) moment, and the Rotation (R) moment at the right knee joint, and the 3D GRFs.
Sensors 21 07709 g010
Figure 11. Error histograms of the whole test set. (a) Histogram of the errors on the prediction of the right Ground Reaction Forces; (b) Histogram of the errors on the prediction of the moments at the right knee joint.
Figure 11. Error histograms of the whole test set. (a) Histogram of the errors on the prediction of the right Ground Reaction Forces; (b) Histogram of the errors on the prediction of the moments at the right knee joint.
Sensors 21 07709 g011
Figure 12. Adduction/abduction moments. (a) Measured right knee AA moment; (b) Predicted right knee AA moment. Each curve corresponds to a jump performed by a subject and as it can be noticed, the prediction errors accumulated at the beginning of the curves, in the part enclosed by the dotted rectangle.
Figure 12. Adduction/abduction moments. (a) Measured right knee AA moment; (b) Predicted right knee AA moment. Each curve corresponds to a jump performed by a subject and as it can be noticed, the prediction errors accumulated at the beginning of the curves, in the part enclosed by the dotted rectangle.
Sensors 21 07709 g012
Table 1. Data and anthropometric characteristics of the participants.
Table 1. Data and anthropometric characteristics of the participants.
Subject IDAge (Years)GenderHeight (cm)Weight (kg)
129M18077.5
224M17880.0
324F16561.8
424F15050.8
526F17057.4
624M18160.1
723M17384.3
823M18065.9
924F16158.0
1024F16051.5
1125M17867.8
Mean24.5 170.565.0
SD1.7 10.311.3
Table 2. Conventions used to represent knee moments.
Table 2. Conventions used to represent knee moments.
AxisMovementPositiveNegativeLabel
xFlexion/ExtensionFlexionExtensionFE
yAdduction/AbductionVarusValgusAA
zRotationInternalExternalR
Table 3. Signals fed to the network (inputs or predictors) and outputs (or targets).
Table 3. Signals fed to the network (inputs or predictors) and outputs (or targets).
Predictor/Target #Real Data
Predictor 1Frame reference indexes
Predictor 2Angular velocity x (gyroscope, SA)
Predictor 3Angular velocity y (gyroscope, SA)
Predictor 4Angular velocity z (gyroscope, SA)
Predictor 5Acceleration x (accelerometer, SA)
Predictor 6Acceleration y (accelerometer, SA)
Predictor 7Acceleration z (accelerometer, SA)
Predictor 8Angular velocity x (gyroscope, RT)
Predictor 9Angular velocity y (gyroscope, RT)
Predictor 10Angular velocity z (gyroscope, RT)
Predictor 11Acceleration x (accelerometer, RT)
Predictor 12Acceleration y (accelerometer, RT)
Predictor 13Acceleration z (accelerometer, RT)
Predictor 14Angular velocity x (gyroscope, RL)
Predictor 15Angular velocity y (gyroscope, RL)
Predictor 16Angular velocity z (gyroscope, RL)
Predictor 17Acceleration x (accelerometer, RL)
Predictor 18Acceleration y (accelerometer, RL)
Predictor 19Acceleration z (accelerometer, RL)
Target 1Right Knee Flexion/Extension (FE) Moment
Target 2Right Knee Adduction/Abduction (AA) Moment
Target 3Right Knee Rotation (R) Moment
Target 4Ground Reaction Force x
Target 5Ground Reaction Force y
Target 6Ground Reaction Force z
Table 4. Mean, maximum, and minimum RMSE values corresponding to each target variable predicted by the network according to cut-off frequency of the Butterworth Low-pass filter applied to IMU data. The predicted variables are the 3D GRFs, the Flexion/Extension (FE) moment, the Adduction/Abduction (AA) moment, and the Rotation (R) moment at the right knee joint.
Table 4. Mean, maximum, and minimum RMSE values corresponding to each target variable predicted by the network according to cut-off frequency of the Butterworth Low-pass filter applied to IMU data. The predicted variables are the 3D GRFs, the Flexion/Extension (FE) moment, the Adduction/Abduction (AA) moment, and the Rotation (R) moment at the right knee joint.
RMSE
Ground Reaction Forces (N/kg)Right Knee Moments (N·m/kg)
Cut-Off FrequencyGRFxGRFyGRFzFEAAR
mmaxminmmaxminmmaxminmmaxminmmaxminmmaxmin
12 Hz0.02900.03810.0192 0.04330.0606 0.0232 0.04090.05870.02090.11590.2161 0.09820.03020.04650.0220 0.01140.01960.0090
24 Hz0.00870.01070.00310.03810.06440.01030.01880.03890.01400.08160.11710.06920.03340.05930.02920.01210.04980.0223
32 Hz0.00810.01020.00720.03500.04600.03380.01840.01980.01520.07480.07760.07250.02950.03480.02480.01170.01300.0106
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cerfoglio, S.; Galli, M.; Tarabini, M.; Bertozzi, F.; Sforza, C.; Zago, M. Machine Learning-Based Estimation of Ground Reaction Forces and Knee Joint Kinetics from Inertial Sensors While Performing a Vertical Drop Jump. Sensors 2021, 21, 7709. https://0-doi-org.brum.beds.ac.uk/10.3390/s21227709

AMA Style

Cerfoglio S, Galli M, Tarabini M, Bertozzi F, Sforza C, Zago M. Machine Learning-Based Estimation of Ground Reaction Forces and Knee Joint Kinetics from Inertial Sensors While Performing a Vertical Drop Jump. Sensors. 2021; 21(22):7709. https://0-doi-org.brum.beds.ac.uk/10.3390/s21227709

Chicago/Turabian Style

Cerfoglio, Serena, Manuela Galli, Marco Tarabini, Filippo Bertozzi, Chiarella Sforza, and Matteo Zago. 2021. "Machine Learning-Based Estimation of Ground Reaction Forces and Knee Joint Kinetics from Inertial Sensors While Performing a Vertical Drop Jump" Sensors 21, no. 22: 7709. https://0-doi-org.brum.beds.ac.uk/10.3390/s21227709

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop