Hostname: page-component-54dcc4c588-54gsr Total loading time: 0 Render date: 2025-09-27T08:08:35.383Z Has data issue: false hasContentIssue false

Using a wearable system combining inertial and force sensing for simultaneous detection of limb motion and grasping actions in the workplace

Published online by Cambridge University Press:  25 September 2025

Matteo Musso*
Affiliation:
Department of Materials and Production, https://ror.org/04m5j1k67 Aalborg University , Aalborg, Denmark
Shaoping Bai
Affiliation:
Department of Materials and Production, https://ror.org/04m5j1k67 Aalborg University , Aalborg, Denmark
Anderson Oliveira
Affiliation:
Department of Materials and Production, https://ror.org/04m5j1k67 Aalborg University , Aalborg, Denmark
*
Corresponding author: Matteo Musso; Email: mmus@mp.aau.dk

Abstract

Using wearable sensors to evaluate workers’ performance is challenging with existing sensor techniques. It requires detecting not only limb motions but also the onset and offset of specific actions. Commonly used inertial measurement units (IMUs) can be combined with surface electromyography (sEMG) to detect muscular activity. However, sEMG requires skin preparation and careful sensor placement, and can be affected by sweat or motion artifacts. To address these limitations, we used a wearable system combining IMUs and force-sensing resistors (FSRs), where IMUs capture joint kinematics and FSRs detect grasping actions. The system included three IMUs (on the trunk, upper arm, and forearm) and two FSR arrays (on the upper and lower arms). The system was first validated in a laboratory setting against an optical motion capture system with 10 healthy young adults performing isolated upper limb movements and mimicking lifting tasks. The results showed high agreement in joint angle estimation (coefficient of multiple correlation = 0.95 $ \pm $ 0.04), with a maximum root mean square error of 8.7 $ \pm $ 2.92°, and a mean absolute timing error for grasp detection of −0.59 seconds. To evaluate its applicability in real-world scenarios, a pilot in-field test was then conducted with two manufacturing workers (using and not using a passive shoulder exoskeleton) during a repetitive panel-packing task. The test shows highly consistent grasping detection, which allowed segmenting the task with a small variability in task duration (maximum coefficient of variation = 5.16$ \% $). These findings demonstrate the feasibility of using the proposed method in industrial environments to analyze upper limb motion and grasping activity.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press

1. Introduction

There has been a growing use of wearable sensors to evaluate workers’ performance and maintain or improve their health, safety, and productivity (Patel et al., Reference Patel, Chesmore, Legner and Pandey2021). In particular, inertial measurement units (IMUs) have gained popularity among researchers and industry entities seeking to track human motion in real-world working environments (Digo et al., Reference Digo, Pastorelli and Gastaldi2022). IMU-based measurements offer a more practical solution for ergonomic assessments in the field compared to gold-standard techniques such as optical motion capture. IMUs require a lower number of sensors to determine body segment motion and do not require large spaces for recordings. Moreover, optical motion capture systems experience marker occlusion, which can compromise motion tracking by causing gaps in segment trajectories. Although IMUs offer several advantages, the technology also presents limitations related to magnetic interferences, which may be amplified in ergonomic settings, as well as the high susceptibility for data drift in long recordings (Digo et al., Reference Digo, Pastorelli and Gastaldi2022). Nevertheless, IMUs have been widely used to evaluate human motion under real-world conditions (Digo et al., Reference Digo, Pastorelli and Gastaldi2022).

IMU systems have been used for in-field risk assessments, collaborative robotics, motion tracking in industries, action recognition (Digo et al., Reference Digo, Pastorelli and Gastaldi2022), and assessing the influence of exoskeletons on human performance (Amandels et al., Reference Amandels, het Eyndt, Daenen and Hermans2019; Seiferheld et al., Reference Seiferheld, Frost, Krog and Samani2022). Because IMUs generate only kinematic data, researchers often combine them with physiological or mechanical measures such as surface electromyography (sEMG), spiroergometry systems, and questionnaires to enhance the assessment of physical demands. sEMG, in particular, is used to assess the muscle electrical activity during tasks. sEMG has been used for ergonomic risk score evaluation (Peppoloni et al., Reference Peppoloni, Filippeschi, Ruffaldi and Avizzano2016; Giannini et al., Reference Giannini, Bassani, Avizzano and Filippeschi2020; Bangaru et al., Reference Bangaru, Wang and Aghazadeh2022; Hubaut et al., Reference Hubaut, Guichard, Greenfield and Blandeau2022), human–robot collaboration (Wang et al., Reference Wang, Li, Diekel, Chen, Zhang and Jia2019; Chico et al., Reference Chico, Cruz, Vásconez, Benalcázar, Álvarez, Barona and Valdivieso2021; Lin and Peng, Reference Lin and Peng2022), evaluation of aid provided by new technologies (Amandels et al., Reference Amandels, het Eyndt, Daenen and Hermans2019; Seiferheld et al., Reference Seiferheld, Frost, Krog and Samani2022), and activity recognition (Tao et al., Reference Tao, Lai, Leu and Yin2018; Al-Amin et al., Reference Al-Amin, Tao, Doell, Lingard, Yin, Leu and Qin2019; Kubota et al., Reference Kubota, Iqbal, Shah and Riek2019; Bangaru et al., Reference Bangaru, Wang, Busam and Aghazadeh2021). In these studies, sEMG sensors were typically placed around the forearm to detect grasping activities during upper limb movements. However, conventional sEMG measurements require adequate skin preparation, and data quality might be affected by sweat and movement artifacts. Moreover, sEMG sensor placement plays a crucial role in data quality, as signals from adjacent muscles (i.e., crosstalk) can contaminate recordings from the target muscle. Additionally, electrodes may shift due to skin movement during dynamic contractions, leading to erroneous interpretation of the sEMG data (Oliveira and Negro, Reference Oliveira and Negro2021). As a result, assessing sEMG during real-world ergonomic tasks is highly challenging and often impractical, considering the need for lengthy setup and potentially inaccurate data.

Force-sensing resistors (FSRs) have been explored as an alternative method for estimating muscle activation to address the limitations imposed by surface sEMG. Esposito et al. (Reference Esposito, Andreozzi, Fratini, Gargiulo, Savino, Niola and Bifulco2018) reported a strong association between FSR and sEMG signals captured during flexor carpi ulnaris muscle activation during various grasping actions. Additional studies have successfully used FSR bands placed on the forearm to distinguish arm and forearm position and gestures (Xiao and Menon, Reference Xiao and Menon2014; Xiao and Menon, Reference Xiao and Menon2017a; Islam and Bai, Reference Islam and Bai2020; Rehman et al., Reference Rehman, Shah, Haq and Khurshid2022), as well as to accurately detect grip and release actions (Wininger et al., Reference Wininger, Kim and Craelius2008; Xiao and Menon, Reference Xiao and Menon2017b). In industrial contexts, FSRs have been used in shoe insoles to function as weighing sensors, to detect object picking and drive a back-support exoskeleton accordingly (Mateos, Reference Mateos2017) and monitor the weight of objects carried by construction workers (Lee and Son, Reference Lee and Son2021). However, FSRs have not yet been used to sense or analyze forearm muscle activation during tasks performed in real work environments.

In this study, we use a wearable sensor system (WSS) that combines IMU and FSR sensors to track upper limb motion in real-world ergonomic settings. The IMU sensors provide kinematic data, while the FSR sensors detect muscle activation in the arm and forearm to identify the onset and offset of movement cycles. The WSS allows simultaneous tracking of upper limb motion and hand actions. We first conducted a controlled laboratory validation to assess the accuracy of the WSS in estimating joint angles and detecting grasping onset and offset. Subsequently, a preliminary in-field test was conducted with two factory workers to examine their motion patterns using the WSS and explore potential differences in movement with and without the use of an exoskeleton.

2. IMU/FSR WSS

2.1. Sensors

The WSS consists of two instrumented armbands (Biox ApS, Aalborg, Denmark) and one trunk-mounted instrumented band. Each of the two armbands and the trunk band includes one IMU (BNO055, Robert Bosch GmbH, Gerlingen, Germany), which contains a gyroscope ( $ \pm $ 2,000°/s), an accelerometer (up to $ \pm $ 16 G), and a magnetometer ( $ \pm $ 1.3 mT x and y axes, $ \pm $ 2.5 mT z axis). Additionally, each armband contains eight FSRs distributed equidistantly around the band, with a sensing range of 0.2–20 N. Both the armbands and the trunk band are equipped with an ESP32 chip, which acquires data from the IMU and FSR sensors and transmits it wirelessly to a tablet (Tab M10 HD, Lenovo, Beijing, China). An Android application has been developed to establish Bluetooth communication between the ESP32 and the tablet. The application enables real-time acquisition of IMU and FSR data, initial joint angle computation, and data storage on a cloud server at a frequency of 18 Hz. A schematic of the system is shown in Figure 1.

Figure 1. Illustration of the components used to create the WSS. Two BioX armbands containing IMUs and FSRs were used to acquire upper limb movement and muscle activation. The data are transmitted via Bluetooth to a tablet, which saves it to a cloud service and computes real-time joint angles.

In this study, the armbands were positioned on the anterior sides of the forearm and upper arm, firmly fixed where these segments have the greatest cross-sectional area. We ensured that once placed, the bands did not slip on the arm and did not cause discomfort to the user, but the initial armband tension was not standardized among participants. For the trunk, the IMU was placed between the shoulder blades using a customized, adjustable tank top (Figure 2), with the sensor securely attached.

Figure 2. IMU sensor placement and orientation of the sensors’ internal reference systems.

2.2. Estimation of shoulder and elbow joint angles

The internal reference coordinate system of each IMU is defined as follows: the x-axis points upward, the y-axis is aligned with the human frontal axis, and the z-axis is parallel to the sagittal axis. Accordingly, rotations around the y-axis correspond to flexion/extension in the sagittal plane, rotations around the x-axis correspond to abduction/adduction, and rotations around the z-axis correspond to internal/external rotation. The quaternions representing the orientation of each IMU are extracted and used to estimate shoulder and elbow angles using custom scripts (MATLAB R2023a, The MathWorks, Inc., Natick).

To calculate shoulder angles, the quaternion representing the orientation of the back sensor ( $ {q}_b $ ) is conjugated and multiplied by the quaternion of the upper arm sensor ( $ {q}_u $ ), to compute the shoulder quaternion ( $ {q}_{sh} $ ):

.(2.1) $$ {q}_{sh}={q}_b^{-1}\cdot {q}_u $$

The shoulder angles are then obtained by converting ( $ {q}_{sh} $ ) into Euler angles using the “YZX” rotation sequence. The angle around the y-axis corresponds to flexion/extension (and is negated), while the angle around the z-axis corresponds to abduction/adduction.

To estimate elbow angles, the quaternion from the upper arm sensor ( $ {q}_u $ ) is conjugated and multiplied by the forearm sensor quaternion ( $ {q}_f $ ) to obtain the elbow quaternion ( $ {q}_{el} $ ):

.(2.2) $$ {q}_{el}={q}_u^{-1}\cdot {q}_f $$

The resulting quaternion $ {q}_{el} $ is then converted to joint angles using the “YZX” rotation sequence. No negation of the y-axis angle is needed as both sensors’ axes are aligned with the same orientation.

3. Laboratory validation test

The WSS was tested on 10 healthy young volunteers (nine males, one female; 28.2 $ \pm $ 3 years; height 179.15 $ \pm $ 7 cm; body mass 78.9 $ \pm $ 14 kg). The exclusion criteria included any impediments to arm motion and/or pain in the upper limb joints. All participants received written and verbal instructions about the experiment and provided written informed consent. The experimental procedures were approved by the local ethics committee (case number 2024-505-00159).

3.1. Experimental design

Participants were initially equipped with two armbands on the right arm and the trunk-mounted tank top as described in Section 2.2. A tablet was used to verify the initial alignment of the IMU sensors and ensure data quality. Retroreflective markers were subsequently attached for simultaneous optical motion capture using a 12-camera system (Qualisys AB, Göteborg, Sweden; 100 Hz sampling frequency). The Plug-in Gait Full-Body model (Vicon, 2024) was used to define the trunk, right forearm, and upper arm segments based on 11 retroreflective markers. Subsequently, a musculoskeletal model was created using the marker trajectory data in Visual3D (HAS-Motion, Kingston, ON, Canada) to replicate the trunk and right forearm, upper arm segments (Has-Motion, 2024). Before experimental trials, a single 5-second recording of each participant in neutral reference posture (N-pose) was performed to scale the Visual3D model to the participants’ size.

Participants performed the following upper limb movements:

T1 Shoulder flexion (from 0 to 90°) in the sagittal plane.

T2 Shoulder flexion (from 0 to 90°) at a 45° angle between the sagittal and frontal planes.

T3 Shoulder abduction (from 0 to 90°).

T4 Elbow flexion (from 0 to 90°) in the sagittal plane.

T5 Combined shoulder and elbow flexion/extension in the sagittal plane. Participants were asked to first flex their elbows from 0 to 90°, then performed a shoulder flexion from 0 to 90° while maintaining the 90° elbow angle. Subsequently, the participant extended the elbow to 0° and then the shoulder back to 0°.

T6 Shoulder flexion (from 0 to 160°) in the sagittal plane, performed with the torso flexed approximately 40° forward.

T7 Simulated work task: Participants simulated picking up an object from the ground, walking 6 m, placing it down, and standing upright for 2 seconds. The participants did not have an actual object. After 2 seconds, the task was repeated. During walking, participants were instructed to slightly face a side wall, as if carrying the object with another person. This sequence was repeated 10 times.

Participants performed all tasks at a self-selected speed. Tasks T1–T6 were repeated in four sets of eight repetitions. After each set, participants were asked to move and rotate around the vertical axis to change the sensor orientation, allowing us to evaluate the versatility of the motion-tracking system in estimating joint angles. Figure 3 shows sample shoulder flexion/extension angles derived from both the optical motion capture and the IMU system.

Figure 3. Computed shoulder flexion/extension angles from the two methods and segmentation results.

3.2. Data processing and analysis

3.2.1. IMU data

All data were processed using custom-made MATLAB scripts (The MathWorks, Inc., Natick). The tridimensional optical marker position data were labeled, gap-filled (Qualisys Track Manager 2023.1, Qualisys, Göteborg, Sweden), and exported as C3D files. The subjects’ Visual3D models were created from N-pose C3D files. Visual3D models were created from the N-pose recordings and then applied to movement tasks to extract shoulder and elbow joint angles.

IMU-derived joint angles were upsampled to 100 Hz using MATLAB’s resample function (default linear method) to match the sampling frequency of the optical motion capture data and synchronized using cross-correlation. Shoulder and elbow joint angles from the optical motion capture system were segmented by identifying the minimum point immediately before or after the maximum flexion angle of each task repetition. These segmentation points were used to define movement cycles in the joint angle data of both the optical motion capture system and the WSS (Figure 3).

The WSS’s accuracy in tracking upper limb kinematics was evaluated using the coefficient of multiple correlation (CMC) (Ferrari et al., Reference Ferrari, Cutti and Cappello2010), where values between 0.75 and 0.84 indicate good, 0.85 and 0.94 very good, and 0.95 and 1 excellent agreement (Ferrari et al., Reference Ferrari, Cutti and Cappello2010; Zhang et al., Reference Zhang, Novak, Brouwer and Li2013). In addition, the absolute (RMSE) and relative root mean square error (RRMSE) were also calculated (Ren et al., Reference Ren, Jones and Howard2008). CMC, RMSE, and RRMSE were computed for each movement cycle and then averaged per participant and task. The 95% confidence interval and standard deviation were calculated. Finally, the agreement between the two measurement systems was assessed using Bland–Altman plots (Bland and Altman, Reference Bland and Altman1986), a commonly used method for comparing the performance of a new technology against that of the gold standard one (Mihcin, Reference Mihcin2019, Reference Mihcin2022). The maximum and minimum joint angles during each task repetition, obtained from both methods, were compared.

3.2.2. FSR data

To evaluate the use of the FSRs for identifying grasping and releasing events, the simulated work task recordings from the optical system were segmented by identifying the minima in the vertical position of the right wrist marker A (RWRA; Figure 4). These minima corresponded to the box grasping and releasing events and were therefore used to define complete task cycles. The eight FSR signals from the forearm band were averaged and segmented based on the minima occurring before and after signal bursts. We used the averaged signal due to the simplicity of the target movement. The grasping and lifting movement involved all the fingers as well as the wrist, and therefore strongly engaged the forearm muscles. For more complex movement, a classification algorithm or spatial analysis would probably be required. These bursts were detectable due to increased activation of wrist and finger flexor muscles when the box was picked up and moved during the task. The minima delimiting each burst were considered the start and end points of the grasping actions. We selected the minima as they indicate no pressure on the hand and hand open, corresponding to the minimum recruitment of the forearm muscles. Using the burst peak values to identify grasping would likely be more indicative of the lifting phase, when the full weight of the box is supported by the hand, rather than the grasping action itself. The timing of wrist marker minima was then compared with the corresponding FSR minima to assess the accuracy of grasp/release event detection. The average absolute and relative errors across all 10 participants were reported separately for grasping and releasing actions.

Figure 4. Identified minima (red circles) in the average FSR signal (a) and RWRA marker trajectory (b) during the simulated work task.

3.3. Validation results

3.3.1. Limb motion tracking

The joint angles estimated using the WSS showed excellent agreement with those obtained from the gold standard, as indicated by high CMC values for both the shoulder (Figure 5a) and elbow joints (Figure 5d) across all isolated tasks. RMSE values remained below 10° across all tasks and joints. For the shoulder, the highest RMSE and intersubject variability were observed in tasks T2 and T6 (Figure 5b). Similarly, the highest RMSE for the elbow was observed during task T5 (Figure 5e). RRMSE ranged from 7 to 25% for the shoulder (Figure 5c), with consistent patterns across tasks. The average RRMSE exceeded 16% in two tasks: T3 (16.9 $ \pm $ 5.98 $ \% $ ) and T7 (16.3 $ \pm $ 3.87 $ \% $ ). For the elbow (Figure 5f), the RRMSE was notably higher during T7. A summary table of the CMC, RMSE, and RRMSE values is provided in Supplementary Table S1. Figure 6 and Supplementary Figure S1 present the correlation and the Bland–Altman plots for shoulder and elbow angles. The Bland–Altman plots also report the mean difference and limits of agreement. The largest mean difference was observed during task T2 on the shoulder, with an absolute value of 3.9°. The smallest was 0.33° during task T7 on the elbow.

Figure 5. Violin plots of the coefficient of multiple correlation (CMC; a and d), root mean square error (RMSE; b and e), and relative root mean square error (RRMSE; c and f) computed from the comparison between optical motion capture and IMU-derived joint angle calculations for the shoulder (upper row) and elbow (bottom row). Each colored dot in a violin represents the average value for one participant, the white dot represents the median across participants, the gray vertical bar indicates the first and third quartiles, and the shaded colored area depicts the density curve. The tasks are described in Section 3.1.

Figure 6. Correlation and Bland–Altman plots for three selected tasks: T1, T4, and T7.

3.3.2. Grasping action recognition

For the grasping events, the average absolute and relative errors between the timings of the identified minima in the FSR average signals and the RWRA trajectory were −0.59 seconds and −4.39%, respectively. For releasing events, the values were 0.54 seconds and 1.49%.

4. In-field test

Following laboratory validation of the WSS, in-field testing was conducted to demonstrate the feasibility of the new technology. The field tests were conducted at a company producing kitchen furniture. Two females (32 $ \pm $ 12 years, height 171.5 $ \pm $ 2 cm, body mass 63 $ \pm $ 1 kg) participated in the field test. We evaluated the WSS’s sensitivity in detecting changes in motion patterns when workers performed their tasks with and without an exoskeleton.

The workers were responsible for packing kitchen cabinet parts. Worker 1 was responsible for packing four 2.4 kg panels together, while worker 2 was responsible for packing two 4.67 kg panels (Figure 7). Both picked panels from a stack behind them and placed them in pairs into a box, with side tasks such as placing the box on a conveyor belt and adding a thinner panel. The study was conducted in conjunction with a pilot trial to evaluate the feasibility of using a specific passive exoskeleton (Skelex 360-XFR, Rotterdam, the Netherlands, weight: 2.3 kg, support between 0.5 kg and 4 kg/arm) to assist the workers. This test followed a laboratory test evaluating support provided by the device in a similar manual handling activity (Musso et al., Reference Musso, Oliveira and Bai2024). Therefore, we assessed the motion patterns of the two workers during the packing task with (“Exo” condition) and without (“Free” condition) the use of the passive exoskeleton, using the WSS.

Figure 7. Illustrative image of a worker performing the task of moving a panel for packing. The workers performed the packing task with and without using a passive exoskeleton (Skelex 360-XFR). The task involved shoulder abduction and flexion, combined with elbow flexion. WSS was used to identify the grasping moments for subsequent segmentation of joint angles.

We used the WSS to track the workers’ movements and identify the start and end of each packing cycle. For worker 2, a full cycle was defined as the part of the motion between two consecutive panel packings. For worker 1, it was the interval between packing the thinner panels, which she placed after the main four panels as the final step of her work cycle. Within each cycle, the key event of interest was the panel-packing action, which began when a pair of panels was grabbed from the stack and ended when they were placed on the conveyor belt. This motion involved sequential shoulder and elbow flexion and extension. The signals from the forearm FSR were used to detect grasp and release moments, which were detectable due to the increase in activation of the wrist and finger flexor muscles (see Figure 8 for illustration). Only cycles without missing data points were included in the analysis.

Figure 8. Averaged force-sensing resistor (FSR) signals from the two workers and the identified segments. Dashed lines indicate the start and end of the main events; solid lines indicate the limits of full task cycles. For worker 2, these two sets of lines coincide.

The start and end times were used to segment joint angle and angular velocity time series. The angular velocity of the arm IMUs around their internal y-axes was interpreted as the flexion velocity of the proximal joint. Worker 1 performed 32 movement cycles (16 for the Free condition and 16 for the Exo condition), while worker 2 performed 53 movement cycles (21 for the Free condition and 32 for the Exo condition). The cycles under the same conditions were not all recorded together but in up to 3 subsequent sessions. Subsequently, data from each cycle were time-normalized to 80 points, representing 0–100% of the cycle. Subsequently, a median representative curve was generated from the multiple cycles for each worker. A median curve was generated for each worker, and cycles with a correlation coefficient <0.6 with the median were excluded to remove outliers and non-standard movements, such as discarding a defective panel.

For each cycle, the main event duration, total cycle duration, shoulder and elbow range of motion (ROM), and both average and peak angular velocities were calculated for the Free and Exo conditions. The coefficient of variation (CV = $ std/ mean\cdot 100 $ ) was also calculated for some parameters to evaluate segmentation consistency.

Due to the small sample size, kinematic results are reported separately for the two workers (Table 1). For worker 1, 12 cycles were included under the Free condition and 14 under the Exo condition. For worker 2, 19 Free and 28 Exo cycles were analyzed.

Table 1. Summary of the movement cycle parameters for the two workers under the two conditions, along with the percentage differences between the “Free” and “Exo” conditions

Note: Results for each condition are reported as the average value $ \pm $ standard deviation. Range of motion results are presented as the average maximum and minimum shoulder flexion angles.

The main event and cycle durations were longer for worker 1 when using the exoskeleton. Additionally, both shoulder and elbow ROM decreased by at least 19% with the exoskeleton. Worker 1’s average shoulder flexion velocity decreased by −15%, while the average elbow velocity increased by −17% when using the exoskeleton. In contrast, worker 2 showed minimal differences (<3%) in average angular velocities between conditions. Regarding peak velocities, worker 1 showed decreased shoulder and increased elbow flexion velocities (−17% and +21%), while worker 2 had a 5% increase in peak shoulder flexion velocity and an 18% reduction in peak elbow flexion velocity.

5. Discussion

This study demonstrated that a WSS based on IMUs and FSRs provides accurate measurements of upper limb kinematics and detection of grasping onsets and offsets. Altogether, our findings support the feasibility of using the IMU/FSR WSS, highlighting the potential of such systems for in-field monitoring of movement patterns aimed at reducing the risk of work-related musculoskeletal disorders.

We observed excellent agreement between the joint angle time series from the WSS and the gold-standard optical system, with CMC values exceeding 0.95 across all tasks and joints. The average RMSE across all tasks was approximately 6°, with the highest RMSE being 8.7 $ \pm $ 2.92° for task T6. These results are consistent with those reported for other IMU systems, as reviewed in Fang et al. (Reference Fang, Woodford, Senanayake and Ackland2023). Previous studies have reported RMSE values between 5° and 10° during various movements (Bouvier et al., Reference Bouvier, Duprey, Claudon, Dumas and Savescu2015; Serra-Hsu and Taboga, Reference Serra-Hsu and Taboga2022; Wu et al., Reference Wu, Tao, Chen, Tian and Sun2022; Zhu et al., Reference Zhu, Li, Wang, Chen, Shi, Zheng and Li2022). RMSE values exceeding 10° have also been reported (Pérez et al., Reference Pérez, Costa, Torrent, Solana, Opisso, Cáceres, Tormos, Medina and Gómez2010; Bouvier et al., Reference Bouvier, Duprey, Claudon, Dumas and Savescu2015; Bessone et al., Reference Bessone, Höschele, Schwirtz and Seiberl2022; Henschke et al., Reference Henschke, Kaplick, Wochatz and Engel2022) when validating IMU-based systems for joint angle estimation during isolated shoulder flexion/extension or abduction movements. In more complex movements such as box lifting (Robert-Lachaine et al., Reference Robert-Lachaine, Mecheri, Larue and Plamondon2017; Robert-Lachaine et al., Reference Robert-Lachaine, Mecheri, Muller, Larue and Plamondon2020; Humadi et al., Reference Humadi, Nazarahari, Ahmad and Rouhani2021), shoulder flexion RMSEs up to 35.8° have been reported (Robert-Lachaine et al., Reference Robert-Lachaine, Mecheri, Larue and Plamondon2017), whereas we observed substantially lower errors in similar conditions (RMSE 7°). For elbow flexion, our RMSE results align with previous studies, which reported approximately 6° RMSE (El-Gohary and McNames, Reference El-Gohary and McNames2012; Pérez et al., Reference Pérez, Costa, Torrent, Solana, Opisso, Cáceres, Tormos, Medina and Gómez2010; Z. Q. Zhang et al., Reference Zhang, Wong and Wu2011). Our system shows smaller RMSEs than those reported in (Bouvier et al., Reference Bouvier, Duprey, Claudon, Dumas and Savescu2015; Robert-Lachaine et al., Reference Robert-Lachaine, Mecheri, Muller, Larue and Plamondon2020; Bessone et al., Reference Bessone, Höschele, Schwirtz and Seiberl2022) for similar movements. Therefore, the proposed WSS demonstrates comparable or superior joint angle estimation performance relative to previously reported systems for upper limb kinematic assessment.

These results were achieved without using a sensor-to-segment calibration step to establish a relationship between the IMU and the anatomical reference system. We ensured that the user maintained the arms along the body and hand in a neutral position while checking the sensor alignment in the tablet, which displayed shoulder and elbow angles close to 0°. A similar approach was used by Chan et al. (Reference Chan, Chua, Chou, Seah, Huang, Luo, Dacy and Razak2022), who achieved RMSE values comparable to ours. Nevertheless, more accurate results might be achievable using sensor-to-segment calibration, as shown by studies that reported RMSEs below 1° during shoulder movements using MT9B (Cutti et al., Reference Cutti, Giovanardi, Rocchi, Davalli and Sacchetti2008) and MARG sensors (Madrigal et al., Reference Madrigal, Cardiel, Rogeli, Salas and Guerrero2016).

Another factor that may have affected our results is the soft tissue artifact. It refers to the different movements of the markers relative to the underlying bones due to the presence of soft tissues in between (Mihcin, Reference Mihcin2019).

The Bland–Altman plots show good agreement between systems. Table 2 shows the percentage of points in every Bland–Altman within the upper and lower limits of agreement (mean $ \pm $ 1.96 SD of the difference). According to Bland and Altman (Reference Bland and Altman1986), 95% of the data points should fall within these limits. This threshold was reached for tasks 1 and 2; for the remaining tasks, percentages ranged from 91.3 to 94.7%.

Table 2. Percentage of the data in the Bland–Altman plots falling within the upper and lower limits of agreement (mean $ \pm $ 1.96 SD of difference)

Previous studies have used full-body IMU systems to evaluate ergonomic tasks (Skals et al., Reference Skals, Bláfoss, Andersen, de Zee and Andersen2021; Seiferheld et al., Reference Seiferheld, Frost, Krog and Samani2022), but this technology is often costly and requires the handling of numerous sensors. Therefore, further development of simplified WSS approaches is needed to facilitate motion tracking in real-world ergonomic applications.

The FSR component of the WSS successfully detected grasp and release events during panel handling, with both workers exhibiting consistent joint angles and low inter-trial variability throughout the task cycle (see Figure 9). Considering the full cycle duration, the highest CV was 5.16 $ \% $ for worker 2 under the Exo condition, while the lowest was 2.96 $ \% $ for worker 1 under the Free condition. The consistency of shoulder flexion kinematics across repetitions confirms the potential of the WSS to monitor task execution and identify specific task-related events. The use of an exoskeleton slightly changed the duration of the task ( $ \Delta \% $ < 10% for both workers). The differences in ROM might have been influenced more by the change in panel stack height during task performance than by the use of the exoskeleton.

Figure 9. Worker 2 shoulder and flexion angles during full cycles.

The WSS proved easy to handle. Fitting and setting up the sensor bands required only a few minutes per worker and did not interfere with their regular work pace. Moreover, the workers did not report any discomfort when using the WSS throughout their working routines. Similarly, they experienced no issues using the wearable sensor system in combination with assistive tools, in particular, the upper limb exoskeleton in this study. The exoskeleton cuff was placed around the upper arm band, with ample space to fix both devices. This suggests that the WSS can be used for in-field testing of exoskeleton technologies. The limited sample size and the specific task performed in this study limit the generalization of the results obtained. Moreover, our in-field testing was performed only by females, further limiting the generalization of the presented results. Sensor placement in males and females may slightly differ due to anthropometric characteristics, especially around the torso area. Therefore, future studies comprising a larger sample size and the inclusion of a balanced gender distribution will be crucial to establish the system’s applicability and fully validate it for movement analysis in the industrial field.

In addition to the scope of this study, the WSS could be applied to risk assessments Tahir et al., Reference Tahir, Bai, Shen, Tahir, Bai and Shen2023) using tools like the Rapid Upper Limb Assessment (McAtamney and Corlett, Reference McAtamney and Corlett1993), which evaluates joint postures and load handling for ergonomic risk scoring. Another potential application is in human–robot interaction, such as adjusting the level of an exoskeleton based on shoulder angle and object manipulation. This approach is similar to that described in Mateos (Reference Mateos2017), but it also includes information on joint angles. Nonetheless, applications beyond those tested here, such as fast-paced movements in sports or tasks requiring more precise estimation of joint angles or grasping timing, will require further validation as the average errors obtained (RMSE = 8.7° and $ \Delta $ t = 0.5 s) may not be sufficiently small. Field testing also revealed occasional data transmission issues from the WSS to the tablet, likely due to Bluetooth interference. Future updates to the Bluetooth module could address this limitation.

6. Conclusions

In this work, we developed a wearable motion tracking system that combines inertial and force sensing to simultaneously estimate upper limb motion and detect grasping events. Laboratory-based validation demonstrated strong agreement and satisfactory accuracy for shoulder and elbow kinematics. Moreover, the system identified grasping and releasing actions with a consistent timing error. Finally, the in-field test illustrated the feasibility of using the motion tracking system under real-world conditions by comparing workers’ performance with and without exoskeletons. Future studies should focus on improving joint angle estimation through sensor-to-segment calibration and expanding the system’s application in the human–robot interaction and in-field risk assessment. Evaluating the system in different industrial settings and with more subjects will also be necessary to fully validate it for use in the industrial field.

Supplementary material

The supplementary material for this article can be found at http://doi.org/10.1017/wtc.2025.10026.

Data availability statement

Data can be made available to interested researchers upon request by email to the corresponding author.

Acknowledgments

The authors acknowledge support from Dorthe Lind from Kvik A/S during the in-field pilot tests.

Authorship contribution

Conceptualization, M.M., A.S.O., and S.B.; Methodology, M.M., A.S.O., and S.B.; Data curation, M.M.; Data visualisation, M.M; Writing-original draft, M.M.; Writing-review and editing, M.M., A.S.O., and S.B. All authors approved the final submitted draft.

Funding statement

This research project was funded as part of the Interreg North Sea Region project EXSKALLERATE (www.northsearegion.eu/exskallerate) funded by the EU Interreg Council.

Competing interests

S.B. is a co-founder of BioX ApS, a start-up company to commercialize wearable technologies.

Ethical standard

The research meets all ethical guidelines, including adherence to the legal requirements of the study country. Both of the studies were approved by the local ethics committee. The laboratory test case number is 2024-505-00159 and the field test case number is 2023-505-00142.

References

Al-Amin, M, Tao, W, Doell, D, Lingard, R, Yin, Z, Leu, MC and Qin, R (2019) Action recognition in manufacturing assembly using multimodal sensor fusion. Procedia Manufacturing 39, 158167. https://doi.org/10.1016/J.PROMFG.2020.01.288.CrossRefGoogle Scholar
Amandels, S, het Eyndt, HO, Daenen, L and Hermans, V (2019) Introduction and testing of a passive exoskeleton in an industrial working environment. Advanced in Intelligent Systems and Computing 820, 387392. https://doi.org/10.1007/978-3-319-96083-8_51.CrossRefGoogle Scholar
Bangaru, SS, Wang, C and Aghazadeh, F (2022) Automated and continuous fatigue monitoring in construction workers using forearm EMG and IMU wearable sensors and recurrent neural network. Sensors 22, 9729. https://doi.org/10.3390/S22249729.CrossRefGoogle ScholarPubMed
Bangaru, SS, Wang, C, Busam, A and Aghazadeh, F (2021) Ann-based automated scaffold builder activity recognition through wearable EMG and IMU sensors. Automation in Construction 126, 9265805. https://doi.org/10.1016/j.autcon.2021.103653.CrossRefGoogle Scholar
Bessone, V, Höschele, N, Schwirtz, A and Seiberl, W (2022) Validation of a new inertial measurement unit system based on different dynamic movements for future in-field applications. Sports Biomechanics 21, 685700. https://doi.org/10.1080/14763141.2019.1671486.CrossRefGoogle ScholarPubMed
Bland, JM and Altman, DG (1986) Statistical methods for assessing agreement between two methods of clinical measurement. The Lancet 327, 307310. https://doi.org/10.1016/S0140-6736(86)90837-8.CrossRefGoogle Scholar
Bouvier, B, Duprey, S, Claudon, L, Dumas, R and Savescu, A (2015) Upper limb kinematics using inertial and magnetic sensors: Comparison of sensor-to-segment calibrations. Sensors 15, 1881318833. https://doi.org/10.3390/S150818813.CrossRefGoogle ScholarPubMed
Chan, LYT, Chua, CS, Chou, SM, Seah, RYB, Huang, Y, Luo, Y, Dacy, L and Razak, HRBA (2022) Assessment of shoulder range of motion using a commercially available wearable sensor—A validation study. mHealth 8. https://doi.org/10.21037/MHEALTH-22-7.CrossRefGoogle ScholarPubMed
Chico, A, Cruz, PJ, Vásconez, JP, Benalcázar, ME, Álvarez, R, Barona, L and Valdivieso, ÁL (2021). Hand gesture recognition and tracking control for a virtual UR5 robot manipulator. ETCM 2021 - 5th Ecuador Technical Chapters Meeting. https://doi.org/10.1109/ETCM53643.2021.9590677.CrossRefGoogle Scholar
Cutti, AG, Giovanardi, A, Rocchi, L, Davalli, A and Sacchetti, R (2008) Ambulatory measurement of shoulder and elbow kinematics through inertial and magnetic sensors. Medical and Biological Engineering and Computing 46, 169178. https://doi.org/10.1007/S11517-007-0296-5.CrossRefGoogle ScholarPubMed
Digo, E, Pastorelli, S and Gastaldi, L (2022) A narrative review on wearable inertial sensors for human motion tracking in industrial scenarios. Robotics 11, 138. https://doi.org/10.3390/ROBOTICS11060138.CrossRefGoogle Scholar
El-Gohary, M and McNames, J (2012) Shoulder and elbow joint angle tracking with inertial sensors. IEEE Transactions on Biomedical Engineering 59, 26352641. https://doi.org/10.1109/TBME.2012.2208750.CrossRefGoogle ScholarPubMed
Esposito, D, Andreozzi, E, Fratini, A, Gargiulo, GD, Savino, S, Niola, V and Bifulco, P (2018) A piezoresistive sensor to measure muscle contraction and mechanomyography. Sensors 18, 2553. https://doi.org/10.3390/S18082553.CrossRefGoogle ScholarPubMed
Fang, Z, Woodford, S, Senanayake, D and Ackland, D (2023) Conversion of upper-limb inertial measurement unit data to joint angles: A systematic review. Sensors 23, 6535. https://doi.org/10.3390/S23146535.CrossRefGoogle ScholarPubMed
Ferrari, A, Cutti, AG and Cappello, A (2010) A new formulation of the coefficient of multiple correlation to assess the similarity of waveforms measured synchronously by different motion analysis protocols. Gait & Posture 31, 540542. https://doi.org/10.1016/J.GAITPOST.2010.02.009.CrossRefGoogle ScholarPubMed
Giannini, P, Bassani, G, Avizzano, CA and Filippeschi, A (2020) Wearable sensor network for biomechanical overload assessment in manual material handling. Sensors 20, 3877. https://doi.org/10.3390/S20143877.CrossRefGoogle Scholar
Has-Motion. (2024). Tutorial: Plug-in Gait Full-Body - Software Product Documentation. Available at https://wiki.has-motion.com/index.php?title=Tutorial:_Plug-In_Gait_Full-Body (accessed 21 May 2024).Google Scholar
Henschke, J, Kaplick, H, Wochatz, M and Engel, T (2022) Assessing the validity of inertial measurement units for shoulder kinematics using a commercial sensor-software system: A validation study. Health Science Reports 5, e772. https://doi.org/10.1002/HSR2.772.CrossRefGoogle ScholarPubMed
Hubaut, R, Guichard, R, Greenfield, J and Blandeau, M (2022) Validation of an embedded motion-capture and EMG setup for the analysis of musculoskeletal disorder risks during manhole cover handling. Sensors 22, 436. https://doi.org/10.3390/S22020436.CrossRefGoogle ScholarPubMed
Humadi, A, Nazarahari, M, Ahmad, R and Rouhani, H (2021) Instrumented ergonomic risk assessment using wearable inertial measurement units: Impact of joint angle convention. IEEE Access 9, 72937305. https://doi.org/10.1109/ACCESS.2020.3048645.CrossRefGoogle Scholar
Islam, MRU and Bai, S (2020) Effective multi-mode grasping assistance control of a soft hand exoskeleton using force myography. Frontiers in Robotics and AI 7, 567491. https://doi.org/10.3389/FROBT.2020.567491.CrossRefGoogle ScholarPubMed
Kubota, A, Iqbal, T, Shah, JA and Riek, LD (2019) Activity recognition in manufacturing: The roles of motion capture and sEMG+inertial wearables in detecting fine vs gross motion. Proceedings - IEEE International Conference on Robotics and Automation, 2019-May, 65336539. https://doi.org/10.1109/ICRA.2019.8793954.CrossRefGoogle Scholar
Lee, SH and Son, J (2021) Development of a safety management system tracking the weight of heavy objects carried by construction workers using fsr sensors. Applied Sciences 11, 1378. https://doi.org/10.3390/APP11041378.CrossRefGoogle Scholar
Lin, CJ and Peng, HY (2022) A study of the human-robot synchronous control based on IMU and EMG sensing of an upper limb. ASCC 2022–2022 13th Asian Control Conference, Proceedings, 14741479. https://doi.org/10.23919/ASCC56756.2022.9828042.CrossRefGoogle Scholar
Madrigal, JAB, Cardiel, E, Rogeli, P, Salas, LL and Guerrero, RM (2016) Evaluation of suitability of a micro-processing unit of motion analysis for upper limb tracking. Medical Engineering Physics 38, 793800. https://doi.org/10.1016/J.MEDENGPHY.2016.04.011.CrossRefGoogle Scholar
Mateos, LA (2017) Characterizing lifting and lowering activities with insole fsr sensors in industrial exoskeletons. https://doi.org/10.48550/arXiv.1706.05440.CrossRefGoogle Scholar
McAtamney, L and Corlett, EN (1993) Rula: A survey method for the investigation of work-related upper limb disorders. Applied Ergonomics 24, 9199. https://doi.org/10.1016/0003-6870(93)90080-S.CrossRefGoogle Scholar
Mihcin, S (2019) Methodology on co-registration of MRI and optoelectronic motion capture marker sets: In-vivo wrist case study. Hittite Journal of Science Engineering 6, 99107. https://doi.org/10.17350/HJSE19030000134.CrossRefGoogle Scholar
Mihcin, S (2022) Simultaneous validation of wearable motion capture system for lower body applications: Over single plane range of motion (ROM) and gait activities. Biomedical Engineering / Biomedizinische Technik 67 (3), 185199. https://doi.org/10.1515/bmt-2021-0429.CrossRefGoogle ScholarPubMed
Musso, M, Oliveira, AS and Bai, S (2024) Influence of an upper limb exoskeleton on muscle activity during various construction and manufacturing tasks. Applied Ergonomics 114, 104158. https://doi.org/10.1016/j.apergo.2023.104158.CrossRefGoogle ScholarPubMed
Oliveira, AS and Negro, F (2021) Neural control of matched motor units during muscle shortening and lengthening at increasing velocities. Journal of Applied Physiology 130, 17981813. https://doi.org/10.1152/JAPPLPHYSIOL.00043.2021.CrossRefGoogle ScholarPubMed
Patel, V, Chesmore, A, Legner, CM and Pandey, S (2021) Trends in Workplace Wearable Technologies and Connected-Worker Solutions for Next-Generation Occupational Safety, Health, and Productivity. https://doi.org/10.1002/aisy.202100099.CrossRefGoogle Scholar
Peppoloni, L, Filippeschi, A, Ruffaldi, E and Avizzano, CA (2016) A novel wearable system for the online assessment of risk for biomechanical load in repetitive efforts. International Journal of Industrial Ergonomics 52, 111. https://doi.org/10.1016/J.ERGON.2015.07.002.CrossRefGoogle Scholar
Pérez, R, Costa, Ú, Torrent, M, Solana, J, Opisso, E, Cáceres, C, Tormos, JM, Medina, J and Gómez, EJ (2010) Upper limb portable motion analysis system based on inertial technology for neurorehabilitation purposes. Sensors 10, 1073310751. https://doi.org/10.3390/S101210733.CrossRefGoogle ScholarPubMed
Rehman, MU, Shah, K, Haq, IU and Khurshid, H (2022) A force myography based HMI for classification of upper extremity gestures. 2nd IEEE International Conference on Artificial Intelligence, ICAI 2022, 100104. https://doi.org/10.1109/ICAI55435.2022.9773429.CrossRefGoogle Scholar
Ren, L, Jones, RK and Howard, D (2008) Whole body inverse dynamics over a complete gait cycle based only on measured kinematics. Journal of Biomechanics 41, 27502759. https://doi.org/10.1016/J.JBIOMECH.2008.06.001.CrossRefGoogle Scholar
Robert-Lachaine, X, Mecheri, H, Larue, C and Plamondon, A (2017) Validation of inertial measurement units with an optoelectronic system for whole-body motion analysis. Medical and Biological Engineering and Computing 55, 609619. https://doi.org/10.1007/S11517-016-1537-2.CrossRefGoogle ScholarPubMed
Robert-Lachaine, X, Mecheri, H, Muller, A, Larue, C and Plamondon, A (2020) Validation of a low-cost inertial motion capture system for whole-body motion analysis. Journal of Biomechanics 99, 109520. https://doi.org/10.1016/J.JBIOMECH.2019.109520.CrossRefGoogle ScholarPubMed
Seiferheld, B, Frost, J, Krog, M and Samani, A (2022) A field evaluation of a passive upper-extremity exoskeleton for manual material handling in blade manufacturing [14th Annual Meeting of the Danish Society of Biomechanics. Conference date: 11-11-2022 Through 11-11-2022]. In Program Abstracts: 14th Annual Meeting of the Danish Society of Biomechanics. pp. 10.Google Scholar
Serra-Hsu, E and Taboga, P (2022) Validation of fuze IMU system for ergonomics assessments. bioRxiv, https://doi.org/10.1101/2022.12.05.519202.CrossRefGoogle Scholar
Skals, S, Bláfoss, R, Andersen, MS, de Zee, M and Andersen, LL (2021) Manual material handling in the supermarket sector. Part 1: Joint angles and muscle activity of trapezius descendens and erector spinae longissimus. Applied Ergonomics 92, 103340. https://doi.org/10.1016/j.apergo.2020.103340.CrossRefGoogle ScholarPubMed
Tahir, A, Bai, S, Shen, MA, Tahir, A, Bai, S and Shen, M (2023) A wearable multi-modal digital upper limb assessment system for automatic musculoskeletal risk evaluation. Sensors 23, 4863. https://doi.org/10.3390/S23104863.CrossRefGoogle ScholarPubMed
Tao, W, Lai, ZH, Leu, MC and Yin, Z (2018) Worker activity recognition in smart manufacturing using IMU and sEMG signals with convolutional neural networks. Procedia Manufacturing 26, 11591166. https://doi.org/10.1016/J.PROMFG.2018.07.152.CrossRefGoogle Scholar
Vicon. (2024). Plug-in Gait Full-Body. Available at https://help.vicon.com/space/Nexus216/11607226/Full+body+modeling+with+Plugin+Gait (accessed: 3 June 2024).Google Scholar
Wang, W, Li, R, Diekel, ZM, Chen, Y, Zhang, Z and Jia, Y (2019) Controlling object hand-over in human-robot collaboration via natural wearable sensing. IEEE Transactions on Human-Machine Systems 49, 5971. https://doi.org/10.1109/THMS.2018.2883176.CrossRefGoogle Scholar
Wininger, M, Kim, NH and Craelius, W (2008) Pressure signature of forearmas predictor of grip force. Journal of Rehabilitation Research and Development 45, 883892. https://doi.org/10.1682/JRRD.2007.11.0187.CrossRefGoogle ScholarPubMed
Wu, Y, Tao, K, Chen, Q, Tian, Y and Sun, L (2022) A comprehensive analysis of the validity and reliability of the perception neuron studio for upper-body motion capture. Sensors 22 (18). https://doi.org/10.3390/s22186954.Google ScholarPubMed
Xiao, ZG and Menon, C (2014) Towards the development of a wearable feedback system for monitoring the activities of the upper-extremities. Journal of Neuroengineering and Rehabilitation 11, 113. https://doi.org/10.1186/1743-0003-11-2.CrossRefGoogle ScholarPubMed
Xiao, ZG and Menon, C (2017a) Performance of forearm FMG and sEMG for estimating elbow, forearm and wrist positions. Journal of Bionic Engineering 14, 284295. https://doi.org/10.1016/S1672-6529(16)60398-0.CrossRefGoogle Scholar
Xiao, ZG and Menon, C (2017b) Counting grasping action using forcemyography: An exploratory study with healthy individuals. JMIR Rehabil Assist Technol 4 ( 1), e6901. https://doi.org/10.2196/REHAB.6901.CrossRefGoogle Scholar
Zhang, JT, Novak, AC, Brouwer, B and Li, Q (2013) Concurrent validation of Xsens MVN measurement of lower limb joint angular kinematics. Physiological Measurement 34. https://doi.org/10.1088/0967-3334/34/8/N63.CrossRefGoogle ScholarPubMed
Zhang, ZQ, Wong, WC and Wu, JK (2011) Ubiquitous human upper-limb motion estimation using wearable sensors. IEEE Transactions on Information Technology in Biomedicine 15, 513521. https://doi.org/10.1109/TITB.2011.2159122.CrossRefGoogle ScholarPubMed
Zhu, H, Li, X, Wang, L, Chen, Z, Shi, Y, Zheng, S and Li, M (2022) IMU motion capture method with adaptive tremor attenuation in teleoperation robot system. Sensors 22(9). https://doi.org/10.3390/s22093353.Google ScholarPubMed
Figure 0

Figure 1. Illustration of the components used to create the WSS. Two BioX armbands containing IMUs and FSRs were used to acquire upper limb movement and muscle activation. The data are transmitted via Bluetooth to a tablet, which saves it to a cloud service and computes real-time joint angles.

Figure 1

Figure 2. IMU sensor placement and orientation of the sensors’ internal reference systems.

Figure 2

Figure 3. Computed shoulder flexion/extension angles from the two methods and segmentation results.

Figure 3

Figure 4. Identified minima (red circles) in the average FSR signal (a) and RWRA marker trajectory (b) during the simulated work task.

Figure 4

Figure 5. Violin plots of the coefficient of multiple correlation (CMC; a and d), root mean square error (RMSE; b and e), and relative root mean square error (RRMSE; c and f) computed from the comparison between optical motion capture and IMU-derived joint angle calculations for the shoulder (upper row) and elbow (bottom row). Each colored dot in a violin represents the average value for one participant, the white dot represents the median across participants, the gray vertical bar indicates the first and third quartiles, and the shaded colored area depicts the density curve. The tasks are described in Section 3.1.

Figure 5

Figure 6. Correlation and Bland–Altman plots for three selected tasks: T1, T4, and T7.

Figure 6

Figure 7. Illustrative image of a worker performing the task of moving a panel for packing. The workers performed the packing task with and without using a passive exoskeleton (Skelex 360-XFR). The task involved shoulder abduction and flexion, combined with elbow flexion. WSS was used to identify the grasping moments for subsequent segmentation of joint angles.

Figure 7

Figure 8. Averaged force-sensing resistor (FSR) signals from the two workers and the identified segments. Dashed lines indicate the start and end of the main events; solid lines indicate the limits of full task cycles. For worker 2, these two sets of lines coincide.

Figure 8

Table 1. Summary of the movement cycle parameters for the two workers under the two conditions, along with the percentage differences between the “Free” and “Exo” conditions

Figure 9

Table 2. Percentage of the data in the Bland–Altman plots falling within the upper and lower limits of agreement (mean $ \pm $1.96 SD of difference)

Figure 10

Figure 9. Worker 2 shoulder and flexion angles during full cycles.

Supplementary material: File

Musso et al. supplementary material

Musso et al. supplementary material
Download Musso et al. supplementary material(File)
File 312.1 KB