Next Article in Journal
Space-Time Block Coded Cooperative MIMO Systems
Next Article in Special Issue
A Cost-Effective Inertial Measurement System for Tracking Movement and Triggering Kinesthetic Feedback in Lower-Limb Prosthesis Users
Previous Article in Journal
A Review of Indirect Tool Condition Monitoring Systems and Decision-Making Methods in Turning: Critical Analysis and Trends
Previous Article in Special Issue
Fluid Intake Monitoring System Using a Wearable Inertial Sensor for Fluid Intake Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Functional Shoulder Task Identification and Sub-Task Segmentation Using Wearable Inertial Measurement Units for Frozen Shoulder Assessment

1
Department of Physical Medicine and Rehabilitation, Tri-Service General Hospital, School of Medicine, National Defense Medical Center, Taipei 114, Taiwan
2
Department of Physical Therapy and Assistive Technology, National Yang-Ming University, Taipei 112, Taiwan
3
Department of Biomedical Engineering, National Yang-Ming University, Taipei 112, Taiwan
4
Research Center for Information Technology Innovation, Academia Sinica, Taipei 115, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(1), 106; https://doi.org/10.3390/s21010106
Submission received: 3 December 2020 / Revised: 22 December 2020 / Accepted: 22 December 2020 / Published: 26 December 2020
(This article belongs to the Special Issue Wearable Inertial Sensors)

Abstract

:
Advanced sensor technologies have been applied to support frozen shoulder assessment. Sensor-based assessment tools provide objective, continuous and quantitative information for evaluation and diagnosis. However, the current tools for assessment of functional shoulder tasks mainly rely on manual operation. It may cause several technical issues to the reliability and usability of the assessment tool, including manual bias during the recording and additional efforts for data labeling. To tackle these issues, this pilot study aims to propose an automatic functional shoulder task identification and sub-task segmentation system using inertial measurement units to provide reliable shoulder task labeling and sub-task information for clinical professionals. The proposed method combines machine learning models and rule-based modification to identify shoulder tasks and segment sub-tasks accurately. A hierarchical design is applied to enhance the efficiency and performance of the proposed approach. Nine healthy subjects and nine frozen shoulder patients are invited to perform five common shoulder tasks in the lab-based and clinical environments, respectively. The experimental results show that the proposed method can achieve 87.11% F-score for shoulder task identification, and 83.23% F-score and 427 mean absolute time errors (milliseconds) for sub-task segmentation. The proposed approach demonstrates the feasibility of the proposed method to support reliable evaluation for clinical assessment.

1. Introduction

Frozen shoulder (FS) is a common joint condition that causes stiffness and pain among people aged from 40 to 65 years [1], especially in women [2]. The stiffness and pain of shoulder joints lead the limitation to the range of motion in all movement planes of the shoulder joints. FS has great impacts on the quality of daily life and activities of daily living (ADL) performance [2,3]. The common treatments in FS patients involving physical therapy and joint shoulder injection aim to relieve pain, improve joint mobility, and increase the independent ability. In order to support clinical decisions, there is a requirement of objective assessment for clinical evaluations and follow-up progresses [4].
Goniometry measurements [5] and questionnaires [6] are common evaluation tools for clinical FS assessment. However, these traditional assessment approaches have several challenges and limitations related to inter-rater reliability, respondent interpretation, and cultural diversity [7,8,9]. In recent years, inertial measurement units (IMUs) have been used to develop objective evaluation systems. Joint evaluation systems using IMUs have advantages in simplification of implementation, cost, and computation complexity. They have the potential to continuously and accurately measure dynamic and static range of motion of shoulder joints, including flexion, extension and rotation [10]. Previous studies have shown the reliability of measurement systems with inertial sensors for elbow and shoulder movement in laboratory environments [10,11,12,13].
For FS patients, wearable IMUs are also implemented to objectively measure functional abilities while the questionnaires can only provide subjective scores from the patients (e.g., shoulder pain and disability index [14] and simple shoulder score [15]). These works extracted movement features and parameters to evaluate the performance of functional shoulder tasks. However, the whole measurement still relies on manual operation. For example, researchers or clinical professionals have to manually label the starting and ending time of the shoulder tasks from the continuous signals. Then, they label the spotted shoulder task with the correct task information. These additional efforts may decrease the feasibility and usability of the IMU-based evaluation systems in the clinical setting.
To tackle the aforementioned challenges, this pilot study aims to propose an automatic functional shoulder task identification and sub-task segmentation system using wearable IMUs for FS assessment. We hypothesized that the proposed wearable-based systems would be reliable and feasible to automatically provide shoulder task information for clinical evaluation and assessment. Several typical pattern recognition and signal processing techniques (e.g., symmetry-weighted moving average, sliding window and principal component analysis), machine learning models (e.g., support vector machine, k-nearest-neighbor and classification and regression tree), and rule-based modification are applied to the proposed system to accurately identify shoulder tasks and segment sub-tasks from continuous sensing signal. Moreover, a hierarchical approach is applied to enhance the reliability and efficiency of the proposed system. The novelty and contribution of this pilot study are listed as follows:
  • This work firstly proposes a functional shoulder task identification system for automatic shoulder task labeling while the traditional functional measurement in clinical setting still relies on manual operation.
  • The proposed approach can provide not only shoulder task information (e.g., cleaning head) but also sub-task information (e.g., lifting hands to head, washing head and putting heads down). Such sub-task information has the potential to support clinical professionals for further analysis and examination.
  • The feasibility and the effectiveness of the proposed shoulder identification and sub-task segmentation is validated on nine FS patients and nine healthy subjects.

2. Related Works

In recent years, automatic movement identification and segmentation algorithms have been proposed to clinical evaluation and healthcare applications [16,17,18,19,20]. The main objective of identification and segmentation algorithms is to spot the starting and ending points of target activates precisely. For example, previous studies have developed diverse approaches to automatically and objectively obtain detailed lower limb and trunk movement information, such as sitting, standing, walking and turning [16]. Such reliable segmentation approaches can assist clinical professionals for various disease assessment, involving Parkinson’s disease [17], fall prediction [18] and dementia [19]. Similar approaches are also applied to upper limb assessments in stroke patients. Biswas et al. [20] proposed segmentation algorithms using a single inertial sensor to gather three basic movements from the complicated forearm activities in healthy and stroke patients, involving extension, flexion and rotation. However, few studies focus on the development of automatic systems in FS patients [11]. Most evaluation tools for FS assessment still relied on manual operation [10,21,22,23,24].
Various machine learning (ML) approaches are applied to automatically identify human movements for healthcare applications [25,26,27,28,29]. Generally, there are two categories for ML: discriminative and generative approaches. The typical discriminative approaches involving k- nearest-neighbors (k-NN) [25], classification and regression tree (CART) [26] and support vector machine (SVM) [26] aim to optimize the rules or decision boundaries to separate classes. Such approaches have shown the high- speed processing and reliable detection performance for movement segmentation. Another approach is generative models, such as hidden Markov models (HMM) [27], which are built based on probabilistic models to identify continuous movements. Generative approaches have better abilities to the more complicated activities and temporal order problems. Additionally, diverse deep learning approaches are widely applied to movement segmentation [28] and human activity recognition [29], e.g., convolutional neural networks (CNN) and recurrent neural networks (RNN). They have superior classification ability compared to traditional ML approaches. However, generative and deep learning approaches require a large dataset to ensure the performance of the detection model, while the data requirements for discriminative approaches are comparatively low.

3. Methods

The framework of the proposed automatic shoulder task identification and sub-task segmentation is shown in Figure 1. The brief introduction of the whole training and testing stages for the identification and segmentation is as follows:
  • Input and pre-processing: In the beginning, accelerometers and gyroscopes are utilized to collect shoulder task sequences (Input). Then, the sensing sequences are pre-processed with the moving average technique to filter the noises. These pre-processed sequences are spilt into the training set and testing set for the training and testing stages, respectively.
  • Training for shoulder task identification: The feature extraction process with 12 feature types is firstly applied to the pre-processed sequences. Then, the principal component analysis is employed to reduce the size of the features and select the critical features for training machine learning models. Next, the machine learning model is trained with the selected features of the training set for shoulder task identification. Various machine learning techniques, including SVM, CART, and kNN, are investigated in this work. The parameter optimization for each technique is executed in this stage.
  • Training for sub-task segmentation: First, the sliding window technique divides the pre-processed sequences into segments. Then, the feature extraction and dimension reduction techniques are employed to obtain the critical features from the segments. Lastly, the machine learning model for ML-based sub-task segmentation is built with the critical features. During the training stage, several machine learning techniques (e.g., SVM, CART, and kNN) and their optimized parameters are also explored.
  • Testing for shoulder task identification: Initially, the selected features are extracted from the shoulder task sequence of the testing set. Then, these features are identified using the trained machine learning model to output the shoulder task information (output 1).
  • Testing for sub-task segmentation: After the testing stage of the shoulder task identification, the sliding window technique is firstly applied to the shoulder task sequence to gather a sequence of segments. Secondly, the feature extraction process is employed to the segments to obtain selected features. Thirdly, the process of ML-based sub-task segmentation classifies these segments and the corresponding features using the trained machine learning models and outputs a sequence of the identified class labels. Fourthly, the rule-based modification is utilized to modify the output of the ML-based sub-task segmentation. Finally, the sub-task information generator generates a sequence of sub-task labels based on the classified and modified class labels and outputs it as the sub-task information (output 2).

3.1. Participants

Participants were outpatients at a rehabilitation department of Tri-service general hospital who were diagnosed with primary FS between June 2020 and September 2020. The patients were included if they have shoulder pain with a limited range of motion more than 3 months and age from 20 to 70 years old. Participants were diagnosed with primary FS according to standardized history, physical examination, and ultrasonographic evaluation by an experiment physiatrist. Patients were excluded if they had any of the following: full or massive thickness tear of the rotator cuff on ultrasonography or magnetic resonance imaging (MRI); secondary FS (secondary to other causes, including metabolic, rheumatic, or infectious arthritis; stroke; tumor; or fracture); and acute cervical radiculopathy.
The study was approved by the institutional review board (TSGHIRB No.: A202005024) at the university hospital, and all participants gave written informed consent. Our research procedure followed the Helsinki Declaration. All participants were assured that their participation was entirely voluntary and that they could withdraw at any time. Nine healthy adults (height: 170.6 ± 7.9 cm, weight: 75.1 ± 17.0 kg, age: 27.0 ± 5.0 years old) and nine FS patients (height: 164.3 ± 11.1 cm, weight: 66.3 ± 14.4 kg, age: 56.4 ± 9.9 years old) participated in the experiments.

3.2. Experimental Protocol and Data Collection

Two IMUs placed on the arm and wrist are employed to sense the upper limb movement, as shown in Figure 2. Similar sensor placements have been selected in previous works [20,21]. The sensors placed on the arm and wrist can catch information of upper limb movement while performing shoulder tasks. The used IMU (APDM Inc., Portland, OR, USA) involves a tri-axial accelerometer, tri-axial gyroscope, and tri-axial magnetometer. In this study, only the tri-axial accelerometer (range: ±16 g; resolution: 14 bits) and tri-axial gyroscope (range: ±2000°/s; resolution: 16 bits) work for the data. The data is collected with a sampling frequency of 128 Hz.
The experiment is executed in the lab-based and clinical environments for healthy and FS subjects, respectively. Each subject is asked to perform five shoulder tasks once, including cleaning head, cleaning upper back and shoulder, cleaning lower back, placing an object on a high shelf, and putting/removing an object from the back pocket. These shoulder tasks have been widely adopted for shoulder function assessment and evaluation in previous works [21,22]. The performed shoulder tasks and the corresponding three sub-tasks are listed in Table 1. Each task consists of three sub-tasks. Totally, there are 90 shoulder task sequences (18 subjects × 5 shoulder tasks). The participants are free to execute tasks in their ways with basic manual instruction. The sub-tasks are performed continuously within the same shoulder task. Mean sub-task time performed by healthy and FS patients is listed in Table 2.
The external camera synchronized with inertial sensors is applied to provide reference information for the ground truth labeling, including starting and ending points of shoulder tasks. During the experiment, the camera is put in front of the subjects. The frame per second of the camera is 30 Hz.

3.3. Data Pre-Processing

This study applies the symmetry-weighted moving average (SWMA) technique to the sensing signals to reduce the noise and artifacts for shoulder task identification and segmentation. This pre-processing technique has been applied to other applications while the sensors are placed on the upper limbs, including eating activity recognition and daily activity recognition [30,31]. SWMA technique determines different weights to sample points within the determined ranges. The data points closer to the central point are assigned with higher weights.
Suppose the sensing data of any shoulder task sequence is defined as S = { s i | i = 1 ,   2 , , n R } , where n R is the total number of the data samples from the sequence. The pre-processed sensing data point s ˜ t at time t with the determined range m is defined as follows:
s ˜ t = 1 T o t a l δ ( δ 0 s i + ( i = 1 m + 1 2 1 δ i ( s t + i + s t i ) ) ) ,
T o t a l δ = i = 0 m + 1 2 1 δ i ,
where T o t a l δ is the sum of all determined weights, δ 0 is m + 1 2 and δ i = { δ o i | i = 1 ,   2 , , δ 0 } . For example, if m is 5, s 3 = [ ( s 3 × 3 + s 2 × 1 + s 4 × 1 + s 1 × 0.5 + s 5 × 0.5 ) / ( 3 + 1 + 1 + 0.5 + 0.5 ) ] . The SWMA with m = 9 is applied to this study.

3.4. Shoulder Task Identification

3.4.1. Feature Extraction

The main objective of the feature extraction process is to extract movement characteristics from the continuous sensing data for shoulder task identification. There are two feature categories that have been applied to catch motion features, such as statistical and kinematic features. The common statistical features involving mean, standard deviation (StD), variance (var), maximum (max), minimum (min), range, kurtosis, skewness, and correlation coefficient (CorrCoef) have been applied to the field of activity recognition applications [32]. These nine statistical features are applied to this work. Also, kinematic features have been applied to upper limb movement recognition systems in several clinical applications, such as stroke rehabilitation and assessment [33]. This study employs three general kinematic features, such as the number of velocity peaks (NVP), zero crossing (NZR), and mean crossing (NMR) for shoulder task identification.
Suppose a sequence of data from a sensor is defined as S ˜ = { s ˜ i | i = 1 ,   2 , , n R } , where n R is the total number of the data samples from the sequence. Any sample point s i includes data collected from a tri-axial sensor s ˜ i = { r ˜ x i ,   r ˜ y i , r ˜ z i } . Then, the feature extraction process is applied to the shoulder sequence. The utilized features are listed in Table 3.
In this work, the sensing data of the shoulder task sequence from two IMUs is defined as S ˜ s e q = { s ˜ i | i = 1 ,   2 , , n s e q } , where n s e q is the total number of S ˜ s e q . Any sample point s ˜ i of S ˜ s e q is defined as:
S ˜ i = { a ˜ x i w r i s t , a ˜ y i w r i s t , a ˜ z i w r i s t , g ˜ x i w r i s t , g ˜ y i w r i s t , g ˜ z i w r i s t , a ˜ x i a r m , a ˜ y i a r m , a ˜ z i a r m , g ˜ x i a r m , g ˜ y i a r m , g ˜ z i a r m }
The formation of the extracted features from S ˜ s e q is show in Figure 3. There are two IMUs, four sensors (2 accelerometers + 2 gyroscopes), and a total of 144 features (4 sensor units × 36 features) are obtained.

3.4.2. Feature Selection

During the training stage, the feature selection process is applied to all extracted features after the feature extraction. This is because the size of all features (144 features) is quite big for the systems.
Using a suitable feature selection technique can simplify the computing processes, which is beneficial for training and testing stages. This study utilizes principal component analysis (PCA) [34] to select critical features and reduce the number of features in dealing with multi-dimensional time sequence data. PCA aims to find a linear transformation matrix that transforms the raw feature vectors F ˜ = [ f ˜ 1 , f ˜ 2 , , f ˜ k ] to lower dimensional feature vectors F ^ = [ f ^ 1 , f ^ 2 , , f ^ l ] , where k = 144 is the number of the raw feature vectors and l is the number of the transformed feature vectors.
Firstly, the covariance matrix C f is calculated based on the variance maximization of the projected data. Then, the eigenvalues λ = ( λ 1 , λ 2 , , λ k ) and eigenvectors ν = ( ν 1 , ν 2 , , ν k ) can be determined based on C f . Note that the eigenvectors ν are the principal components, where that first eigenvector has the largest variance.
In the dimension reduction process, the l eigenvectors with the most explained components are kept, where l k . A threshold t h r e s = 0.99 is set to keep 99% variance information of the raw feature vectors. The minimum value of l is determined as Equation (3):
i = 1 l λ i i = 1 k λ i t h r e s
For the shoulder task identification, the number of features is reduced from 144 to 35 after PCA and dimension reduction processes. Compared to the original raw feature vectors, the system using the transformed feature sets has the potential to reduce computational complexity for the classification of the shoulder task.

3.4.3. Shoulder Task Identification Using Machine Learning

Suppose there is a set of class labels C = ( c 1 , c 2 , , c n c ) , where n C is the number of the class labels. The training set Γ t r a i n = { ( F ^ i t r a i n , c i ) | i = 1 ,   2 , , n t r a i n } has n t r a i n pairs of feature vectors F ^ i t r a i n and the corresponding label c i . In the training stage, the machine learning technique can optimize the parameters θ of a classification model by minimizing the classification loss on Γ t r a i n . For the shoulder task identification, n C = 5 is the number of the shoulder tasks.
In the testing stage, given that the testing set Γ t e s t = { F ^ i t e s t | i = 1 ,   2 , , n t e s t } has n t e s t feature vectors. Each F ^ i t e s t is mapped to a set of class labels C with the corresponding confidence score P i = { p i j | j = 1 ,   2 , , n C } using the trained classification model Η with the optimized parameters θ :
p i ( c | F ^ i t e s t , θ ) = Η ( F ^ i t e s t , θ ) ,
where c C . Then, we select the class label with the maximum confidence score as the final classification output:
c i = a r g m a x c C ,       p P i p ( c | F ^ i t e s t , θ ) .
There are various machine learning models have been applied to segment human movements and recognize activities in other clinical applications [16,17,18,19,20]. At this moment, several machine learning techniques requiring a lot of data volume for model training are not considered in this work, involving HMM, CNN, and RNN. Therefore, we focus on exploring the feasibility of the following machine learning models for shoulder task identification:
  • Support vector machine (SVM): The main objective of the SVM model is to find a hyperplane to separate two classes. It maximizes the margin between two classes to support distinct classification with more confidence. Since the number of the classes are more than two, we employ one-vs-all techniques to multi-class classification with a radial basis kernel function.
  • K-nearest-neighbors (kNN): kNN approach is also called as a lazy classifier as this approach actually does not require any training process. The main idea of this approach is to determine the class of the testing data based on the major voting of nearest k neighbors. The determination of the value k is application-dependent, which have critical influences on the performance of the classifier. In this work, a range of k from 1 to 9 is explored. The results show that k = 7 achieves the best detection performance.
  • Classification and regression tree (CART): The CART approach is a binary tree that can tackle classification and regression problems. The branch size and the process of the splitting is determined by measure of the Gini impurity. This approach has advantages in easy implementation and high processing speed.
The feasibility and reliability of the explored techniques have been validated in the field of activity recognition [29].

3.5. Sub-Task Segmentation

3.5.1. Sliding Window

There are several windowing approaches that have been proposed to divide the continuous data into chunks [35], involving sliding window, event-defined window and activity-defined window techniques. This work uses the sliding window to segment the data into small segments. This windowing approach is very popular in the field of activity recognition due to its simple realization and fast processing speed.
Suppose the pre-processed sensing data of the shoulder task sequence from two IMUs is defined as S ˜ s e q = { s ˜ i | i = 1 ,   2 , , n s e q } , where n s e q is the total number of S ˜ s e q . The sliding window technique is applied to S ˜ s e q with several parameters, including window size w s , the starting point of the segment s p , ending point of the segment e p , sliding samples s s . The pseudocode of the sliding window is described in Algorithm 1 and illustrated in Figure 4.
Algorithm 1: Sliding Window.
Input:the pre-processed sensing data S ˜ s e q = { s ˜ i | i = 1 , 2 , , n s e q   } , window size w s , the starting point of the segment s p , ending point of the segment e p , sliding samples s s
Output:a set of segments W = { w j | j = 1 , 2 , , n s l   }
1:Begin
2:initialize s p 1 , e p w s and j 1
3:while e p n s e q do
4: w j { s ˜ s p j , s ˜ s p + 1 j , , s ˜ e p 1 j , s ˜ e p j }
5: W w j
6: j j + 1
7: s p s p + s s
8:e p s p + w s 1
9:endwhile
10:End
After the process of sliding window, a set of segments obtained from the shoulder task sequences S ˜ s e q is defined as W = { w j | j = 1 ,   2 , , n s l } , where n s l is the total number of segments obtained from S ˜ s e q . Any segment is defined as w j = { s ˜ s p j , s ˜ s p + 1 j , , s ˜ e p 1 j , s ˜ e p j } , where s ˜ s p and s ˜ e p are the starting and ending points of the segment. Note that o p is defined as the overlapping percentage between w j and w j + 1 , where j + 1 n s l . o p can be calculated as follows:
o p = o s w s ,
o s = w s s s ,
where o s is the overlapped samples.
The window size has great impact on the system performance while using the sliding window technique. A range of window sizes from 0.1 to 1.5 s with a fixed overlapping of 50% is tested to explore the reliability of the proposed automatic sub-task segmentation.

3.5.2. Training Stage for Sub-Task Segmentation

Given that there is a set of segments W T r S e t = { w j t r a i n | j = 1 ,   2 , , n T r S e t } obtained from the pre-processed shoulder task sequences using sliding window, where w j t r a i n = { s ˜ j t r a i n | j = 1 ,   2 , , n w s } contains n w s sample points. Any s ˜ i t r a i n containing the sensing data collected from the wrist and arm is defined as S ˜ i t r a i n = { a ˜ x j w r i s t , a ˜ y j w r i s t , a ˜ z j w r i s t , g ˜ x j w r i s t , g ˜ y j w r i s t , g ˜ z j w r i s t , a ˜ x j a r m , a ˜ y j a r m , a ˜ z j a r m , g ˜ x j a r m , g ˜ y j a r m , g ˜ z j a r m } . The training process is as follows:
  • Firstly, w T r S e t are initially extracted with nine types of statistical features and three types of kinematic features to obtain training feature vectors U ˜ t r a i n = { F ˜ j t r a i n | j = 1 ,   2 , n T r S e t } , where F ˜ j t r a i n = { f ˜ j , 1 t r a i n , f ˜ j , 2 t r a i n , , f ˜ j , k t r a i n } and k = 144 .
  • Then, PCA is also applied to U ˜ t r a i n   t o obtain dimensionless feature vectors U ^ t r a i n = { F ^ j t r a i n | j = 1 ,   2 , n T r S e t } , where F ^ j t r a i n = { f ^ j , 1 t r a i n , f ^ j , 2 t r a i n , , f ^ j , l ´ t r a i n } and l ´ k . In this paper, the size of the utilized feature vectors l ´ for different windows is reduced from 144 to less than 50.
  • After the processes of feature extraction and selection, a training set Γ ´ t r a i n = { ( F ^ j t r a i n , c j t r a i n ) | i = 1 ,   2 , , n T r S e t } is created, where n T r S e t is the number of feature vectors, and c i t r a i n is the corresponding label of F ^ i t r a i n . In this work, there is a set of class labels C ´ = ( c ´ 1 , c ´ 2 , , c ´ n c ´ ) , where n C ´ is 3, including sub-task A, B, and C.
  • Finally, using a machine learning technique learns the parameters θ ´ of the machine learning model Η ´ from Γ ´ t r a i n . Several typical ML approaches are also explored for sub-task segmentation, such as SVM, CART, and kNN.

3.5.3. Testing Stage for Sub-Task Segmentation Using Machine Learning Models, Rule-Based Modification and Sub-Task Information Generator

There are three main processes for sub-task segmentation: ML-based identification, rule-based modification and sub-task information generator. The first process is to employ ML approaches to segment and identify sub-tasks. Several typical machine learning approaches are tested, such as SVM, CART, and kNN. However, mis-segmentation and mis-identification is unavoidable during the process. Therefore, the second process is to correct the errors from the ML-based approach. The modification process modifies fragmentation errors as the identified results are irrational to the context. For example, a continuous data stream identified as sub-task B “washing head” should not involve other sub-tasks (e.g., lifting hands or putting hands down). Finally, the generator generates the sub-task information based on the outputs of the rule-based modification.
Given that a set of segments W T e S e t = { w 1 t e s t , w 2 t e s t , , w n S t e s t } and the corresponding feature vectors U ^ t e s t = { F ^ 1 t e s t , F ^ 2 t e s t , , F ^ n S t e s t } are obtained from a pre-processed shoulder task sequence of the testing set S ˜ T e S e q = { s ˜ 1 , s ˜ 2 , , s ˜ n T e S e q } by using the sliding window technique and feature extraction with the selected features, where n T e S e q and n S are the total number of S ˜ T e S e q and W T e S e t , respectively. The detailed ML-based sub-task segmentation and rule-based modification processes in the testing stage is described as follows:
  • Firstly, the mapped confidence score P ´ i = { p ´ i 1 ,   p ´ i 2 , , p ´ i n c ´ } of a set of class labels C ´ = ( c ´ 1 , c ´ 2 , , c ´ n c ´ ) from each F ^ i t e s t is calculated, where n c ´ is the total number of C ´ .
  • Secondly, each F ^ i t e s t maps to a class label c ´ M L with the maximum confidence score by using the trained machine learning model Η ´ and the optimized parameters θ ´ . A sequence of classified class labels D M L = { c ´ 1 M L , c ´ 2 M L , , c ´ n S M L } is generated from U ^ t e s t using Η ´ and θ ´ .
  • Thirdly, the rule-based modification is applied to D M L to obtain a sequence of modified class labels D r = { c ´ 1 r , c ´ 2 r , , c ´ n S r } . If c ´ t M L is different from c ´ t 1 M L and c ´ t + 1 M L , and c ´ t 1 M L is equal to c ´ t + 1 M L then c ´ t M L would be modified as the sub-task of c ´ t 1 M L and c ´ t + 1 M L , where c ´ t M L D M L and 2 t n S 1 . An example to illustrate the modification process is shown in Figure 5.
  • Finally, a generator generates a sequence of sub-task labels D g = { c ´ 1 g , c ´ 2 g , , c ´ n g g } based on D r , where n g is the total number of D g and determined as:
    n g = w s + s s × ( n S 1 ) ,
    where w s and s s are window size and sliding samples, respectively. The processes of the sub-task information generator are illustrated in Figure 6 and the corresponding pseudocode is shown in Algorithm 2.
Algorithm 2: Sub-task Information Generator.
Input:a sequence of modified class labels D r = { c ´ j r | j = 1 , 2 , , n s } , window size w s , sliding samples s s
Output:a sequence of sub-task labels D g = { c ´ i g | i = 1 , 2 , , n g }
1:Begin
2:  initializei 1
3:  for  j = 1  to  n s 1  do //for the first n s 1 modified class labels
4:   while  i ( j × s s )
5: c ´ i g c ´ j r
6: D g c ´ i g
7: i i + 1
8:   endwhile
9: endfor
10: for  i  to  i + w s  do //for the last n s modified class labels
11: c ´ i g c ´ n s r
12: D g c ´ i g
13: endfor
14:End

3.6. Performance Evaluation and Statistical Analysis

The whole system implementation and statistical analysis are done using the Statistics and Machine Learning Toolbox in Matlab 2017b (MathWorks Inc., Natick, MA, USA).
This study utilizes a leave-one-subject-out cross-validation approach [32] to validate the system performance of the proposed shoulder task identification and sub-task segmentation. This validation approach divides the dataset into k folds based on the subjects, where k is the number of subjects; one fold is kept as the testing set and the remaining k-1 folds are utilized for the training. The whole process repeats k times until each fold is used as the testing set. Finally, the system outputs the average results of k folds.
In order to evaluate the reliability of the shoulder task identification, several typical metrics are utilized for performance evaluation, including sensitivity, precision and F-score [36] as shown in Equations (9)–(11):
sensitivity = TP TP   +   FN
precision = TP TP   +   FP
F score = 2 × sensitivity   ×   precision sensitivity   +   precision
where TP, FP, TN, and FN are true positive, false positive, true negative, and false negative. F-Score is the harmonic mean of precision and recall, which is a common approach to evaluate the reliability and performance of classification systems.
There are two evaluation and analysis approaches applied for the evaluation of sub-task segmentation: the sample-based approach [36] and mean absolute time errors (MATE) [37,38,39]. An illustration of the evaluation approaches for sub-task segmentation is shown in Figure 7. The first one is to calculate the number of TP, FP, TN, and FN based on the sample-by-sample mapping between the ground truth and system outputs. Then, the sensitivity, precision and F-score are applied to assess the system reliability based on the mapping results. The second approach is to calculate the average of the absolute time errors between the reference and identified boundaries, where the boundary is the edge between two sub-tasks. There are two MATE values calculated for the proposed sub-task segmentation:
  • MATE A , B : MATE of the boundaries between sub-task A and sub-task B.
  • MATE B , C : MATE of the boundaries between sub-task B and sub-task C.
  • MATE overall : MATE of all boundaries between sub-task A and sub-task B and between sub-task B sub-task C.

4. Results

The experimental results of the shoulder task identification are shown in Table 4. The results show that the shoulder task identification using SVM model can achieve 87.06% sensitivity, 88.43% precision and 87.11% F-score, and outperform that using other ML models. However, the proposed approach using SVM model is still weak to tackle several shoulder tasks such as T3 (cleaning lower back) and T5 (putting/removing an object in/form the back pocket) while the F-score of other shoulder tasks can achieve over 90%.
The sensitivity, precision, and F-score of the sub-task segmentation using different ML approaches and window sizes are presented in Table 5, Table 6 and Table 7, respectively. Generally, the sub-task segmentation using SVM and kNN models have the similar performance in sensitivity, precision, and F-score, which outperforms that using kNN model. The experimental results show that the proposed segmentation approach with SVM model can achieve the best overall performance in sensitivity (82.27%), precision (85.07%) and F-score (83.23%) while the worst performance is with CART model. Furthermore, using SVM model has the best F-score of 86.53%, 82.75%, and 82.42% in the sub-task A, sub-task B, and sub-task C, respectively.
The results also reveal that the F-score of the sub-task segmentation model using SVM and kNN models significantly decreases when the window is larger than 1.0 s. Most of them achieve the best performance as window sizes are 0.2 and 0.3 s. However, the performance using CART model achieves the best F-score with the window size of 1.5 s.
Table 8, Table 9 and Table 10 presents the sub-task segmentation performance of MATE A , B , MATE B , C and MATE overall using different machine learning models and window sizes for all subjects, healthy subject and FS patients, respectively. Overall, the proposed segmentation using kNN achieves the lowest MATE A , B , MATE B , C and MATE overall in most subject groups. However, the best machine learning models for MATE overall and MATE B , C of FS patients are SVM and CART, respectively. The lowest MATE overall of all subjects, healthy subjects and FS patients are 427, 273, and 517 ms, respectively. Also, the experimental results reveal that the MATE of healthy subjects is lower than that of the FS patients.
The impact of window sizes in the sub-task segmentation performance of MATE A , B , MATE B , C and MATE overall is similar to that of sensitivity, precision and F-score. The proposed segmentation approach with different machine learning models have the lowest MATE values when the window size is smaller or equal to 1.0 s. Particularly, the results show that the proposed segmentation system using window sizes of 0.1 and 1.0 s can achieve the lowest MATE A , B , MATE B , C and MATE overall .
An example to demonstrate the processes of ML-based identification and rule-based modification for sub-task segmentation on the healthy subject is shown in Figure 8. It presents that a complete segment is often divided into fragments when the system used ML-based segmentation only, as shown in Figure 8c. For example, a segment of sub-task B is divided into 4 fragments. The proposed rule-based modification can correct the segmentation errors caused by ML-based sub-task segmentation, as presented in Figure 8d. After the processes of ML-based sub-task segmentation and rule-based modification, the segmentation errors of this work mainly occur in the boundaries between two sub-tasks, which decrease the performance of the proposed sub-task segmentation approach.

5. Discussion

Various sensor technologies have been applied to develop objective evaluation systems, including range of motion measurement and function evaluation. To tackle the issues in labeling errors and bias during the measurement, we propose an automatic functional shoulder task identification and sub-task segmentation system using wearable IMUs for FS assessment. The proposed approach can achieve 87.11% F-score for shoulder task identification, and 83.23% F-score, 387 MATE A , B and 403 MATE B , C for sub-task segmentation. The proposed system has the potential to support clinical professionals in automatic shoulder task labeling and sub-task information obtainment.
The results show that the proposed shoulder task identification has poor performance on T3 and T5 as the F-score on them are lower than 80%. This is because several FS patients are unable to move hands to the lower back but they can reach the back pocket while performing T3. The execution of T3 and T5 performed by the patients have very similar movement patterns. Such a situation confuses the models for identification of T3 and T5, even for SVM model.
Several machine learning models have been applied in this work, including SVM, CART and kNN. Previous works have shown the feasibility and the effectiveness of these models in movement identification and segmentation [16,17,18,19,20]. The proposed segmentation approach using SVM and kNN models can achieve the best performance in F-score and MATE, respectively. However, the differences between their segmentation performance are very close in the two evaluation performance approaches. Considering that the kNN model has the advantages of less computation complexity and simple implementation, the kNN model is more suitable for the proposed system.
Previous studies have shown that the sliding window approach is sensitive to the window sizes [35]. The proposed sub-task segmentation approach has similar experimental results as the segmentation performance with different window sizes ranges over 10%. This is because the larger sizes of the window may smooth the movement characteristics that confuse the identification models and lead to misidentification. Also, using too larger window sizes may lead to early or late segmentation of the sub-tasks, which increases the segmentation errors of the proposed system. An illustration of the segmentation performance using smaller and larger window sizes is shown in Figure 9.
Figure 10 shows the signal of T2 “clean upper back and shoulder task” collected from the FS patient and healthy subject using a wrist-worn sensor. Due to stiffness and pain of the shoulder, the FS patients perform the shoulder task slowly and carefully with a limited range of motion. Obviously, the movement patterns of the three sub-tasks performed by the FS patient are significantly different from those performed by the healthy subject. It means the shoulder task can be performed in diverse ways according to the health status and the function of the shoulder, which leads to identification challenges of variability and similarity to the shoulder task identification and sub-task segmentation [32].
To our best knowledge, this is the first study aiming to identify and segment upper limb movements of shoulder tasks using machine learning approaches in FS patients, especially for FS assessment. Machine learning models have been successfully applied to automatic movement identification and recognition models to analyze lower limb movements in other clinical applications [16,17,18,19,20]. However, most IMU-based shoulder function assessment systems still rely on manual operation [10,21,22,23,24]. Our results demonstrate the feasibility and effectiveness of the ML-based functional shoulder task identification for supporting clinical assessment and proof of concept. Moreover, the proposed system can obtain sub-task information from continuous signals, which has the potential for further analysis and investigation of functional performance.
Some technical challenges still limit the performance of the proposed system to shoulder task identification and sub-task segmentation, including gesture time, variability, similarity, and boundary decision. We plan to test other powerful machine learning models to improve identification and segmentation performance, such as CNN, LSTM, longest common subsequence (LCSS) dynamic time warping (DTW), hidden Markov model (HMM) and conditional random field (CRF). Another limitation is that the proposed automatic system is validated on five shoulder tasks only. More shoulder tasks from other clinical tests and questionnaires are going to be explored for validation of the proposed system, e.g., simple and shoulder score [14], American Shoulder and Elbow Surgeons score [40], and so on. Furthermore, there are only nine FS patients and nine healthy subjects participating in this work. More FS patients with different functional disabilities, the different ages of healthy subjects and different disease groups will be recruited for validation and investigation.

6. Conclusions

In order to support FS assessment in the clinical setting, we propose a functional shoulder task identification system using IMUs for shoulder task identification and sub-task segmentation. We use several typical pattern recognition techniques, machine learning models and rule-based modification to automatically identify five shoulder tasks and segment three sub-tasks. The feasibility and reliability of this study are validated with healthy and FS subjects. The experimental results show that the proposed system has the potential to provide automatic labeling of the shoulder task and sub-task information for clinical professionals.

Author Contributions

Conceptualization, C.-Y.C. and K.-C.L.; software, C.-Y.H. and H.-Y.H.; investigation, C.-Y.C. and K.-C.L.; writing—original draft preparation, C.-Y.C. and K.-C.L.; writing—review and editing, C.-Y.H. and H.-Y.H.; resources, Y.-T.W., C.-T.C., L.-C.C.; supervision, Y.-T.W., C.-T.C., L.-C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Tri-Service General Hospital (protocol code: A202005024, April 13, 2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kelley, M.J.; McClure, P.W.; Leggin, B.G. Frozen shoulder: Evidence and a proposed model guiding rehabilitation. J. Orthop. Sports Phys. Ther. 2009, 39, 135–148. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Kelley, M.J.; Shaffer, M.A.; Kuhn, J.E.; Michener, L.A.; Seitz, A.L.; Uhl, T.L.; Godges, J.J.; McClure, P.W. Shoulder pain and mobility deficits: Adhesive capsulitis. J. Orthop. Sports Phys. Ther. 2013, 43, A1–A31. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Neviaser, A.S.; Neviaser, R.J. Adhesive capsulitis of the shoulder. J. Am. Acad. Orthop. Surg. 2011, 19, 536–542. [Google Scholar] [CrossRef] [PubMed]
  4. Fayad, F.; Roby-Brami, A.; Gautheron, V.; Lefevre-Colau, M.M.; Hanneton, S.; Fermanian, J.; Poiraudeau, S.; Revel, M. Relationship of glenohumeral elevation and 3-dimensional scapular kinematics with disability in patients with shoulder disorders. J. Rehabil. Med. 2008, 40, 456–460. [Google Scholar] [CrossRef] [Green Version]
  5. Struyf, F.; Meeus, M. Current evidence on physical therapy in patients with adhesive capsulitis: What are we missing? Clin. Rheumatol. 2014, 33, 593–600. [Google Scholar] [CrossRef] [PubMed]
  6. Roy, J.-S.; MacDermid, J.C.; Woodhouse, L.J. Measuring shoulder function: A systematic review of four questionnaires. Arthritis Care Res. 2009, 61, 623–632. [Google Scholar] [CrossRef]
  7. Olley, L.; Carr, A. The Use of a Patient-Based Questionnaire (The Oxford Shoulder Score) to Assess Outcome after Rotator Cuff Repair. Ann. R. Coll. Surg. Engl. 2008, 90, 326–331. [Google Scholar] [CrossRef] [Green Version]
  8. Ragab, A.A. Validity of self-assessment outcome questionnaires: Patient-physician discrepancy in outcome interpretation. Biomed. Sci. Instrum. 2003, 39, 579–584. [Google Scholar]
  9. Muir, S.W.; Corea, C.L.; Beaupre, L. Evaluating change in clinical status: Reliability and measures of agreement for the assessment of glenohumeral range of motion. N. Am. J. Sports Phys. Ther. 2010, 5, 98–110. [Google Scholar]
  10. De Baets, L.; Vanbrabant, S.; Dierickx, C.; van der Straaten, R.; Timmermans, A. Assessment of Scapulothoracic, Glenohumeral, and Elbow Motion in Adhesive Capsulitis by Means of Inertial Sensor Technology: A Within-Session, Intra-Operator and Inter-Operator Reliability and Agreement Study. Sensors 2020, 20, 876. [Google Scholar] [CrossRef] [Green Version]
  11. Coley, B.; Jolles, B.M.; Farron, A.; Pichonnaz, C.; Bassin, J.P.; Aminian, K. Estimating dominant upper-limb segments during daily activity. Gait Posture 2008, 27, 368–375. [Google Scholar] [CrossRef] [PubMed]
  12. Luinge, H.J.; Veltink, P.H.; Baten, C.T. Ambulatory measurement of arm orientation. J. Biomech. 2007, 40, 78–85. [Google Scholar] [CrossRef] [PubMed]
  13. Rundquist, P.J.; Anderson, D.D.; Guanche, C.A.; Ludewig, P.M. Shoulder kinematics in subjects with frozen shoulder. Arch. Phys. Med. Rehabil. 2003, 84, 1473–1479. [Google Scholar] [CrossRef]
  14. Breckenridge, J.D.; McAuley, J.H. Shoulder Pain and Disability Index (SPADI). J. Physiother. 2011, 57, 197. [Google Scholar] [CrossRef] [Green Version]
  15. Schmidt, S.; Ferrer, M.; González, M.; González, N.; Valderas, J.M.; Alonso, J.; Escobar, A.; Vrotsou, K. Evaluation of shoulder-specific patient-reported outcome measures: A systematic and standardized comparison of available evidence. J. Shoulder Elbow Surg. 2014, 23, 434–444. [Google Scholar] [CrossRef]
  16. Sprint, G.; Cook, D.J.; Weeks, D.L. Toward Automating Clinical Assessments: A Survey of the Timed Up and Go. IEEE Rev. Biomed. Eng. 2015, 8, 64–77. [Google Scholar] [CrossRef]
  17. Palmerini, L.; Mellone, S.; Avanzolini, G.; Valzania, F.; Chiari, L. Quantification of motor impairment in Parkinson’s disease using an instrumented timed up and go test. IEEE Trans. Neural Syst. Rehabil. Eng. 2013, 21, 664–673. [Google Scholar] [CrossRef]
  18. Greene, B.R.; O’Donovan, A.; Romero-Ortuno, R.; Cogan, L.; Scanaill, C.N.; Kenny, R.A. Quantitative falls risk assessment using the timed up and go test. IEEE Trans. Biomed. Eng. 2010, 57, 2918–2926. [Google Scholar] [CrossRef]
  19. Greene, B.R.; Kenny, R.A. Assessment of cognitive decline through quantitative analysis of the timed up and go test. IEEE Trans. Biomed. Eng. 2012, 59, 988–995. [Google Scholar] [CrossRef]
  20. Biswas, D.; Cranny, A.; Gupta, N.; Maharatna, K.; Achner, J.; Klemke, J.; Jöbges, M.; Ortmann, S. Recognizing upper limb movements with wrist worn inertial sensors using k-means clustering classification. Hum. Mov. Sci. 2015, 40, 59–76. [Google Scholar] [CrossRef] [Green Version]
  21. Coley, B.; Jolles, B.M.; Farron, A.; Bourgeois, A.; Nussbaumer, F.; Pichonnaz, C.; Aminian, K. Outcome evaluation in shoulder surgery using 3D kinematics sensors. Gait Posture 2007, 25, 523–532. [Google Scholar] [CrossRef] [PubMed]
  22. Pichonnaz, C.; Aminian, K.; Ancey, C.; Jaccard, H.; Lécureux, E.; Duc, C.; Farron, A.; Jolles, B.M.; Gleeson, N. Heightened clinical utility of smartphone versus body-worn inertial system for shoulder function B-B score. PLoS ONE 2017, 12, e0174365. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Körver, R.J.; Heyligers, I.C.; Samijo, S.K.; Grimm, B. Inertia based functional scoring of the shoulder in clinical practice. Physiol. Meas. 2014, 35, 167–176. [Google Scholar] [CrossRef] [PubMed]
  24. Bavan, L.; Wood, J.; Surmacz, K.; Beard, D.; Rees, J. Instrumented assessment of shoulder function: A study of inertial sensor based methods. Clin. Biomech. 2020, 72, 164–171. [Google Scholar] [CrossRef]
  25. Parate, A.; Chiu, M.C.; Chadowitz, C.; Ganesan, D.; Kalogerakis, E. RisQ: Recognizing Smoking Gestures with Inertial Sensors on a Wristband. MobiSys 2014, 2014, 149–161. [Google Scholar] [CrossRef] [Green Version]
  26. Reinfelder, S.; Hauer, R.; Barth, J.; Klucken, J.; Eskofier, B.M. Timed Up-and-Go phase segmentation in Parkinson’s disease patients using unobtrusive inertial sensors. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2015, 2015, 5171–5174. [Google Scholar] [CrossRef]
  27. Lin, J.F.; Kulić, D. Online Segmentation of Human Motion for Automated Rehabilitation Exercise Analysis. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 168–180. [Google Scholar] [CrossRef]
  28. Panwar, M.; Biswas, D.; Bajaj, H.; Jobges, M.; Turk, R.; Maharatna, K.; Acharyya, A. Rehab-Net: Deep Learning Framework for Arm Movement Classification Using Wearable Sensors for Stroke Rehabilitation. IEEE Trans. Biomed. Eng. 2019, 66, 3026–3037. [Google Scholar] [CrossRef]
  29. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef] [Green Version]
  30. Chernbumroong, S.; Cang, S.; Atkins, A.; Yu, H. Elderly activities recognition and classification for applications in assisted living. Expert Syst. Appl. 2013, 40, 1662–1674. [Google Scholar] [CrossRef]
  31. Dong, Y.; Scisco, J.; Wilson, M.; Muth, E.; Hoover, A. Detecting Periods of Eating During Free-Living by Tracking Wrist Motion. IEEE J. Biomed. Health Inform. 2014, 18, 1253–1260. [Google Scholar] [CrossRef] [PubMed]
  32. Bulling, A.; Blanke, U.; Schiele, B. A tutorial on human activity recognition using body-worn inertial sensors. ACM Comput. Surv. 2014, 46, 33. [Google Scholar] [CrossRef]
  33. De los Reyes-Guzmán, A.; Dimbwadyo-Terrer, I.; Trincado-Alonso, F.; Monasterio-Huelin, F.; Torricelli, D.; Gil-Agudo, A. Quantitative assessment based on kinematic measures of functional impairments during upper extremity movements: A review. Clin. Biomech. 2014, 29, 719–727. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Ringnér, M. What is principal component analysis? Nat. Biotechnol. 2008, 26, 303–304. [Google Scholar] [CrossRef]
  35. Banos, O.; Galvez, J.M.; Damas, M.; Pomares, H.; Rojas, I. Window size impact in human activity recognition. Sensors 2014, 14, 6474–6499. [Google Scholar] [CrossRef] [Green Version]
  36. Ward, J.A.; Lukowicz, P.; Gellersen, H.W. Performance metrics for activity recognition. ACM Trans. Intell. Syst. Technol. 2011, 2, 6. [Google Scholar] [CrossRef]
  37. Savoie, P.; Cameron, J.A.D.; Kaye, M.E.; Scheme, E.J. Automation of the Timed-Up-and-Go Test Using a Conventional Video Camera. IEEE J. Biomed. Health Inform. 2020, 24, 1196–1205. [Google Scholar] [CrossRef]
  38. Ortega-Bastidas, P.; Aqueveque, P.; Gómez, B.; Saavedra, F.; Cano-de-la-Cuerda, R. Use of a Single Wireless IMU for the Segmentation and Automatic Analysis of Activities Performed in the 3-m Timed Up & Go Test. Sensors 2019, 19, 1647. [Google Scholar]
  39. Beyea, J.; McGibbon, C.A.; Sexton, A.; Noble, J.; O’Connell, C. Convergent Validity of a Wearable Sensor System for Measuring Sub-Task Performance during the Timed Up-and-Go Test. Sensors 2017, 17, 934. [Google Scholar] [CrossRef] [Green Version]
  40. Sallay, P.I.; Reed, L. The measurement of normative American Shoulder and Elbow Surgeons scores. J. Shoulder Elbow Surg. 2003, 12, 622–627. [Google Scholar] [CrossRef]
Figure 1. The framework of the automatic shoulder task identification and sub-task segmentation.
Figure 1. The framework of the automatic shoulder task identification and sub-task segmentation.
Sensors 21 00106 g001
Figure 2. An illustration of the sensor placements.
Figure 2. An illustration of the sensor placements.
Sensors 21 00106 g002
Figure 3. A formation of the extracted features from the IMUs.
Figure 3. A formation of the extracted features from the IMUs.
Sensors 21 00106 g003
Figure 4. An illustration of the sliding window on the sensing signals. s ˜ s p j and s ˜ e p j are the starting and ending points of the segment w j , where j = 1 , 2 , 3 ,   . The sliding samples is the distances from s ˜ s p j to s ˜ s p j + 1 . The overlapping samples is the number of the overlapping data samples between segments w j and w j + 1 .
Figure 4. An illustration of the sliding window on the sensing signals. s ˜ s p j and s ˜ e p j are the starting and ending points of the segment w j , where j = 1 , 2 , 3 ,   . The sliding samples is the distances from s ˜ s p j to s ˜ s p j + 1 . The overlapping samples is the number of the overlapping data samples between segments w j and w j + 1 .
Sensors 21 00106 g004
Figure 5. An illustration of the modification process to the fragmentation errors. The “sub-task A” of c ´ t is the misidentified result that is modified as “sub-task B” according to the proposed rule-based modification.
Figure 5. An illustration of the modification process to the fragmentation errors. The “sub-task A” of c ´ t is the misidentified result that is modified as “sub-task B” according to the proposed rule-based modification.
Sensors 21 00106 g005
Figure 6. An illustration of the sub-task information generator. A sequence of sub-task labels D g = { c ´ i g | i = 1 ,   2 , , n g } is obtained based on a sequence of modified class labels D r = { c ´ j r | j = 1 ,   2 , , n s } . For the first n s 1 modified class labels, each c ´ j r maps to a sequence of sub-task labels { c ´ i g ,   c ´ i + 1 g ,…,   c ´ i + s s 2 g , c ´ i + s s 1 g }, where i = 1 + s s × ( j 1 ) . Finally, a sequence of sub-task labels { c ´ i g ,   c ´ i + 1 g ,…,   c ´ i + w s 2 g , c ´ i + w s 1 g } is obtained from the last modified class label c ´ n s r .
Figure 6. An illustration of the sub-task information generator. A sequence of sub-task labels D g = { c ´ i g | i = 1 ,   2 , , n g } is obtained based on a sequence of modified class labels D r = { c ´ j r | j = 1 ,   2 , , n s } . For the first n s 1 modified class labels, each c ´ j r maps to a sequence of sub-task labels { c ´ i g ,   c ´ i + 1 g ,…,   c ´ i + s s 2 g , c ´ i + s s 1 g }, where i = 1 + s s × ( j 1 ) . Finally, a sequence of sub-task labels { c ´ i g ,   c ´ i + 1 g ,…,   c ´ i + w s 2 g , c ´ i + w s 1 g } is obtained from the last modified class label c ´ n s r .
Sensors 21 00106 g006
Figure 7. An illustration of the annotation for the performance evaluation of sub-task segmentation, including true positive, true negative, false positive, false negative, and absolute time error.
Figure 7. An illustration of the annotation for the performance evaluation of sub-task segmentation, including true positive, true negative, false positive, false negative, and absolute time error.
Sensors 21 00106 g007
Figure 8. An example of the signal performed by the health subject and the processes of the proposed sub-task segmentation. (a) The accelerometer and gyroscope signals collected from the IMUs placed on the wrist and arm. (b) The divided segments obtained from the process of sliding window. In this example, there are 23 segments (c) The classification results for sub-task A, B and C after the processes of feature extraction and ML-based sub-task segmentation, where the TP, TN, FP, and FN are annotated. (d) The classification results after the processes of rule-based modification, where the modified sliding segments are highlighted in red square (e.g., w 4 , w 10 , w 15 , w 22 ) and the successful modification results are annotated.
Figure 8. An example of the signal performed by the health subject and the processes of the proposed sub-task segmentation. (a) The accelerometer and gyroscope signals collected from the IMUs placed on the wrist and arm. (b) The divided segments obtained from the process of sliding window. In this example, there are 23 segments (c) The classification results for sub-task A, B and C after the processes of feature extraction and ML-based sub-task segmentation, where the TP, TN, FP, and FN are annotated. (d) The classification results after the processes of rule-based modification, where the modified sliding segments are highlighted in red square (e.g., w 4 , w 10 , w 15 , w 22 ) and the successful modification results are annotated.
Sensors 21 00106 g008
Figure 9. An illustration of the segmentation performance using smaller and larger window sizes. (a) The classification results using smaller window size, where the window size is 0.5 s. (b) The classification results using larger window size, where the window size is 1.5 s.
Figure 9. An illustration of the segmentation performance using smaller and larger window sizes. (a) The classification results using smaller window size, where the window size is 0.5 s. (b) The classification results using larger window size, where the window size is 1.5 s.
Sensors 21 00106 g009
Figure 10. An example of the data of the T2 task “cleaning upper back and shoulder” collected from the wrist-worn sensor, which are performed by (a) the FS patient and (b)the healthy subject.
Figure 10. An example of the data of the T2 task “cleaning upper back and shoulder” collected from the wrist-worn sensor, which are performed by (a) the FS patient and (b)the healthy subject.
Sensors 21 00106 g010
Table 1. A list of shoulder task and sub-task.
Table 1. A list of shoulder task and sub-task.
Task IDShoulder TaskSub-Task ASub-Task BSub-Task C
T1Cleaning headLifting hands toward headWashing headPutting hands down
T2Cleaning upper back and shoulderLifting hands toward upper back and shoulderWashing upper back and shoulderPutting hands down
T3Cleaning lower backLifting hands towards lower backWashing lower backPutting hands down
T4Placing an object on a high shelfLifting the object toward the shelfHolding the hands on the shelf for few secondsPutting hands down
T5Putting/Removing an object into/from the back pocketPutting an object into the back pocketHolding the hands in the back pocket for few secondsRemoving an object from the back pocket
Table 2. Mean sub-task time performed by healthy and FS patients (s).
Table 2. Mean sub-task time performed by healthy and FS patients (s).
Healthy SubjectsFS PatientsAll Subjects
Sub-task A0.86 ± 0.151.6 ± 0.731.22 ± 0.62
T1Sub-task B3.13 ± 1.185.66 ± 24.46 ± 2.08
Sub-task C0.9 ± 0.21.18 ± 0.161.04 ± 0.22
Sub-task A1.16 ± 0.251.82 ± 0.851.45 ± 0.68
T2Sub-task B2.58 ± 0.757.81 ± 3.754.93 ± 3.63
Sub-task C1.18 ± 0.21.35 ± 0.211.29 ± 0.24
Sub-task A0.78 ± 0.121.1 ± 0.440.92 ± 0.35
T3Sub-task B3.09 ± 1.156.32 ± 4.194.94 ± 3.57
Sub-task C0.99 ± 0.210.94 ± 0.250.95 ± 0.24
Sub-task A1.53 ± 0.352.02 ± 0.651.81 ± 0.59
T4Sub-task B0.97 ± 0.491.98 ± 0.531.45 ± 0.7
Sub-task C1.52 ± 0.411.25 ± 0.451.45 ± 0.49
Sub-task A1.47 ± 0.461.79 ± 0.751.61 ± 0.62
T5Sub-task B0.9 ± 0.650.89 ± 0.680.93 ± 0.66
Sub-task C2.42 ± 2.441.37 ± 0.351.88 ± 1.82
Table 3. A list of statistical and kinematic feature types from a single sensor.
Table 3. A list of statistical and kinematic feature types from a single sensor.
No.Description
f ˜ 1 f ˜ 3 Mean of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 4 f ˜ 6 Standard Deviation of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 7 f ˜ 9 Variance of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 10 f ˜ 12 Maximum of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 13 f ˜ 15 Minimum of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 16 f ˜ 18 Range of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 19 f ˜ 21 Kurtosis of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 22 f ˜ 24 Skewness of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 25 f ˜ 27 Correlation coefficient between each pair of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 28 f ˜ 30 Number of velocity peaks of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 31 f ˜ 33 Number of zero crossing of r ˜ x i ,   r ˜ y i , r ˜ z i
f ˜ 34 f ˜ 36 Number of mean crossing of r ˜ x i ,   r ˜ y i , r ˜ z i
Note. r ˜ x i ,   r ˜ y i , r ˜ z i are the sample points of x-axis, y-axis and z-axis collected from a tri-axial sensor node.
Table 4. The results of the shoulder task identification using machine learning approaches (%).
Table 4. The results of the shoulder task identification using machine learning approaches (%).
Shoulder TaskSensitivityPrecisionF-Score
SVMkNNCARTSVMkNNCARTSVMkNNCART
T194.1282.3570.59100.0060.8766.6796.9770.0068.57
T2100.0064.7164.7185.0091.6778.5791.8975.8670.97
T388.2476.4782.3571.4386.6777.7878.9581.2580.00
T482.3582.3576.47100.0082.3568.4290.3282.3572.22
T570.5988.2476.4785.7183.3381.2577.4285.7178.79
Overall87.0678.8274.1288.4380.9874.5487.1179.0474.11
Note. SVM: support vector machine; kNN: k-nearest-neighbors; CART: classification and regression tree.
Table 5. The sensitivity of the sub-task segmentation using machine learning approaches (%) vs. different window sizes (s).
Table 5. The sensitivity of the sub-task segmentation using machine learning approaches (%) vs. different window sizes (s).
Window Size (s)Sub-Task ASub-Task BSub-Task COverall
SVMkNNCARTSVMkNNCARTSVMkNNCARTSVMkNNCART
0.177.9483.8894.3990.1587.4364.0578.7161.4048.3282.27 a77.5768.92
0.283.0281.4688.8088.8187.1259.8174.8671.3970.8182.2379.9973.14
0.375.9182.3183.5687.8083.9863.0979.4376.2775.2881.0580.8573.98
0.473.2078.0481.7383.3876.3350.9283.2880.6078.6479.9678.3270.43
0.574.0679.6180.5787.4580.9456.9282.4679.5080.6181.3280.0272.70
0.671.5076.3473.7086.2779.3056.1382.2180.6381.9179.9978.7670.58
0.773.7374.6480.1689.6684.4668.1581.9177.6572.3381.7778.9273.55
0.867.5069.9870.9786.1384.8766.2780.8078.0780.4878.1477.6472.57
0.965.0071.3476.2290.3979.7168.8775.1274.7974.0076.8475.2873.03
166.0670.5271.2687.8780.2075.2980.2872.5475.7078.0774.4274.08
1.166.6168.7777.6589.7184.9669.3570.6165.6067.9775.6473.1171.66
1.266.6068.4178.8086.4177.8075.3175.0067.4561.8276.0071.2271.98
1.366.5566.7073.1090.0683.4073.8171.3065.8260.8175.9771.9769.24
1.467.8565.2567.3494.3591.5883.0156.9454.2656.0173.0570.3668.79
1.569.8666.6672.7992.0490.4079.4860.4155.8054.9174.1070.9569.06
Note. The best performance of the column is highlighted in bold; SVM: support vector machine; kNN: k-nearest-neighbors; CART: classification and regression tree. a: The best overall performance.
Table 6. The precision of the sub-task segmentation using machine learning approaches (%) vs. different window sizes (s).
Table 6. The precision of the sub-task segmentation using machine learning approaches (%) vs. different window sizes (s).
Window Size (s)Sub-Task ASub-Task BSub-Task COverall
SVMkNNCARTSVMkNNCARTSVMkNNCARTSVMkNNCART
0.193.7381.3657.9477.6773.2276.9782.9693.7276.2684.7982.7770.39
0.290.4685.7171.3578.5776.2677.5086.2090.9456.5985.07 a84.3068.48
0.391.0689.2378.2876.1976.1574.4380.9782.0854.4282.7482.4969.04
0.490.4889.3178.9475.4076.3175.8576.5669.1949.6580.8178.2768.14
0.590.8488.6678.1076.2676.2875.0779.8077.8353.1982.3080.9268.78
0.689.3885.1481.0275.2576.3872.5476.7274.2149.9780.4578.5867.85
0.791.4787.5169.7276.8674.8177.9983.5386.0063.0083.9682.7770.24
0.893.0687.9081.6673.0274.2373.8778.0884.2955.6181.3882.1470.38
0.994.7386.0978.8671.5572.7071.5086.6077.9661.2184.2978.9270.52
192.4986.3083.4973.8570.4072.3780.4381.0565.8882.2679.2573.91
1.193.8688.2273.8370.8369.1771.3179.9881.5063.3181.5579.6369.49
1.292.2578.6376.4771.4069.1974.5779.1279.6873.1080.9275.8374.71
1.396.1683.5672.3270.9368.9270.4283.3579.8565.1883.4877.4469.31
1.496.9391.7975.9868.7567.0167.3489.4089.0284.1485.0382.6175.82
1.596.2987.8775.8269.3467.4368.9186.3188.4580.7283.9881.2575.15
Note. The best performance of the column is highlighted in bold; SVM: support vector machine; kNN: k-nearest-neighbors; CART: classification and regression tree. a: The best overall performance.
Table 7. The F-score of the sub-task segmentation using machine learning approaches (%) vs. different window sizes (s).
Table 7. The F-score of the sub-task segmentation using machine learning approaches (%) vs. different window sizes (s).
Window Size (s)Sub-Task ASub-Task BSub-Task COverall
SVMkNNCARTSVMkNNCARTSVMkNNCARTSVMkNNCART
0.184.7982.2981.1882.7578.8174.0380.3073.4571.2082.6178.1875.47
0.286.5383.0879.5482.4281.0668.9680.7479.8665.3383.23a81.3371.27
0.382.6885.5483.2380.7779.5673.2379.2078.9070.1480.8881.3375.53
0.480.6883.1980.6578.5175.9259.5678.6873.0062.3279.2977.3767.51
0.581.0683.8481.9380.7378.2265.7380.1377.7464.5780.6479.9370.74
0.678.9680.4283.8779.1177.3867.1478.5976.2764.6478.8978.0271.89
0.781.3480.4582.2482.1478.9973.9582.4280.8369.4981.9780.0975.23
0.878.3277.8381.3878.4678.9673.8578.6980.4469.5478.4979.0874.92
0.977.0677.5779.2078.8375.8468.4079.9374.8760.3678.6176.0969.32
176.8477.5178.4779.6474.7973.3180.1076.3470.0378.8676.2173.94
1.177.9677.0277.0377.8175.4570.9274.5372.0569.6676.7774.8472.53
1.277.2572.5476.5777.3373.0075.7676.4871.9272.4377.0272.4974.92
1.378.6074.1174.7978.3874.4774.3276.6270.9471.0377.8773.1773.38
1.478.8775.9475.0577.7875.8574.1069.2665.5768.5575.3072.4572.57
1.580.4075.3978.2877.1175.4676.6769.1864.6371.7375.5771.8375.56
Note. The best performance of the column is highlighted in bold; SVM: support vector machine; kNN: k-nearest-neighbors; CART: classification and regression tree. a: The best overall performance.
Table 8. The MATE A , B , MATE B , C and MATE overall of all subject using different machine learning models vs. difference window sizes (s).
Table 8. The MATE A , B , MATE B , C and MATE overall of all subject using different machine learning models vs. difference window sizes (s).
Window Size (sec) M A T E A , B   ( ms ) M A T E B , C   ( ms ) M A T E o v e r a l l   ( ms )
SVMkNNCARTSVMkNNCARTSVMkNNCART
0.1393387a569496466819445427c694
0.23924385904724731238433456914
0.3468502481522422959495462720
0.450552553856757715495365511044
0.54895594885144471419502503954
0.65365544395554041379546479909
0.74785864964304061017454496757
0.85435514955544031024549477760
0.95605555444994111425530483985
1.0556527558500403b909528465734
1.1537579612624515916581547764
1.2561581616591541739576561678
1.3533494691564550723549522707
1.4523490676679560816601525746
1.5498497594698599665598548630
Note. The best performance of the column is highlighted in bold; SVM: support vector machine; kNN: k-nearest-neighbors; CART: classification and regression tree. a: The lowest MATE value of MATE A , B ; b: The lowest MATE value of MATE B , C ; c: The lowest MATE value of MATE overall .
Table 9. The MATE A , B , MATE B , C and MATE overall of healthy subject using different machine learning models vs. difference window sizes (s).
Table 9. The MATE A , B , MATE B , C and MATE overall of healthy subject using different machine learning models vs. difference window sizes (s).
Window Size (sec) M A T E A , B   ( ms ) M A T E B , C   ( ms ) M A T E o v e r a l l   ( ms )
SVMkNNCARTSVMkNNCARTSVMkNNCART
0.1250223a458328397389289310424
0.2314282450431301501373292476
0.3304274442365271514335273c478
0.4384311438470388796427350617
0.53253014263423161026334309726
0.6359321359281267b544320294452
0.7371355486353341481362348484
0.8436343427490313655463328541
0.9474361415410332756442347586
1.0458381398357353635408367517
1.1427409592573474653500442623
1.2433380568586574608510477588
1.3424299695538562786481431741
1.4411328676778609703595469690
1.5423357556812633804618495680
Note. The best performance of the column is highlighted in bold; SVM: support vector machine; kNN: k-nearest-neighbors; CART: classification and regression tree. a: The lowest MATE value of MATE A , B ; b: The lowest MATE value of MATE B , C ; c: The lowest MATE value of MATE overall .
Table 10. The MATE A , B , MATE B , C and MATE overall of FS patients using different machine learning models vs. difference window sizes (s).
Table 10. The MATE A , B , MATE B , C and MATE overall of FS patients using different machine learning models vs. difference window sizes (s).
Window Size (s) M A T E A , B   ( ms ) M A T E B , C   ( ms ) M A T E overall   ( ms )
SVMkNNCARTSVMkNNCARTSVMkNNCART
0.1535551a6806175361250576544965
0.24725947295626461975517c6201352
0.36317305206795741404655652962
0.462673963866476723026457531470
0.565381754968657818126706981181
0.671478852083054122157726651368
0.758581750550647015545466441030
0.86507595636184941394634627979
0.964674867358849120946176201384
1.0654673718643453b1184649563951
1.16467496316755571178661653905
1.2689783665596507870643645768
1.3642690687589538660616614674
1.4635652675580510929608581802
1.5573637632584566526579602579
Note. The best performance of the column is highlighted in bold; SVM: support vector machine; kNN: k-nearest-neighbors; CART: classification and regression tree. a: The lowest MATE value of MATE A , B ; b: The lowest MATE value of MATE B , C ; c: The lowest MATE value of MATE overall .
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chang, C.-Y.; Hsieh, C.-Y.; Huang, H.-Y.; Wu, Y.-T.; Chen, L.-C.; Chan, C.-T.; Liu, K.-C. Automatic Functional Shoulder Task Identification and Sub-Task Segmentation Using Wearable Inertial Measurement Units for Frozen Shoulder Assessment. Sensors 2021, 21, 106. https://doi.org/10.3390/s21010106

AMA Style

Chang C-Y, Hsieh C-Y, Huang H-Y, Wu Y-T, Chen L-C, Chan C-T, Liu K-C. Automatic Functional Shoulder Task Identification and Sub-Task Segmentation Using Wearable Inertial Measurement Units for Frozen Shoulder Assessment. Sensors. 2021; 21(1):106. https://doi.org/10.3390/s21010106

Chicago/Turabian Style

Chang, Chih-Ya, Chia-Yeh Hsieh, Hsiang-Yun Huang, Yung-Tsan Wu, Liang-Cheng Chen, Chia-Tai Chan, and Kai-Chun Liu. 2021. "Automatic Functional Shoulder Task Identification and Sub-Task Segmentation Using Wearable Inertial Measurement Units for Frozen Shoulder Assessment" Sensors 21, no. 1: 106. https://doi.org/10.3390/s21010106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop