A Semi-Automatic Annotation Approach for Human Activity Recognition
<p>Schematic representation of an Human Activity Recognition (HAR) system architecture.</p> "> Figure 2
<p>A dataset of 3000 samples illustrating the working principles of Active Learning (AL) and Semi-Supervised Active Learning (SSAL). The samples are illustrated with colours identifying their respective class. The samples selected by the AL for expert annotation are depicted by the ×’s. The grey vertical line denotes the decision boundary between the two classes. (<b>a</b>) Active Learning. (<b>b</b>) Semi-Supervised Active Learning.</p> "> Figure 3
<p>In (<b>a</b>) it is shown the AL performance of the initial 300 iterations. The horizontal red line denotes the accuracy average score of Supervised Learning (SL). In (<b>b</b>) it is shown the classifier Least Confidence score and Classifier Overall Uncertainty score throughout 300 iterations. (<b>a</b>) Least Confidence Uncertainty Score. (<b>b</b>) Overall Uncertainty Score.</p> "> Figure 4
<p>Example of 1-Nearest Neighbour (NN) label propagation, in which the sample <math display="inline"><semantics> <msup> <mi>x</mi> <mo>*</mo> </msup> </semantics></math> propagates its label (in this example represented by the colour green) to its 1-NN, the sample <span class="html-italic">B</span>. Each circle represents a sample whose colours, green and red, represent two different classes. The samples in grey denote unlabelled samples.</p> "> Figure 5
<p>Example of 1-reverse Nearest Neighbour (rNN) label propagation, in which the sample <math display="inline"><semantics> <msup> <mi>x</mi> <mo>*</mo> </msup> </semantics></math> propagates its label (in this example represented by the colour green) to the samples to which, regarding the labelled samples, it is their NN, the samples <span class="html-italic">A</span>, <span class="html-italic">B</span> and <span class="html-italic">C</span>. Each circle represents a sample whose colours (green and red) represent two different classes. Samples in grey denote unlabelled samples.</p> "> Figure 6
<p>Horizon Plot showing the features and their behaviour along some of the dataset activities. In the y axis, it is presented the information about the sensor, its signal axis and the feature name. The green and red colours denote the signal’s positive and negative values, respectively, with its intensity increasing with the feature’s normalised absolute value and decreasing otherwise. (<b>a</b>) UCI dataset. (<b>b</b>) CADL dataset.</p> "> Figure 7
<p>Average increase of the AL classifier’s accuracy for the Query Strategies (QSs) throughout the cycle of iterations. The horizontal lines denote the average accuracy for SL (in red) and UL (in blue). LD denotes the Local Density Sampling and LC the Least Confident Sampling. (<b>a</b>) UCI dataset. (<b>b</b>) CADL dataset.</p> "> Figure 8
<p>Principal Component Analysis (PCA) of the CADL dataset samples after performing AL for 50 iterations. The classifier’s training set samples are depicted by the ×’s, whose colour identifies their respective class. The darker grey dots represent the unselected samples existent in the validation set. (<b>a</b>) Local Density * Margin Sampling. (<b>b</b>) Local Density * Least Confident Sampling.</p> "> Figure 9
<p>Classifier’s accuracy for the SSAL methods throughout the AL iterations. The horizontal lines denote the 10-CV average accuracy for SL (in red), and UL ARI score (in blue). Following the underscore in the NN and rNN methods: Euc, Cos, Dynamic Time Warping (DTW) and Time Alignment Metric (TAM), denote the distances used. (<b>a</b>) UCI dataset. (<b>b</b>) CADL dataset.</p> "> Figure 10
<p>Evolution on the percentage of the validation set unlabelled samples for the SSAL methods throughout the AL cycle iterations. Following the underscore in the NN and rNN methods: Euc, Cos, DTW and TAM, denote the similarity distances used in the respective method. (<b>a</b>) UCI dataset. (<b>b</b>) CADL dataset.</p> "> Figure 11
<p>Percentage of correctly automatically annotated samples throughout the AL iterations, for the SSAL methods, using the UCI and the CADL datasets. Following the underscore in the NN and rNN methods: Euc, Cos, DTW and TAM, denote the similarity distances used in the respective method. (<b>a</b>) UCI dataset. (<b>b</b>) CADL dataset.</p> "> Figure 12
<p>Confusion matrix for the Self-Training Semi-Supervised Active Learning (ST-SSAL) method using the Overall Uncertainty Classification-Change Stopping Criterion (Over-CC SC). (<b>a</b>) UCI dataset. (<b>b</b>) CADL dataset.</p> ">
Abstract
:1. Introduction
2. Related Work
3. Methods
3.1. General Active Learning Strategy
Algorithm 1 General Active Learning |
Input: initial train set L, unlabelled validation set U, independent test set T |
Output: predicted labels for the test set |
|
3.1.1. Initial Train Set
3.1.2. Sample Selection Strategy
- Margin Sampling [33,34]: Selects the sample with the minimum difference (margin) between the prediction probabilities of the first and second most likely classes, according the following equation.
- Entropy Sampling [33,34]: Selects the sample with the greatest entropy value, according to the following equation.Additionally, in order to create a homogeneous initial training set, a weight was introduced to the previously mentioned QSs while the training set was less than 1% of the validation set, according to the following equation [36].According to the literature [28], a sample with high uncertainty will most likely be an outlier. Thus, to overcome this issue, we tested the Local Density Sampling and the Uncertainty and Local Density Sampling QSs.
- Local Density Sampling: Selects the sample with higher representation on the feature space, i.e., located in a high-density region, which is measured by the amount of NNs surrounding the sample, according to the following equation.
- Uncertainty and Local Density Sampling [28,30,36]: Obtained through the linear combination between the previously mentioned QSs according to the following equation.
3.1.3. Stopping Criterion
- Max-Confidence SC (Max-Conf) [38]: As previously described, in the Least Confident Sampling it is selected for the oracle to annotate the sample with the highest uncertainty, i.e., the sample that the classifier is least confident in its classification. Moreover, if the selected sample has a low uncertainty score, it is possible to presume that the classifier is able to confidently classify that sample, as well as the remaining samples. Hence, the AL process can be stopped.
- Overall Uncertainty SC (Over-Unc) [38]: Similar to Max-Confidence SC, but instead of stopping the AL system if the least confident score is low, it is used the average of the least confident score computed on the remaining unlabelled samples. That is, if this value, denominated overall uncertainty score, shows insignificant low values, we can assume that the classifier has sufficient confidence in the classification of the remaining unlabelled samples and, therefore, the AL cycle can stop.Figure 3 shows that the stabilisation of the AL performance overlaps the stabilisation of both the least confident score and the overall uncertainty score. Hence, it was developed a condition to automatically detect whether the scores stabilised based on the mean and standard deviation over a given number of consecutive iterations S. The AL process stops when both the two following conditions are verified: ≤ and ≤; , and N = number of iterations. The threshold was obtained through a calibration based on the stabilisation of the classifier accuracy score using the ground truth data.
- Classification-Change SC (CC) [37,38]: As discussed in Section 2, uncertainty-based QSs aim to select the most informative samples for the classification, which should correspond to the ones located near decision boundaries. Thus, dictating the class to which each sample is allocated to, therefore, significantly changing the classifier’s performance and its prediction output. Hence, in the CC SC the AL is stopped once decision boundaries samples have been annotated and added to the classifier’s training set. Under these assumptions, alterations in the classifier’s prediction of the unlabelled data labels can be used to infer if the decision boundaries have been changed. Thus, if in two consecutive iterations the classifier’s labels prediction has been constant, then, we can assume that the newly annotated samples are not near a decision boundary but rather inside it, hence, the AL process can be put to an end.
- Combination Strategy SC: Consists in a multi-createria-based strategy that combines the prior SCs, namely Overall Uncertainty SC and Classification-Change SC (Over-CC) SC and Max-Confidence Uncertainty SC and Classification-Change SC (Max-CC) SC. The AL is stopped only if both SCs are verified. This method is justified in the cases where the uncertainty score quickly drops to insignificant low values, however, there are inconsistencies in the classifier’s prediction. Thus, the annotation of new samples may result in changes on the decision boundaries and, therefore, on an improvement of the classifier’s performance.
3.2. Semi-Supervised Active Learning Framework
Algorithm 2 Semi-Supervised Active Learning |
Input: initial train set L, unlabelled validation set U, independent test set T |
Output: predicted labels for the test set |
|
- Self-Training Semi-Supervised Active Learning (ST-SSAL) [31,32,40]: A classifier is trained on the available labelled data and posteriorly tested on the unlabelled data. Validation set samples having the highest prediction confidence score are added to the classifier’s training set and removed from the unlabelled dataset. This process is repeated iteratively as the classifier is re-trained on an increasingly larger and larger training set. Therefore, under the assumption that higy confident predicted labels are correct, the learner uses its own predictions to iteratively teach himself, consequently improving its performance. Hence, a sample will get annotated with if >= . The threshold will influence the amount of propagation and its accuracy. A larger will increase the automatic annotation but decrease its accuracy, since the model is less certain in the annotated sample’s label. On the other hand, a smaller will decrease the amount of annotation but increase its accuracy, since the few annotations are performed with high certainty. In the present work, was empirically set to 0.98 in order to obtain a significant automatic annotation while maintaining a good certainty in the annotation.
- k-Nearest Neighbour Semi-Supervised Active Learning (k-NN-SSAL) [30]: The sample selected by the QS () propagates its label to its k-NNs. The definition of k (the number of NNs to propagate ’s label) requires a trade-off between the amount of automatic annotation and the addition of error to the system. With a small k, few samples are automatically annotated, but, the ones annotated are done so with a good confidence, as they are close in the feature space. On the other hand, with a higher k, more samples are annotated, however, at the cost of possibly adding error to the classifier, as is giving its label to samples at a further distance and therefore, may be wrongly annotated. In the present work, k empirically was set to 5, in order to obtain a significant amount of automatic annotation without compromising the classification accuracy and execution time. Figure 4 depicts the 1-NN label propagation step. Each circle represents a sample whose colours (green and red) represent two different classes. Samples in grey denote unlabelled samples. In this example, the sample propagates its label to its 1-NN the sample B.
- k-rNN Semi-Supervised Active Learning (k-rNN-SSAL) [30]: The sample selected by the QS () propagates its label to all the samples to which, regarding the labelled samples, it is their NN and it is within a empirically set distance. For the rNN method, as in [30], k was set to 1 to enhance the label propagation performance. Figure 5 illustrates 1-rNN label propagation step. Thus, in this example, the sample propagates its label to the samples A, B and C.
Distance Measures
- Euclidean Distance: Measures the length of the straight line distance between two samples ( and , with dimension m) according to the following equation.
- Cosine Similarity Distance: Measures the cosine of the angle between two samples ( and ) according to the following equation. Cosine similarity ranges between -1 and 1, for opposite and coincident samples, respectively, with the distance value becoming larger as the samples become less similar.
- Dynamic Time Warping (DTW): Measures the similarity between two time-dependent sequences through a non-linear alignment minimising the distance between both. Moreover, the minimal distance is obtained through the computation of a local cost measure , where {, , }, {, , } are two time series of length N and M∈, respectively, producing a cost matrix. Where each element corresponds to the Euclidean distance, between each pair of elements in the both sequences. Thus, , will hold a small value (low cost) if and are similar, or a larger value (high cost) otherwise. Hence, the DTW finds the warping path (W) yielding the minimum total cost amount all possible warping paths, by going through the low cost values in the local cost matrix [41,42].
- Time Alignment Metric (TAM) [41]: Uses the optimal time alignment obtained by the DTW to infer the intervals when two time series are in phase, advance or in delay in relation to each other. TAM returns a distance metric benefiting series in phase, and penalising when signals are in advance or delay with each other. Thus, resulting in an output value decreasing as the similarity between the two signals increases and increasing otherwise between 0 and 3, the former for signals constantly in phase and the latter for completely out of phase signals. Considering again, two time sequences {, , …, } and {, , …, } of length N and M; N, M∈. Assuming is delayed in relation to , by a total time , advanced a total time and in phase by a time . The TAM () is given by:
4. Results
4.1. Datasets
4.2. Signal Processing
4.3. Model Selection
4.4. QS Analysis
- Accuracy: The obtained accuracy values from the QSs are very similar and tend to the value obtained by the SL algorithm. This is supported by Figure 7, where it is presented the classifier’s accuracy for the QSs throughout the AL iterations. As expected, overall, the learner becomes more reliable as its training set size increases, resulting in the continuous increase of its accuracy value throughout the iterations. Margin Sampling and Local Density * Least Confident Sampling attain the highest classification’s performances, outperforming PL. However, the difference between AL and PL is low due to two reasons: (1) an initial biased prediction probability due to the classifier very small initial training set; (2) both UCI and CADL datasets are equally balanced. In this circumstance a random selection of samples is enough to create a representative dataset with a few samples from each class. Local Density and Local Density * Margin Sampling, attain the lowest score, not achieving a reasonable performance. These QSs’ low performances are explained by the biased training set, non-representative of the entire dataset distribution under which the classifier operates. As observed in Figure 8a, the density weight causes the preferential selection of activities located in high-density regions, for the deterioration of the remaining as they become unknown for the classifier. Under these circumstances, the classifier does not have a homogeneous training set with sufficient amount of samples from all the class labels from which it can learn to be able to correctly predict all the samples’ labels. Still, with the exception of the aforementioned QSs, the remaining results are in accordance with the literature review [29,30] with the introduction of a density weight to the uncertainty sampling functions avoiding the selection of outliers as observed in Figure 8.
- QS Execution Time: With the exception of the density weighted QSs, in general, the selective sampling functions hold a low execution time. For the density weighted QSs it is observed a significant increase in the QS execution time due to the calculation of the density weight which requires the calculation of each sample’s NNs. This process ultimately increases the algorithm’s computational complexity and execution time. The execution times were obtained using a E3-1285 v6 @ 4.10GHz CPU and 16 GB of RAM.Due to the coherent high accuracy performance, surpassing PL, and its low execution time and computational complexity, Margin Sampling was selected as the most suitable QS to be included in both the AL and SSAL frameworks. Hence, forthcoming result presentations on this section were achieved using Margin Sampling. Besides the algorithm’s performance analysis, it is also worth to mention a comparison between the amount of labelled data for SL and AL. From 100% of the validation set annotated in SL, to, approximately 2.8 (0.1)% and 13.9 (0.5)%, for the UCI and CADL dataset, corresponding to the annotation of 250 samples and a reduction of 97.2 (0.1)% and 86.1 (0.5)% in the validation set annotation cost, respectively. These results confirm the applicability of AL in the context of HAR and its efficiency in reducing the annotation effort required to construct a higy confident classifier.
4.5. Active Learning Semi-Supervised Analysis
- Accuracy: Experimental results demonstrated that with the exception of the SSAL methods using the DTW or TAM distance, the accuracy of the proposed methods converges to the results of the SL technique. Figure 9 presents the classifier’s accuracy for the SSAL methods throughout the AL iterations. For each method, in every iteration the model training set grows, resulting in the increase of the classification’s accuracy.
- Automated Annotation Percentage (Aut Ann): Consists of the percentage of samples automatically annotated in relation to the total validation set size. Figure 10 displays the evolution on the percentage of the validation set unlabelled samples for the SSAL methods throughout the AL cycle iterations. In the AL and PL, the oracle annotates one sample per iteration, therefore, in Figure 10, both present an overlapping linear decline in the number of unlabelled samples. The NN-SSAL methods annotate six samples per iteration, one by the oracle and five by the automatic annotator, therefore, these show in Figure 10 an overlapping linear decline with higher slope than AL and PL. On the other hand, rNN-SSAL presents a curved decline in the number of unlabelled samples, outperforming the remaining during the first iterations. ST-SSAL displays during the initial iterations an automatic annotation percentage similar to AL and PL, with only the expert annotator labelling new samples and no automatic annotation, since the 0.98 prediction confidence threshold required for automatic annotation is not reached due to the classifier small labelled training set. Once the labelled set becomes representative of the dataset, the 0.98 threshold is reached and ST-SSAL automatic annotation increases exponentially until the unlabelled dataset becomes exhausted, easily surpassing the 5 constantly automatically annotated by the NN-SSAL. On the whole, ST-SSAL attains the highest performance for the UCI dataset, and NN-SSAL for the CADL dataset, the latter closely followed by rNN-SSAL.
- Automated Annotation Accuracy (Ann Acc): Consists of the percentage of correctly automatically annotated samples. Moreover, Figure 11 presents the evolution throughout the AL process of the automated annotation accuracy for the SSAL methods. As observed, ST-SSAL outperforms the remaining, attaining high results, especially for the latter iterations. ST-SSAL high annotation accuracy on the latter iterations results from the threshold required for the automatic annotation to be performed. As noted, this threshold is only reached during the latter iterations when the model training set becomes representative of the dataset and predictions can be performed with high certainty. This fact contrasts with the remaining methods, where higher results are obtained during the first iterations. For the NN-SSAL methods, this is justified by the queried sample propagating its label to closer samples during the first iterations. Whereas in the latter iterations, its closest neighbours start to be already annotated so the sample’s label is given to further away samples. The same is applied to rNN-SSAL, with the stabilisation of the propagation accuracy being accompanied by the stabilisation of the amount of automatic propagation (Figure 10). Additionally, this metric allows to discriminate between the performance of the different distance functions. As it can be seen, the Euclidean distance and Cosine similarity obtained similar results. In contrast to DTW and TAM, presenting a poor percentage of correctly annotated samples, explaining their low classification performance.
- Execution time: AL shows the fastest execution time. The algorithm execution time, allows to favour between the different similarity measures, since the DTW and TAM expensive time and computational complexity, render those algorithms non-applicable to a viable solution. Furthermore, comparing the presented four distance metrics, Euclidean distance presents the lowest time expense and, therefore, was chosen as the most suitable distance metric.
4.6. Stopping Criterion Analysis
5. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Patel, S.; Park, H.; Bonato, P.; Chan, L.; Rodgers, M. A review of wearable sensors and systems with application in rehabilitation. J. NeuroEng. Rehabil. 2012, 9, 21. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ramasamy Ramamurthy, S.; Roy, N. Recent trends in machine learning for human activity recognition—A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1254. [Google Scholar] [CrossRef]
- Organisation, W.H. Global Strategy on Diet, Physical Activity and Health. Available online: http://www.who.int/dietphysicalactivity/pa/en/ (accessed on 19 August 2018).
- del Rosario, M.B.; Redmond, S.J.; Lovell, N.H. Tracking the Evolution of Smartphone Sensing for Monitoring Human Movement. Sensors 2015, 15, 18901–18933. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Lara, O.D.; Labrador, M.A. A Survey on Human Activity Recognition using Wearable Sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef] [Green Version]
- Zhang, W.; Zhang, Z.; Qi, D.; Liu, Y. Automatic Crack Detection and Classification Method for Subway Tunnel Safety Monitoring. Sensors 2014, 14, 19307–19328. [Google Scholar] [CrossRef] [Green Version]
- Bulling, A.; Blanke, U.; Schiele, B. A Tutorial on Human Activity Recognition Using Body-worn Inertial Sensors. ACM Comput. Surv. 2014, 46, 33:1–33:33. [Google Scholar] [CrossRef]
- Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Attal, F.; Mohammed, S.; Dedabrishvili, M.; Chamroukhi, F.; Oukhellou, L.; Amirat, Y. Physical Human Activity Recognition Using Wearable Sensors. Sensors 2015, 15, 31314–31338. [Google Scholar] [CrossRef] [Green Version]
- Silva, J.; Monteiro, M.; Sousa, F. Human activity classification with inertial sensors. Stud. Health Technol. Inform. 2014, 200, 101–104. [Google Scholar]
- Figueira, C.; Matias, R.; Gamboa, H. Body Location Independent Activity Monitoring. In Proceedings of the BIOSIGNALS, Rome, Italy, 21–23 February 2016. [Google Scholar]
- Romera-Paredes, B.; Aung, M.; Bianchi-Berthouze, N. A One-vs-One Classifier Ensemble with Majority Voting for Activity Recognition. In Proceedings of the ESANN 2013 proceedings 21st European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013; pp. 443–448. [Google Scholar]
- Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. A Public Domain Dataset for Human Activity Recognition using Smartphones. In Proceedings of the ESANN 2013 proceedings 21st European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
- Machado, I.P.; Gomes, A.L.; Gamboa, H.; Paixão, V.; Costa, R.M. Human activity data discovery from triaxial accelerometer sensor: Non-supervised learning sensitivity to feature extraction parametrization. Inf. Process. Manag. 2015, 51, 204–214. [Google Scholar] [CrossRef]
- Luísa Gomes, A.; Paixão, V.; Gamboa, H. Human Activity Recognition Based on Novel Accelerometry Features and Hidden Markov Models Application. In Proceedings of the International Joint Conference on Biomedical Engineering Systems and Technologies (BIOSTEC 2015), Lisbon, Portugal, 12–15 January 2015; SCITEPRESS—Science and Technology Publications, Lda: Setúbal, Portugal, 2015; Volume 4, pp. 76–85. [Google Scholar] [CrossRef]
- Murad, A.; Pyun, J.Y. Deep Recurrent Neural Networks for Human Activity Recognition. Sensors 2017, 17, 2556. [Google Scholar] [CrossRef]
- Ordóñez, F.J.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed]
- Li, F.; Shirahama, K.; Nisar, M.A.; Köping, L.; Grzegorzek, M. Comparison of Feature Learning Methods for Human Activity Recognition Using Wearable Sensors. Sensors 2018, 18, 679. [Google Scholar] [CrossRef] [PubMed]
- Stikic, M.; Larlus, D.; Ebert, S.; Schiele, B. Weakly Supervised Recognition of Daily Life Activities with Wearable Sensors. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2521–2537. [Google Scholar] [CrossRef] [PubMed]
- Hossain, H.M.; Khan, M.A.A.H.; Roy, N. Active learning enabled activity recognition. Pervasive Mob. Comput. 2017, 38, 312–330. [Google Scholar] [CrossRef]
- Liu, R.; Chen, T.; Huang, L. Research on human activity recognition based on active learning. In Proceedings of the 2010 International Conference on Machine Learning and Cybernetics, Qingdao, China, 11–14 July 2010; Volume 1, pp. 285–290. [Google Scholar] [CrossRef]
- Cruciani, F.; Cleland, I.; Nugent, C.; McCullagh, P.; Synnes, K.; Hallberg, J. Automatic Annotation for Human Activity Recognition in Free Living Using a Smartphone. Sensors 2018, 18, 2203. [Google Scholar] [CrossRef] [PubMed]
- Tonkin, E.L.; Burrows, A.; Woznowski, P.; Laskowski, P.; Yordanova, K.Y.; Twomey, N.; Craddock, I.J. Talk, Text, Tag? Understanding Self-Annotation of Smart Home Data from a User’s Perspective. Sensors 2018, 18, 2365. [Google Scholar] [CrossRef] [PubMed]
- Schröder, M.; Yordanova, K.; Bader, S.; Kirste, T. Tool Support for the Online Annotation of Sensor Data. In Proceedings of the 3rd International Workshop on Sensor-Based Activity Recognition and Interaction ( iWOAR ’16), Rostock, Germany, 23–24 June 2016; ACM: New York, NY, USA, 2016; pp. 9:1–9:7. [Google Scholar] [CrossRef]
- Cleland, I.; Han, M.; Nugent, C.; Lee, H.; McClean, S.; Zhang, S.; Lee, S. Evaluation of Prompted Annotation of Activity Data Recorded from a Smart Phone. Sensors 2014, 14, 15861–15879. [Google Scholar] [CrossRef] [Green Version]
- Shahmohammadi, F.; Hosseini, A.; King, C.E.; Sarrafzadeh, M. Smartwatch Based Activity Recognition Using Active Learning. In Proceedings of the Second IEEE/ACM International Conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE ’17), Philadelphia, PA, USA, 17–19 July 2017; IEEE Press: Piscataway, NJ, USA, 2017; pp. 321–329. [Google Scholar] [CrossRef]
- Alemdar, H.; van Kasteren, T.L.M.; Ersoy, C. Using Active Learning to Allow Activity Recognition on a Large Scale. In Proceedings of the International Joint Conference on Ambient Intelligence, Amsterdam, The Netherlands, 16–18 November 2011; Keyson, D.V., Maher, M.L., Streitz, N., Cheok, A., Augusto, J.C., Wichert, R., Englebienne, G., Aghajan, H., Kröse, B.J.A., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 105–114. [Google Scholar]
- Zhu, J.; Wang, H.; Tsou, B.K.; Ma, M. Active Learning With Sampling by Uncertainty and Density for Data Annotations. IEEE Trans. Audio Speech Lang. Process. 2010, 18, 1323–1331. [Google Scholar] [CrossRef]
- Zhu, J.; Wang, H.; Yao, T.; Tsou, B.K. Active Learning with Sampling by Uncertainty and Density for Word Sense Disambiguation and Text Classification. In Proceedings of the 22nd International Conference on Computational Linguistics (COLING ’08), Manchester, UK, 18–22 August 2008; Association for Computational Linguistics: Stroudsburg, PA, USA, 2008; Volume 1, pp. 1137–1144. [Google Scholar]
- He, G.; Li, Y.; Zhao, W. An uncertainty and density based active semi-supervised learning scheme for positive unlabeled multivariate time series classification. Knowl.-Based Syst. 2017, 124, 80–92. [Google Scholar] [CrossRef]
- Stikic, M.; Van Laerhoven, K.; Schiele, B. Exploring semi-supervised and active learning for activity recognition. In Proceedings of the 12th IEEE International Symposium on Wearable Computers, Pittaburgh, PA, USA, 28 September–1 October 2008; Volume 1, pp. 81–88. [Google Scholar] [CrossRef]
- Zhu, X.; Goldberg, A.B. Introduction to semi-supervised learning. Synth. Lect. Artif. Intell. Mach. Learn. 2009, 3, 1–130. [Google Scholar] [CrossRef]
- Fu, Y.; Zhu, X.; Li, B. A survey on instance selection for active learning. Knowl. Inf. Syst. 2013, 35, 249–283. [Google Scholar] [CrossRef]
- Settles, B. Active Learning Literature Survey; Computer Sciences Technical Report 1648; University of Wisconsin–Madison: Madison, WI, USA, 2009. [Google Scholar]
- Aggarwal, C.C.; Kong, X.; Gu, Q.; Han, J.; Yu, P.S. Active Learning: A Survey; Technical Report; CRC Press: Boca Raton, FL, USA, 2014. [Google Scholar]
- Chen, Y.; Mani, S.; Xu, H. Applying active learning to assertion classification of concepts in clinical text. J. Biomed. Inform. 2012, 45, 265–272. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Bloodgood, M.; Vijay-Shanker, K. A Method for Stopping Active Learning Based on Stabilizing Predictions and the Need for User-adjustable Stopping. In Proceedings of the Thirteenth Conference on Computational Natural Language Learning (CoNLL ’09), Boulder, CO, USA, 4–5 June 2009; Association for Computational Linguistics: Stroudsburg, PA, USA, 2009; pp. 39–47. [Google Scholar]
- Zhu, J.; Wang, H.; Hovy, E. Multi-criteria-based Strategy to Stop Active Learning for Data Annotation. In Proceedings of the 22nd International Conference on Computational Linguistics (COLING ’08), Manchester, UK, 18–22 August 2008; Association for Computational Linguistics: Stroudsburg, PA, USA, 2008; Volume 1, pp. 1129–1136. [Google Scholar]
- He, G.; Duan, Y.; Li, Y.; Qian, T.; He, J.; Jia, X. Active Learning for Multivariate Time Series Classification with Positive Unlabeled Data. In Proceedings of the 2015 IEEE 27th International Conference on Tools with Artificial Intelligence (ICTAI), Vietri sul Mare, Italy, 9–11 November 2015; pp. 178–185. [Google Scholar] [CrossRef]
- Wei, L.; Keogh, E. Semi-supervised Time Series Classification. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’06), Philadelphia, PA, USA, 20–23 August 2006; ACM: New York, NY, USA, 2006; pp. 748–753. [Google Scholar] [CrossRef]
- Folgado, D.; Barandas, M.; Matias, R.; Martins, R.; Carvalho, M.; Gamboa, H. Time Alignment Measurement for Time Series. Pattern Recognit. 2018, 81, 268–279. [Google Scholar] [CrossRef]
- Müller, M. Dynamic Time Warping; Springer: Berlin/Heidelberg, Germany, 2007; pp. 69–84. [Google Scholar]
- Danka, T.; Horvath, P. modAL: A Modular Active Learning Framework for Python. Available online: https://github.com/cosmic-cortex/modAL (accessed on 26 November 2018).
- Reyes-Ortiz, J.L.; Ghio, A.; Parra, X.; Anguita, D.; Cabestany, J.; Català, A. Human Activity and Motion Disorder Recognition: Towards smarter Interactive Cognitive Environments. In Proceedings of the ESANN 2013 proceedings 21st European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
Datasets | ||
---|---|---|
UCI HAR Using Smartphones | Continuous Activities of Daily Living | |
N° of Users | 30 | 12 |
Activities | Laying, sitting, standing, walking, upstairs and downstairs | Laying, sitting, standing, running, walking, upstairs and downstairs |
Sensors | Accelerometer (50 Hz), gyroscope (50 Hz) | Accelerometer (100 Hz), gyroscope (100 Hz), barometer (30 Hz) |
Device | Samsung Galaxy S2 (waist) | Samsung S5 (right hand) and wearable sensor (left hand, right ankle and right side of the waist) |
N° of Samples | 10,299 | 2047 |
(a) Supervised Learning | ||
Dataset | ||
Supervised Learning Method | UCI | CADL |
Nearest Neighbours | 91.0 (1.9) | 83.6 (3.4) |
Decision Tree | 87.4 (3.5) | 83.6 (4.3) |
Random Forest | 91.4 (2.4) | 89.1 (4.0) |
SVM | 90.7 (2.6) | 77.6 (3.3) |
AdaBoost | 40.8 (6.8) | 54.1 (4.6) |
Naive Bayes | 88.9 (2.9) | 75.9 (3.1) |
QDA | 90.8 (2.7) | 79.0 (3.7) |
(b) Unsupervised Learning | ||
Dataset | ||
Unsupervised Learning Method | UCI | CADL |
K-Means | 52.1 (4.3) | 50.9 (6.1) |
Mini Batch K-Means | 50.7 (5.5) | 50.5 (5.3) |
Spectral Clustering | 57.8 (3.5) | 61.9 (8.9) |
Gaussian Mixture | 49.8 (2.7) | 58.9 (6.6) |
DBSCAN | 16.4 (7.2) | 13.9 (6.5) |
UCI Dataset | CADL Dataset | |||
---|---|---|---|---|
Query Strategy | Accuracy in % | QS Time in s | Accuracy in % | QS Time in s |
Local Density * Least Confident | 87.6 (4.0) | 27.2 (5.1) | 83.5 (6.7) | 0.9 (0.1) |
Least Confident | 87.9 (3.7) | 0.1 (0.1) | 72.0 (5.0) | 0.1 (0.1) |
Local Density * Entropy | 85.7 (3.7) | 22.5 (0.4) | 80.3 (7.9) | 0.9 (0.1) |
Entropy | 87.1 (2.5) | 0.1 (0.1) | 70.9 (6.0) | 0.1 (0.1) |
Local Density * Margin | 52.5 (8.6) | 22.6 (0.5) | 32.8 (8.0) | 0.9 (0.1) |
Margin | 88.4 (2.8) | 0.1 (0.1) | 84.8 (7.0) | 0.1 (0.1) |
Local Density | 63.6 (5.9) | 22.5 (0.4) | 68.9 (8.5) | 0.9 (0.1) |
Passive Learning | 87.0 (4.5) | 0.1 (0.1) | 82.0 (7.6) | 0.1 (0.1) |
Supervised Learning | 91.4 (2.4) | 89.1 (4.0) | ||
Unsupervised Learning | 57.8 (3.5) | 61.9 (8.9) |
UCI Dataset | CADL Dataset | |||||||
---|---|---|---|---|---|---|---|---|
Method | Accuracy in % | Aut Ann in % | Ann Acc in % | Time in s | Accuracy in % | Aut Ann in % | Ann Acc in % | Time in s |
NN_Euc | 88.1 (2.7) | 13.5 (0.1) | 76.8 (1.0) | 92.3 (9.4) | 82.5 (5.7) | 68.2 (2.7) | 68.2 (1.0) | 44.3 (4.8) |
NN_Cos | 89.4 (3.0) | 13.5 (0.1) | 75.3 (1.4) | 860.9 (7.4) | 82.8 (8.2) | 68.2 (2.7) | 64.6 (1.7) | 79.1 (4.2) |
NN_DTW | 70.3 (6.0) | 13.5 (0.1) | 39.7 (1.7) | 6772.0 (4.5) | 68.4 (5.4) | 68.2 (2.7) | 30.3 (3.1) | 6735.4 (1.3) |
NN_TAM | 75.2 (4.8) | 13.5 (0.1) | 44.1 (4.0) | 6771.4 (5.4) | 68.8 (7.3) | 68.2 (2.7) | 31.6 (1.2) | 6735.4 (1.9) |
rNN_Euc | 85.4 (2.7) | 33.3 (2.0) | 77.3 (3.8) | 437.9 (61.9) | 74.9 (7.8) | 56.6 (1.3) | 66.5 (2.7) | 60.0 (10.8) |
rNN_Cos | 82.5 (3.6) | 37.7 (4.0) | 74.7 (5.2) | 1165.3 (86.4) | 71.3 (8.1) | 61.7 (2.5) | 59.5 (5.6) | 89.7 (11.2) |
rNN_DTW | 65.1 (4.9) | 13.7 (1.6) | 41.8 (2.9) | 6995.4 (74.9) | 77.3 (7.1) | 20.8 (0.8) | 39.9 (1.9) | 6739.2 (7.6) |
rNN_TAM | 64.5 (8.4) | 11.9 (3.9) | 45.7 (4.9) | 6968.3 (81.2) | 81.8 (8.4) | 11.2 (2.0) | 41.3 (4.6) | 6733.8 (7.3) |
ST-SSAL | 84.0 (6.3) | 56.7 (11.6) | 86.1 (10.5) | 99.3 (17.5) | 84.8 (7.0) | 20.9 (6.9) | 92.5 (2.7) | 11.3 (1.0) |
AL | 88.4 (2.8) | 23.6 (5.0) | 84.8 (7.0) | 12.0 (1.0) | ||||
PL | 87.0 (4.5) | 23.2 (1.4) | 82.0 (7.6) | 10.3 (1.0) | ||||
SL | 91.4 (2.4) | 0.7 (0.1) | 89.1 (4.0) | 0.1 (0.1) | ||||
UL | 57.8 (3.5) | 0.3 (0.2) | 61.9 (8.9) | 0.1 (0.1) |
UCI | CADL | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Method | SP | Max-Conf | Over-Unc | CC | Max-CC | Over-CC | SP | Max-Conf | Over-Unc | CC | Max-CC | Over-CC | |
NN_Euc | Acc N.it | 85.5 (3.3) 68.5 (26.1) | 69.2 (12.5) 20.0 (11.4) | 78.7 (8.7) 37.5 (9.5) | 73.6 (11.9) 28.9 (12.4) | 82.2 (8.2) 68.0 (39.3) | 81.5 (9.2) 82.5 (44.1) | 81.6 (5.1) 69.6 (38.4) | 68.2 (8.9) 44.5 (23.4) | 76.7 (6.3) 64.5 (14.3) | 57.3 (14.1) 21.2 (8.2) | 77.9 (6.0) 123.0 (97.5) | 80.2 (6.4) 103.0 (73.8) |
NN_Cos | Acc N.it | 79.6 (7.3) 61.1 (29.7) | 70.3 (11.3) 22.0 (13.2) | 79.0 (9.6) 41.5 (11.9) | 68.8 (13.3) 16.8 (7.6) | 81.5 (10.4) 76.5 (47.5) | 82.7 (10.8) 67.5 (27.0) | 74.6 (15.6) 80.5 (31.4) | 62.8 (16.9) 28.0 (10.0) | 69.8 (9.9) 52.5 (15.8) | 44.2 (20.5) 13.5 (6.8) | 74.9 (11.0) 88.5 (65.1) | 78.4 (10.5) 88.5 (56.9) |
NN_DTW | Acc N.it | 67.0 (4.3) 250.0 (0.0) | 54.3 (6.0) 26.0 (7.4) | 47.8 (12.4) 16.0 (2.2) | 57.6 (244.7) 244.7 (155.0) | 67.4 (7.4) 297.7 (105.9) | 58.1 (17.5) 220.4 (158.8) | 58.7 (12.5) 117.0 (95.1) | 43.9 (7.6) 22.5 (4.5) | 33.8 (7.3) 16.0 (5.5) | 52.0 (17.0) 215.1 (134.5) | 62.8 (4.0) 278.3 (79.9) | 64.5 (7.3) 305.3 (11.2) |
NN_TAM | Acc N.it | 77.0 (4.9) 313.2 (78.3) | 59.5 (9.0) 37.5 (17.0) | 68.0 (7.9) 53.5 (10.8) | 63.8 (20.0) 244.6 (153.2) | 73.1 (7.0) 318.1 (95.7) | 67.3 (12.4) 260.8 (136.7) | 62.0 (6.0) 153.0 (86.3) | 49.5 (8.0) 47.5 (23.9) | 50.2 (8.4) 47.5 (14.8) | 60.9 (12.7) 274.9 (90.0) | 63.9 (5.2) 305.3 (11.2) | 65.8 (5.4) 305.3 (11.2) |
rNN_Euc | Acc N.it | 84.2 (2.3) 92.5 (23.0) | 72.6 (12.4) 37.5 (17.6) | 83.7 (3.9) 97.0 (37.6) | 61.9 (13.4) 15.9 (8.7) | 72.6 (13.9) 45.5 (27.4) | 84.9 (2.7) 129.4 (52.1) | 77.2 (5.9) 198.3 (133.3) | 59.9 (14.5) 32.5 (11.8) | 76.9 (9.0) 174.3 (97.2) | 45.0 (10.8) 14.0 (2.6) | 74.9 (7.0) 209.6 (125.4) | 78.1 (5.4) 319.1 (92.7) |
rNN_Cos | Acc N.it | 65.5 (9.9) 37.0 (10.4) | 60.6 (11.6) 31.0 (19.8) | 66.0 (8.2) 37.5 (15.2) | 52.7 (10.4) 9.7 (3.5) | 65.9 (13.3) 63.5 (71.9) | 72.0 (12.1) 66.5 (37.6) | 52.7 (16.1) 43.0 (21.5) | 53.4 (14.7) 32.0 (11.6) | 60.6 (11.6) 72.5 (34.2) | 34.2 (17.5) 12.3 (5.6) | 53.7 (9.4) 87.4 (101.3) | 60.5 (14.4) 165.7 (136.7) |
rNN_DTW | Acc N.it | 38.7 (15.4) 39.0 (27.2) | 44.6 (8.9) 21.0 (5.9) | 43.0 (6.8) 31.5 (13.2) | 29.2 (9.7) 7.4 (0.9) | 56.4 (13.9) 156.6 (158.0) | 56.1 (14.6) 133.2 (142.5) | 41.2 (14.8) 51.0 (32.9) | 35.4 (6.9) 25.0 (7.3) | 35.9 (6.3) 30.0 (13.0) | 23.4 (5.8) 9.3 (4.3) | 56.9 (20.9) 97.0 (94.2) | 61.2 (18.2) 132.0 (87.2) |
rNN_TAM | Acc N.it | 68.6 (9.4) 290.7 (118.6) | 43.1 (10.1) 17.0 (3.7) | 48.2 (7.0) 21.0 (5.5) | 40.5 (10.2) 12.2 (7.9) | 59.7 (16.7) 124.3 (137.2) | 50.1 (15.8) 128.2 (146.3) | 56.9 (19.6) 77.0 (34.4) | 29.8 (7.8) 18.0 (4.6) | 27.7 (7.4) 14.5 (5.5) | 33.1 (22.6) 22.0 (24.7) | 53.2 (19.0) 54.0 (29.6) | 47.9 (20.3) 128.2 (146.2) |
ST-SSAL | Acc N.it | 85.2 (3.5) 201.5 (83.8) | 48.8 (15.3) 22.5 (9.5) | 66.3 (9.9) 61.0 (29.2) | 49.1 (18.9) 66.4 (107.1) | 75.3 (10.7) 81.0 (41.6) | 84.5 (4.1) 214.0 (46.5) | 82.8 (6.6) 164.0 (70.1) | 33.0 (13.0) 25.0 (10.2) | 61.9 (12.1) 67.0 (21.1) | 15.0 (0.6) 12.0 (0.0) | 76.2 (12.7) 120.0 (28.4) | 84.7 (7.2) 182.0 (53.9) |
AL | Acc N.it | 86.1 (2.6) 109.5 (24.4) | 60.3 (11.7) 28.5 (10.1) | 65.7 (10.5) 38.0 (16.4) | 37.4 (11.1) 10.9 (3.9) | 84.0 (4.7) 98.0 (26.1) | 86.0 (5.6) 97.0 (40.8) | 84.6 (7.4) 193.0 (51.0) | 51.2 (15.5) 58.0 (17.9) | 64.0 (15.0) 118.0 (59.5) | 15.0 (0.6) 12.0 (0.0) | 76.8 (9.7) 116.0 (42.0) | 76.6 (16.5) 139.0 (43.1) |
SL | 91.4 (2.4) | 89.1 (4.0) | |||||||||||
UL | 57.8 (3.5) | 61.9 (8.9) |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Bota, P.; Silva, J.; Folgado, D.; Gamboa, H. A Semi-Automatic Annotation Approach for Human Activity Recognition. Sensors 2019, 19, 501. https://doi.org/10.3390/s19030501
Bota P, Silva J, Folgado D, Gamboa H. A Semi-Automatic Annotation Approach for Human Activity Recognition. Sensors. 2019; 19(3):501. https://doi.org/10.3390/s19030501
Chicago/Turabian StyleBota, Patrícia, Joana Silva, Duarte Folgado, and Hugo Gamboa. 2019. "A Semi-Automatic Annotation Approach for Human Activity Recognition" Sensors 19, no. 3: 501. https://doi.org/10.3390/s19030501