4.8 Article

Physical Activity Recognition With Statistical-Deep Fusion Model Using Multiple Sensory Data for Smart Health

期刊

IEEE INTERNET OF THINGS JOURNAL
卷 8, 期 3, 页码 1533-1543

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JIOT.2020.3013272

关键词

Feature extraction; Support vector machines; Activity recognition; Medical services; Correlation; Internet of Things; Machine learning; Deep learning; intermediate fusion; physical activity (PA) recognition; wearable sensor system

资金

  1. National Research Foundation of Korea (NRF) through the Creativity Challenge Research-Based Project [2019R1I1A1A01063781]
  2. Priority Research Centers Program through the NRF of Korea - Ministry of Education, Science, and Technology [2018R1A6A1A03024003]
  3. National Research Foundation of Korea [2019R1I1A1A01063781] Funding Source: Korea Institute of Science & Technology Information (KISTI), National Science & Technology Information Service (NTIS)

向作者/读者索取更多资源

This study proposes a fusion model for activity recognition by combining deep convolutional neural networks with traditional handcrafted features, achieving high accuracy in activity recognition in multisensor systems, outperforming other state-of-the-art approaches.
Nowadays, enhancing the living standard with smart healthcare via the Internet of Things is one of the most critical goals of smart cities, in which artificial intelligence plays as the core technology. Many smart services, deployed according to wearable sensor-based physical activity recognition, have been able to early detect unhealthy daily behaviors and further medical risks. Numerous approaches have studied shallow handcrafted features coupled with traditional machine learning (ML) techniques, which find it difficult to model real-world activities. In this work, by revealing deep features from deep convolutional neural networks (DCNNs) in fusion with conventional handcrafted features, we learn an intermediate fusion framework of human activity recognition (HAR). According to transforming the raw signal value to pixel intensity value, segmentation data acquired from a multisensor system are encoded to an activity image for deep model learning. Formulated by several novel residual triple convolutional blocks, the proposed DCNN allows extracting multiscale spatiotemporal signal-level and sensor-level correlations simultaneously from the activity image. In the fusion model, the hybrid feature merged from the handcrafted and deep features is learned by a multiclass support vector machine (SVM) classifier. Based on several experiments of performance evaluation, our fusion approach for activity recognition has achieved the accuracy over 96.0% on three public benchmark data sets, including Daily and Sport Activities, Daily Life Activities, and RealWorld. Furthermore, the method outperforms several state-of-the-art HAR approaches and demonstrates the superiority of the proposed intermediate fusion model in multisensor systems.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据