论文标题
合并短期空间/频率特征提取和长期INDRNN的框架以进行活动识别
A Framework of Combining Short-Term Spatial/Frequency Feature Extraction and Long-Term IndRNN for Activity Recognition
论文作者
论文摘要
如今,随着智能手机的普及,基于智能手机传感器的人类活动识别正在吸引日益增长的兴趣。由于智能手机传感器的采样率很高,这是一个高度长时间的时间识别问题,尤其是在较大的阶层距离的情况下,例如在不同位置(例如在袋子中或身体上)带有的智能手机,以及诸如乘坐火车或地铁之类的小阶层距离。为了解决这个问题,我们提出了一个新的框架,以结合短期空间/频率特征提取和长期独立复发的神经网络(INDRNN),以进行活动识别。考虑到传感器数据的周期性特征,首先在空间和频域中提取短期时间特征。然后,Indrnn能够捕获长期模式,用于进一步获得分类的长期特征。鉴于在不同位置携带智能手机时存在较大的差异,首先开发基于组的位置识别,以查明智能手机的位置。 SHL挑战的Sussex-Huawei运动(SHL)数据集用于评估。提出的方法的较早版本已在2020年SHL Challenge赢得了第二名(如果不考虑多种型号融合方法,则获得了第一名)。本文进一步改进了所提出的方法,并实现了80.72 $ \%$的精度,比使用单个模型的现有方法更好。
Smartphone sensors based human activity recognition is attracting increasing interests nowadays with the popularization of smartphones. With the high sampling rates of smartphone sensors, it is a highly long-range temporal recognition problem, especially with the large intra-class distances such as the smartphones carried at different locations such as in the bag or on the body, and the small inter-class distances such as taking train or subway. To address this problem, we propose a new framework of combining short-term spatial/frequency feature extraction and a long-term Independently Recurrent Neural Network (IndRNN) for activity recognition. Considering the periodic characteristics of the sensor data, short-term temporal features are first extracted in the spatial and frequency domains. Then the IndRNN, which is able to capture long-term patterns, is used to further obtain the long-term features for classification. In view of the large differences when the smartphone is carried at different locations, a group based location recognition is first developed to pinpoint the location of the smartphone. The Sussex-Huawei Locomotion (SHL) dataset from the SHL Challenge is used for evaluation. An earlier version of the proposed method has won the second place award in the SHL Challenge 2020 (the first place if not considering multiple models fusion approach). The proposed method is further improved in this paper and achieves 80.72$\%$ accuracy, better than the existing methods using a single model.