Human fall detection using segment-level CNN features and sparse dictionary learning
Paper i proceeding, 2017

This paper addresses issues in human fall detection from videos. Unlike using handcrafted features in the conventional machine learning, we extract features from Convolutional Neural Networks (CNNs) for human fall detection. Similar to many existing work using two stream inputs, we use a spatial CNN stream with raw image difference and a temporal CNN stream with optical flow as the inputs of CNN. Different from conventional two stream action recognition work, we exploit sparse representation with residual-based pooling on the CNN extracted features, for obtaining more discriminative feature codes. For characterizing the sequential information in video activity, we use the code vector from long-range dynamic feature representation by concatenating codes in segment-levels as the input to a SVM classifier. Experiments have been conducted on two public video databases for fall detection. Comparisons with six existing methods show the effectiveness of the proposed method.

E-healthcare

Deep learning

sparse dictionary learning

human fall detection

residual-based pooling

assisted living.

Convolutional Network

automatic feature learning

Författare

Chenjie Ge

Chalmers, Signaler och system, Signalbehandling och medicinsk teknik

Irene Yu-Hua Gu

Chalmers, Signaler och system, Signalbehandling och medicinsk teknik

Jie Yang

IEEE International Workshop on Machine Learning for Signal Processing, MLSP

21610363 (ISSN) 21610371 (eISSN)

6-

Styrkeområden

Transport

Livsvetenskaper och teknik (2010-2018)

Ämneskategorier

Människa-datorinteraktion (interaktionsdesign)

Signalbehandling

Datorseende och robotik (autonoma system)

DOI

10.1109/MLSP.2017.8168185

Mer information

Senast uppdaterat

2023-03-21