• Login
    • Library Home
    View Item 
    •   BracU IR
    • School of Data and Sciences (SDS)
    • Department of Computer Science and Engineering (CSE)
    • Thesis & Report, MSc (Computer Science and Engineering)
    • View Item
    •   BracU IR
    • School of Data and Sciences (SDS)
    • Department of Computer Science and Engineering (CSE)
    • Thesis & Report, MSc (Computer Science and Engineering)
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Efficient Spatio-temporal feature extraction for human action recognition

    Thumbnail
    View/Open
    19366007_CSE.pdf (15.50Mb)
    Date
    2021-11
    Publisher
    Brac University
    Author
    Ghosh, Dipon Kumar
    Metadata
    Show full item record
    URI
    http://hdl.handle.net/10361/15946
    Abstract
    Human actuation recognition (HAR) has been performed using current deep learning (DL) algorithms using a variety of input formats, including video footage, optical flow, and even skeleton points, which may be acquired via depth sensors or pose estimation technologies. Recent techniques, on the other hand, are computationally costly and have a high memory footprint, making them unsuitable for use in realworld environments. Furthermore, the design of existing techniques does not allow for the full extraction of spatial and temporal characteristics of an action, and as a result, information is lost throughout the recognition process. Here, we present a novel framework for action recognition that extracts spatial and temporal characteristics separately while reducing the amount of information lost by a substantial amount. The multi-dimensional convolutional network (MDCN) and the redefined spatio-temporal graph convolutional network (RSTCN) are two models developed in accordance with this framework. In both cases, spatial and temporal information are extracted irrespective of the precise spatio-temporal location. Our approach was evaluated in two particular aspects of human action recognition, namely violence detection and skeleton-based action recognition, in order to ensure that our models were accurate and reliable. In spite of being cost e↵ective and having less parameters, our proposed MDCN achieved 87.5% accuracy in the largest violence detection benchmark dataset and RST-GCN obtained 92.2% accuracy on the skeleton dataset. The performance of our models edge devices with limited resources, which are suitable for deploying at real-world environments is also also analyze and compare, such as surveillance system and smart healthcare system. The proposed MDCN model processes 80 frames per second on edge device such as, Nvidia Jetson Nano and RST-GCN performs at a speed of 993 frames per second. Our proposed methods o↵er a strong balance between accuracy, memory consumption, and processing time, which make them suitable for deploying at real-world environments.
    Keywords
    Human action recognition (HAR); Surveillance systems; Violence detection; Skeleton-based human action recognition; Convolutional neural network (CNN); Graph convolutional networks (GCN); Feature fusion
     
    LC Subject Headings
    Human activity recognition; Neural network (Computer Science)
     
    Description
    This thesis is submitted in partial fulfilment of the requirements for the degree of Master of Engineering in Computer Science and Engineering, 2021.
     
    Cataloged from PDF version of thesis.
     
    Includes bibliographical references (pages 67-75).
    Department
    Department of Computer Science and Engineering, Brac University
    Collections
    • Thesis & Report, MSc (Computer Science and Engineering)

    Copyright © 2008-2019 Ayesha Abed Library, Brac University 
    Contact Us | Send Feedback
     

     

    Policy Guidelines

    • BracU Policy
    • Publisher Policy

    Browse

    All of BracU Institutional RepositoryCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

    My Account

    LoginRegister

    Statistics

    View Usage Statistics

    Copyright © 2008-2019 Ayesha Abed Library, Brac University 
    Contact Us | Send Feedback