SlideShare a Scribd company logo
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2324
Recurrent Neural Network for Human Action Recognition using
Star Skeletonization
Nithin Paulose1, M. Muthukumar2, S. Swathi3, M. Vignesh4
1,2B.E.Computer Science and Engineering, Dhaanish Ahmed Institute of Technology, Coimbatore, Tamil Nadu, India
3,4Assistant Professor, Dept. of Computer Science and Engineering, Dhaanish Ahmed Institute of Technology,
Coimbatore, Tamil Nadu, India
---------------------------------------------------------------------***----------------------------------------------------------------------
Abstract - This project presents a Recurrent Neural
Network (RNN) methodology for Human Action Recognition
using star skeleton as a representative descriptor of human
posture. Star skeleton is a fast skeletonization technique by
connecting from geometric center of target object to contour
extremes. For the action recognition using the feature star
skeleton, we clearly define the feature as a five-dimensional
vector in star fashion because the head and four limbs are
usually local extremes of human shape. In our project we
assumed an action is composed of a series of star skeletons
overtime. Therefore, the images whicharetime-sequential are
expressing human action that is transformed into a feature
vector sequence. Then the feature vector sequence must be
transformed into symbol sequence so that RNN can model the
action. We used RNN because the features extracted are time
dependent.
Key Words: Recurrent Neural Network (RNN), star skeleton,
contour extremes, Human Action Recognition, five-
dimensional vector, time-sequential.
1. INTRODUCTION
Human activity recognition is an important task for
Ambient Intelligence systems. The state of a person is to be
recognized, which provides us with valuable information
that is been used as input for other systems. For example, in
health care, fall detection can be used to alert the medical in
case of an accident; in security, abnormal behavior can be
detected and thus used to prevent a burglary or other
criminal activities. Human motion analysis is currently
receiving increasing attention from computer vision
researchers. For example, Human body segmentation in an
image, the movement of joints are tracked in an image
sequence, and the analysisofathletic performanceisdone by
recovering underlying 3D body structure , also used for
medical diagnostics. Other applications include building
man-machine user interfaces and video conferencing.
The goal of human activity recognition is to
automatically analyses on -goingactivitiesfromanunknown
video. The objective of the system is to correctly classify the
video into its activity category, for example where a video is
segmented to contain only one executionofa humanactivity.
The starting and ending times of all occurring activitiesfrom
an input video is detected, from which the continuous
recognition of human activities are performed. The
constructions of several important applications are
constructed from the videos which has the ability to
recognize complex humanactivities.Automatedsurveillance
systems in public places like airports and subway stations
require detection of abnormal and suspicious activities, as
opposed to normal activities. For example, an automatic
recognize of suspicious activities like “a personleavesa bag”
or “a person places his/her bag in a trash bin” in an airport
surveillance systemmustberecognized.Usingrecognitionof
human activity the real-time monitoring of patients,
children, and elderly persons can be done. By using activity
recognition the construction of gesture-based human
computer interfaces and vision-based intelligent
environments becomes possible.
There are various types of human activities.
Depending on their complexity, they can be conceptually
categorized into four different levels: gestures, actions,
interactions, and group activities.
1.1 HUMAN ACTIVITY RECOGNITION FROM VIDEO
SEQUENCES
Human activity recognition role is that human-to-
human interaction and interpersonal relations. HAR
provides information about the identity, personality, and
psychological state that is difficult to extract. In various
classification techniques two main questions emerge:
“Where it is in the video?” and “What is the action?” To
recognize human activities one must determine the active
states of a person, to recognize the efficiency. “Walking” is
the daily human activity to recognize. The complexactivities
such as “peeling an apple” are more difficult to identify. The
easier way to recognize is to simplify the complex activities
into other simpler activities. For the better understanding of
human activities the detectionofobjects mayprovideuseful
information about the ongoing event.
The human activity recognition assumes a figure-
centric scene of uncluttered background, where the actor is
free to perform an activity. The challenging task to classifya
person’s activities with low error in a fully automated
human activity recognition system are background clutter,
partial occlusion, changes in scale, viewpoint, lighting and
appearance, and frame resolution. Moreover, commenting
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2325
behavioral roles is time consuming and requires knowledge
of the specific event. However, intra- and inter-class
similarities make the problem more challenging, that is ,the
same action done by two people may finding difficult to
detect their individual action . The human activity is based
on their habits, and this make s so difficulttodeterminetheir
activity. In real time the challenging tasks is that, the
construction of visual model for learning and analyzing
human movements with inadequate benchmark datasets for
evaluation.
Fig 1: Architecture diagram
1.2 HUMAN ACTIVITY CATEGORIZATION
The human activity recognition methods are
classified into two main categories: (i) uni-modal and (ii)
multimodal activity recognition methods according to the
nature of sensor data they employ. Then, these two
categories is further break down into sub-categories
depending on how they model human activities.
Uni-modal methodsrepresenthumanactivitiesfrom
data of a single modality, such as images, and they are
further categorized as: (i) space-time, (ii) stochastic, (iii)
rule-based, and (iv) shape based methods. Space-time
methods involve activity recognition methods, which
represent human activities as a set of spatio-temporal
features or trajectories. By applying statistical models to
represent human actions stochastic methods are used. The
modelings the motion of human body parts are shape-based
methods efficiently represent activities with high level
reasoning.
Multimodal methods combine features collected
from different sources and are classified into three
categories: (i) affective, (ii) behavioral, and (iii) social
networking methods. Emotional communications and the
affective state of a person represent human activity.
Behavioral methods aim to recognize behavioral attributes,
non-verbal multimodal cues such as gestures, facial
expressions and auditory cues.Thecharacteristicsaremodel
by social networking methods .It model the characteristics
and the behavior of humans in several layers of human-to-
human interactions in social events from gestures, body
motion, and speech. The activity and behavior are the two
terms which are used interchangeably. These two terms
activity and behavior is used to describe a sequence of
actions that correspond to specific body motion and to
characterize both activities and events that are associated
with facial expression, emotional states, and gestures with
single person auditory cues.
1.3 HUMAN ACTIVITY RECOGNITION MODEL
The human activity recognition has largely focused
on statistical methods using spatio-temporal features. The
typical model consists of spatio-temporal interest-points
which are detected in the video sequence and the local
maxima become the centerpointofa spatio-temporal region.
Features are then extracted from the spatio-temporal region
(such as features based on optical flow or gradient values)
and summarized or histogram med to form a feature
descriptor. The feature descriptors are used to form a code
book, typically followed by a bag of visual words model
adapted from statistical natural language processing. While
methods based on spatio-temporal features are the most
common, other methods make use of other video features
such as medium-term tracking, volumetric representations
and graph-based features.
2. LITERATURE REVIEW
[1] A computational efficient action recognition
framework using depth motion maps (DMMs)-based local
binary patterns (LBPs) and kernel-based extreme learning
machine (KELM). In depth video sequence depth frames are
projected onto three orthogonal Cartesianplanestoformthe
projected images corresponding to three projection views
[front (f), side (s), and top (t) views].For calculating LBP
histogram an LBP operator is applied to each block and the
DMMs are divided into overlapped blocks. Feature-level
fusion and decision-level fusion approachesareinvestigated
using KELM [2] in this paper, we propose to use human
limbs to augment constraints between neighboring human
joints. We model a limb as a wide line to represent its shape
information. Instead of estimating its length and rotation
angle, we calculate as per-pixel likelihood for each human
limb by a ConvNet. [3] Video based action recognition is one
of the important and challenging problems in computer
vision re- search. The several realistic datasets are the
HMDB51, UCF50, and UCF101. BoVW is a general pipeline to
construct a global representation from local features
composed of five steps; (i) feature extraction, (ii) feature
pre-processing, (iii) codebook generation, (iv)feature
encoding, and (v) pooling and normalization.[4]For efficient
video representation on action recognition is shown by
dense and achieved state-of-the-art results on a variety of
datasets. The performance is corrected bytakingthecamera
motion into account. The feature points between frames
using SURF descriptors and dense optical flow are used to
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2326
estimate camera motion. With the help of RANSAC
homography is estimated. The performance can be
significantly improved by removing backgroundtrajectories
and warping optical flow with a robustly estimated
homograph approximating the camera motion. [5] The
discriminative features for action recognition Convolutional
Neural Networks (ConvNets) are employed into color
texture images that are referred to as skeleton optical
spectra. The learning of suitable dynamic features and
ConvNet architecture from skeleton sequences is possible
without training millions of parameters from this kind of
spectrum views.
3. DESIGN METHODS
There are some methods used for Human Action
Recognition using RNN.
3.1 HUMAN SILHOUETTE EXTRACTION
In this project we extract human body contour from
given image. In Frame Videos: To obtain the human body we
should take the direct difference between the background
and the current frame. Out Frame Videos: To extract the
human body from frames of the videos we used the inbuilt
Gaussian Mixture Model based Foreground detection
method.
Fig 3.1 Background Subtraction
3.2 HUMAN CONTOUR EXTRACTION
To extract the contour of a detected human body, at
thresholding and morphological method, the important
approaches in the field of image segmentation are to choose
a correct threshold and that is difficult under irregular
illumination.
Fig 3.2 Extraction of Human Body Contour
3.3 STAR SKELETONIZATION
The concept of star skeleton is to connect from
centroid to gross extremities of a humancontour.Tofindthe
gross extremities of human contour, the distances from the
centroid to each border point areprocessed ina clockwiseor
counter-clockwise order. The star skeletonisconstructed by
connecting the points to the target centroid.
Fig 3.3 a walk action is a series of postures over time
3.4 TRAINING THE MODEL USING RNN
From the class of artificial neural network the
connection between nodes form a sequence of directed
graph. To learn the temporal dynamics of sequential data
that contains the cyclic connectionsfromtheneural network
architecture. To process sequence of inputs RNN use their
internal memory.
4. CONCLUSION
A Recurrent neural Network (RNN) methodology
for Human Action Recognition using star skeleton as a
representative descriptor of human posture. Star skeleton
could be a quick skeletonization technique by connecting
from center of mass of target object to contour extremes. To
use star skeleton as feature for action recognition, wehavea
tendency to clearly outline the feature as a five-dimensional
vector in star fashion as a result of the top and 4 limbs are
typically local extremes of human shape. In our project we
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2327
assumed an action is composed of a series of star skeletons
over time. Therefore, time-sequential pictures expressing
human activity square measure remodeled into a feature
vector sequence. Then the feature vector sequence must be
transformed into symbol sequence so that RNN can model
the action.
5. FUTURE WORK
Human Activity Recognition (HAR) mistreatment
smart phones dataset associated an LSTM RNN. Classifying
the type of movement amongst Five categories:
• WALKING_UPSTAIRS,
• WALKING_DOWNSTAIRS,
• SITTING,
• STANDING,
• LAYING
6. REFERENCES
[1] C. Chen, R. Jafari, and N. Kehtarnavaz, “Action
recognition from depth sequences using depth motion
maps-based local binary patterns,” in Proc. IEEE Win.
Conf. Appl. Comput. Vis., 2015, pp. 1092–1099.
[2] G. Liang, X. Lan, J. Wang, J. Wang, and N. Zheng, “A
limb-based graphical model for human pose
estimation,” IEEE Trans. Syst., Man, Cybern., Syst., vol.
48, no. 7, pp. 1080–1092, Jul. 2018
[3] X. Peng, L. Wang, X. Wang, and Y. Qiao, “Bag of
visual words and fusion methods for action
recognition,” Comput. Vis. ImageUnderstand.,vol.150,
pp. 109–125, Sep. 2016.
[4] H. Wang and C. Schmid, “Action recognition with
improved trajectories,” in Proc.IEEEInt.Conf.Comput.
Vis., 2013, pp. 3551–3558.
[5] Y. Hou, Z. Li, P. Wang, and W. Li, “Skeleton optical
spectra-based action recognition using convolutional
neural networks,” IEEE Trans. Circuits Syst. Video
Technol., vol. 28, no. 3, pp. 807–811, Mar. 2018.

More Related Content

PDF
BIOMETRIC AUTHORIZATION SYSTEM USING GAIT BIOMETRY
IJCSEA Journal
 
PDF
1879 1885
Editor IJARCET
 
PDF
Human activity detection based on edge point movements and spatio temporal fe...
IAEME Publication
 
PDF
Facial image classification and searching –a survey
Zac Darcy
 
PDF
ADVANCED FACE RECOGNITION FOR CONTROLLING CRIME USING PCA
IAEME Publication
 
PDF
Ts 2 b topic
Harini Vemula
 
PDF
Hand Gesture Recognition System for Human-Computer Interaction with Web-Cam
ijsrd.com
 
PDF
Iris Encryption using (2, 2) Visual cryptography & Average Orientation Circul...
AM Publications
 
BIOMETRIC AUTHORIZATION SYSTEM USING GAIT BIOMETRY
IJCSEA Journal
 
1879 1885
Editor IJARCET
 
Human activity detection based on edge point movements and spatio temporal fe...
IAEME Publication
 
Facial image classification and searching –a survey
Zac Darcy
 
ADVANCED FACE RECOGNITION FOR CONTROLLING CRIME USING PCA
IAEME Publication
 
Ts 2 b topic
Harini Vemula
 
Hand Gesture Recognition System for Human-Computer Interaction with Web-Cam
ijsrd.com
 
Iris Encryption using (2, 2) Visual cryptography & Average Orientation Circul...
AM Publications
 

What's hot (16)

PDF
Human gait recognition using preprocessing and classification techniques
IJECEIAES
 
PDF
40120140501006
IAEME Publication
 
PDF
Review of facial expression recognition system and used datasets
eSAT Journals
 
PDF
Review of facial expression recognition system and
eSAT Publishing House
 
PDF
IRJET- Survey Paper on Vision based Hand Gesture Recognition
IRJET Journal
 
PDF
Vision Based Gesture Recognition Using Neural Networks Approaches: A Review
Waqas Tariq
 
PDF
Vehicle Monitoring System based On IOT, Using 4G/LTE
Dr. Amarjeet Singh
 
PDF
IRJET- Facial Expression Recognition using GPA Analysis
IRJET Journal
 
PDF
Activity recognition using histogram of
ijcseit
 
PDF
G0333946
iosrjournals
 
PDF
Face Recognition Using Simplified Fuzzy Artmap
sipij
 
PDF
Volume 2-issue-6-1960-1964
Editor IJARCET
 
PDF
Face recogition from a single sample using rlog filter and manifold analysis
acijjournal
 
PDF
Dq4301702706
IJERA Editor
 
PDF
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...
ijceronline
 
PDF
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...
IRJET Journal
 
Human gait recognition using preprocessing and classification techniques
IJECEIAES
 
40120140501006
IAEME Publication
 
Review of facial expression recognition system and used datasets
eSAT Journals
 
Review of facial expression recognition system and
eSAT Publishing House
 
IRJET- Survey Paper on Vision based Hand Gesture Recognition
IRJET Journal
 
Vision Based Gesture Recognition Using Neural Networks Approaches: A Review
Waqas Tariq
 
Vehicle Monitoring System based On IOT, Using 4G/LTE
Dr. Amarjeet Singh
 
IRJET- Facial Expression Recognition using GPA Analysis
IRJET Journal
 
Activity recognition using histogram of
ijcseit
 
G0333946
iosrjournals
 
Face Recognition Using Simplified Fuzzy Artmap
sipij
 
Volume 2-issue-6-1960-1964
Editor IJARCET
 
Face recogition from a single sample using rlog filter and manifold analysis
acijjournal
 
Dq4301702706
IJERA Editor
 
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...
ijceronline
 
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...
IRJET Journal
 
Ad

Similar to IRJET- Recurrent Neural Network for Human Action Recognition using Star Skeletonization (20)

PDF
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
IRJET Journal
 
PDF
Human Activity Recognition
IRJET Journal
 
DOCX
Chapter 1_Introduction.docx
KISHWARYA2
 
PDF
IRJET- Survey on Detection of Crime
IRJET Journal
 
PDF
IRJET- Recognition of Human Action Interaction using Motion History Image
IRJET Journal
 
PDF
Survey on Human Behavior Recognition using CNN
IRJET Journal
 
PPTX
seminar Islideshow.pptx
rakeshjain627294
 
PDF
A Framework for Human Action Detection via Extraction of Multimodal Features
CSCJournals
 
PDF
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET Journal
 
PDF
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
IRJET Journal
 
PDF
Gait Recognition using MDA, LDA, BPNN and SVM
IJEEE
 
PDF
IRJET- Tracking and Recognition of Multiple Human and Non-Human Activites
IRJET Journal
 
PDF
A Review on Human Activity Recognition System
vivatechijri
 
PDF
A Review on Human Activity Recognition System
vivatechijri
 
PDF
Detection of abnormal human behavior using deep learning
BOHRInternationalJou1
 
PDF
HUMAN IDENTIFIER WITH MANNERISM USING DEEP LEARNING
IRJET Journal
 
PDF
Intelligent Video Surveillance System using Deep Learning
IRJET Journal
 
PDF
Comparison of feed forward and cascade forward neural networks for human acti...
nooriasukmaningtyas
 
PDF
Paper id 25201468
IJRAT
 
PDF
Continuous hand gesture segmentation and acknowledgement of hand gesture path...
International Journal of Reconfigurable and Embedded Systems
 
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
IRJET Journal
 
Human Activity Recognition
IRJET Journal
 
Chapter 1_Introduction.docx
KISHWARYA2
 
IRJET- Survey on Detection of Crime
IRJET Journal
 
IRJET- Recognition of Human Action Interaction using Motion History Image
IRJET Journal
 
Survey on Human Behavior Recognition using CNN
IRJET Journal
 
seminar Islideshow.pptx
rakeshjain627294
 
A Framework for Human Action Detection via Extraction of Multimodal Features
CSCJournals
 
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET Journal
 
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
IRJET Journal
 
Gait Recognition using MDA, LDA, BPNN and SVM
IJEEE
 
IRJET- Tracking and Recognition of Multiple Human and Non-Human Activites
IRJET Journal
 
A Review on Human Activity Recognition System
vivatechijri
 
A Review on Human Activity Recognition System
vivatechijri
 
Detection of abnormal human behavior using deep learning
BOHRInternationalJou1
 
HUMAN IDENTIFIER WITH MANNERISM USING DEEP LEARNING
IRJET Journal
 
Intelligent Video Surveillance System using Deep Learning
IRJET Journal
 
Comparison of feed forward and cascade forward neural networks for human acti...
nooriasukmaningtyas
 
Paper id 25201468
IJRAT
 
Continuous hand gesture segmentation and acknowledgement of hand gesture path...
International Journal of Reconfigurable and Embedded Systems
 
Ad

More from IRJET Journal (20)

PDF
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
IRJET Journal
 
PDF
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
IRJET Journal
 
PDF
Kiona – A Smart Society Automation Project
IRJET Journal
 
PDF
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
IRJET Journal
 
PDF
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
IRJET Journal
 
PDF
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
IRJET Journal
 
PDF
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
IRJET Journal
 
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
PDF
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
IRJET Journal
 
PDF
BRAIN TUMOUR DETECTION AND CLASSIFICATION
IRJET Journal
 
PDF
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
IRJET Journal
 
PDF
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
IRJET Journal
 
PDF
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
IRJET Journal
 
PDF
Breast Cancer Detection using Computer Vision
IRJET Journal
 
PDF
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
PDF
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
PDF
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
IRJET Journal
 
PDF
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
PDF
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
IRJET Journal
 
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
IRJET Journal
 
Kiona – A Smart Society Automation Project
IRJET Journal
 
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
IRJET Journal
 
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
IRJET Journal
 
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
IRJET Journal
 
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
IRJET Journal
 
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
IRJET Journal
 
BRAIN TUMOUR DETECTION AND CLASSIFICATION
IRJET Journal
 
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
IRJET Journal
 
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
IRJET Journal
 
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
IRJET Journal
 
Breast Cancer Detection using Computer Vision
IRJET Journal
 
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
IRJET Journal
 
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 

Recently uploaded (20)

PDF
Machine Learning All topics Covers In This Single Slides
AmritTiwari19
 
PDF
All chapters of Strength of materials.ppt
girmabiniyam1234
 
PPTX
Victory Precisions_Supplier Profile.pptx
victoryprecisions199
 
PPTX
Civil Engineering Practices_BY Sh.JP Mishra 23.09.pptx
bineetmishra1990
 
PPTX
Module2 Data Base Design- ER and NF.pptx
gomathisankariv2
 
PDF
Introduction to Ship Engine Room Systems.pdf
Mahmoud Moghtaderi
 
PDF
Advanced LangChain & RAG: Building a Financial AI Assistant with Real-Time Data
Soufiane Sejjari
 
PPTX
22PCOAM21 Session 2 Understanding Data Source.pptx
Guru Nanak Technical Institutions
 
PDF
FLEX-LNG-Company-Presentation-Nov-2017.pdf
jbloggzs
 
PDF
20ME702-Mechatronics-UNIT-1,UNIT-2,UNIT-3,UNIT-4,UNIT-5, 2025-2026
Mohanumar S
 
PDF
The Effect of Artifact Removal from EEG Signals on the Detection of Epileptic...
Partho Prosad
 
PPTX
Chapter_Seven_Construction_Reliability_Elective_III_Msc CM
SubashKumarBhattarai
 
PDF
settlement FOR FOUNDATION ENGINEERS.pdf
Endalkazene
 
PPT
Understanding the Key Components and Parts of a Drone System.ppt
Siva Reddy
 
PDF
LEAP-1B presedntation xxxxxxxxxxxxxxxxxxxxxxxxxxxxx
hatem173148
 
PPTX
quantum computing transition from classical mechanics.pptx
gvlbcy
 
PDF
AI-Driven IoT-Enabled UAV Inspection Framework for Predictive Maintenance and...
ijcncjournal019
 
PDF
Biodegradable Plastics: Innovations and Market Potential (www.kiu.ac.ug)
publication11
 
PDF
Cryptography and Information :Security Fundamentals
Dr. Madhuri Jawale
 
PDF
Unit I Part II.pdf : Security Fundamentals
Dr. Madhuri Jawale
 
Machine Learning All topics Covers In This Single Slides
AmritTiwari19
 
All chapters of Strength of materials.ppt
girmabiniyam1234
 
Victory Precisions_Supplier Profile.pptx
victoryprecisions199
 
Civil Engineering Practices_BY Sh.JP Mishra 23.09.pptx
bineetmishra1990
 
Module2 Data Base Design- ER and NF.pptx
gomathisankariv2
 
Introduction to Ship Engine Room Systems.pdf
Mahmoud Moghtaderi
 
Advanced LangChain & RAG: Building a Financial AI Assistant with Real-Time Data
Soufiane Sejjari
 
22PCOAM21 Session 2 Understanding Data Source.pptx
Guru Nanak Technical Institutions
 
FLEX-LNG-Company-Presentation-Nov-2017.pdf
jbloggzs
 
20ME702-Mechatronics-UNIT-1,UNIT-2,UNIT-3,UNIT-4,UNIT-5, 2025-2026
Mohanumar S
 
The Effect of Artifact Removal from EEG Signals on the Detection of Epileptic...
Partho Prosad
 
Chapter_Seven_Construction_Reliability_Elective_III_Msc CM
SubashKumarBhattarai
 
settlement FOR FOUNDATION ENGINEERS.pdf
Endalkazene
 
Understanding the Key Components and Parts of a Drone System.ppt
Siva Reddy
 
LEAP-1B presedntation xxxxxxxxxxxxxxxxxxxxxxxxxxxxx
hatem173148
 
quantum computing transition from classical mechanics.pptx
gvlbcy
 
AI-Driven IoT-Enabled UAV Inspection Framework for Predictive Maintenance and...
ijcncjournal019
 
Biodegradable Plastics: Innovations and Market Potential (www.kiu.ac.ug)
publication11
 
Cryptography and Information :Security Fundamentals
Dr. Madhuri Jawale
 
Unit I Part II.pdf : Security Fundamentals
Dr. Madhuri Jawale
 

IRJET- Recurrent Neural Network for Human Action Recognition using Star Skeletonization

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2324 Recurrent Neural Network for Human Action Recognition using Star Skeletonization Nithin Paulose1, M. Muthukumar2, S. Swathi3, M. Vignesh4 1,2B.E.Computer Science and Engineering, Dhaanish Ahmed Institute of Technology, Coimbatore, Tamil Nadu, India 3,4Assistant Professor, Dept. of Computer Science and Engineering, Dhaanish Ahmed Institute of Technology, Coimbatore, Tamil Nadu, India ---------------------------------------------------------------------***---------------------------------------------------------------------- Abstract - This project presents a Recurrent Neural Network (RNN) methodology for Human Action Recognition using star skeleton as a representative descriptor of human posture. Star skeleton is a fast skeletonization technique by connecting from geometric center of target object to contour extremes. For the action recognition using the feature star skeleton, we clearly define the feature as a five-dimensional vector in star fashion because the head and four limbs are usually local extremes of human shape. In our project we assumed an action is composed of a series of star skeletons overtime. Therefore, the images whicharetime-sequential are expressing human action that is transformed into a feature vector sequence. Then the feature vector sequence must be transformed into symbol sequence so that RNN can model the action. We used RNN because the features extracted are time dependent. Key Words: Recurrent Neural Network (RNN), star skeleton, contour extremes, Human Action Recognition, five- dimensional vector, time-sequential. 1. INTRODUCTION Human activity recognition is an important task for Ambient Intelligence systems. The state of a person is to be recognized, which provides us with valuable information that is been used as input for other systems. For example, in health care, fall detection can be used to alert the medical in case of an accident; in security, abnormal behavior can be detected and thus used to prevent a burglary or other criminal activities. Human motion analysis is currently receiving increasing attention from computer vision researchers. For example, Human body segmentation in an image, the movement of joints are tracked in an image sequence, and the analysisofathletic performanceisdone by recovering underlying 3D body structure , also used for medical diagnostics. Other applications include building man-machine user interfaces and video conferencing. The goal of human activity recognition is to automatically analyses on -goingactivitiesfromanunknown video. The objective of the system is to correctly classify the video into its activity category, for example where a video is segmented to contain only one executionofa humanactivity. The starting and ending times of all occurring activitiesfrom an input video is detected, from which the continuous recognition of human activities are performed. The constructions of several important applications are constructed from the videos which has the ability to recognize complex humanactivities.Automatedsurveillance systems in public places like airports and subway stations require detection of abnormal and suspicious activities, as opposed to normal activities. For example, an automatic recognize of suspicious activities like “a personleavesa bag” or “a person places his/her bag in a trash bin” in an airport surveillance systemmustberecognized.Usingrecognitionof human activity the real-time monitoring of patients, children, and elderly persons can be done. By using activity recognition the construction of gesture-based human computer interfaces and vision-based intelligent environments becomes possible. There are various types of human activities. Depending on their complexity, they can be conceptually categorized into four different levels: gestures, actions, interactions, and group activities. 1.1 HUMAN ACTIVITY RECOGNITION FROM VIDEO SEQUENCES Human activity recognition role is that human-to- human interaction and interpersonal relations. HAR provides information about the identity, personality, and psychological state that is difficult to extract. In various classification techniques two main questions emerge: “Where it is in the video?” and “What is the action?” To recognize human activities one must determine the active states of a person, to recognize the efficiency. “Walking” is the daily human activity to recognize. The complexactivities such as “peeling an apple” are more difficult to identify. The easier way to recognize is to simplify the complex activities into other simpler activities. For the better understanding of human activities the detectionofobjects mayprovideuseful information about the ongoing event. The human activity recognition assumes a figure- centric scene of uncluttered background, where the actor is free to perform an activity. The challenging task to classifya person’s activities with low error in a fully automated human activity recognition system are background clutter, partial occlusion, changes in scale, viewpoint, lighting and appearance, and frame resolution. Moreover, commenting
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2325 behavioral roles is time consuming and requires knowledge of the specific event. However, intra- and inter-class similarities make the problem more challenging, that is ,the same action done by two people may finding difficult to detect their individual action . The human activity is based on their habits, and this make s so difficulttodeterminetheir activity. In real time the challenging tasks is that, the construction of visual model for learning and analyzing human movements with inadequate benchmark datasets for evaluation. Fig 1: Architecture diagram 1.2 HUMAN ACTIVITY CATEGORIZATION The human activity recognition methods are classified into two main categories: (i) uni-modal and (ii) multimodal activity recognition methods according to the nature of sensor data they employ. Then, these two categories is further break down into sub-categories depending on how they model human activities. Uni-modal methodsrepresenthumanactivitiesfrom data of a single modality, such as images, and they are further categorized as: (i) space-time, (ii) stochastic, (iii) rule-based, and (iv) shape based methods. Space-time methods involve activity recognition methods, which represent human activities as a set of spatio-temporal features or trajectories. By applying statistical models to represent human actions stochastic methods are used. The modelings the motion of human body parts are shape-based methods efficiently represent activities with high level reasoning. Multimodal methods combine features collected from different sources and are classified into three categories: (i) affective, (ii) behavioral, and (iii) social networking methods. Emotional communications and the affective state of a person represent human activity. Behavioral methods aim to recognize behavioral attributes, non-verbal multimodal cues such as gestures, facial expressions and auditory cues.Thecharacteristicsaremodel by social networking methods .It model the characteristics and the behavior of humans in several layers of human-to- human interactions in social events from gestures, body motion, and speech. The activity and behavior are the two terms which are used interchangeably. These two terms activity and behavior is used to describe a sequence of actions that correspond to specific body motion and to characterize both activities and events that are associated with facial expression, emotional states, and gestures with single person auditory cues. 1.3 HUMAN ACTIVITY RECOGNITION MODEL The human activity recognition has largely focused on statistical methods using spatio-temporal features. The typical model consists of spatio-temporal interest-points which are detected in the video sequence and the local maxima become the centerpointofa spatio-temporal region. Features are then extracted from the spatio-temporal region (such as features based on optical flow or gradient values) and summarized or histogram med to form a feature descriptor. The feature descriptors are used to form a code book, typically followed by a bag of visual words model adapted from statistical natural language processing. While methods based on spatio-temporal features are the most common, other methods make use of other video features such as medium-term tracking, volumetric representations and graph-based features. 2. LITERATURE REVIEW [1] A computational efficient action recognition framework using depth motion maps (DMMs)-based local binary patterns (LBPs) and kernel-based extreme learning machine (KELM). In depth video sequence depth frames are projected onto three orthogonal Cartesianplanestoformthe projected images corresponding to three projection views [front (f), side (s), and top (t) views].For calculating LBP histogram an LBP operator is applied to each block and the DMMs are divided into overlapped blocks. Feature-level fusion and decision-level fusion approachesareinvestigated using KELM [2] in this paper, we propose to use human limbs to augment constraints between neighboring human joints. We model a limb as a wide line to represent its shape information. Instead of estimating its length and rotation angle, we calculate as per-pixel likelihood for each human limb by a ConvNet. [3] Video based action recognition is one of the important and challenging problems in computer vision re- search. The several realistic datasets are the HMDB51, UCF50, and UCF101. BoVW is a general pipeline to construct a global representation from local features composed of five steps; (i) feature extraction, (ii) feature pre-processing, (iii) codebook generation, (iv)feature encoding, and (v) pooling and normalization.[4]For efficient video representation on action recognition is shown by dense and achieved state-of-the-art results on a variety of datasets. The performance is corrected bytakingthecamera motion into account. The feature points between frames using SURF descriptors and dense optical flow are used to
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2326 estimate camera motion. With the help of RANSAC homography is estimated. The performance can be significantly improved by removing backgroundtrajectories and warping optical flow with a robustly estimated homograph approximating the camera motion. [5] The discriminative features for action recognition Convolutional Neural Networks (ConvNets) are employed into color texture images that are referred to as skeleton optical spectra. The learning of suitable dynamic features and ConvNet architecture from skeleton sequences is possible without training millions of parameters from this kind of spectrum views. 3. DESIGN METHODS There are some methods used for Human Action Recognition using RNN. 3.1 HUMAN SILHOUETTE EXTRACTION In this project we extract human body contour from given image. In Frame Videos: To obtain the human body we should take the direct difference between the background and the current frame. Out Frame Videos: To extract the human body from frames of the videos we used the inbuilt Gaussian Mixture Model based Foreground detection method. Fig 3.1 Background Subtraction 3.2 HUMAN CONTOUR EXTRACTION To extract the contour of a detected human body, at thresholding and morphological method, the important approaches in the field of image segmentation are to choose a correct threshold and that is difficult under irregular illumination. Fig 3.2 Extraction of Human Body Contour 3.3 STAR SKELETONIZATION The concept of star skeleton is to connect from centroid to gross extremities of a humancontour.Tofindthe gross extremities of human contour, the distances from the centroid to each border point areprocessed ina clockwiseor counter-clockwise order. The star skeletonisconstructed by connecting the points to the target centroid. Fig 3.3 a walk action is a series of postures over time 3.4 TRAINING THE MODEL USING RNN From the class of artificial neural network the connection between nodes form a sequence of directed graph. To learn the temporal dynamics of sequential data that contains the cyclic connectionsfromtheneural network architecture. To process sequence of inputs RNN use their internal memory. 4. CONCLUSION A Recurrent neural Network (RNN) methodology for Human Action Recognition using star skeleton as a representative descriptor of human posture. Star skeleton could be a quick skeletonization technique by connecting from center of mass of target object to contour extremes. To use star skeleton as feature for action recognition, wehavea tendency to clearly outline the feature as a five-dimensional vector in star fashion as a result of the top and 4 limbs are typically local extremes of human shape. In our project we
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 03 | Mar 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2327 assumed an action is composed of a series of star skeletons over time. Therefore, time-sequential pictures expressing human activity square measure remodeled into a feature vector sequence. Then the feature vector sequence must be transformed into symbol sequence so that RNN can model the action. 5. FUTURE WORK Human Activity Recognition (HAR) mistreatment smart phones dataset associated an LSTM RNN. Classifying the type of movement amongst Five categories: • WALKING_UPSTAIRS, • WALKING_DOWNSTAIRS, • SITTING, • STANDING, • LAYING 6. REFERENCES [1] C. Chen, R. Jafari, and N. Kehtarnavaz, “Action recognition from depth sequences using depth motion maps-based local binary patterns,” in Proc. IEEE Win. Conf. Appl. Comput. Vis., 2015, pp. 1092–1099. [2] G. Liang, X. Lan, J. Wang, J. Wang, and N. Zheng, “A limb-based graphical model for human pose estimation,” IEEE Trans. Syst., Man, Cybern., Syst., vol. 48, no. 7, pp. 1080–1092, Jul. 2018 [3] X. Peng, L. Wang, X. Wang, and Y. Qiao, “Bag of visual words and fusion methods for action recognition,” Comput. Vis. ImageUnderstand.,vol.150, pp. 109–125, Sep. 2016. [4] H. Wang and C. Schmid, “Action recognition with improved trajectories,” in Proc.IEEEInt.Conf.Comput. Vis., 2013, pp. 3551–3558. [5] Y. Hou, Z. Li, P. Wang, and W. Li, “Skeleton optical spectra-based action recognition using convolutional neural networks,” IEEE Trans. Circuits Syst. Video Technol., vol. 28, no. 3, pp. 807–811, Mar. 2018.