SlideShare a Scribd company logo
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3130
Chatbot with Gesture based User Input
Rajkumar G1, Gayathri M2
1Student/CSE, SRM Institute of Science and Technology, Chennai-603203
2Assistant Professor/CSE, SRM institute of Science and Technology, Chennai
---------------------------------------------------------------------***----------------------------------------------------------------------
Abstract: The paper describes a approach towards
building a chatbot with gesture based input. A gesture table
will be described here which is a mapping sequence of the
appropriate gestures to the appropriate commands they
mean. After successful gesture recognition, the classified or
recognized gesture is looked up in the table and is
converted accordingly into text and then that text is fed
into the chatbot program as input and appropriate
command is executed. The paper selects existing known
algorithms and aims to explores the overall effects it has
and if such a application is suitable for areas such as deaf
and dumb sign language or in public galleries, to quickly
get required information
1. Introduction
Gesture recognition is a topic in computer science with the
goal of interpreting human gestures algorithms. Gestures
can be created from any bodily motion but commonly
originate from the face or hand. Quite a few approaches
have been suggested using cameras and computer vision
algorithms to interpret sign language.
Gesture recognition applications in various fields such as
automotive, consumer electronics, Transit, gaming,
smartphones, defence, sign language translation sectors. Its
used to provide a more natural interaction with the UI to
save time and make things more intuitive. In
entertainment, mainly gaming ,it can be used to add a new
way of interaction to attract players. In smartphones, its
used for faster unlock. For deaf and dumb, its used to for
translating sign language into text.
Gesture recognition has spawned several correlated
technology such as touchless interface. Touchless user
interface (TUI) is a interface through which a computer can
be given commands or operations in form of body motion
and gestures without touching a keyboard, mouse, or
screen. For example, Microsoft's Kinect is a touchless game
interface; This is basically another major application of
gesture recognition. Many companies are invested in
development of this technology.
Intel corporation is researching how touch multi-factor
authentication can help healthcare organizations minimize
security risks and at the same time improve clinical
efficiency.Here Multi-factor authentication means multiple
layers/levels to authorise a particular transaction.
Microsoft corporation is researching the use of touchless
interaction within surgical settings, allowing images to be
viewed and manipulated without any physical contact with
traditional computer hardware through gesture
recognition.
Gesture recognition is linked to the field of digital image
processing. It uses many algorithms and concepts of digital
image processing along with concepts and algorithms of AI.
Some of these include thresholding, otsu algorithm,
anisotropic diffusion, Hidden Markov model, image editing,
image restoration etc(belonging to digital image
processing), machine learning ,computer vision involving
automatic inspection, Assisting humans in identification
tasks, Controlling processes, detecting events, interaction,
modeling objects, navigation, organizing information.
2. Literature Survey
Rupesh Prajapati, Vedant Pandey, Nupur Jamindar, Neeraj
Yadav, Prof. Neelam Phadnis have published a research
paper titled “Hand Gesture Recognition and Voice
Conversion for Deaf and Dumb”. In it they accept the input
in form of Video feed through webcam, they have proposed
a method of Create a database of images for training and
use PCA for creating a relationship out of its linearly
uncorrelated characterstics and then use classfication
algorithm like KNN and SVN.[1]
Anchal Sood and Anju Mishra have published a research
paper titled “AAWAAZ: A communication system for deaf
and dumb-”.They have proposed a sign recognition system
based on Harris algorithm for feature extraction and then it
is stored in a matrix. This matrix is further used to match
the image from the database. The system does have a few
limitations. Binary imaging technique doesn’t perfectly
seperates discrete objects from backgrounds of similar
colors due to inaccurate thresholding algorithm. But the
results are efficient.[2]
Ms R. Vinitha and Ms A. Theerthana have published a
research paper titled “Design And Development of Hand
Gesture Recognition System For Speech Impaired People.”
In it they accept input in form of Data glove(consisting of
sensors,accelometer and PIC-microcontroller). Gesture
recognition is based on hardware approach[3]
Shangeetha, R. K., V. Valliammai, and S. Padmavathi have
published a research paper titled “Computer vision based
approach for Indian Sign Language character recognition”.
In it they accept input in form of Live video feed from
webcam. They have also proposed a method of using HIS
model and then the feature extraction is done by distance
transform method. The results are as follows,The gestures
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3131
stored and recognized and coverted to appropriate voice
output.[4]
R. Aravind, D.Anbasaran, K.Alice have published a research
paper titled “GRS-Gesture Based Recognition System for
Indian Sign Language Recognition System for Deaf and
Dumb people”. In it they accept input in form of webcam.
They have proposed a method of gesture recognition using
color intensity variation and edge detection and
background subtraction. The results were as follows,
gesture symbols extracted from the image alone The gaps
identified were problems identifying concurrent
symbols[5]
Tripathi K. and Nandi have published a research paper
titled “Continuous indian sign language gesture recognition
and sentence”. In they accept input in form of webcam.
They have proposed a method using Orientation Histogram
and PCA. The results were that highly accurate gesture
recognition by reducing dimension features after
extraction.[6]
Daniel Simões Lopes have developed a gesture based
recognition system as a basis for touchless user interface.
In it they accept input in form of Hand gestures and body
postures. They have proposed a method of Developed a
gesture recognition system where each hand is can make
gestures and body postures can also be used. They found
out that that using touchless interface is indeed better than
relying on 2-D devices as it allows for faster retrieval and
greater flexibility in the sensitive surgical enviornment.
The gaps that were identified were Precision tasks such as
clipping plane and visualization and tagging are better
performed with mouse due to not good gesture
recognition,[7]
Another interesting method proposed was relying on
passive props.3D medical visualization relied on passive
interface props. Hardware interface placed in a position
where the user needs manipulate the control devices. The
result was that the system facilitates a natural interaction
between UI and user. The gaps that were identified were
that It was not easy to stop or use in several situations, the
proposed mechanism is not flexible enough[8]
-Chatbot -Gonzalas Cenelia Article Internet
Gonzalas Cenelia, gives a excellent approach to building a
chatbot based on keyword concept. It involves finding
keywords in the given user input and match it with a
already predefine knowledge base which has keywords
and assoiciated responses to it. Revolving around the
concept of keyword, things like context of the conversation
,learning from feedback to add knowledge base, keyword
ranking etc are also introduced [9]
3. Proposed System:-
Note the high level view of the proposed system in the
given diagram.
Input: Imagestream from webcam
Output: Appropriate textual output or execution of a
command like search
A. Gesture Recognition Module:-
Input: Imagestream from webcam representing correct
gestures
Output:Successfully recognized gesture label(See gesture
mapping sequence)
The recognized gestures can be divided into two types.
Counting gestures
-Counting gestures refer to the gestures that indicate
numbers like using only your index finger to count 1 ,or
any finger for that matter.
Implementation used:Convexity Defect
General gestures
-General gestures are the set of gestures that can be
classified by a AI model which is trained on a database of
images.
Implementation used:CNN model
For each of these types, A separate implementation is
provided which is considered more optimum for
recognizing those gestures. Using the convexity defect
approach, we don’t have to train the AI using large number
of images for each number representational gesture.
Besides the convexity defect approach is very flexible ,it
simply counts the number of fingers so we don’t have to
exact representational gesture like for example using the
pinky finger to represent 1 rather than the index finger.(To
its recommended to use the normal ones).
I. Convexity Defect Implementation:-
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3132
1)Main GUI Control Window
-This window is the main navigation window used to
activate all the modules /algorithms for sucessful gesture
recognition
2)GUI Input window
-This window activates the webcam and gives you input
video feed to manipulate on
3)Thresholding(Background substraction)
As the video feed continues ,every frame is continously
getting converted to grayscale image.
A grayscale image is a image without saturation aka color
.The only information its pixels have is on the intensity of
light falling on that pixel. So basically the image appears in
shades of grey.
After this process, thresholding is applied. Thresholding
converts the grayscale image to a binary color image which
typically are black and white. A value called the threshold
point can be manually provided or using an algorithm can
be automated . The threshold function will make the pixels
whose intensity are greater than the threshold value white
and vice versa for black.
In our case,we have use otsu algorithm to calculate
threshold point.
4)Finding contours and convex hull and defects
Contours are the areas of the discrete objects that we
capture from the image. We assume that the contour which
has the maximum area in the binary image will be of our
hand, thats why a white screen should be used to hide the
background objects.
We calculate then the convex hull which is the smallest
polygon that encloses a set of points. Then the convexity
defect is defined as the deepest point in the said binary
image between two vertices of the convex hull. The actual
definition of convexity defect in terms of physical
boundaries of the object is given below
Convexity defect is a cavity in an object (blob, contour)
segmented out from an image. That means an area that do
not belong to the object but located inside of its outer
boundary -convex hull.
II. Deep Learning Implementation:
(Contd)
The input consists of 20,000 images . The first Module
Image URL loader and reader is a simple algorithm which
reads all the appropriate pathnames of the files and
converts them into arrays using opencv functions .It
extracts all the images required for training and stores the
labels for each image in a separate list. One list all the
images and the other list contains all the corresponding
labels of each image. During the process the images are
converted to grayscale and resized to smaller size so that
the training phase of CNN model will be shorter. Then its
fed into the CNN model. It splits its data into 70% as
training and 30% as testing.
(In the below table ,the gestures can be mapped to the
appropriate text according to the usage, for example thumb
down to what or L to how etc, say if the application is to be
used in a deaf/dumb language or a custom gesture system
required etc.)
It takes 30 mins approximately for five iterations over the
entire database of images for training as well as the
validating(testing) to complete. After that model is ready to
predict furthur images.
(A sample tableis given if this application is to be used in a
museum)
Gesture Label
Thumb down How many
artifacts are
here?
palm(Horizontal) What is the name
of the place
L Where is it
located
Fist(Horizontal) When was it
founded
Fist(Vertical) What is the most
famous artifact
here.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3133
Tools used:
-Python(Anaconda package)
-Spyder(IDE)
-Keras,tensorflow and opencv as python libraries
B.Chatbot
Template Response:- We put up ready-made templates as
replies and those templates will have parameters to which
will be replaced by the appropriate data relevent to the
conversion
Keywords:-The concept on the basis of which the chatbot
will be implemented .The chatbot will scan the user input
and see if a keyword exists in it. Theses keywords are
stored in the knowledge base. And each of the keyword
have a set of responses linked to it. So a response is chosen
and given as reply
Knowledge base:-The knowledge base is a 2-D array
storing a set of keywords in the first column and its
associated responses in the next columns.
Keyword Ranking:-This is used when there are multiple
keyword match and based on the furthur corresponding
keyword match and repetitions, the appropriate keyword
is selected then the corresponding reply is given
The chatbot will be written in java based module which
will accept the input as text which was converted
appropriately from the gesture input.
Screenshots:
Deep learning implementation
(Note:Predict prints a command line out put.
The image is first captured and then saved
As png file and then is fed into the deep
Learning model)
Convexity defect to count fingers to display numbers
The output is planned to be written to
Text file which will be fed into a java based
chatbot
Chatbot Output:(Separate module)
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3134
Example sentences are given as input, it can be customised
according to the table created by simple adding those
responses to the knowledge base of the chatbot.
References
[1] Rupesh Prajapati, Vedant Pandey, Nupur Jamindar,
Neeraj Yadav, Prof. Neelam Phadnis,”Hand Gesture
Recognition and Voice Conversion for Deaf and
Dumb”,IRJET,2018
[2] Anchal Sood and Anju Mishra, “AAWAAZ:
communication system for deaf and dumb-”.IEEE,2016
[3]Ms R. Vinitha and Ms A. Theerthana,"Design And
Development Of Hand Gesture Recognition System For
Speech Impaired People.",2016
[4]Shangeetha, R. K., V. Valliammai, and S.
Padmavathi,"Computer vision based approach for Indian
Sign Language character recognition." Machine Vision and
Image Processing (MVIP), IEEE,2012
[5] R.Aravind, D.Anbasaran, K.Alice,”GRS-Gesture Based
Recognition System for Indian Sign Language Recognition
System for Deaf and Dumb people”, International Journal of
Trend in Scientific Research and Development -
IJTSRD,2018
[6]Tripathi K. and Nandi, “Continuous indian sign language
gesture recognition and sentence”,Procedia
Computer,2015
[7]Daniel Simões Lopes “On the utility of 3D hand cursors
to explore medical volume datasets with a touchless
interface”, Journal of biomedical informatics,2017 Passive
real-world interface props for neurosurgical visualization,
Hinckley et al.CHI '94,
[9]Otsu Algorithm-(Algorithm used for Thresholding,
grayscale to binary),See OpenCv 3.4.2 documentation for
algorithm implementation
[10]Peter-Draukr algorithm(For calculation of contours)
[11]Satoshi Suzuki and others. Topological structural
analysis of digitized binary images by border following.
Computer Vision, Graphics, and Image Processing,
30(1):32–46, 1985.(Algorithm used for finding contours)
[12]CNN(For training the AI model)

More Related Content

What's hot (20)

PDF
IRJET - Hand Gestures Recognition using Deep Learning
IRJET Journal
 
PDF
Gesture recognition using artificial neural network,a technology for identify...
NidhinRaj Saikripa
 
PDF
Paper id 25201413
IJRAT
 
PDF
IRJET- Real-Time Object Detection System using Caffe Model
IRJET Journal
 
PDF
IRJET - NETRA: Android Application for Visually Challenged People to Dete...
IRJET Journal
 
PDF
Development of Real Time Face Recognition System using OpenCV
IRJET Journal
 
PDF
Real Time Vision Hand Gesture Recognition Based Media Control via LAN & Wirel...
IJMER
 
PDF
Smart Door Access using Facial Recognition
ijtsrd
 
PDF
IRJET- Applications of Object Detection System
IRJET Journal
 
PDF
IRJET- Vision Based Sign Language by using Matlab
IRJET Journal
 
PDF
Profile Identification through Face Recognition
ijtsrd
 
PDF
Paper id 21201494
IJRAT
 
PDF
Iris Encryption using (2, 2) Visual cryptography & Average Orientation Circul...
AM Publications
 
PDF
IRJET-Human Face Detection and Identification using Deep Metric Learning
IRJET Journal
 
PDF
Implementation of video tagging identifying characters in video
eSAT Publishing House
 
PDF
Final Year Project-Gesture Based Interaction and Image Processing
Sabnam Pandey, MBA
 
PDF
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET Journal
 
PDF
Authentication Schemes for Session Passwords using Color and Images
IJNSA Journal
 
PDF
IRJET - Mutecom using Tensorflow-Keras Model
IRJET Journal
 
IRJET - Hand Gestures Recognition using Deep Learning
IRJET Journal
 
Gesture recognition using artificial neural network,a technology for identify...
NidhinRaj Saikripa
 
Paper id 25201413
IJRAT
 
IRJET- Real-Time Object Detection System using Caffe Model
IRJET Journal
 
IRJET - NETRA: Android Application for Visually Challenged People to Dete...
IRJET Journal
 
Development of Real Time Face Recognition System using OpenCV
IRJET Journal
 
Real Time Vision Hand Gesture Recognition Based Media Control via LAN & Wirel...
IJMER
 
Smart Door Access using Facial Recognition
ijtsrd
 
IRJET- Applications of Object Detection System
IRJET Journal
 
IRJET- Vision Based Sign Language by using Matlab
IRJET Journal
 
Profile Identification through Face Recognition
ijtsrd
 
Paper id 21201494
IJRAT
 
Iris Encryption using (2, 2) Visual cryptography & Average Orientation Circul...
AM Publications
 
IRJET-Human Face Detection and Identification using Deep Metric Learning
IRJET Journal
 
Implementation of video tagging identifying characters in video
eSAT Publishing House
 
Final Year Project-Gesture Based Interaction and Image Processing
Sabnam Pandey, MBA
 
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET Journal
 
Authentication Schemes for Session Passwords using Color and Images
IJNSA Journal
 
IRJET - Mutecom using Tensorflow-Keras Model
IRJET Journal
 

Similar to IRJET - Chatbot with Gesture based User Input (20)

PDF
IRJET- Survey on Sign Language and Gesture Recognition System
IRJET Journal
 
DOC
Test
theextraaedge
 
PDF
Sign Language Identification based on Hand Gestures
IRJET Journal
 
PDF
Controlling Computer using Hand Gestures
IRJET Journal
 
PDF
IRJET - Sign Language Recognition using Neural Network
IRJET Journal
 
PDF
IRJET- Hand Gesture Recognition for Deaf and Dumb
IRJET Journal
 
PDF
Accessing Operating System using Finger Gesture
IRJET Journal
 
PDF
Sign Language Recognition with Gesture Analysis
paperpublications3
 
PDF
DHWANI- THE VOICE OF DEAF AND MUTE
IRJET Journal
 
PDF
DHWANI- THE VOICE OF DEAF AND MUTE
IRJET Journal
 
PDF
Hand gesture recognition using machine learning algorithms
CSITiaesprime
 
PDF
Gesture recognition using artificial neural network,a technology for identify...
NidhinRaj Saikripa
 
PDF
A Survey Paper on Controlling Computer using Hand Gestures
IRJET Journal
 
PDF
Gesture Recognition System
IRJET Journal
 
PDF
D0361015021
ijceronline
 
PPTX
ppt of gesture recognition
Aayush Agrawal
 
PDF
Sign Language Recognition with Gesture Analysis
paperpublications3
 
PDF
Human Computer Interaction Based HEMD Using Hand Gesture
IJAEMSJORNAL
 
PPTX
Indian Sign Language Recognition Method For Deaf People
Takrim Ul Islam Laskar
 
PDF
SLIDE PRESENTATION BY HAND GESTURE RECOGNITION USING MACHINE LEARNING
IRJET Journal
 
IRJET- Survey on Sign Language and Gesture Recognition System
IRJET Journal
 
Sign Language Identification based on Hand Gestures
IRJET Journal
 
Controlling Computer using Hand Gestures
IRJET Journal
 
IRJET - Sign Language Recognition using Neural Network
IRJET Journal
 
IRJET- Hand Gesture Recognition for Deaf and Dumb
IRJET Journal
 
Accessing Operating System using Finger Gesture
IRJET Journal
 
Sign Language Recognition with Gesture Analysis
paperpublications3
 
DHWANI- THE VOICE OF DEAF AND MUTE
IRJET Journal
 
DHWANI- THE VOICE OF DEAF AND MUTE
IRJET Journal
 
Hand gesture recognition using machine learning algorithms
CSITiaesprime
 
Gesture recognition using artificial neural network,a technology for identify...
NidhinRaj Saikripa
 
A Survey Paper on Controlling Computer using Hand Gestures
IRJET Journal
 
Gesture Recognition System
IRJET Journal
 
D0361015021
ijceronline
 
ppt of gesture recognition
Aayush Agrawal
 
Sign Language Recognition with Gesture Analysis
paperpublications3
 
Human Computer Interaction Based HEMD Using Hand Gesture
IJAEMSJORNAL
 
Indian Sign Language Recognition Method For Deaf People
Takrim Ul Islam Laskar
 
SLIDE PRESENTATION BY HAND GESTURE RECOGNITION USING MACHINE LEARNING
IRJET Journal
 
Ad

More from IRJET Journal (20)

PDF
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
IRJET Journal
 
PDF
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
IRJET Journal
 
PDF
Kiona – A Smart Society Automation Project
IRJET Journal
 
PDF
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
IRJET Journal
 
PDF
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
IRJET Journal
 
PDF
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
IRJET Journal
 
PDF
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
IRJET Journal
 
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
PDF
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
IRJET Journal
 
PDF
BRAIN TUMOUR DETECTION AND CLASSIFICATION
IRJET Journal
 
PDF
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
IRJET Journal
 
PDF
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
IRJET Journal
 
PDF
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
IRJET Journal
 
PDF
Breast Cancer Detection using Computer Vision
IRJET Journal
 
PDF
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
PDF
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
PDF
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
IRJET Journal
 
PDF
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
PDF
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
IRJET Journal
 
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
IRJET Journal
 
Kiona – A Smart Society Automation Project
IRJET Journal
 
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
IRJET Journal
 
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
IRJET Journal
 
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
IRJET Journal
 
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
IRJET Journal
 
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
IRJET Journal
 
BRAIN TUMOUR DETECTION AND CLASSIFICATION
IRJET Journal
 
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
IRJET Journal
 
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
IRJET Journal
 
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
IRJET Journal
 
Breast Cancer Detection using Computer Vision
IRJET Journal
 
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
IRJET Journal
 
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Ad

Recently uploaded (20)

PDF
A presentation on the Urban Heat Island Effect
studyfor7hrs
 
PDF
13th International Conference of Networks and Communications (NC 2025)
JohannesPaulides
 
PPTX
Introduction to Neural Networks and Perceptron Learning Algorithm.pptx
Kayalvizhi A
 
PPTX
Benefits_^0_Challigi😙🏡💐8fenges[1].pptx
akghostmaker
 
PPTX
Break Statement in Programming with 6 Real Examples
manojpoojary2004
 
PPT
Oxygen Co2 Transport in the Lungs(Exchange og gases)
SUNDERLINSHIBUD
 
PPT
04 Origin of Evinnnnnnnnnnnnnnnnnnnnnnnnnnl-notes.ppt
LuckySangalala1
 
PDF
MOBILE AND WEB BASED REMOTE BUSINESS MONITORING SYSTEM
ijait
 
PPTX
Pharmaceuticals and fine chemicals.pptxx
jaypa242004
 
PDF
Detailed manufacturing Engineering and technology notes
VIKKYsing
 
PDF
monopile foundation seminar topic for civil engineering students
Ahina5
 
PPTX
Types of Bearing_Specifications_PPT.pptx
PranjulAgrahariAkash
 
PPTX
Mining Presentation Underground - Copy.pptx
patallenmoore
 
PDF
UNIT-4-FEEDBACK AMPLIFIERS AND OSCILLATORS (1).pdf
Sridhar191373
 
PDF
1_ISO Certifications by Indian Industrial Standards Organisation.pdf
muhammad2010960
 
PDF
Book.pdf01_Intro.ppt algorithm for preperation stu used
archu26
 
PPTX
ISO/IEC JTC 1/WG 9 (MAR) Convenor Report
Kurata Takeshi
 
PDF
Set Relation Function Practice session 24.05.2025.pdf
DrStephenStrange4
 
PPTX
Presentation on Foundation Design for Civil Engineers.pptx
KamalKhan563106
 
PPTX
File Strucutres and Access in Data Structures
mwaslam2303
 
A presentation on the Urban Heat Island Effect
studyfor7hrs
 
13th International Conference of Networks and Communications (NC 2025)
JohannesPaulides
 
Introduction to Neural Networks and Perceptron Learning Algorithm.pptx
Kayalvizhi A
 
Benefits_^0_Challigi😙🏡💐8fenges[1].pptx
akghostmaker
 
Break Statement in Programming with 6 Real Examples
manojpoojary2004
 
Oxygen Co2 Transport in the Lungs(Exchange og gases)
SUNDERLINSHIBUD
 
04 Origin of Evinnnnnnnnnnnnnnnnnnnnnnnnnnl-notes.ppt
LuckySangalala1
 
MOBILE AND WEB BASED REMOTE BUSINESS MONITORING SYSTEM
ijait
 
Pharmaceuticals and fine chemicals.pptxx
jaypa242004
 
Detailed manufacturing Engineering and technology notes
VIKKYsing
 
monopile foundation seminar topic for civil engineering students
Ahina5
 
Types of Bearing_Specifications_PPT.pptx
PranjulAgrahariAkash
 
Mining Presentation Underground - Copy.pptx
patallenmoore
 
UNIT-4-FEEDBACK AMPLIFIERS AND OSCILLATORS (1).pdf
Sridhar191373
 
1_ISO Certifications by Indian Industrial Standards Organisation.pdf
muhammad2010960
 
Book.pdf01_Intro.ppt algorithm for preperation stu used
archu26
 
ISO/IEC JTC 1/WG 9 (MAR) Convenor Report
Kurata Takeshi
 
Set Relation Function Practice session 24.05.2025.pdf
DrStephenStrange4
 
Presentation on Foundation Design for Civil Engineers.pptx
KamalKhan563106
 
File Strucutres and Access in Data Structures
mwaslam2303
 

IRJET - Chatbot with Gesture based User Input

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3130 Chatbot with Gesture based User Input Rajkumar G1, Gayathri M2 1Student/CSE, SRM Institute of Science and Technology, Chennai-603203 2Assistant Professor/CSE, SRM institute of Science and Technology, Chennai ---------------------------------------------------------------------***---------------------------------------------------------------------- Abstract: The paper describes a approach towards building a chatbot with gesture based input. A gesture table will be described here which is a mapping sequence of the appropriate gestures to the appropriate commands they mean. After successful gesture recognition, the classified or recognized gesture is looked up in the table and is converted accordingly into text and then that text is fed into the chatbot program as input and appropriate command is executed. The paper selects existing known algorithms and aims to explores the overall effects it has and if such a application is suitable for areas such as deaf and dumb sign language or in public galleries, to quickly get required information 1. Introduction Gesture recognition is a topic in computer science with the goal of interpreting human gestures algorithms. Gestures can be created from any bodily motion but commonly originate from the face or hand. Quite a few approaches have been suggested using cameras and computer vision algorithms to interpret sign language. Gesture recognition applications in various fields such as automotive, consumer electronics, Transit, gaming, smartphones, defence, sign language translation sectors. Its used to provide a more natural interaction with the UI to save time and make things more intuitive. In entertainment, mainly gaming ,it can be used to add a new way of interaction to attract players. In smartphones, its used for faster unlock. For deaf and dumb, its used to for translating sign language into text. Gesture recognition has spawned several correlated technology such as touchless interface. Touchless user interface (TUI) is a interface through which a computer can be given commands or operations in form of body motion and gestures without touching a keyboard, mouse, or screen. For example, Microsoft's Kinect is a touchless game interface; This is basically another major application of gesture recognition. Many companies are invested in development of this technology. Intel corporation is researching how touch multi-factor authentication can help healthcare organizations minimize security risks and at the same time improve clinical efficiency.Here Multi-factor authentication means multiple layers/levels to authorise a particular transaction. Microsoft corporation is researching the use of touchless interaction within surgical settings, allowing images to be viewed and manipulated without any physical contact with traditional computer hardware through gesture recognition. Gesture recognition is linked to the field of digital image processing. It uses many algorithms and concepts of digital image processing along with concepts and algorithms of AI. Some of these include thresholding, otsu algorithm, anisotropic diffusion, Hidden Markov model, image editing, image restoration etc(belonging to digital image processing), machine learning ,computer vision involving automatic inspection, Assisting humans in identification tasks, Controlling processes, detecting events, interaction, modeling objects, navigation, organizing information. 2. Literature Survey Rupesh Prajapati, Vedant Pandey, Nupur Jamindar, Neeraj Yadav, Prof. Neelam Phadnis have published a research paper titled “Hand Gesture Recognition and Voice Conversion for Deaf and Dumb”. In it they accept the input in form of Video feed through webcam, they have proposed a method of Create a database of images for training and use PCA for creating a relationship out of its linearly uncorrelated characterstics and then use classfication algorithm like KNN and SVN.[1] Anchal Sood and Anju Mishra have published a research paper titled “AAWAAZ: A communication system for deaf and dumb-”.They have proposed a sign recognition system based on Harris algorithm for feature extraction and then it is stored in a matrix. This matrix is further used to match the image from the database. The system does have a few limitations. Binary imaging technique doesn’t perfectly seperates discrete objects from backgrounds of similar colors due to inaccurate thresholding algorithm. But the results are efficient.[2] Ms R. Vinitha and Ms A. Theerthana have published a research paper titled “Design And Development of Hand Gesture Recognition System For Speech Impaired People.” In it they accept input in form of Data glove(consisting of sensors,accelometer and PIC-microcontroller). Gesture recognition is based on hardware approach[3] Shangeetha, R. K., V. Valliammai, and S. Padmavathi have published a research paper titled “Computer vision based approach for Indian Sign Language character recognition”. In it they accept input in form of Live video feed from webcam. They have also proposed a method of using HIS model and then the feature extraction is done by distance transform method. The results are as follows,The gestures
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3131 stored and recognized and coverted to appropriate voice output.[4] R. Aravind, D.Anbasaran, K.Alice have published a research paper titled “GRS-Gesture Based Recognition System for Indian Sign Language Recognition System for Deaf and Dumb people”. In it they accept input in form of webcam. They have proposed a method of gesture recognition using color intensity variation and edge detection and background subtraction. The results were as follows, gesture symbols extracted from the image alone The gaps identified were problems identifying concurrent symbols[5] Tripathi K. and Nandi have published a research paper titled “Continuous indian sign language gesture recognition and sentence”. In they accept input in form of webcam. They have proposed a method using Orientation Histogram and PCA. The results were that highly accurate gesture recognition by reducing dimension features after extraction.[6] Daniel Simões Lopes have developed a gesture based recognition system as a basis for touchless user interface. In it they accept input in form of Hand gestures and body postures. They have proposed a method of Developed a gesture recognition system where each hand is can make gestures and body postures can also be used. They found out that that using touchless interface is indeed better than relying on 2-D devices as it allows for faster retrieval and greater flexibility in the sensitive surgical enviornment. The gaps that were identified were Precision tasks such as clipping plane and visualization and tagging are better performed with mouse due to not good gesture recognition,[7] Another interesting method proposed was relying on passive props.3D medical visualization relied on passive interface props. Hardware interface placed in a position where the user needs manipulate the control devices. The result was that the system facilitates a natural interaction between UI and user. The gaps that were identified were that It was not easy to stop or use in several situations, the proposed mechanism is not flexible enough[8] -Chatbot -Gonzalas Cenelia Article Internet Gonzalas Cenelia, gives a excellent approach to building a chatbot based on keyword concept. It involves finding keywords in the given user input and match it with a already predefine knowledge base which has keywords and assoiciated responses to it. Revolving around the concept of keyword, things like context of the conversation ,learning from feedback to add knowledge base, keyword ranking etc are also introduced [9] 3. Proposed System:- Note the high level view of the proposed system in the given diagram. Input: Imagestream from webcam Output: Appropriate textual output or execution of a command like search A. Gesture Recognition Module:- Input: Imagestream from webcam representing correct gestures Output:Successfully recognized gesture label(See gesture mapping sequence) The recognized gestures can be divided into two types. Counting gestures -Counting gestures refer to the gestures that indicate numbers like using only your index finger to count 1 ,or any finger for that matter. Implementation used:Convexity Defect General gestures -General gestures are the set of gestures that can be classified by a AI model which is trained on a database of images. Implementation used:CNN model For each of these types, A separate implementation is provided which is considered more optimum for recognizing those gestures. Using the convexity defect approach, we don’t have to train the AI using large number of images for each number representational gesture. Besides the convexity defect approach is very flexible ,it simply counts the number of fingers so we don’t have to exact representational gesture like for example using the pinky finger to represent 1 rather than the index finger.(To its recommended to use the normal ones). I. Convexity Defect Implementation:-
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3132 1)Main GUI Control Window -This window is the main navigation window used to activate all the modules /algorithms for sucessful gesture recognition 2)GUI Input window -This window activates the webcam and gives you input video feed to manipulate on 3)Thresholding(Background substraction) As the video feed continues ,every frame is continously getting converted to grayscale image. A grayscale image is a image without saturation aka color .The only information its pixels have is on the intensity of light falling on that pixel. So basically the image appears in shades of grey. After this process, thresholding is applied. Thresholding converts the grayscale image to a binary color image which typically are black and white. A value called the threshold point can be manually provided or using an algorithm can be automated . The threshold function will make the pixels whose intensity are greater than the threshold value white and vice versa for black. In our case,we have use otsu algorithm to calculate threshold point. 4)Finding contours and convex hull and defects Contours are the areas of the discrete objects that we capture from the image. We assume that the contour which has the maximum area in the binary image will be of our hand, thats why a white screen should be used to hide the background objects. We calculate then the convex hull which is the smallest polygon that encloses a set of points. Then the convexity defect is defined as the deepest point in the said binary image between two vertices of the convex hull. The actual definition of convexity defect in terms of physical boundaries of the object is given below Convexity defect is a cavity in an object (blob, contour) segmented out from an image. That means an area that do not belong to the object but located inside of its outer boundary -convex hull. II. Deep Learning Implementation: (Contd) The input consists of 20,000 images . The first Module Image URL loader and reader is a simple algorithm which reads all the appropriate pathnames of the files and converts them into arrays using opencv functions .It extracts all the images required for training and stores the labels for each image in a separate list. One list all the images and the other list contains all the corresponding labels of each image. During the process the images are converted to grayscale and resized to smaller size so that the training phase of CNN model will be shorter. Then its fed into the CNN model. It splits its data into 70% as training and 30% as testing. (In the below table ,the gestures can be mapped to the appropriate text according to the usage, for example thumb down to what or L to how etc, say if the application is to be used in a deaf/dumb language or a custom gesture system required etc.) It takes 30 mins approximately for five iterations over the entire database of images for training as well as the validating(testing) to complete. After that model is ready to predict furthur images. (A sample tableis given if this application is to be used in a museum) Gesture Label Thumb down How many artifacts are here? palm(Horizontal) What is the name of the place L Where is it located Fist(Horizontal) When was it founded Fist(Vertical) What is the most famous artifact here.
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3133 Tools used: -Python(Anaconda package) -Spyder(IDE) -Keras,tensorflow and opencv as python libraries B.Chatbot Template Response:- We put up ready-made templates as replies and those templates will have parameters to which will be replaced by the appropriate data relevent to the conversion Keywords:-The concept on the basis of which the chatbot will be implemented .The chatbot will scan the user input and see if a keyword exists in it. Theses keywords are stored in the knowledge base. And each of the keyword have a set of responses linked to it. So a response is chosen and given as reply Knowledge base:-The knowledge base is a 2-D array storing a set of keywords in the first column and its associated responses in the next columns. Keyword Ranking:-This is used when there are multiple keyword match and based on the furthur corresponding keyword match and repetitions, the appropriate keyword is selected then the corresponding reply is given The chatbot will be written in java based module which will accept the input as text which was converted appropriately from the gesture input. Screenshots: Deep learning implementation (Note:Predict prints a command line out put. The image is first captured and then saved As png file and then is fed into the deep Learning model) Convexity defect to count fingers to display numbers The output is planned to be written to Text file which will be fed into a java based chatbot Chatbot Output:(Separate module)
  • 5. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3134 Example sentences are given as input, it can be customised according to the table created by simple adding those responses to the knowledge base of the chatbot. References [1] Rupesh Prajapati, Vedant Pandey, Nupur Jamindar, Neeraj Yadav, Prof. Neelam Phadnis,”Hand Gesture Recognition and Voice Conversion for Deaf and Dumb”,IRJET,2018 [2] Anchal Sood and Anju Mishra, “AAWAAZ: communication system for deaf and dumb-”.IEEE,2016 [3]Ms R. Vinitha and Ms A. Theerthana,"Design And Development Of Hand Gesture Recognition System For Speech Impaired People.",2016 [4]Shangeetha, R. K., V. Valliammai, and S. Padmavathi,"Computer vision based approach for Indian Sign Language character recognition." Machine Vision and Image Processing (MVIP), IEEE,2012 [5] R.Aravind, D.Anbasaran, K.Alice,”GRS-Gesture Based Recognition System for Indian Sign Language Recognition System for Deaf and Dumb people”, International Journal of Trend in Scientific Research and Development - IJTSRD,2018 [6]Tripathi K. and Nandi, “Continuous indian sign language gesture recognition and sentence”,Procedia Computer,2015 [7]Daniel Simões Lopes “On the utility of 3D hand cursors to explore medical volume datasets with a touchless interface”, Journal of biomedical informatics,2017 Passive real-world interface props for neurosurgical visualization, Hinckley et al.CHI '94, [9]Otsu Algorithm-(Algorithm used for Thresholding, grayscale to binary),See OpenCv 3.4.2 documentation for algorithm implementation [10]Peter-Draukr algorithm(For calculation of contours) [11]Satoshi Suzuki and others. Topological structural analysis of digitized binary images by border following. Computer Vision, Graphics, and Image Processing, 30(1):32–46, 1985.(Algorithm used for finding contours) [12]CNN(For training the AI model)