Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/20923
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorAbbod, M-
dc.contributor.advisorSwash, M-
dc.contributor.authorAlnaim, Norah-
dc.date.accessioned2020-06-02T13:07:38Z-
dc.date.available2020-06-02T13:07:38Z-
dc.date.issued2020-
dc.identifier.urihttp://bura.brunel.ac.uk/handle/2438/20923-
dc.descriptionThis thesis was submitted for the award of Doctor of Philosophy and was awarded by Brunel University Londonen_US
dc.description.abstractHuman Computer Interaction (HCI) is a broad field involving different types of interactions including gestures. Gesture recognition concerns non-verbal motions used as a means of communication in HCI. A system may be utilised to identify human gestures to convey information for device control. This represents a significant field within HCI involving device interfaces and users. The aim of gesture recognition is to record gestures that are formed in a certain way and then detected by a device such as a camera. Hand gestures can be used as a form of communication for many different applications. It may be used by people who possess different disabilities, including those with hearing-impairments, speech impairments and stroke patients, to communicate and fulfil their basic needs. Various studies have previously been conducted relating to hand gestures. Some studies proposed different techniques to implement the hand gesture experiments. For image processing there are multiple tools to extract features of images, as well as Artificial Intelligence which has varied classifiers to classify different types of data. 2D and 3D hand gestures request an effective algorithm to extract images and classify various mini gestures and movements. This research discusses this issue using different algorithms. To detect 2D or 3D hand gestures, this research proposed image processing tools such as Wavelet Transforms and Empirical Mode Decomposition to extract image features. The Artificial Neural Network (ANN) classifier which used to train and classify data besides Convolutional Neural Networks (CNN). These methods were examined in terms of multiple parameters such as execution time, accuracy, sensitivity, specificity, positive predictive value, negative predictive value, positive likelihood, negative likelihood, receiver operating characteristic, area under ROC curve and root mean square. This research discusses four original contributions in the field of hand gestures. The first contribution is an implementation of two experiments using 2D hand gesture video where ten different gestures are detected in short and long distances using an iPhone 6 Plus with 4K resolution. The experiments are performed using WT and EMD for feature extraction while ANN and CNN for classification. The second contribution comprises 3D hand gesture video experiments where twelve gestures are recorded using holoscopic imaging system camera. The third contribution pertains experimental work carried out to detect seven common hand gestures. Finally, disparity experiments were performed using the left and the right 3D hand gesture videos to discover disparities. The results of comparison show the accuracy results of CNN being 100% compared to other techniques. CNN is clearly the most appropriate method to be used in a hand gesture system.en_US
dc.description.sponsorshipImam Abdulrahman bin Faisal Universityen_US
dc.language.isoenen_US
dc.publisherBrunel University Londonen_US
dc.relation.urihttps://bura.brunel.ac.uk/bitstream/2438/20923/1/FulltextThesis.pdf-
dc.subjectConvolutional Neural Networken_US
dc.subjectHoloscopic 3D videoen_US
dc.subjectWavelet Transformen_US
dc.subjectEmpirical Mode Decompositionen_US
dc.subjectHand Trackingen_US
dc.titleHand gesture recognition using deep learning neural networksen_US
dc.typeThesisen_US
Appears in Collections:Electronic and Computer Engineering
Dept of Electronic and Electrical Engineering Theses

Files in This Item:
File Description SizeFormat 
FulltextThesis.pdf6.41 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.