WO2023033725A2 - Sensory glove system and method for sign gesture sentence recognition - Google Patents

Sensory glove system and method for sign gesture sentence recognition Download PDF

Info

Publication number
WO2023033725A2
WO2023033725A2 PCT/SG2022/050609 SG2022050609W WO2023033725A2 WO 2023033725 A2 WO2023033725 A2 WO 2023033725A2 SG 2022050609 W SG2022050609 W SG 2022050609W WO 2023033725 A2 WO2023033725 A2 WO 2023033725A2
Authority
WO
WIPO (PCT)
Prior art keywords
input signal
sentence
fragments
glove
sensor
Prior art date
Application number
PCT/SG2022/050609
Other languages
French (fr)
Other versions
WO2023033725A3 (en
Inventor
Feng Wen
Zixuan Zhang
Tianyiyi HE
Chengkuo Lee
Original Assignee
National University Of Singapore
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by National University Of Singapore filed Critical National University Of Singapore
Publication of WO2023033725A2 publication Critical patent/WO2023033725A2/en
Publication of WO2023033725A3 publication Critical patent/WO2023033725A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/014Hand-worn input/output arrangements, e.g. data gloves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures

Definitions

  • the present invention relates, in general terms, to a sensory glove system for detecting sign gestures, and a method for sign gesture sentence recognition.
  • Sign language recognition is significant in lowering the communication barrier between the hearing/speech impaired (signers) and those who do not use sign language (non-signers).
  • Specially designed gloves are sometimes used to detect motion of the hands.
  • Existing glove solutions can only recognise discrete single gestures (i.e., numbers, letters, or words). This loses context when those single gestures are part of a sentence, and therefore do not satisfy signers' daily communication needs.
  • wearable sensors Hold great promise in various applications spanning from environmental monitoring, personalized healthcare to soft robotics, and human-machine interaction.
  • Typical wearable sensors rely on resistive and capacitive mechanisms for human status tracking and surrounding sensing. Such sensors often use an external power supply to generate excitation signals, inhibiting the self-sustainability and independence of glove solutions.
  • TENGs triboelectric nanogenerator
  • HMI human-machine interface
  • the proposed TENGs-based solutions provide self- powered triboelectric HMIs.
  • TENGs-based HMIs can seamlessly detect the multiple degrees-of-freedom motions of the dexterous hands, enabling sign language interpretation and system control command recognition.
  • an artificial intelligence enabled sign language recognition and communication system comprising sensing gloves, deep learning block, and virtual reality interface.
  • a sensory glove system for recognition of sign gestures comprising: a first glove for wearing on a first hand of a user, the first glove comprising: a first wrist motion sensor for wrist motion perception of a first wrist; a finger bending sensor for each finger (including thumb) of the first glove; and a palm sensor; and a second glove for wearing on a second hand of the user, the second glove comprising: a second wrist motion sensor for wrist motion perception of a second wrist; a finger bending sensor for each finger of the second glove; and two fingertip sensors, one on each of a middle and index finger.
  • Also disclosed is a method for sign gesture sentence recognition from an input signal comprising a sentence comprising: dividing the input signal into fragments; applying a machine learning model to the fragments to recognise individual said fragments; and inversely reconstructing the sentence to recognise the sentence.
  • the fragments may include intact word signals, incomplete word signals and background signals, and wherein the applying the machine learning model to the fragments comprises: determining, based on a predetermined threshold number of data points associated with the fragment, if the fragment corresponds to background noise; discarding the fragment if the fragment is determined to correspond to background noise; and recognising each fragment that has not been discarded.
  • the method may further comprises recognising the fragment as comprising background signals on the basis of the fragment comprising fewer than 100 non-zero data points, and tagging the fragment as "empty".
  • a non-zero data point is a data point with a value of "0" or sufficiently close to zero that the data point does not constitute part of a word.
  • a virtual reality conversation system comprising: a sensory glove system as described above; a signer terminal for use by a first user wearing the sensory glove system ; a non-signer terminal for use by a second user; at least one processor configured to: receive an input signal comprising an output from each wrist motion sensor, bending sensor, palm sensor and fingertip sensor; apply one or more machine learning models to the input signal to recognise a sentence in the input signal; and project the signal into a virtual reality environment, for display at the non-signer terminal.
  • finger shall be construed to include the fingers and thumb of the relevant hand
  • known sentence shall be construed to include words known to a machine learning model due to their appearance in a training data set upon which the machine learning model was trained
  • projected into virtual space shall be construed to include displaying one or more words or a sentence in a virtual reality environment in association with an avatar of the user who inputted the one or more words or sentence
  • inversely reconstruct means after an Al model recognizes all word elements split from sentence samples (i.e. fragments), it (the model) will reversely to recognize the sentences .
  • embodiments employing machine learning use nonsegmentation and segmentation assisted deep learning models to achieve recognition of words and sentences.
  • segmentation approach splits entire sentence signals into word units, referred to herein as "fragments" or "word elements”.
  • the deep learning model recognizes all word units and inversely reconstructs and recognizes sentences.
  • new sentences - i.e. sentences not yet encountered by the machine learning or Al model - created by a new-order or combination of word elements can be recognized with an average accuracy of 86.67%.
  • the sign language recognition results can be projected into virtual space and translated into text and audio, allowing the remote and bidirectional communication between signers and non-signers.
  • Figure 1 schematically illustrates a communication system for facilitating communication between signers (those who use sign language) and non-signers (those who do not use sign language), whether or not those communicating are at a distance from each other;
  • Figure 2 shows the detailed location, area information and channel label of sensors on gloves
  • Figure 3 shows the data analysis for signals of words and sentences
  • Figure 4 demonstrates word and sentence recognition based on the nonsegmentation method
  • Figure 5 shows the train and validation accuracy increase with epochs
  • Figure 6 demonstrates word and sentence recognition based on segmentation method
  • Figure 7 illustrates division of input signal, comprising a sentence, into fragments, identification of fragments and reconstruction (i.e. inverse reconstruction) of the sentence;
  • Figure 8 shows the recognition result of three new sentences
  • Figure 9 is a radar comparison map of non-segmentation and segmentation recognition methods based on their pros and cons; and Figure 10 comprises a demonstration of communication between the speech impaired and the non-signer.
  • Described below are sensory glove systems with sensor configurations that are optimised to minimise the number, and in some cases the area, of sensors required to determine sign language signals.
  • the systems described herein can interpret a large number of sign language words and sentences based on known words and sentences, but may also construct new sentences.
  • TENG gloves have been demonstrated for monitoring finger motions using magnitude analysis or pulse counting.
  • their data analytics are mostly based on direct feature extraction (e.g., amplitude, peak number). This leads to a limited variety of recognizable hand motions/gestures with substantial feature loss. Sophisticated hand gesture discrimination remains challenging. Such limitations are not present in some sensory glove systems of the present disclosure due to the placement of sensors and the manner in which sensor outputs are processed.
  • the system successfully realizes the recognition of 50 words and 20 sentences and can be expanded to a greater number of word sand sentences, even learning from new sentences - i.e. sentences that did not form part of the training dataset but that are recognised by the segmentation approach described below, by identifying all the known words in the new sentence.
  • the recognition results are projected into virtual space in the forms of comprehensible voice and text to facilitate barrier-free communication between signers and non-signers
  • FIG. 1 shows a virtual reality conversation system 100.
  • the system 100 comprises a sensory glove system 102 as described with reference to Figure 2.
  • a user wearing the sensory glove system 102 interacts -by direct typing and/or through commands or input signals issued by the sensory glove system 102 - with a signer terminal 104.
  • Input signals received from the sensors of the sensory glove system 102 are processed by a processor in at least one of (e.g. it may be a distributed processor) the sensory glove system 102 itself, the signer terminal 104 (processor 106), a non-signer terminal 108 that is used by a second user (processor 116), and a central server 110 (processor 118).
  • Instructions (i.e. program code) 112 are sent to the processor from memory 114, which may be in the signer terminal 104 as shown, or elsewhere in the system 100.
  • the processor of which there may be more than one, is configured to implement a method 120.
  • the method 120 involves receiving an input signal comprising an output from each sensor of the sensory glove system 102 (step 122), applying a machine learning model, or multiple models as needed, to the input signal to recognise a sentence in the input signal (step 124), and projecting the signal into a virtual reality environment, for display at the non-signer terminal (step 126).
  • the VR environment of VR space may be hosted by one of terminals 104 and 108, or may central server 110.
  • All terminals may communicate over network 132, which can be any suitable network.
  • a sensor 202 e.g. a triboelectric sensor
  • two sensors 204 are put on wrists for wrist motion perception.
  • the finger motion sensors and/or wrist motion sensors are positioned on the outside of the respective finger and wrist.
  • the index and middle fingertips of the right hand are in frequent use, and the left hand is often used to interact with the right hand and other parts of their body.
  • two sensors 206 are placed on the fingertips of right index and middle of the glove, and one sensor 208 is incorporated on the palm of the left hand.
  • Sensor placement on the gloves will be described with reference to left and right gloves. However, the sensors of the left glove may instead be provide on the right glove and vice versa, depending on user preference and hand dominance.
  • the reduction in sensors assists with minimising the number of sensors and system complexity, while maintaining gesture detection accuracy.
  • the left palm is often open and facing up at the end of a gesture. Therefore, the left palm usage, and thus left palm sensor measurement, is likely more influential in gesture detection than the right palm.
  • the gloves of system 200 are thus configured with 15 triboelectric sensors in total.
  • the area of sensors is optimised based on key parameters that largely influence the triboelectric voltage output performance, such as sensor area, force, bending degree, and bending speed.
  • the voltage output increases when sensor area, force, bending degree and bending speed increase.
  • image (c) where the sensors are triboelectric sensors (noting that capacitive, resistive or other sensors may be used despite being generally less desirable due to power consumption requirements), the triboelectrification layers are composed of ecoflex and wrinkled nitrile. With the conductive textile as electrodes, a flexible and thin triboelectric sensor is fabricated.
  • TENG sensors may comprise 00-30 Ecoflex with a 1: 1 weight ratio of part A and part B coated on a conductive textile (for electrode) for consolidation. Then the positive wrinkled nitrile is attached on the conductive textile contacts with the negative Ecoflex layer to assemble the triboelectric sensor. Finally, the conductive electrodes are encapsulated by non-conductive textiles to shield the ambient electrostatic effect. Different sensors with different areas are fabrication for customized glove configuration.
  • the glove itself may comprise the abovementioned two-layered triboelectric sensors sewed into a small cotton textile pocket.
  • input signals are generated by the sensors - an input signal comprising the output signals of all sensors. This can include sensors of parts of the hand that are not involved in the gesture.
  • the input signal is then received (per step 122) and processed according to step 124. Processing can occur in various ways. Discussed with reference to Figures 3 and 4 is a non-segmentation method, whereby the input signal is processed by a machine learning (i.e. Al) model trained using data comprising various known words and sentences.
  • the non-segmentation method effectively compares the input signal with the training data to determine which known sentence of word is most similar, or correlates best, to the input signal.
  • the non-segmentation non-segmentation Al framework described below achieves high accuracy for 50 words (91.3%) and 20 sentences (95%) by independently identifying word and sentence signals. Naturally, this can be extended to cover a greater number of words and sentences.
  • the developed segmentation Al frame splits the entire sentence signal into word units (fragments) and then recognizes all the signal fragments. Inversely, the sentence can be reconstructed and recognized with an accuracy of 85.58% upon established correlation between word units and sentences. In some embodiments, this requires the fragments to represent known words.
  • the segmentation approach may comprise dividing the input signal into fragments, applying a machine learning or Al model to each fragment to recognise one or more words in the fragment, and reconstructing the sentence from the words reflected in the fragments.
  • the segmentation approach can be employed to recognise new sentences (average correct rate: 86.67%) that are created by word elements or fragments combined in a new order.
  • the new sentence can be stored and, if repeatedly encountered, used as training data for the machine learning model employed in the non-segmentation method.
  • the sensory glove system or terminal to which it is connected may comprise a signal acquisition module (e.g. as part of processor 106) that acquires the generated triboelectric signals generated by different gestures - e.g. using chicken MEGA 2560 microcontroller.
  • a signal acquisition module e.g. as part of processor 106
  • For the non-segmentation method during training 100 samples for each word and sentence are collected where 80 samples (80%) for training and 20 samples (20%) for testing. Thus, there are 5000 samples for total 50 words and 2000 samples for total 20 sentences.
  • For training the segmentation approach since all the data form sentences, there are 50 samples with number series labels for each sentence (17 sentences and 850 samples in total).
  • each data point may be zero or non-zero, and thus constitute background signal or potentially part of a gesture or word.
  • each sentence sample with 800 data points is segmented into 13 elements.
  • the extracted data elements are labelled with numbers of corresponding words or empty.
  • Each number represents a signal fragment that may be an intact gesture signal, background noise, or an incomplete gesture signal (e.g. part of a word in sign language).
  • 60% of samples for each number i.e., 0-19) are used for training, 20% for validation, and 20% for testing.
  • 5 samples for 3 new sentences are employed to verify the capability of new sentence recognition of the segmentation-assisted CNN model without prior training.
  • gestures were selected based on their daily use in a signer's life. These gestures are reflected in Figure 3, image (a). Using the sensory glove system 200, corresponding triboelectric signals of these gestures are given in Figure 3, image (b). These triboelectric signals are the input signals for each respective word.
  • the system 100 then applies one or more machine learning models to the input signal, to recognise a sentence in the input signal.
  • the sentence comprises a single word in each case.
  • the system 100 recognises the words by applying the machine learning model to perform signal similarity and correlation analysis against the original data.
  • there are 100 data samples for each gesture in the dataset where 'Get' is shown in the enlarged view in the below middle of Figure 3, image (b).
  • image (b) shows a new input 300 of 'Get' compared with its own database 302, from which a mean correlation coefficient of 0.53 is calculated using any appropriate means - e.g. least squares error between data points (x-axis) of the new input signal over the average values of the database signal.
  • 0.53 is above a predetermined threshold (presently 0.3) that is the nominal cut-off for a high correlation.
  • the training samples may be averaged to enable correlation with the new input signal.
  • the signal of 'Must' is compared with the 'Get' database.
  • the mean correlation coefficient of 0.37, between the input signal 300 and the database signal for 'Must' 304, is also larger than 0.3. This shows confusion by the system, since both 'Get' and 'Must' indicate high signal similarity (the bottom right of Figure 3, image (c)).
  • the traditional basic recognition algorithm may not be suitable for a discrimination between individual gestures where there is a large number of gestures between which is must discriminate.
  • This also means the recognising individual gestures using traditional methods can result in nonsensical sentences - e.g. gestures representing "Get the doctor” may be recognised as "Must the doctor”.
  • present methods employ advanced data analytics such as machine learning to get a better recognition accuracy, and perform full sentence recognition against a database of known sentences.
  • Figure 3 is a matrix representation of the correlation coefficient between input signals (i.e. gestures corresponding to individual words) of known words and database signals for all other known words.
  • input signals i.e. gestures corresponding to individual words
  • image (d) being a distribution curve of correlation coefficients from which it can be observed that signals of several gestures have a strong correlation. This means there is a high similarity among these gesture signals. These gestures are susceptible to incorrect classification. The same is shown for sentences in daily use, in Figure 3, images (e) and (f).
  • CNN convolutional Neural Networks
  • INN 1-dimensional
  • box plots were used to assess kernel size (Figure 4, image (a)), filter number (Figure 4, image (b)), and number of convolutional layers (Figure 4, image (c)).
  • the box plots indicate median (middle line), 25 th , 75 th percentile (box) and 5 th and 95 th percentile (whiskers) as well as outliers (single points).
  • the machine learning model used to assess similarity or correlation between input signal and known words or sentences may comprise a ID CNN model with 5 kernels, 64 filters, and 4 convolutional layers.
  • Table 1 The detailed CNN structure parameters are presented in Table 1.
  • CNN parameters The detailed parameters for constructing Convolutional Neural Network (CNN).
  • the CNN is schematically represented in Figure 4, image (d), in which the 15- channel signals are served as inputs without segmentation.
  • image (d) in which the 15- channel signals are served as inputs without segmentation.
  • sentence signals are not pieced into word elements and not linked with the basic word units within a non-segmented deep learning framework.
  • the signals of words and sentences are isolated from each other and independently recognized.
  • the present machine learning model may use any dimension reduction method, such as Principal Component Analysis (PCA), to reduce the dimension of the output data while maintaining information.
  • PCA Principal Component Analysis
  • PCA can be performed using eigenvalues and eigenvectors of an analogue signal matrix of gestures, which can point along the major variation directions of data.
  • the dimension of data could be reduced with adjustable principal component matrix which comprises the non-unique eigenvectors of the data matrix.
  • PCA Principal Component Analysis
  • PCA is used to reduce the dimensions whilst retaining maximum information.
  • the principle of PCA relies on the correlation between each dimension and provides a minimum number of variables which keeps the maximum amount of variation about the distribution of original data.
  • PCA employs the eigenvalues and eigenvectors of the data-matrix which can point along the major variation directions of data to achieve the purpose of dimension reduction.
  • the vector X includes the component that is the input signal Xi for each gesture,
  • the eigenvalues A and eigenvectors p of the covariance matrix are defined as,
  • Figure 4 image (e), shows feature clustering result of 50 words for the input layer.
  • Feature clustering can be achieve using any known method. Presently, it was achieved by the t-distributed Stochastic Neighbor Embedding (t-SNE) in the dimension of principal component 1 (PCI) and principal component 2 (PC2).
  • t-SNE t-distributed Stochastic Neighbor Embedding
  • Figure 4 image (e) shows the performance of feature clustering is poor with category overlaps before going through the CNN network.
  • image (f) indicates a desirable classification performance of the developed CNN model with clear boundaries between these 50 classes with less overlap.
  • sentence recognition as reflected in Figure 4, images (h) and (i) respectively.
  • the CNN model upon non-segmentation approach only enables classifying known sentence signals in the dataset. It cannot identify new (i.e. not before seen) sentences even where a sentence consists of known words, if those known words are in an order not previously seen by the system 100. This limitation results from labelling each word or sentence with a single label as a distinct item - i.e. the CNN does not distinguish between word and sentence for analysis purposes.
  • the CNN model employing non-segmentation methods needs to leverage the whole long-data-length signal (generally 200 data points for the word signal and 800 data points for the sentence signal) to trigger the recognition procedure. This makes it undesirable for the real-time sign language translation.
  • the segmentation method involves dividing the input signal into fragments, analysing the fragments to identify words, and inversely reconstructing the sentence based on the identified words.
  • the analysis step may be twofold - firstly, distinguishing between fragments representing background noise and those representing words (i.e. filtering) and, secondly, performing recognition on fragments that represent words.
  • Sentences may be divided into fragments using a data sliding window. Each window may comprise an intact word signal, incomplete word signal or background signal (i.e. noise). These fragments split from all sentence signals are recognized first, and then the CNN model will inversely reconstruct and recognize the sentences. In this way, the segmentation approach can identify known sentences but also identify or interpret new sentences from known words. Since the input signal can be divided as soon as data comprising the 'next' window is received, the input signal can be processed progressively. Consequently, the recognition latency is significantly reduced due to smalldistance sliding window length (50 data points per sliding) as the recognition process will be triggered with each sliding. Any new sentences can be used to expand the sentence database. Therefore, the machine learning models only need to be trained on words, with the segmentation method enabling individual words to be detected and sentences constructed in the database - e.g. memory 132 of Figure 1.
  • Figure 6 image (a) shows 19 words (W01-W19) labelled from 0 to 18 based on usage frequency. Background noise is considered an 'empty' word (W20) labelled 19.
  • the label action is a critical step for the supervised learning of the CNN model.
  • Figure 6, image (b) demonstrates the detailed process of sentence signal division and labelling.
  • the system 100 takes the sentence 'The dog scared me'. It analyses the voltage on each channel (i.e. the output of each sensor, as reflected in the input signal) as detected at many points in time, the x-axis reflected 800 data points for the sentence. Using a 200 data point sliding window and a sliding step 50 data points, the sentence becomes 13 fragments.
  • Each fragment is an intact single word, background signal, and mixture signal (i.e., incomplete word signals and background signals).
  • this fragment signal will be labelled with the number of the corresponding word as shown in Figure 6, image (a).
  • the background noise is labelled as 'empty' (W20) with number 19.
  • the label is assigned based on the principal component. If word signal or background noise accounts for more than 50% of the sliding window size, the label will be the word number of the word in the fragment, or 'empty' 19, respectively.
  • the dog scared me' is represented by the label number series '[5 5 19 19 19 15 15 15 19 19 0 0 0]'.
  • image (c) summarizes the essential information of 20 sentences: sentence category codes (Y01-Y17 and Newl-New3), and word/gesture components, an example label series, and unique number orders derived from the label series minus background signals.
  • sentence category codes Y01-Y17 and Newl-New3
  • word/gesture components an example label series
  • Newl, New2 and New3 are unknown sentences used to verify the potential of recognizing new sentences. There is no corresponding label series for these three sentences.
  • the single classifier was built to identify all the fragments from 17 sentences without preliminarily filtering empty signals (Figure 6, image (d)).
  • the dataset contains 50 samples for each sentence and hence 850 (50*17) sentence samples in total.
  • the word fragments are from 510 samples for training (60%), 170 samples for validation (20%), and 170 samples for testing (20%).
  • the confusion matrix in Figure 6, image (e) shows an accuracy of 81.9% for the single classifier recognizing word units W01-W20.
  • the sentence is inversely reconstructed - i.e. constructed from recognised word fragments to form a text or audio output.
  • image (f) the result mapping of sentence recognition illustrates an accuracy of 79.41% in sentence reconstruction for known sentences. This is too poor an accuracy, particularly for Y01 and Y04. This results from the system endeavouring to classify empty words (i.e. W20).
  • a hierarchical classifier is shown in Figure 6, image (g).
  • the hierarchical classifier 600 has a first-level classifier 602 to separate background signals from intact word signals (i.e. fragments comprising an intact word or at least 50% data points relate to a word), and a second-level classifier 604 for precise identification of word signals.
  • the recognition accuracy of the hierarchical classifier is 82.81% for word fragments - the related confusion map is shown in Figure 6, image (h). Sentence recognition was also higher, at an average accuracy of 85.58% as shown in Figure 6, image (i).
  • the method employed by system 100 may therefore comprise dividing the input signal into fragments that include intact word signals, incomplete word signals and background signals.
  • the machine learning is then applied to the fragments to determine, based on a predetermined threshold number of data points associated with the fragment, if the fragment corresponds to background noise (e.g. 50% of data points relate to word or background noise), discard each fragment if the fragment is determined to correspond to background noise, perform recognition on each fragment that has not been discarded.
  • background noise e.g. 50% of data points relate to word or background noise
  • FIG. 7 illustrates recognition of a new sentence by segmentation, real-time sequential fragment identification, and sentence recognition.
  • the CNN classifier is used, with 5 samples for each new sentence engaged validating the CNN classifier performance in recognizing new sentences.
  • both classifiers render the number series prediction for the total 15 inputs of these 3 new sentences although the CNN model has not had the opportunity to be trained on the true label series.
  • the hierarchy classifier shows a reliable performance for new sentence recognition, indicated by fewer incorrect predictions - bold and underlined figures.
  • Figure 8 shows the results after translation or inverse reconstruction of the new sentence, with Figure 8, image (a) being the accuracy for the single classifier - 60% - and Figure 8, image (b) being the accuracy for the hierarchical classifier
  • Figure 9 illustrates the pros and cons of non-segmentation and segmentation methods.
  • the non-segmentation classifier i.e. non-segmentation approach
  • the segmentation method i.e. segmentation-assisted CNN
  • the method may employ both techniques in sequence - e.g. the input signal may be first processed using the non-segmentation approach. If this results in a correlation between the input signal and known words/sentences below a predetermined threshold (e.g. 0.3), then the segmentation method is employed using a hierarchical classifier. New sentences identified by the segmentation classifier (i.e. segmentation method) can be incorporated into the training dataset.
  • a predetermined threshold e.g. 0.3
  • the sensory glove system and the methods described above can overcome the drawbacks of image recognition systems used in sign language recognition.
  • the sensory glove system is not affected by varying luminance and can work well even under entirely dark conditions with higher environmental tolerance.
  • sensory glove systems do not capture facial and other data that is otherwise inadvertently captured by image recognition systems.
  • the system 100 may implement a virtual reality conversation system.
  • the signer and non-signer terminals in VR. space allow display sign recognition results and the direct typing of non-signers.
  • the VR interface facilitates two-way remote communication, linked with the Al front end for sign language recognition.
  • the VR conversation system may implement a broader VR environment in which more than two parties may interact, regardless of the number of signers and non-signers among them.
  • the deep learning integrated VR interface realizes bidirectional communication.
  • the system 100 used for recognition and communication, therefore comprises five major blocks including triboelectric gloves (sensory glove system 102) for hand motion capture, the printed circuit board (PCB) for signal preprocessing (this may be in the processor 104 or 118), loT module chicken connected PC for data collection (this may be part of the signer terminal 104), deep-learning-based analytics for signal recognition (stored in central server 110, though may be elsewhere in some embodiments), and VR. interface in Unity for interaction.
  • Terminals 104 and 108 and nodes 128 and 130 may each be a VR interface, and may comprise VR headset, speaker for audio output of recognised input signals from terminals operated by signers and of text input from terminals operated by non-signers, and a display.
  • this recognised word or sentence will be projected into cyberspace, in which Al will send corresponding commands based on recognition results and deal with inputs from the nonsigner, to control the communication in VR interface based on Transmission Control Protocol/Internet Protocol (TCP/IP).
  • TCP/IP Transmission Control Protocol/Internet Protocol
  • the deep learning block recognizes the non-mainstream sign language, translates it into the prevalent conversational medium such as text and audio, and transmits such information into the next virtual space.
  • image (b) sequence (i) to (v) the VR interface is designed for communication between the speech impaired and those without speech impairment, with switched host-guest views in cyberspace.
  • the client and server are built in the VR interface based on TCP/IP and are accessible for the signer and non-signer, respectively.
  • the client of VR interface (akin to terminal 104) allows the speech impaired to use the sign language that they are familiar with to engage in the communication. More precisely, the sign language delivered by the speech impaired is recognized and translated into speech and text by deep learning. Then the speech and text are captured and sent to the non-signer-controlled server (akin to terminal 108, or server 110).
  • the healthy user employs the server to type directly responding to the speech disordered user.
  • image (b) demonstrates, a greeting scenario is created to demonstrate the feasible communication between the speech/hearing impaired and healthy user under the identical local area network (LAN), making interaction more realistic with Al integrated VR interface.
  • LAN local area network
  • sequence number (i) the speech-impaired user Lily performs sign language 'How are you?' which is recognized and converted to text and audio by the deep learning model.
  • the client connected with deep learning component receives the text/speech message 'How are you?' and transmit it to the nonsigner Mary controlled server.
  • the signer avatar Lily slightly lifts her hand to greet her friend (i.e., the non-signer Mary).
  • the non-signer types 'Not good. I have a stomach ache.' to respond to the signer.
  • the virtual girl Mary represents the non-signer shakes her head and covers her stomach with hands to show her illness, as shown in Figure 10, image (b), sequence number (ii). Then the speech-impaired user replies to the non-signer Mary with the sign language 'You need a doctor' Figure 10, image (b), sequence number (iii).
  • the VR communication interface linked with Al allows interaction of the speech/hearing-disordered and healthy people closely and even remotely, providing a promising platform for the immense interactions between two population groups.
  • Figure 10 image (a) shows the triboelectric sensory glove system 1000 with 15 sensor channels connected to an
  • Python 1006 - implementing CNN
  • the raw data of different gestures is acquired and recognized by the trained CNN model with Tensorflow and Keras frames, then displaying recognition results on the Python interface.
  • the words or sentences recognized by the Python script (1006) are sent into virtual Unity space via TCP/IP communication and displayed on the VR. interface.
  • the server (for non-signers) and client (for signers) terminal building in Unity also rely on TCP/IP.
  • the speech/hearing impaired and healthy people can remotely communicate via the VR interface where the speech/hearing disordered uses their familiar sign language, and the non-signer types directly.
  • a sign language recognition and communication system comprising a smart triboelectric sensory glove system, Al block (implemented by a processor on one or more of the terminal and nodes shown in Figure 1), and the back-end VR interface (server 110 of Figure 1).
  • the system enables the separate and independent recognition of words (i.e., single gestures) and sentences (i.e., continuous multiple gestures) with high accuracies of 91.3% and 95% within non-segmentation frame.
  • the segmentation method is proposed. It divides all the sentence signals into word fragments while Al learns and memorizes all split elements.
  • the embedded VR interface can act as the bridge of user terminals, transmitting messages back and forth, configuring a closed-loop communication system with gloves and Al component.
  • the speech/hearing impaired can directly perform sign language to interact with the non-signer (i.e., human-to- human interaction), while the non-signer directly involves in the communication process via direct typing.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Image Analysis (AREA)
  • User Interface Of Digital Computer (AREA)
  • Machine Translation (AREA)

Abstract

A system for recognition of sign gestures and barrier communication between signers and non-signers. The system includes a first glove and second glove, each including a motion sensor for wrist motion perception, a finger bending sensor for each finger, a palm sensor on the first glove and two fingertip sensors, one on each of a middle and index finger of the second glove. For segmentation assisted sentence recognition, the data sliding window splits sentence signals in dataset into word elements. The deep learning (DL) classifier first recognizes all word units as well as some kinds of relations among word signals. Then DL model can recognize new/never-seen sentences which are composed of new-order recombination of word elements. Besides, mutual communication at server and client terminals of VR interface is enabled by Transmission Control Protocol/Internet Protocol (TCP/IP).

Description

SENSORY GLOVE SYSTEM AND METHOD FOR SIGN GESTURE SENTENCE RECOGNITION
Technical Field
The present invention relates, in general terms, to a sensory glove system for detecting sign gestures, and a method for sign gesture sentence recognition.
Background
Sign language recognition, especially sentence recognition, is significant in lowering the communication barrier between the hearing/speech impaired (signers) and those who do not use sign language (non-signers). Specially designed gloves are sometimes used to detect motion of the hands. Existing glove solutions can only recognise discrete single gestures (i.e., numbers, letters, or words). This loses context when those single gestures are part of a sentence, and therefore do not satisfy signers' daily communication needs.
Benefiting from a number of attractive features such as light weight, good compliance, and comfort, wearable sensors hold great promise in various applications spanning from environmental monitoring, personalized healthcare to soft robotics, and human-machine interaction. Typical wearable sensors rely on resistive and capacitive mechanisms for human status tracking and surrounding sensing. Such sensors often use an external power supply to generate excitation signals, inhibiting the self-sustainability and independence of glove solutions.
Being closely associated with sophisticated hand gestures and as an essential part of biomedical care, sign language interpretation is of substantial importance. This is particularly the case since the general public typically cannot sign. Generally, visual images/videos, surface electromyography (sEMG) electrodes, and inertial sensors are conventional means to reconstruct hand gesture information towards sign language recognition applications. These communication solutions are limited by light conditions and privacy concerns, electromagnetic noise and crosstalk with other biopotentials, or a huge amount of data.
It would be desirable to provide a system that can recognise sign gestures in a more comprehensive and natural way than prior art solutions, that can recognise and translate signed sentences and overcomes the drawbacks of, or at least to provide a useful alternative to, the above systems.
Summary
The present disclosure leverages triboelectric nanogenerator (TENG) based wearable sensors. TENGs are simple to fabricate, allow wide material choice, and expeditious dynamic response, are power-compatible and self-sustainable. The present disclosure uses TENGs as the basis for a human-machine interface (HMI). HMIs are often resistive or capacitive, which results in the need for a power supply. However, the proposed TENGs-based solutions provide self- powered triboelectric HMIs. When incorporated into a glove as taught herein, TENGs-based HMIs can seamlessly detect the multiple degrees-of-freedom motions of the dexterous hands, enabling sign language interpretation and system control command recognition.
Complex gestures and sentences are interpreted using the present TENGs gloves by applying artificial intelligence (Al) to outputs of TENGs sensors. The use of Al, when combined with the present sensor arrangement, allows extraction of features from which more diversified and complex gesture monitoring is achievable when compared with direct feature extraction - e.g. peak and amplitude detection.
Following from this, presently disclosed is an artificial intelligence enabled sign language recognition and communication system comprising sensing gloves, deep learning block, and virtual reality interface.
Disclosed is a sensory glove system for recognition of sign gestures, comprising: a first glove for wearing on a first hand of a user, the first glove comprising: a first wrist motion sensor for wrist motion perception of a first wrist; a finger bending sensor for each finger (including thumb) of the first glove; and a palm sensor; and a second glove for wearing on a second hand of the user, the second glove comprising: a second wrist motion sensor for wrist motion perception of a second wrist; a finger bending sensor for each finger of the second glove; and two fingertip sensors, one on each of a middle and index finger.
Also disclosed is a method for sign gesture sentence recognition from an input signal comprising a sentence, comprising: dividing the input signal into fragments; applying a machine learning model to the fragments to recognise individual said fragments; and inversely reconstructing the sentence to recognise the sentence.
The fragments may include intact word signals, incomplete word signals and background signals, and wherein the applying the machine learning model to the fragments comprises: determining, based on a predetermined threshold number of data points associated with the fragment, if the fragment corresponds to background noise; discarding the fragment if the fragment is determined to correspond to background noise; and recognising each fragment that has not been discarded.
For each fragment comprising background signals, the method may further comprises recognising the fragment as comprising background signals on the basis of the fragment comprising fewer than 100 non-zero data points, and tagging the fragment as "empty". In this context, a non-zero data point is a data point with a value of "0" or sufficiently close to zero that the data point does not constitute part of a word. Also disclosed herein is a virtual reality conversation system comprising: a sensory glove system as described above; a signer terminal for use by a first user wearing the sensory glove system ; a non-signer terminal for use by a second user; at least one processor configured to: receive an input signal comprising an output from each wrist motion sensor, bending sensor, palm sensor and fingertip sensor; apply one or more machine learning models to the input signal to recognise a sentence in the input signal; and project the signal into a virtual reality environment, for display at the non-signer terminal.
Unless context dictates otherwise, the following terms shall be given the meaning set out here: "finger" shall be construed to include the fingers and thumb of the relevant hand; "known sentence", "known word" and similar shall be construed to include words known to a machine learning model due to their appearance in a training data set upon which the machine learning model was trained; "projected into virtual space" shall be construed to include displaying one or more words or a sentence in a virtual reality environment in association with an avatar of the user who inputted the one or more words or sentence; and "inversely reconstruct" means after an Al model recognizes all word elements split from sentence samples (i.e. fragments), it (the model) will reversely to recognize the sentences .
Advantageously, embodiments employing machine learning use nonsegmentation and segmentation assisted deep learning models to achieve recognition of words and sentences. Significantly, the segmentation approach splits entire sentence signals into word units, referred to herein as "fragments" or "word elements". Then the deep learning model recognizes all word units and inversely reconstructs and recognizes sentences. Furthermore, new sentences - i.e. sentences not yet encountered by the machine learning or Al model - created by a new-order or combination of word elements can be recognized with an average accuracy of 86.67%. Finally, in embodiments providing a VR. conversation system the sign language recognition results can be projected into virtual space and translated into text and audio, allowing the remote and bidirectional communication between signers and non-signers.
Brief description of the drawings
Embodiments of the present invention will now be described, by way of nonlimiting example, with reference to the drawings in which:
Figure 1 schematically illustrates a communication system for facilitating communication between signers (those who use sign language) and non-signers (those who do not use sign language), whether or not those communicating are at a distance from each other;
Figure 2 shows the detailed location, area information and channel label of sensors on gloves;
Figure 3 shows the data analysis for signals of words and sentences;
Figure 4 demonstrates word and sentence recognition based on the nonsegmentation method;
Figure 5 shows the train and validation accuracy increase with epochs;
Figure 6 demonstrates word and sentence recognition based on segmentation method;
Figure 7 illustrates division of input signal, comprising a sentence, into fragments, identification of fragments and reconstruction (i.e. inverse reconstruction) of the sentence;
Figure 8 shows the recognition result of three new sentences;
Figure 9 is a radar comparison map of non-segmentation and segmentation recognition methods based on their pros and cons; and Figure 10 comprises a demonstration of communication between the speech impaired and the non-signer.
Detailed description
Described below are sensory glove systems with sensor configurations that are optimised to minimise the number, and in some cases the area, of sensors required to determine sign language signals. When combined with Al models for interpreting the outputs of the TENGs sensors, the systems described herein can interpret a large number of sign language words and sentences based on known words and sentences, but may also construct new sentences.
TENG gloves have been demonstrated for monitoring finger motions using magnitude analysis or pulse counting. However, their data analytics are mostly based on direct feature extraction (e.g., amplitude, peak number). This leads to a limited variety of recognizable hand motions/gestures with substantial feature loss. Sophisticated hand gesture discrimination remains challenging. Such limitations are not present in some sensory glove systems of the present disclosure due to the placement of sensors and the manner in which sensor outputs are processed.
In addition, previous attempts at using Al technology for interpreting sign signals have been limited in the number of sign language gestures they can understand. Previous approaches lack an effective and practical approach for real-time sentence recognition of sign language which is more significant for the practical communication of signers and non-signers. Besides, the interfaces (e.g., mobile phone or PC) that sign language recognition results are projected into or displayed on usually do not allow the signer's interaction with nonsigners - i.e. they are one-way communication systems. Some systems disclosed herein enable two-way communication through a VR inspired communication system. Here, we show a sign language recognition and communication system comprising triboelectric sensor integrated gloves, Al block, and the VR. interaction interface. The system successfully realizes the recognition of 50 words and 20 sentences and can be expanded to a greater number of word sand sentences, even learning from new sentences - i.e. sentences that did not form part of the training dataset but that are recognised by the segmentation approach described below, by identifying all the known words in the new sentence. The recognition results are projected into virtual space in the forms of comprehensible voice and text to facilitate barrier-free communication between signers and non-signers
Figure 1 shows a virtual reality conversation system 100. The system 100 comprises a sensory glove system 102 as described with reference to Figure 2. A user wearing the sensory glove system 102 interacts -by direct typing and/or through commands or input signals issued by the sensory glove system 102 - with a signer terminal 104. Input signals received from the sensors of the sensory glove system 102 are processed by a processor in at least one of (e.g. it may be a distributed processor) the sensory glove system 102 itself, the signer terminal 104 (processor 106), a non-signer terminal 108 that is used by a second user (processor 116), and a central server 110 (processor 118). Instructions (i.e. program code) 112 are sent to the processor from memory 114, which may be in the signer terminal 104 as shown, or elsewhere in the system 100.
The processor, of which there may be more than one, is configured to implement a method 120. The method 120 involves receiving an input signal comprising an output from each sensor of the sensory glove system 102 (step 122), applying a machine learning model, or multiple models as needed, to the input signal to recognise a sentence in the input signal (step 124), and projecting the signal into a virtual reality environment, for display at the non-signer terminal (step 126). The VR environment of VR space may be hosted by one of terminals 104 and 108, or may central server 110. Notably, particularly where the VR environment involves a large number of parties in addition to those communication through terminals 104 and 108 - as represented by nodes 128, 130, which are two of potentially very many - it can be useful for the VR space to be hosted by a central server 110 to facilitate flexible resourcing. All terminals may communicate over network 132, which can be any suitable network.
Glove configuration and sensor characterization To measure as many gestures as possible with a small number of sensors - i.e. parsimonious sensor usage - the sensor positions should be optimized. By referring to the frequently used sign language in the American Sign Language guide book, an analysis was conducted for involved motions in daily sign expressions of the speech/hearing impaired. It was determined that sign language includes three major motion dynamics, including elbow/shoulder motions, face muscle activities, and hand movements. Hand motion dominated the other motion dynamics, accounting for 43% of communication. Analysis of individual finger, wrist, palm and fingertip motions were then analysed, with communication relying on finger bending (56%), wrist motion (18%), touch with fingertips (16%), and interaction with palm (10%). With reference to Figure 2, image (b), sensors were therefore placed to optimise communications while minimising the number of sensors. Correspondingly, in the glove system 200, a sensor 202 (e.g. a triboelectric sensor) is mounted on each finger for finger bending measurement, and two sensors 204 are put on wrists for wrist motion perception. The finger motion sensors and/or wrist motion sensors are positioned on the outside of the respective finger and wrist. The index and middle fingertips of the right hand are in frequent use, and the left hand is often used to interact with the right hand and other parts of their body. Thus, two sensors 206 are placed on the fingertips of right index and middle of the glove, and one sensor 208 is incorporated on the palm of the left hand. Sensor placement on the gloves will be described with reference to left and right gloves. However, the sensors of the left glove may instead be provide on the right glove and vice versa, depending on user preference and hand dominance.
While a palm sensor could be placed on each glove, only one palm sensor is allocated for present purposes. The reduction in sensors assists with minimising the number of sensors and system complexity, while maintaining gesture detection accuracy. The left palm is often open and facing up at the end of a gesture. Therefore, the left palm usage, and thus left palm sensor measurement, is likely more influential in gesture detection than the right palm.
With reference to Figure 2, image (a), the gloves of system 200 are thus configured with 15 triboelectric sensors in total. The area of sensors is optimised based on key parameters that largely influence the triboelectric voltage output performance, such as sensor area, force, bending degree, and bending speed. The voltage output increases when sensor area, force, bending degree and bending speed increase. With reference to Figure 2, image (c), where the sensors are triboelectric sensors (noting that capacitive, resistive or other sensors may be used despite being generally less desirable due to power consumption requirements), the triboelectrification layers are composed of ecoflex and wrinkled nitrile. With the conductive textile as electrodes, a flexible and thin triboelectric sensor is fabricated.
More particularly, TENG sensors may comprise 00-30 Ecoflex with a 1: 1 weight ratio of part A and part B coated on a conductive textile (for electrode) for consolidation. Then the positive wrinkled nitrile is attached on the conductive textile contacts with the negative Ecoflex layer to assemble the triboelectric sensor. Finally, the conductive electrodes are encapsulated by non-conductive textiles to shield the ambient electrostatic effect. Different sensors with different areas are fabrication for customized glove configuration.
The glove itself may comprise the abovementioned two-layered triboelectric sensors sewed into a small cotton textile pocket. The finalized encapsulated triboelectric sensors for each hand position attached to the glove by E7000 textile glue for a seamless fit with fingers.
As the user makes gestures with the sensory glove system 200, input signals are generated by the sensors - an input signal comprising the output signals of all sensors. This can include sensors of parts of the hand that are not involved in the gesture. The input signal is then received (per step 122) and processed according to step 124. Processing can occur in various ways. Discussed with reference to Figures 3 and 4 is a non-segmentation method, whereby the input signal is processed by a machine learning (i.e. Al) model trained using data comprising various known words and sentences. The non-segmentation method effectively compares the input signal with the training data to determine which known sentence of word is most similar, or correlates best, to the input signal. The non-segmentation non-segmentation Al framework described below achieves high accuracy for 50 words (91.3%) and 20 sentences (95%) by independently identifying word and sentence signals. Naturally, this can be extended to cover a greater number of words and sentences. To overcome the issue of unable to recognizing new sentences, the developed segmentation Al frame splits the entire sentence signal into word units (fragments) and then recognizes all the signal fragments. Inversely, the sentence can be reconstructed and recognized with an accuracy of 85.58% upon established correlation between word units and sentences. In some embodiments, this requires the fragments to represent known words. Thus, the segmentation approach may comprise dividing the input signal into fragments, applying a machine learning or Al model to each fragment to recognise one or more words in the fragment, and reconstructing the sentence from the words reflected in the fragments. The segmentation approach can be employed to recognise new sentences (average correct rate: 86.67%) that are created by word elements or fragments combined in a new order. The new sentence can be stored and, if repeatedly encountered, used as training data for the machine learning model employed in the non-segmentation method.
To train the sensory glove system 102, system 100 or terminal 104, to perform the non-segmentation approach and segmentation approach, the sensory glove system or terminal to which it is connected may comprise a signal acquisition module (e.g. as part of processor 106) that acquires the generated triboelectric signals generated by different gestures - e.g. using Arduino MEGA 2560 microcontroller. For the non-segmentation method, during training 100 samples for each word and sentence are collected where 80 samples (80%) for training and 20 samples (20%) for testing. Thus, there are 5000 samples for total 50 words and 2000 samples for total 20 sentences. For training the segmentation approach, since all the data form sentences, there are 50 samples with number series labels for each sentence (17 sentences and 850 samples in total). The division window in the Matlab workspace with the length of 200 data points slides at a step of 50 data points to extract the signal fragments - in some embodiments each data point may be zero or non-zero, and thus constitute background signal or potentially part of a gesture or word. Thus, each sentence sample with 800 data points is segmented into 13 elements. The extracted data elements are labelled with numbers of corresponding words or empty. Each number represents a signal fragment that may be an intact gesture signal, background noise, or an incomplete gesture signal (e.g. part of a word in sign language). 60% of samples for each number (i.e., 0-19) are used for training, 20% for validation, and 20% for testing. Finally, 5 samples for 3 new sentences are employed to verify the capability of new sentence recognition of the segmentation-assisted CNN model without prior training.
Data analysis of signals of 50 words and 20 sentences
The tentative data analysis is beneficial for the preliminary understanding of the raw data. Firstly, gestures were selected based on their daily use in a signer's life. These gestures are reflected in Figure 3, image (a). Using the sensory glove system 200, corresponding triboelectric signals of these gestures are given in Figure 3, image (b). These triboelectric signals are the input signals for each respective word.
The system 100 then applies one or more machine learning models to the input signal, to recognise a sentence in the input signal. In the present example, the sentence comprises a single word in each case. The system 100 recognises the words by applying the machine learning model to perform signal similarity and correlation analysis against the original data. In the example, there are 100 data samples for each gesture in the dataset, where 'Get' is shown in the enlarged view in the below middle of Figure 3, image (b). For correlation analysis, the below left of Figure 3, image (b) shows a new input 300 of 'Get' compared with its own database 302, from which a mean correlation coefficient of 0.53 is calculated using any appropriate means - e.g. least squares error between data points (x-axis) of the new input signal over the average values of the database signal. 0.53 is above a predetermined threshold (presently 0.3) that is the nominal cut-off for a high correlation. The same can apply to processing multi-word sentences. To produce the database 'Get' or database word in each case, the training samples may be averaged to enable correlation with the new input signal. Also, the signal of 'Must' is compared with the 'Get' database. The mean correlation coefficient of 0.37, between the input signal 300 and the database signal for 'Must' 304, is also larger than 0.3. This shows confusion by the system, since both 'Get' and 'Must' indicate high signal similarity (the bottom right of Figure 3, image (c)). Therefore, the traditional basic recognition algorithm (e.g., correlation analysis mentioned here) may not be suitable for a discrimination between individual gestures where there is a large number of gestures between which is must discriminate. This also means the recognising individual gestures using traditional methods can result in nonsensical sentences - e.g. gestures representing "Get the doctor" may be recognised as "Must the doctor". To solve this issue, present methods employ advanced data analytics such as machine learning to get a better recognition accuracy, and perform full sentence recognition against a database of known sentences.
Figure 3, image (c), is a matrix representation of the correlation coefficient between input signals (i.e. gestures corresponding to individual words) of known words and database signals for all other known words. The same information is reflected in Figure 3, image (d), being a distribution curve of correlation coefficients from which it can be observed that signals of several gestures have a strong correlation. This means there is a high similarity among these gesture signals. These gestures are susceptible to incorrect classification. The same is shown for sentences in daily use, in Figure 3, images (e) and (f).
Word and sentence recognition upon non-segmentation method
For sequence modelling of signals, hidden Markov model (HMM), recurrent neural network (RNN), and more recently developed long short-term memory (LSTM) have been widely utilized owing to enabling memorizing the output of the last moment for circulated self-updating and adapting. Convolutional Neural Networks (CNN) have also been designed to process data in the form of multiple arrays. In particular, a 1-dimensional (ID) CNN is a simple and feasible solution for recognizing the time-series signals of human motions from triboelectric sensors. Because the positions of features in the segment are not highly correlated. To optimize the CNN model toward more effective recognition behaviours, adjustment of kernel size, the number of filters, and convolutional layers is implemented. To optimise the CNN for use with the present sensory glove system in detecting and interpreting gestures signals, box plots were used to assess kernel size (Figure 4, image (a)), filter number (Figure 4, image (b)), and number of convolutional layers (Figure 4, image (c)). The box plots indicate median (middle line), 25th, 75th percentile (box) and 5th and 95th percentile (whiskers) as well as outliers (single points). The machine learning model used to assess similarity or correlation between input signal and known words or sentences may comprise a ID CNN model with 5 kernels, 64 filters, and 4 convolutional layers. The detailed CNN structure parameters are presented in Table 1.
No Layer Type No. of Kernel/ Stride Input Size Output Size Padding
Filters Pool Size
Figure imgf000015_0001
Table 1 : CNN parameters. The detailed parameters for constructing Convolutional Neural Network (CNN).
The CNN is schematically represented in Figure 4, image (d), in which the 15- channel signals are served as inputs without segmentation. In other words, the sentence signals are not pieced into word elements and not linked with the basic word units within a non-segmented deep learning framework. The signals of words and sentences are isolated from each other and independently recognized.
To better understand the clustering performance of the proposed CNN structure, the results of last fully-connected layer is extracted in which each data sample with 3000 points finally has been stretched to 3584 after convolution and pooling to compare with the raw data. The present machine learning model may use any dimension reduction method, such as Principal Component Analysis (PCA), to reduce the dimension of the output data while maintaining information. PCA can be performed using eigenvalues and eigenvectors of an analogue signal matrix of gestures, which can point along the major variation directions of data. The dimension of data could be reduced with adjustable principal component matrix which comprises the non-unique eigenvectors of the data matrix. Eventually, the data of input layer and the last fully-connected layer is reduced to 30 dimensionalities after PCA. The detailed mathematics behind PCA
PCA is used to reduce the dimensions whilst retaining maximum information. The principle of PCA relies on the correlation between each dimension and provides a minimum number of variables which keeps the maximum amount of variation about the distribution of original data. PCA employs the eigenvalues and eigenvectors of the data-matrix which can point along the major variation directions of data to achieve the purpose of dimension reduction. For the detailed mathematics of PCA, the vector X includes the component that is the input signal Xi for each gesture,
X = {x1,x2,x3, ...,xn} (1)
Then the mean value Xmean of X is calculated as,
Xmean = ^ SiLl xi (2)
Determining the difference di between the input and mean value, dj — Xj — Xmean (3)
Based on Equation (3), we get the covariance matrix S,
S = iS?=1dldl T (4)
According to linear algebra, the eigenvalues A and eigenvectors p of the covariance matrix are defined as,
Sp = Ap (5)
Because there are more than one the eigenvalues and eigenvectors, the principal component matrix P is,
P = {Pl> P2> P3> ■■■ Pk) (6) Thus, the input signal can be projected into a new output matrix Y with reduced dimensions,
Y = {yi,y2,y3— yp -,yn} (7)
Where y; = PT(Xj - xmean) and k controls the principal component matrix P and hence controls the dimension of output Y.
Figure 4, image (e), shows feature clustering result of 50 words for the input layer. Feature clustering can be achieve using any known method. Presently, it was achieved by the t-distributed Stochastic Neighbor Embedding (t-SNE) in the dimension of principal component 1 (PCI) and principal component 2 (PC2). Figure 4, image (e) shows the performance of feature clustering is poor with category overlaps before going through the CNN network. After undergoing the feature extraction and classification of CNN, the visualization result realized by t-SNE in Figure 4, image (f) indicates a desirable classification performance of the developed CNN model with clear boundaries between these 50 classes with less overlap. The same can be said of sentence recognition, as reflected in Figure 4, images (h) and (i) respectively.
Further in this regard, as shown in Figure 5 after 50 training epochs, the accuracy reaches almost 90%. As a consequence, upon 80 training samples (80%) and 20 test samples (20%) of each word in the dataset, a high recognition accuracy of these 50 words is achieved with the value of 91.3%, as shown in the confusion matrix in Figure 4, image (g) for words and image (j) for sentences. The long-length signal of sentence provides more distinguishable features than the single word signal. Therefore, a high sentence recognition accuracy of 95% is obtained.
Though with high accuracies in discriminating words and sentences, the CNN model upon non-segmentation approach only enables classifying known sentence signals in the dataset. It cannot identify new (i.e. not before seen) sentences even where a sentence consists of known words, if those known words are in an order not previously seen by the system 100. This limitation results from labelling each word or sentence with a single label as a distinct item - i.e. the CNN does not distinguish between word and sentence for analysis purposes. In addition, there is time latency in processing, particularly for large sentences, because the CNN model employing non-segmentation methods needs to leverage the whole long-data-length signal (generally 200 data points for the word signal and 800 data points for the sentence signal) to trigger the recognition procedure. This makes it undesirable for the real-time sign language translation.
Word and sentence recognition upon segmentation method
To explore the feasibility of recognizing new/never-seen sentences, a segmentation method is described herein. The segmentation method involves dividing the input signal into fragments, analysing the fragments to identify words, and inversely reconstructing the sentence based on the identified words. The analysis step may be twofold - firstly, distinguishing between fragments representing background noise and those representing words (i.e. filtering) and, secondly, performing recognition on fragments that represent words.
Sentences may be divided into fragments using a data sliding window. Each window may comprise an intact word signal, incomplete word signal or background signal (i.e. noise). These fragments split from all sentence signals are recognized first, and then the CNN model will inversely reconstruct and recognize the sentences. In this way, the segmentation approach can identify known sentences but also identify or interpret new sentences from known words. Since the input signal can be divided as soon as data comprising the 'next' window is received, the input signal can be processed progressively. Consequently, the recognition latency is significantly reduced due to smalldistance sliding window length (50 data points per sliding) as the recognition process will be triggered with each sliding. Any new sentences can be used to expand the sentence database. Therefore, the machine learning models only need to be trained on words, with the segmentation method enabling individual words to be detected and sentences constructed in the database - e.g. memory 132 of Figure 1.
Figure 6, image (a) shows 19 words (W01-W19) labelled from 0 to 18 based on usage frequency. Background noise is considered an 'empty' word (W20) labelled 19. The label action is a critical step for the supervised learning of the CNN model. Figure 6, image (b) demonstrates the detailed process of sentence signal division and labelling. The system 100 takes the sentence 'The dog scared me'. It analyses the voltage on each channel (i.e. the output of each sensor, as reflected in the input signal) as detected at many points in time, the x-axis reflected 800 data points for the sentence. Using a 200 data point sliding window and a sliding step 50 data points, the sentence becomes 13 fragments. Each fragment is an intact single word, background signal, and mixture signal (i.e., incomplete word signals and background signals). When the sliding window contains the intact word signal, then this fragment signal will be labelled with the number of the corresponding word as shown in Figure 6, image (a). The background noise is labelled as 'empty' (W20) with number 19. For mixed signals, the label is assigned based on the principal component. If word signal or background noise accounts for more than 50% of the sliding window size, the label will be the word number of the word in the fragment, or 'empty' 19, respectively. 'The dog scared me' is represented by the label number series '[5 5 19 19 19 15 15 15 19 19 0 0 0]'. The table in Figure 6, image (c) summarizes the essential information of 20 sentences: sentence category codes (Y01-Y17 and Newl-New3), and word/gesture components, an example label series, and unique number orders derived from the label series minus background signals. Newl, New2 and New3 are unknown sentences used to verify the potential of recognizing new sentences. There is no corresponding label series for these three sentences.
There are two ways in which the segmentation method was employed. The first involved detecting words from fragments in a single pass. The second involved a hierarchical filter for separating background signals from word fragments, and then performing recognition on the word fragments. The single classifier was built to identify all the fragments from 17 sentences without preliminarily filtering empty signals (Figure 6, image (d)). The dataset contains 50 samples for each sentence and hence 850 (50*17) sentence samples in total. The word fragments are from 510 samples for training (60%), 170 samples for validation (20%), and 170 samples for testing (20%). The confusion matrix in Figure 6, image (e) shows an accuracy of 81.9% for the single classifier recognizing word units W01-W20. In the segmentation method, the sentence is inversely reconstructed - i.e. constructed from recognised word fragments to form a text or audio output. As indicated in Figure 6, image (f) the result mapping of sentence recognition illustrates an accuracy of 79.41% in sentence reconstruction for known sentences. This is too poor an accuracy, particularly for Y01 and Y04. This results from the system endeavouring to classify empty words (i.e. W20). A hierarchical classifier is shown in Figure 6, image (g). The hierarchical classifier 600 has a first-level classifier 602 to separate background signals from intact word signals (i.e. fragments comprising an intact word or at least 50% data points relate to a word), and a second-level classifier 604 for precise identification of word signals. The recognition accuracy of the hierarchical classifier is 82.81% for word fragments - the related confusion map is shown in Figure 6, image (h). Sentence recognition was also higher, at an average accuracy of 85.58% as shown in Figure 6, image (i).
The method employed by system 100 may therefore comprise dividing the input signal into fragments that include intact word signals, incomplete word signals and background signals. The machine learning is then applied to the fragments to determine, based on a predetermined threshold number of data points associated with the fragment, if the fragment corresponds to background noise (e.g. 50% of data points relate to word or background noise), discard each fragment if the fragment is determined to correspond to background noise, perform recognition on each fragment that has not been discarded.
Recognizing new sentences upon segmentation approach
Taking the advantages of segmentation, the never-seen sentences Newl-New3 are successfully recognized where these sentences are created by new-order word recombination. The process in Figure 7 illustrates recognition of a new sentence by segmentation, real-time sequential fragment identification, and sentence recognition. The CNN classifier is used, with 5 samples for each new sentence engaged validating the CNN classifier performance in recognizing new sentences. As provided in Table 2, both classifiers render the number series prediction for the total 15 inputs of these 3 new sentences although the CNN model has not had the opportunity to be trained on the true label series. The hierarchy classifier shows a reliable performance for new sentence recognition, indicated by fewer incorrect predictions - bold and underlined figures. Figure 8 shows the results after translation or inverse reconstruction of the new sentence, with Figure 8, image (a) being the accuracy for the single classifier - 60% - and Figure 8, image (b) being the accuracy for the hierarchical classifier
- 86.67%.
Figure imgf000022_0001
Figure 9 illustrates the pros and cons of non-segmentation and segmentation methods. The non-segmentation classifier (i.e. non-segmentation approach) has better recognition accuracy but with higher latency and inability to recognise new sentences. Each word or sentence is recognised individually and treated as a separate class. The segmentation method (i.e. segmentation-assisted CNN) model can recognise new sentences and can perform real-time or near real-time recognition (i.e. translation) of sentences. The method may employ both techniques in sequence - e.g. the input signal may be first processed using the non-segmentation approach. If this results in a correlation between the input signal and known words/sentences below a predetermined threshold (e.g. 0.3), then the segmentation method is employed using a hierarchical classifier. New sentences identified by the segmentation classifier (i.e. segmentation method) can be incorporated into the training dataset.
The sensory glove system and the methods described above can overcome the drawbacks of image recognition systems used in sign language recognition. On the one hand, the sensory glove system is not affected by varying luminance and can work well even under entirely dark conditions with higher environmental tolerance. On the other hand, sensory glove systems do not capture facial and other data that is otherwise inadvertently captured by image recognition systems.
VR communication interface for the signer and non-signer
As mentioned with reference to Figure 1, the system 100 may implement a virtual reality conversation system. The signer and non-signer terminals in VR. space allow display sign recognition results and the direct typing of non-signers. The VR interface facilitates two-way remote communication, linked with the Al front end for sign language recognition. The VR conversation system may implement a broader VR environment in which more than two parties may interact, regardless of the number of signers and non-signers among them. The deep learning integrated VR interface realizes bidirectional communication.
To meet the requirement of high recognition accuracy in practical applications, the non-segmented deep learning model is used to achieve communication in the virtual space. In some embodiments, however, it will be appreciated that the segmentation approach or sequence of non-segmentation and segmentation approaches could be used to improve flexibility of communication in the VR space. The system 100, used for recognition and communication, therefore comprises five major blocks including triboelectric gloves (sensory glove system 102) for hand motion capture, the printed circuit board (PCB) for signal preprocessing (this may be in the processor 104 or 118), loT module Arduino connected PC for data collection (this may be part of the signer terminal 104), deep-learning-based analytics for signal recognition (stored in central server 110, though may be elsewhere in some embodiments), and VR. interface in Unity for interaction. Terminals 104 and 108 and nodes 128 and 130 may each be a VR interface, and may comprise VR headset, speaker for audio output of recognised input signals from terminals operated by signers and of text input from terminals operated by non-signers, and a display.
The sensory glove system (1000) and thus the method performed thereby (1002, though steps of the method are performed by other components of a system such as system 100), will output a signal corresponding to the recognised words or sentences. In the VR context, this recognised word or sentence will be projected into cyberspace, in which Al will send corresponding commands based on recognition results and deal with inputs from the nonsigner, to control the communication in VR interface based on Transmission Control Protocol/Internet Protocol (TCP/IP).
The deep learning block recognizes the non-mainstream sign language, translates it into the prevalent conversational medium such as text and audio, and transmits such information into the next virtual space. As shown in the sequence in Figure 10, image (b), sequence (i) to (v) the VR interface is designed for communication between the speech impaired and those without speech impairment, with switched host-guest views in cyberspace. The client and server are built in the VR interface based on TCP/IP and are accessible for the signer and non-signer, respectively. Owing to the assistance of deep learning enabled sentence recognition and translation, the client of VR interface (akin to terminal 104) allows the speech impaired to use the sign language that they are familiar with to engage in the communication. More precisely, the sign language delivered by the speech impaired is recognized and translated into speech and text by deep learning. Then the speech and text are captured and sent to the non-signer-controlled server (akin to terminal 108, or server 110). Next, the healthy user employs the server to type directly responding to the speech disordered user.
The sequence (i) to (v) in Figure 10, image (b) demonstrates, a greeting scenario is created to demonstrate the feasible communication between the speech/hearing impaired and healthy user under the identical local area network (LAN), making interaction more realistic with Al integrated VR interface. Two virtual characters in the VR. space, Lily and Mary, represent the created avatars for the signer and non-signer, with controllable and programmable multiple degrees-of-freedom motions. In the first step Figure 10, image (b), sequence number (i), the speech-impaired user Lily performs sign language 'How are you?' which is recognized and converted to text and audio by the deep learning model. By means of TCP/IP, the client connected with deep learning component receives the text/speech message 'How are you?' and transmit it to the nonsigner Mary controlled server. Projecting to the virtual space, the signer avatar Lily slightly lifts her hand to greet her friend (i.e., the non-signer Mary). Correspondingly, the non-signer types 'Not good. I have a stomach ache.' to respond to the signer. The virtual girl Mary represents the non-signer shakes her head and covers her stomach with hands to show her illness, as shown in Figure 10, image (b), sequence number (ii). Then the speech-impaired user replies to the non-signer Mary with the sign language 'You need a doctor' Figure 10, image (b), sequence number (iii). Figure 10, image (b), sequence numbers (iv) and (v), similarly show exchange of greetings between the speech impaired (Lily) and the non-signer (Mary). The conversation of Figure 10, image (b), is summarized as shown in Figure 10, image (c). The VR communication interface linked with Al allows interaction of the speech/hearing-disordered and healthy people closely and even remotely, providing a promising platform for the immense interactions between two population groups.
Demonstration of pure words and sentences recognition While various systems can be employed for word and sentence recognition, Figure 10, image (a) shows the triboelectric sensory glove system 1000 with 15 sensor channels connected to an Arduino MEGA 2560 (1004). By serial communication with Python (1006 - implementing CNN) in PC, the raw data of different gestures is acquired and recognized by the trained CNN model with Tensorflow and Keras frames, then displaying recognition results on the Python interface.
VR communication interface building and demonstration
The words or sentences recognized by the Python script (1006) are sent into virtual Unity space via TCP/IP communication and displayed on the VR. interface. The server (for non-signers) and client (for signers) terminal building in Unity also rely on TCP/IP. Within the identical LAN, the speech/hearing impaired and healthy people can remotely communicate via the VR interface where the speech/hearing disordered uses their familiar sign language, and the non-signer types directly.
Described herein are a sign language recognition and communication system comprising a smart triboelectric sensory glove system, Al block (implemented by a processor on one or more of the terminal and nodes shown in Figure 1), and the back-end VR interface (server 110 of Figure 1). The system enables the separate and independent recognition of words (i.e., single gestures) and sentences (i.e., continuous multiple gestures) with high accuracies of 91.3% and 95% within non-segmentation frame. Furthermore, to overcome the limitation of incapability of recognizing new sentences, the segmentation method is proposed. It divides all the sentence signals into word fragments while Al learns and memorizes all split elements. Then the deep learning architectures, single and hierarchy classifiers, inversely infer, reconstruct and recognize the whole sentence (recognition accuracy: 85.58%) benefitting from the established correlation of basic word units and sentences. Finally, the embedded VR interface can act as the bridge of user terminals, transmitting messages back and forth, configuring a closed-loop communication system with gloves and Al component. On the VR platform, the speech/hearing impaired can directly perform sign language to interact with the non-signer (i.e., human-to- human interaction), while the non-signer directly involves in the communication process via direct typing.
It will be appreciated that many further modifications and permutations of various aspects of the described embodiments are possible. Accordingly, the described aspects are intended to embrace all such alterations, modifications, and variations that fall within the spirit and scope of the appended claims or statements.
Throughout this specification and the claims or statements that follow, unless the context requires otherwise, the word "comprise", and variations such as "comprises" and "comprising", will be understood to imply the inclusion of a stated integer or step or group of integers or steps but not the exclusion of any other integer or step or group of integers or steps.
The reference in this specification to any prior publication (or information derived from it), or to any matter which is known, is not, and should not be taken as an acknowledgment or admission or any form of suggestion that that prior publication (or information derived from it) or known matter forms part of the common general knowledge in the field of endeavour to which this specification relates.

Claims

- 26 -
Claims
1. Sensory glove system for recognition of sign gestures, comprising: a first glove for wearing on a first hand of a user, the first glove comprising: a first wrist motion sensor for wrist motion perception of a first wrist; a finger bending sensor for each finger of the first glove; and a palm sensor; and a second glove for wearing on a second hand of the user, the second glove comprising: a second wrist motion sensor for second motion perception of a second wrist; a finger bending sensor for each finger of the second glove; and two fingertip sensors, one on each of a middle and index finger.
2. The sensory glove system claim 1, further comprising a processor, the processor being configured to: receive an input signal comprising an output from each wrist motion sensor, bending sensor, palm sensor and fingertip sensor; apply one or more machine learning models to the input signal to recognise a sentence in the input signal.
3. The sensory glove system of claim 2, wherein at least one said machine learning model is configured to: recognise the sentence based on a similarity between the sentence and one or more known sentences.
4. The sensory glove system of claim 2, wherein the processor is further configured to: divide the input signal into fragments and apply the machine learning model to the input signal by applying at least one said machine learning model to the fragments to recognise one or more words in the fragments; and inversely reconstruct the sentence from the one or more words. The sensory glove system of claim 2, wherein the processor is further configured to: calculate a similarity score between the sentence recognised from the input signal and one or more known sentences; and if the similarity score is at or above a predetermined threshold, output the sentence recognised from the input signal; or if the similarity score is below a predetermined threshold: divide the input signal into fragments and apply at least one said machine learning model to the fragments to recognise one or more words in the fragments; and inversely reconstruct the sentence in the input signal from the one or more words. The sensory glove system of any one of claims 1 to 5, wherein each sensor is a triboelectric sensor. The sensory glove system of claim 6, wherein triboelectrification layers of the triboelectric sensors comprise an additional cured silicon layer. The sensory glove system of claim 6 or 7, wherein triboelectrification layers of the triboelectric sensors comprise a wrinkled nitrile layer. The sensory glove system of any one of claims 1 to 8, comprising 15 said sensors in total. The sensory glove system of any one of claims 1 to 9, wherein the finger bending sensors are disposed on a front side of the respective finger. The sensory glove system of any one of claims 1 to 10, wherein the first wrist motion sensor and second wrist motion sensor are position on a back side of the respective first wrist and second wrist. The sensory glove system of claim 11, wherein the processor is configured to divide the input signal into fragments as the input signal is received. A method for sign gesture sentence recognition from an input signal comprising a sentence, comprising: dividing the input signal into fragments; applying a machine learning model to the fragments to recognise individual said fragments; and inversely reconstructing the sentence to recognise the sentence. The method of claim 13, wherein the input signal is received from a sensory glove system according to any one of claims 1 to 12. The method of claim 13 or 14, wherein the fragments include intact word signals, incomplete word signals and background signals, and wherein the applying the machine learning model to the fragments comprises: determining, based on a predetermined threshold number of data points associated with the fragment, if the fragment corresponds to background noise; discarding the fragment if the fragment is determined to correspond to background noise; and recognising each fragment that has not been discarded. The method of claim 15, wherein, for each fragment comprising an intact word, the method further comprises labelling the fragment with a reference to the intact word. The method of claim 15 or 16, wherein, for each fragment comprising background signals, the method further comprises recognising the fragment as comprising background signals on the basis of the fragment comprising fewer than 100 non-zero data points, and tagging the fragment as "empty". - 29 -
18. The method of any one of claims 13 to 17, wherein dividing the input signal into fragments comprises receiving the input signal and dividing the input signal into fragments as the input signal is received.
19. A virtual reality conversation system comprising: a sensory glove system according to claim 1; a signer terminal for use by a first user wearing the sensory glove system; a non-signer terminal for use by a second user; at least one processor configured to: receive an input signal comprising an output from each wrist motion sensor, bending sensor, palm sensor and fingertip sensor; apply one or more machine learning models to the input signal to recognise a sentence in the input signal; and project the signal into a virtual reality environment, for display at the non-signer terminal.
20. The virtual reality conversation system of claim 19, wherein the processor is further configured to: receive an input signal from the non-signer terminal, the input signal comprising one or more words; and project the input signal into the virtual reality environment for display at the signer terminal.
PCT/SG2022/050609 2021-09-02 2022-08-25 Sensory glove system and method for sign gesture sentence recognition WO2023033725A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
SG10202109596U 2021-09-02
SG10202109596U 2021-09-02

Publications (2)

Publication Number Publication Date
WO2023033725A2 true WO2023033725A2 (en) 2023-03-09
WO2023033725A3 WO2023033725A3 (en) 2023-05-11

Family

ID=85413169

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/SG2022/050609 WO2023033725A2 (en) 2021-09-02 2022-08-25 Sensory glove system and method for sign gesture sentence recognition

Country Status (1)

Country Link
WO (1) WO2023033725A2 (en)

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3289304B2 (en) * 1992-03-10 2002-06-04 株式会社日立製作所 Sign language conversion apparatus and method
JP4922504B2 (en) * 2001-06-29 2012-04-25 株式会社アミテック Glove-type input device
US10446059B2 (en) * 2016-12-07 2019-10-15 Thomas William Pryor Hand motion interpretation and communication apparatus
CN108256458B (en) * 2018-01-04 2020-08-04 东北大学 Bidirectional real-time translation system and method for deaf natural sign language
CN109696963A (en) * 2019-01-23 2019-04-30 深圳大学 Sign language interpretation system, glove for sign language translation and sign language interpretation method
CN110189590A (en) * 2019-06-18 2019-08-30 合肥工业大学 A kind of adaptively correcting formula sign language mutual translation system and method
CN211827191U (en) * 2019-11-08 2020-10-30 桂林电子科技大学 Gesture recognition device fixed knot constructs and wearable gesture recognition device thereof

Also Published As

Publication number Publication date
WO2023033725A3 (en) 2023-05-11

Similar Documents

Publication Publication Date Title
Wen et al. AI enabled sign language recognition and VR space bidirectional communication using triboelectric smart glove
Kudrinko et al. Wearable sensor-based sign language recognition: A comprehensive review
Sagayam et al. Hand posture and gesture recognition techniques for virtual reality applications: a survey
Sruthi et al. Signet: A deep learning based indian sign language recognition system
Chen et al. Hand gesture recognition using Haar-like features and a stochastic context-free grammar
Das et al. Smart glove for sign language communications
Yasir et al. Bangla sign language recognition using convolutional neural network
More et al. Sign language recognition using image processing
Cabrera et al. Glove-based gesture recognition system
Lamari et al. Hand alphabet recognition using morphological PCA and neural networks
Angona et al. Automated Bangla sign language translation system for alphabets by means of MobileNet
Kumar et al. A hybrid gesture recognition method for American sign language
Singh et al. A Review For Different Sign Language Recognition Systems
Ji et al. Recognition of Japanese sign language by sensor-based data glove employing machine learning
WO2023033725A2 (en) Sensory glove system and method for sign gesture sentence recognition
Suganya et al. Design Of a Communication aid for physically challenged
Zheng et al. Review of lip-reading recognition
Robert et al. A review on computational methods based automated sign language recognition system for hearing and speech impaired community
Avadut et al. A deep learning based iot framework for assistive healthcare using gesture based interface
Sandberg Gesture recognition using neural networks
Dhamanskar et al. Human computer interaction using hand gestures and voice
Pradeep et al. Advancement of sign language recognition through technology using python and OpenCV
Al-Mohimeed et al. Dynamic Sign Language Recognition Based on Real-Time Videos.
Caliwag et al. Korean sign language translation using machine learning
Maurya et al. Analysis on hand gesture recognition using artificial neural network

Legal Events

Date Code Title Description
NENP Non-entry into the national phase

Ref country code: DE