NL2020989B1 - Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots. - Google Patents

Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots. Download PDF

Info

Publication number
NL2020989B1
NL2020989B1 NL2020989A NL2020989A NL2020989B1 NL 2020989 B1 NL2020989 B1 NL 2020989B1 NL 2020989 A NL2020989 A NL 2020989A NL 2020989 A NL2020989 A NL 2020989A NL 2020989 B1 NL2020989 B1 NL 2020989B1
Authority
NL
Netherlands
Prior art keywords
language
body language
message
living
computer program
Prior art date
Application number
NL2020989A
Other languages
Dutch (nl)
Inventor
Meinardus Gerardus Stokman Henricus
Jean Baptist Van Oldenborgh Marc
Alnajar Fares
Original Assignee
Kepler Vision Tech Bv
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Kepler Vision Tech Bv filed Critical Kepler Vision Tech Bv
Priority to NL2020989A priority Critical patent/NL2020989B1/en
Priority to US16/956,485 priority patent/US11443557B2/en
Priority to PCT/NL2019/050301 priority patent/WO2019226051A1/en
Priority to EP19743015.0A priority patent/EP3596656B1/en
Application granted granted Critical
Publication of NL2020989B1 publication Critical patent/NL2020989B1/en
Priority to US17/931,268 priority patent/US11908245B2/en
Priority to US18/444,418 priority patent/US20240193994A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

There is provided a body language system for determining a body language message of a living being, said system comprising an artificial intelligence (AI) system, said AI system running a computer program that:5 - retrieves at least one image of said living being showing body language; - labels said living being in said at least one image, resulting in a labeled living being; - determines a baseline body language of said labeled living being; - adapts said AI system using said baseline body language;10 - applies said adapted AI system to at least one of said at least one image for categorizing said body language resulting in a category, and applying said category for determining said body language message.

Description

Field of the invention
The invention relates to a body language system, a sign language transformation system, a command gesture language transformation system, a spoken-language transformation system, and a computer program product for controlling an Al system.
Background of the invention
Artificial intelligence (Al) is developing rapidly and Al applications are supporting or will support all industries including the aerospace industry, agriculture, chemical industry, computer industry, construction industry, defense industry, education industry, energy industry, entertainment industry, financial services industry, food industry, health care industry, hospitality industry, information industry, manufacturing, mass media, mining, telecommunication industry, transport industry, water industry and direct selling industry.
Human-machine communication becomes more and more important. Machines (such as computers, smartphones, tablets and robots) are penetrating society rapidly. Research reveals that communication between humans for a major part is depending on body language. Therefore understanding body language would not only enhance communications between humans but also between humans and machines. More general, the ability to understanding body language of living beings will improve the communication between living beings, and between living beings and machines. Machine learning can be used to understand and evaluate body language.
In ‘‘Survey on Emotional Body Gesture Recognition”, January 2018, by Fatemeh Noroozi et al. (https://arxiv.org/abs/1801.07481) according to its abstract describes “Automatic emotion recognition has become a trending research topic in the past decade. While works based on facial expressions or speech abound, recognizing affect from body gestures remains a less explored topic. We present a new comprehensive survey hoping to boost research in the field. We first introduce emotional body gestures as a component of what is commonly known as body language and comment general aspects as gender differences and culture dependence. We then define a complete framework for automatic emotional body gesture recognition. We introduce person detection and comment static and dynamic body pose estimation methods both in RGB and 3D. We then comment the recent literature related to representation learning and emotion recognition from images of emotionally expressive gestures. We also discuss multi-modal approaches that combine speech or face with body gestures for improved emotion recognition. While pre-processing methodologies (e.g. human detection and pose estimation) are nowadays mature technologies fully developed for robust large scale analysis, we show that for emotion recognition the quantity of labelled data is scarce, there is no agreement on clearly defined output spaces and the representations are shallow and largely based on naive geometrical representations.”
In ‘‘Learning Personalized Models for Facial Expression Analysis and Gesture Recognition”, January 2016, by Gloria Zen et al. (https://ieeexplore.ieee.org/abstract/document/7394167) according to its abstract describes “Facial expression and gesture recognition algorithms are key enabling technologies for human-computer interaction (HCT) systems. State of the art approaches for automatic detection of body movements and analyzing emotions from facial features heavily rely on advanced machine learning algorithms. Most of these methods are designed for the average user, but the assumption “one-size-fits-all” ignores diversity in cultural background, gender, ethnicity, and personal behavior, and limits their applicability in real-world scenarios. A possible solution is to build personalized interfaces, which practically implies learning person-specific classifiers and usually collecting a significant amount of labeled samples for each novel user. As data annotation is a tedious and time-consuming process, in this paper we present a framework for personalizing classification models which does not require labeled target data. Personalization is achieved by devising a novel transfer learning approach. Specifically, we propose a regression framework which exploits auxiliary (source) annotated data to learn the relation between person-specific sample distributions and parameters of the corresponding classifiers. Then, when considering a new target user, the classification model is computed by simply feeding the associated (unlabeled) sample distribution into the learned regression function. We evaluate the proposed approach in different applications: pain recognition and action unit detection using visual data and gestures classification using inertial measurements, demonstrating the generality of our method with respect to different input data types and basic classifiers. We also show the advantages of our approach in terms of accuracy and computational time both with respect to user-independent approaches and to previous personalization techniques”
In “Automatic Error Analysis of Human Motor Performance for Interactive Coaching in Virtual Reality”, September 2017, by Felix Hülsmann et al. (https://arxiv.org/abs/1709.09131) according to its abstract describes “In the context of fitness coaching or for rehabilitation purposes, the motor actions of a human participant must be observed and analyzed for errors in order to provide effective feedback. This task is normally carried out by human coaches, and it needs to be solved automatically in technical applications that are to provide automatic coaching (e.g. training environments in VR). However, most coaching systems only provide coarse information on movement quality, such as a scalar value per body part that describes the overall deviation from the correct movement. Further, they are often limited to static body postures or rather simple movements of single body parts. While there are many approaches to distinguish between different types of movements (e.g., between walking and jumping), the detection of more subtle errors in a motor performance is less investigated. We propose a novel approach to classify errors in sports or rehabilitation exercises such that feedback can be delivered in a rapid and detailed manner: Homogeneous sub-sequences of exercises are first temporally aligned via Dynamic Time Warping. Next, we extract a feature vector from the aligned sequences, which serves as a basis for feature selection using Random Forests. The selected features are used as input for Support Vector Machines, which finally classify the movement errors. We compare our algorithm to a well established state-of-the-art approach in time series classification, 1-Nearest Neighbor combined with Dynamic Time Warping, and show our algorithm's superiority regarding classification quality as well as computational cost.”
WO 2017/218595 according to its abstract describes “A method or system is capable of detecting operator behavior (OB) utilizing a virtuous cycle containing sensors, machine learning center (MLC), and cloud based network (CBN). In one aspect, the process monitors operator body language captured by interior sensors and captures surrounding information observed by exterior sensors onboard a vehicle as the vehicle is in motion. After selectively recording the captured data in accordance with an OB model generated by MLC, an abnormal OB (AOB) is detected in accordance with vehicular status signals received by the OB model. Upon rewinding recorded operator body language and the surrounding information lea ding up to detection of AOB, labeled data associated with AOB is generated. The labeled data is subsequently uploaded to CBN for facilitating OB model training at MLC via a virtuous cycle.”
US2018082112 according to its abstract describes “A system and method are presented for body language analysis of a video interaction. In a contact center system, the video interaction between an agent and a customer may be monitored and used to determine automatic actions when threshold are met and/or matches are made. Training videos comprising determined metrics may be used for comparison to realtime interactions. Scoring and/or matches may be utilized to determine a threshold to trigger pre-determined actions based on comparison to the training videos.”
In “Learning a bidirectional mapping between human whole-body motion and natural language using deep recurrent neural networks”, May 2017, by Matthias Plappert et al. (https://arxiv.org/abs/1705.06400) according to its abstract describes “Linking human whole-body motion and natural language is of great interest for the generation of semantic representations of observed human behaviors as well as for the generation of robot behaviors based on natural language input. While there has been a large body of research in this area, most approaches that exist today require a symbolic representation of motions (e.g. in the form of motion primitives), which have to be defined a-priori or require complex segmentation algorithms. In contrast, recent advances in the field of neural networks and especially deep learning have demonstrated that sub-symbolic representations that can be learned end-to-end usually outperform more traditional approaches, for applications such as machine translation. In this paper we propose a generative model that learns a bidirectional mapping between human whole-body motion and natural language using deep recurrent neural networks (RNNs) and sequence-to-sequence learning. Our approach does not require any segmentation or manual feature engineering and learns a distributed representation, which is shared for all motions and descriptions. We evaluate our approach on 2,846 human whole-body motions and 6,187 natural language descriptions thereof from the KIT Motion-Language Dataset. Our results clearly demonstrate the effectiveness of the proposed model: We show that our model generates a wide variety of realistic motions only from descriptions thereof in form of a single sentence. Conversely, our model is also capable of generating correct and detailed natural language descriptions from human motions. “
US8,0982,73 according to its abstract describes “In one embodiment, a method determines an indication of a mood for a caller during a service call. The mood may be determined using a facial analysis of the caller's facial expressions. The mood may indicate an emotion of the user, such as the user is angry, happy, etc. The mood may be determined based on a facial expression analysis of the caller during a portion of the service call. The service call may be a call between the caller and a service center, which may provide customer support to a caller for a product, service, etc. One example of a service center may be video contact service center that enables video calls with a caller. An action is then determined based on analysis of the mood invoked during a portion of the call. Once the action is determined, the action may be performed.”
US2017124407 according to its abstract describes “Systems, methods, and devices for predicting driver intent and future movements of a human driven vehicles are disclosed herein. A system for predicting future movements of a vehicle includes a camera system, a boundary component, a body language component, and a prediction component. The camera system is configured to capture an image of a vehicle. The boundary component is configured to identify a sub-portion of the image corresponding to an area where a driver of a vehicle is located. The body language component configured to detect a driver's body language. The prediction component configured to predict future motion of the vehicle based on the driver's body language detected by the body language component.”
US2015120465 according to its abstract describes “Concepts and technologies are disclosed herein for detecting body language via bone conduction. According to one aspect, a device can detect body language of a user. The device can generate a signal and send the signal to a sensor network connected to a user. The device can receive a modified signal from the sensor network and compare the modified signal to a body language reference model. The device can determine the body language of the user based upon comparing the modified signal to the body language reference model.”
In “Embedded Real-Time Fall Detection Using Deep Learning For Elderly Care”, November 2017, by Hyunwoo Lee et al. (https://arxiv.org/abs/1711.11200) according to its abstract describes “This paper proposes a real-time embedded fall detection system using a DVS(Dynamic Vision Sensor) that has never been used for traditional fall detection, a dataset for fall detection using that, and a DVS-TN(DVSTemporal Network). The first contribution is building a DVS Falls Dataset, which made our network to recognize a much greater variety of falls than the existing datasets that existed before and solved privacy issues using the DVS. Secondly, we introduce the DVS-TN : optimized deep learning network to detect falls using DVS. Finally, we implemented a fall detection system which can run on low-computing H/W with real-time, and tested on DVS Falls Dataset that takes into account various falls situations. Our approach achieved 95.5% on the Fl-score and operates at 31.25 FPS on NVIDIA Jetson TX1 board.”
In “Video In Sentences Out”, August 2014, by Andrei Barbu et al. (https://arxiv.org/abs/1408.6418) according to its abstract describes “We present a system that produces sentential descriptions of video: who did what to whom, and where and how they did it. Action class is rendered as a verb, participant objects as noun phrases, properties of those objects as adjectival modifiers in those noun phrases, spatial relations between those participants as prepositional phrases, and characteristics of the event as prepositional-phrase adjuncts and adverbial modifiers. Extracting the information needed to render these linguistic entities requires an approach to event recognition that recovers object tracks, the track-to-role assignments, and changing body posture.”
US2013077820 according to its abstract describes “A virtual skeleton includes a plurality of joints and provides a machine readable representation of a human subject observed with a sensor such as a depth camera. A gesture detection module is trained via machine learning to identify one or more features of a virtual skeleton and indicate if the feature(s) collectively indicate a particular gesture.”
Summary of the invention
The invention allows an apparatus or device or system to understand the body language of living beings with an Al system, and to improve for instance response of decision-making using such understanding.
To that end, there is provided a body language system for determining a body language message of a living being, said system comprising an artificial intelligence (Al) system, said Al system running a computer program that:
- retrieves at least one image of said living being showing body language;
- labels said living being in said at least one image, resulting in a labeled living being;
- determines a baseline body language of said labeled living being;
- adapts said Al system using said baseline body language;
- applies said adapted Al system to at least one of said at least one image for categorizing said body language resulting in a category, and applying said category for determining said body language message.
There is further provided a sign language transformation system comprises the body language system, said sign language transformation system for transforming sign language into a computer-processable sign language message, said sign language transformation system comprising an Al system running a sign language computer program for retrieving at least one image of said living being showing sign language and applying said Al system for transforming said sign language into said sign language message, said computer program of said body language system determining a body language message from said at least one image for validating said sign language message.
There is further provided a command gesture language transformation system comprises the body language system, said command gesture language transformation system for transforming command gesture language into a command gesture language message, said command gesture language transformation system comprising an Al system running a command gesture language computer program for retrieving at least one image of said living being showing command gesture language and applying said Al system for transforming said command gesture language into said command gesture language message, said computer program of said body language system determining a body language message from said at least one image for validating said command gesture language message.
There is further provided a spoken-language transformation system comprises the body language system, said spoken-language transformation system for transforming spoken-language into a spoken-language message, said spoken-language transformation system comprising an Al system running a spoken-language computer program for retrieving a spoken-language recording of said labeled living being and applying said Al system for transforming said spoken-language recording into said spoken-language message, wherein said at least one image is time-synchronized with said spoken-language recording, and said computer program of said body language system determining a body language message from said at least one image for validating said spoken-language message.
There is further provided a computer program product for controlling an Al system, said computer program product when running on a data processor: retrieves at least one image of said living being showing body language; labels said living being in said at least one image, resulting in a labeled living being;
determines a baseline body language of said labeled living being; adapts an Al system using said baseline body language;
applies said adapted Al system to at least one of said at least one image for categorizing said body language resulting in a category, and applying said category for determining said body language message.
It was found that body language of a living being is best understood when taking into account the normal, i.e., a baseline body language.
Typically, body language is an involuntary and unconscious phenomenon that adds to the process of communication.
Body language comprises of movements and postures through which attitudes and feelings are communicated, such as his intent was clearly expressed in his body language. In an embodiment, body language consists of these movements and postures.
Nonverbal communication can be by means of facial expressions, eye behavior, gestures, posture, and the like, and are often thought to be or supposed to be involuntary.
Body language is a type of nonverbal communication in which physical behavior, as opposed to words, is used to express or convey information. Such behavior includes body posture, gestures, touch, breath, facial expressions, eye expression, mouth expression, the use of space and appearance.
Body language comprises touching and how it is used in communication, also referred to as haptic communication. As such, handshakes, holding hands, back slapping, high fives, brushing up against someone or patting someone, this all has meaning in the context of communication.
Body language also comprises spatial relationships, which is also known as “proxemics”. Introduced by Edward T. Hall in 1966, proxemics is the study of measurable distances between people as they interact with one another.
Body language further comprises breathing. Patterns of breathing and the context wherein breathing is evaluated are indicative for the mood and state of mind of humans and in general living beings. As such, deep breathing can indicate a relaxed mood and shallow, excessive rapid breathing as being in a more anxious, nervous or stressed state of mind.
The baseline body language of a living being is the body language the live being is expected to show under normal circumstances, in everyday life. Everyday life comprises the ways in which living beings typically act, move, touch, breath, look, speak, think, and feel on a daily basis. Everyday life may be described as routine, natural, habitual, or normal.
The body language message is derived from an amount of deviation of bodylanguage of a living being from its baseline body language.
Adapting an Al system results in an Al system that is able to recognize body language that expresses a body language message. Adapting an Al system may comprise calibrating an Al system with a baseline body language.
A. baseline body language can be set on a group level or on an individual level. On the group level, in an embodiment the body language system determines the common baseline body language for a group of living beings sharing a common aspect that is typical for the group.
The message in most case relates to a machine-processable message. Such a machine-processable message can be stored, or transformed in a format that can be presented to a human being. In most cases, the computer program provides the message in a machine-processable format.
In an embodiment, the computer program retrieves said at least one image of said living being, and wherein said body language comprises a body language pose.
In an embodiment, the computer program retrieves a plurality of images of said living being, and wherein said body language comprises a body language gesture, in particular said plurality of images is part of a time series of images.
In an embodiment, the at least one image is part of a time series of images. In an embodiment, a time seri es of images can comprise at least part of a movie.
In an embodiment, the after said adapting, said Al system is applied to further retrieved images of said living being.
In an embodiment, determining said body language message of a living being occurs substantially real-time, in an embodiment functionally real-time.
In an embodiment, the body language message is determined within a probability range.
On individual level, an embodiment of a body language system provided determination of the baseline body language per living being.
In an embodiment, determining said baseline body language comprises determining body language characteristics which are specific for a group of living beings including said identified living being.
In an embodiment, determining said baseline body language comprises retrieving information regarding said labeled living being, said information comprising at least one selected from species, culture, race, religion, appearance, state of mental health, state of physical health, and age group.
In an embodiment, retrieving information for said baseline body language regarding said labeled living being comprises categorizing said body language from said retrieved at least one image.
In an embodiment, a body language system provides recognition of groups of individuals and provides determination of a common baseline body language. Such a body language system can provide an estimation of a social tension within a group or between groups; for instance when two groups of football hooligans meet.
In another example, an embodiment of a body language system provides recognition of when people in a group are discussing, and provide determination of whether these people are likely to agree or disagree with each other. An example of agreement is when people are copying each other’s body language. In such a case, these people often express mutual understanding.
In an embodiment, determining said baseline body language comprises determining body language details that are specific for said labeled living being.
In an embodiment, determining said baseline body language that are specific for said labeled living being comprises said computer program:
retrieving at least one further image of said living being showing body language;
labeling said living being in said at least, one further image, resulting in said labeled living being;
applying said Al system to said at least one further image for categorizing the body language of said labeled living being resulting in a category, and applying said category' for determining said baseline body language.
In an embodiment, the at least one image comprises a plurality of living beings each showing said body language, and said computer program labeling at least pan of said living beings in said at least one image, determining a common baseline body language of said labeled living beings, adapting said Al system using said common baseline body language.
In an embodiment hereof the system further comprises using said common baseline body language for grouping said labeled living beings.
In an embodiment, the computer program further determines a context of said labeled living being.
In an embodiment the context is an environment in which a living being is positioned, in particular said environment is set by at least one selected from a location, an event, an object, another living being, weather conditions, light conditions, road conditions, and a combination thereof.
In an embodiment, the computer program retrieves said context from said at leastone image.
In a further example, an embodiment of a body language system provides determination of a context of a living being. A context is an environment a living being is positioned in. For instance context may include information about the location, an event, an object, another living being, weather conditions, light conditions and road conditions. Answers to questions such as “Is the living being in a city or in the countryside?” (location), “Is the living being close to an accident?” (event), “Is the living close to a car?” (object), ”Is the human near a dangerous animal?” (another living being), “Is the living standing in a storm?” (weather condition), “Is the living being walking in the night?” (light condition), and “Is the living being on a busy road, is the road a highway or is the road a pedestrian area?” (road condition) contribute to determine the context.
In an embodiment of the sign language transformation system, the validating comprises comparing said sign language message with said body language message and outputs a machine-processable result indicating either said body language message emphasizes said sign language, or said body language message de-eniphasizes said sign language, or said body language message differs from said sign language message.
In an embodiment of the sign language transformation system:
said sign language computer program determines said sign language message within a probability range;
said sign language computer program of said body language system determines said body language message within a further probability range, and wherein said validating comprises comparing said probability range with said further probability range.
Body language must not be confused with sign language or command gesture language. Sign languages are like spoken and written languages with grammar systems. Sign language can exhibit the fundamental properties that exist in all spoken and written languages. As command gesture languages are a set of well-defined gestures used to send a non-trivial command or non-trivial message to a machine or living being. Body language, on the other hand, does not have a grammar and must be interpreted broadly, instead of having an absolute meaning corresponding with a certain movement. Nevertheless, body language can emphasize, de-emphasize, or even alter a message of another type of language when used in parallel with the other type of language. For example, expressing an opinion in a spoken language while showing as body language a clenched fist, emphasizes the message of the opinion. In another example, a message such as “I hate him” in a sign language while showing as body language a big smile deemphasizes the message, or even alters the message in “I love him”. Body language is often complimentary to other types of language. Body language helps to put a message in the right context. For instance, understanding body language can differentiate a gesture as part of command gesture language from a gesture as part of body language. Such as the raising of an arm which can send a “stop” message as part of command gesture language or a “hello” message as part of body language.
In an embodiment of the command gesture language transformation system said validating comprises comparing said command gesture language message with said body language message and outputs a machine-processable result indicating either said body language message emphasizes said command gesture language, or said body language message de-emphasizes said command gesture language, or said body language message differs from said command gesture language message.
In an embodiment, of the spoken-language transformation system, the validating comprises comparing said spoken-language message with said body language message and outputs a machine-processable result indicating either said body language message emphasizes said spoken-language, or said body language message de-emphasizes said spoken-language, or said body language message differs from said spoken-language message.
In an embodiment of the spoken-language transformation system:
said spoken-language computer program determines said spoken-language message within a probability range;
said spoken-language computer program of said body language system determines said body language message within a further probability range, and wherein said validating comprises comparing said probability range with said further probability range.
In an embodiment of the command gesture language transformation system: said command gesture language computer program determines said command gesture language message within a probability range;
said command gesture language computer program of said body language system determines said body language message within a further probability range, and wherein said validating comprises comparing said probability range with said further probability range.
In order to detect and localize a living being in a scene from a retrieved image an embodiment uses a method to detect living beings. Such a method will use machine learning techniques (mainly deep learning) to design and train a model which detects living beings given an input of the same type (RGB image, infrared, etc.) as the system perceives. The model is trained on a large amount of annotated data; it comprises images with and without living beings and locations of the living beings are annotated. In the case of deep learning, a detection framework such as Faster-RCNN, SSD, R-FCN, Mask-RCNN, or one of their derivatives can be used. A base model structure can be VGG, AlexNet, ResNet, GoogLeNet, adapted from the previous, or a new one. A model can be initialized with weights and trained similar tasks to improve and speedup the training. Optimizing the weights of a model, in case of deep learning, can be done with the help of deep learning frameworks such as Tensorflow, Caffe, or MXNET. To train a model, optimization methods such as Adam or RMSProb can be used. Classification loss functions such Hinge Loss or Softmax Loss can be used. Other approaches which utilize handcrafted features (such as LBP, SIFT, or HOG) and conventional classification methods (such as SVM or Random Forest) can be used.
In an embodiment for determining body language and baseline body language features of a human, a deep learning model can used. Such a deep learning model can be designed and trained to analyze the features over time, for instance using a time-series of even (part of) a movie.
For instance expressions such as stress and irritation can be detected by body language based on bodily features as well as facial features.
Hectic movements, as a sign of stress and irritation, are a bodily feature that can be detected by measuring an activity level.
Contracting the muscles on the forehead, as a sign of anger, dissatisfaction, and irritation, is a facial feature that can be detected by determining a contraction level of these muscles.
To detect bodily features, such as an activity level, the system in an embodiment can determine key points on the body (e.g. hands, shoulders, knees, etc.) of a human and measure the displacement of these key points over a (short) period of time. The initial activity level can set a baseline body language.
To detect the key points on the body of a human, in an embodiment the system comprises a model that is designed and trained for this detection. The training data to train the model comprises an annotation of various key points locations. When a new image is presented, the model allow identification of the locations of such key points. To this end, the system can utilize existing key point detection approaches such as MaskRCNN or CMU Part Affinity Fields. The training procedure and data can be customized to best match the context of the content of the retrieved images. Such context may comprise an outdoor context like streets, parks, parks with changing lighting conditions.
To measure movements over time, the system in an embodiment comprises a further model that is designed to take, as an input, a sequence of images with the predicted locations of key points and to output an activity level. In such an embodiment, the system computes sub-features such as temporal and spatial distances between the key points and an anchor (i.e. the head). The distances between all pairs of key points, the angles, the velocities, and the angular velocities are all examples of such subfeatures. Subsequently, in this embodiment the system may comprise yet a further model, which is designed and trained to take the images, the predicted key points and the computed sub-features to predict an activity level of the human. The activity level can be measured by a continuous real number (regression) or assigned to a category.
To detect facial features, such as a contraction level, in an embodiment the system comprises yet a further model that is designed and trained to detect and localize muscles on the face. The initial contraction level of these muscles can set a baseline body language. An initial state can be computed by averaging the muscle contraction levels over several earlier retrieved images.
To detect the facial muscles, the system in an embodiment comprises yet a further model that is designed and trained for this. The training procedure uses data where the samples have annotations for the faces and the different muscles of the face. When a new image is presented, the model will identify the locations of such muscles and measure a contraction level (state of the muscles). The further model allows detection of the facial muscles for the subsequent images and compare with a baseline body language. The contraction level can be represented by various features such as Convolutional Neural Network (CNN) features.
In case of deep learning, a detection framework such as Faster-RCNN, SSD, RFCN, Mask-RCNN, or one of their derivatives can be used. A base model structure can be VGG, AlexNet, ResNet, GoogLeNet, adapted from the previous, or a new one. A model can be initialized with weights and trained with similar tasks to improve and speedup the training. Optimizing the weights of a model, in case of deep learning, can be done with the help of deep learning frameworks such as Tensorflow, Caffe, or MXNET. To train a model, optimization methods such as Adam or RMSProb can be used. Regression loss functions such Euclidean Loss can be used. Other approaches which use handcrafted features (such as LBP, SIFT, or HOG) and conventional classification methods (such as SVM or Random Forest) can be used.
Body language features, such an activity level and contraction level, can be further compared with a baseline body language using a distance function, such as L2, Manhattan, or Cosine distance. A comparison threshold can be set by an expert or learnt from a validation dataset. Based on this, the body language of a human can be further determined; for instance to determine the level of irritation.
Determining body language can also be computed by learning yet a further model comprising body language features and baseline body language as input.
An image or series of images or time series of images can for instance result from a LIDAR, a visual light camera, a sonar imaging, a radar imaging, a laser imaging, or an infrared camera.
The term “substantially” herein, such as in “substantially all emission” or in “substantially consists”, will be understood by the person skilled in the art. The term “substantially” may also include embodiments with “entirely”, “completely”, “all”, etc. Hence, in embodiments the adjective substantially may also be removed. Where applicable, the term “substantially” may also relate to 90% or higher, such as 95% or higher, especially 99% or higher, even more especially 99.5% or higher, including 100%. The term “comprise” includes also embodiments wherein the term “comprises” means “consists of’.
The term functionally will be understood by, and be clear to, a person skilled in the art. The term “substantially” as well as “functionally” may also include embodiments with “entirely”, “completely”, “all”, etc. Hence, in embodiments the adjective functionally may also be removed. When used, for instance in “functionally parallel”, a skilled person will understand that the adjective “functionally” includes the term substantially as explained above. Functionally in particular is to be understood to include a configuration of features that allows these features to function as if the adjective “functionally” was not present. The term “functionally” is intended to cover variations in the feature to which it refers, and which variations are such that in the functional use of the feature, possibly in combination with other features it relates to in the invention, that combination of features is able to operate or function. For instance, if an antenna is functionally coupled or functionally connected to a communication device, received electromagnetic signals that are receives by the antenna can be used by the communication device. The word “functionally’’ as for instance used in “functionally parallel” is used to cover exactly parallel, but also the embodiments that are covered by the word “substantially” explained above. For instance, “functionally parallel” relates to embodiments that in operation function as if the parts are for instance parallel. This covers embodiments for which it is clear to a skilled person that it operates within its intended field of use as if it were parallel.
Furthermore, the terms first, second, third and the like in the description and in the claims, are used for distinguishing between similar elements and not necessarily for describing a sequential or chronological order. It is to be understood that the terms so used are interchangeable under appropriate circumstances and that the embodiments of the invention described herein are capable of operation in other sequences than described or illustrated herein.
The devices or apparatus herein are amongst others described during operation. As will be clear to the person skilled in the art, the invention is not limited to methods of operation or devices in operation.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. Use of the verb to comprise and its conjugations does not exclude the presence of elements or steps other than those stated in a claim. The article a or an preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the device or apparatus claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
The invention further applies to an apparatus or device comprising one or more of the characterizing features described in the description and/or shown in the attached drawings. The invention further pertains to a method or process comprising one or more of the characterizing features described in the description and/or shown in the attached drawings.
The various aspects discussed in this patent can be combined in order to provide additional advantages. Furthermore, some of the features can form the basis for one or more divisional applications.
Brief description of the drawings
Embodiments of the invention will now be described, by way of example only, with reference to the accompanying schematic drawings in which corresponding reference symbols indicate corresponding parts, and in which:
FIG 1 schematically depicts an embodiment of an artificial intelligence ( Al) system determining body language messages of two men with different baseline body language;
FIG 2 schematically depicts an embodiment of an artificial intelligence (Al) system determining body language messages of two men with specific baseline body language;
FIG 3 schematically depicts an embodiment of an artificial intelligence (Al) system determining a body language message of two groups of football supporters having a social interaction with one another;
FIG 4 schematically depicts an embodiment of an artificial intelligence ( Al) system determining a body language message of two men having a social interaction with one another at a bar;
FIG 5 schematically depicts an embodiment of an artificial intelligence (Al) system determining a body language message of a bicyclist;
FIG 6 schematically depicts an embodiment of a sign language transformation system validating a sign language message of a woman;
FIG 7 schematically depicts an embodiment of a command gesture language transformation system validating a command gesture language message of a woman;
FIG 8 schematically depicts an embodiment of a spoken-language transformation system validating a spoken-language message of a speaker;
FIG 9 schematically depicts an embodiment of a video conference system analyzing body language;
FIG 10 schematically depicts an embodiment of a video analyzer, on a cloud server of a social media network, analyzing body language;
FIG 11 schematically depicts an embodiment of a service robot in a supermarket analyzing body language;
FIG 12 schematically depicts an embodiment of a video chat service application, utilized a customer services, analyzing body language;
FIG 13 schematically depicts an embodiment of a self-driving car analyzing body language;
FIG 14 schematically depicts an embodiment of a computer robot playing poker on a video-driven poker platform analyzing body language;
FIG 15 schematically depicts an embodiment of a surveillance system in a military vehicle;
FIG 16 schematically depicts an embodiment of a surveillance system in a elderly care center, and
FIG 17 schematically depicts an embodiment of a surveillance system in a fast food restaurant.
The drawings are not necessarily on scale.
Description of preferred embodiments
FIG 1A-1B schematically depicts an embodiment of an Al system 1 running a computer program 2 comprising a trained machine learning model (3 and 3’), determining body language messages (10 and 10’) of two men (6 and 7) from different cultures with different baseline body language (8 and 9).
In FIG 1A, Al system 1 running computer program 2 retrieves an image 5 of a man 6 which results in a labeled man 6’ showing body language. Computer program 2 determines -for instance by GPS coordinates, by an IP address, by the image setting and/or by a database entry- a baseline body language of an Italian culture 8. The trained machine learning model 3 adapted for the baseline body language of an Italian culture 8 categorizes the body language of labeled man 6’ which results in a computer-processable body language message 10, messaging “a man talking normally”. The moving arms while talking are matching the baseline body language of individuals belonging to a group of Italian culture with baseline body language 8 who are talking normally with moving arms.
In FIG IB, Al system 1 running computer program 2 retrieves an image 5’ of a man 7 which results in a labeled man 7’ showing body language. Computer program 2 determines -for instance by GPS coordinates, by an IP address, by the image setting and/or by a database entry- a baseline body language of a Japanese culture 9. The trained machine learning model 3’ adapted for the baseline body language of a Japanese culture 9 categorizes the body language of labeled man 7’ which results in a computerprocessable body language message 10’, messaging “an agitated man talking”. The moving arms while talking are not matching the baseline body language of individuals belonging to a group of Japanese culture with baseline body language 9 who are talking normally without moving their arms. The moving of the arms is matching body language expressing agitation.
FIG 2A-2B schematically depicts an embodiment of an Al system 1 running a computer program 2 comprising a trained machine learning model (3” and 3”’), determining body language messages (10 and 10’) of two men (6 and 7) with specific baseline body language (6” and 7”).
In FIG 2A, Al system 1 running computer program 2 retrieves an image 5 of a man 6 which results in a labeled man 6’ showing body language. Computer program 2 determines a baseline body language 6” of the man 6. The trained machine learning model 3” adapted for the baseline body language 6” categorizes the body language of labeled man 6’ which results in a computer-processable body language message 10, messaging “a man talking normally”. The moving arms while talking are matching the individual baseline body language 6” of man 6 who is talking normally while moving his arms.
In FIG 2B, Al system 1 running computer program 2 retrieves an image 5’ of a man 7 which results in a labeled man 7’ showing body language. Computer program 2 determines a baseline body language 7” of the man 7. The trained machine learning model 3’” adapted for the baseline body language 7” categorizes the body language of labeled man 7’ which results in a computer-processable body language message 10’, messaging “an agitated man talking”. The moving arms while talking are not matching the individual baseline body language 7” of man 7 who is talking normally without moving his arms. The moving of the arms is matching body language expressing agitation.
FIG 3 schematically depicts an embodiment of an Al system 1 running a computer program 2 comprising a trained machine learning model 13, determining a body language message 20 of two groups (16 and 17) of football supporters (26 and 27) having a social interaction with one another. This may possibly indicate that a fight is about to take place.
Al system 1 running computer program 2 retrieves an image of football supporters 26 gathered in a first group 16 and further football supporters 27 gathered in a second group 17, which results respectively in first labeled football supporters 26’, a first labeled group 16’, second labeled football supporters 27’ and a second labeled group 17’. The football supporters show body language. Computer program 2 determines a first common baseline body language 26” for the football supporters 26, a second common baseline body language 27” for the further football supporters 27 and a baseline body language 18 of a friendly encounter between football supporters. The trained machine learning model 13 adapted for the first common baseline body language 26” and second common baseline body language 27” categorizes the body language (26’ and 27’) of the two groups of labeled football supporters (16’ and 17’) which results in a computer-processable body language message 20, messaging “there is a tension between two social categories possibly indicating a fight is about to happen”. The moving limbs, attributes (such as clubs, sticks, knives) and posture of the football supporters (26 and 27) are not matching the baseline body language 18 of a friendly encounter between football supporters. The moving limbs, attributes (such as clubs, sticks, knives) and posture of the football supporters (26 and 27) are matching body language expressing aggressive behavior and preparing for a fight.
FIG 4 schematically depicts an embodiment of an Al system 1 running a computer program 2 comprising a trained machine learning model 23, determining a body language message 30 of two men (36 and 37) at a bar 38 having a social interaction with one another indicating the two men (36 and 37) having an argument.
Al system 1 running computer program 2 retrieves an image of a first man 36, a second man 37 and a bar 38, which results respectively in a first labeled man 36’, a second labeled man 37’ and a labeled bar 38’. Computer program 2 determines a first common baseline body language 36” for the first man 36, a second common baseline body language 37” for the second man 37 and a baseline body language 39 of men talking in at a bar. The trained machine learning model 23 adapted for the forst common baseline body language 36” and second common baseline body language 37” categorizes the body language of the two labeled men (36’ and 37’) which results in a computer-processable body language message 30, messaging “there is a tension between 2 social categories possibly indicating two men having an argument”. The location at the bar 38 and body language of the two men (36 and 37) are matching a conversation. The leaning forward and closed fist of the first man 36, and the arrogant looking second man 37 are not matching the baseline body language 39 of a calm and relaxed conversation between two men. The forwardly leaning and closed fist of the first man 36, and the arrogant looking second man 37 are matching body language expressing a conversation between two men having an argument.
FIG 5A-B schematically depicts an embodiment of an Al system 1 running a computer program 2 comprising a trained machine learning model (33 and 33’), determining body language messages (40 and 40’) of a bicyclist 46 when bicycling in a context of different environments (48 and 48’).
In FIG 5A, Al system 1 running computer program 2 retrieves an image 45 of a bicyclist 46 which results in a labeled bicyclist 46’ showing or displaying body language. Computer program 2 determines -for instance by GPS coordinates and data from a weather station, and/or by the image setting- the context 48 of a sunny environment without wind. The trained machine learning model 33 adapted for the baseline body language of a bicyclist 46” bicycling in the context 48 of a sunny environment without wind categorizes the body language of labeled bicyclist 46’ which results in a computer-processable body language message 40, messaging “a bicyclist in a hurry”. The forward bending of the bicyclist 46 is not matching the baseline body language 46” of a bicyclist bicycling in the context 48 of a sunny environment without wind. The forward bending of the bicyclist 46 is matching baseline body language expressing a bicyclist hurrying.
In FIG 5B, Al system 1 running computer program 2 retrieves an image 45’ of a bicyclist 46 which results in a labeled bicyclist 46’ showing body language. Computer program 2 determines -for instance from GPS coordinates and from data from a weather station, and/or by the image setting- the context 48’ of a stormy weather environment.
The trained machine learning model 33’ adapted for the baseline body language of a bicyclist 46’” bicycling in the context 48 of a stormy weather environment categorizes the body language of labeled bicyclist 46’ which results in a computer-processable body language message 40’, messaging “a bicyclist bicycling against the wind”. The forward bending of the bicyclist 46 is matching the baseline body language 46’” of a bicyclist bicycling in the context 48’ of a stormy weather environment.
FIG 6 schematically depicts an embodiment of a sign language transformation system 101 running a computer program 102 comprising a trained machine learning model 104, determining a sign language message 110, and a trained machine learning model 103, determining a body language message 120, for validating the sign language message 110 of a woman 106 into a value 150.
Al system 101 running computer program 102 retrieves an image 105 of a woman 106 which results in a labeled woman 106’ showing sign language and body language. The trained machine learning model 104 categorizes the sign language of the labeled woman 106’ by which the computer program 102 transforms the sign language 110 into a computer-processable sign language message 110’ messaging “I am happy”. The trained machine learning model 103 adapted for a baseline body language 108 categorizes the body language of the labeled woman 106’ by which the computer program 102 transforms the body language 120, of the woman 106, in a computerprocessable body language message 120’ messaging “a sad woman” and validates the sign language message 110’ by a value 150 indicating that the sign language 110 of the woman 106 contradicts the body language 120 of the woman 106.
FIG 7 schematically depicts an embodiment of a command gesture language transformation system 201 running a computer program 202 comprising a trained machine learning model 204, determining a command gesture language message 210, and a trained machine learning model 203, determining a body language message 220, for validating the command gesture language message 210 of a woman 206 into a value 150.
Al system 201 running computer program 202 retrieves an image 205 of a woman 206 which results in a labeled woman 206’ showing command gesture language and body language. The trained machine learning model 204 categorizes the command gesture language of the labeled woman 206’ by which the computer program 202 transforms the command gesture language 210 into a computer-processable command gesture language message 210’ messaging a “Stop” command. The trained machine learning model 203 adapted for a baseline body language 208 and context 209 categorizes the body language of the labeled woman 206’ by which the computer program 202 transforms the body language 220, of the woman 206, in a computerprocessable body language message 220’ messaging “hello” and validates the sign language message 210’ by a value 250 indicating that the sign language 210 of the woman 106 in not supported the body language 220 of the woman 206.
FIG 8 schematically depicts an embodiment of a spoken-language transformation system 301 running a computer program 302 comprising a trained machine learning model 304, determining a spoken-language message 310, and a trained machine learning model 303, determining a body language message 320, for validating the spokenlanguage message 310 of a speaker 306 into a value 350.
Al system 301 running computer program 302 retrieves an image 305 of a speaker 306 which results in a labeled speaker 306’ expressing spoken-language and body language. The trained machine learning model 304 categorizes the spoken-language of the labeled speaker 106’ by which the computer program 302 transforms the spokenlanguage 310 into a computer-processable spoken-language message 310’ messaging the words spoken by speaker 106. The trained machine learning model 303 adapted for a baseline body language 308 categorizes the body language of the labeled speaker 306’ by which the computer program 302 transforms the body language 320, of the speaker 306, in a computer-processable body language message 320’ messaging “a self-assured man” and validates the spoken-language message 310’ by a value 350 indicating that the spoken-language 310 of the speaker 306 is supported by the body language 320 of the speaker 306.
FIG 9 schematically depicts an embodiment of a video conference Al system 1 which is installed on a private computer and attached to a webcam. The video conference Al system 1 analyzes body language (505 and 505’) of user 501 and provides coaching feedback through messages 502 and 502’. In the left view, user 501 with body language 505 is participating in a job interview and is instructed by message 502 “sit up straight so you make a more self-certain impression”. In the right view, user 501 changed his posture and shows the body language 505’, and message 502’ provides feedback “you’re doing well now”.
A video conference Al system 1 is useful in many other video conferencing situations, such as during a first date or while giving a presentation. A video conference Al system 1 can also analyze the body language of other participants of the video conference and give feedback about this to user 501. The feedback can be provided in various ways such as by a written message, a spoken message, a sound, a moving body wearable, a punch, a light (flickering, color) and any other visual sign.
Another application of a video conference Al system 1 is to act as a coach for a user preparing a presentation. Feedback could comprise messages such as “don’t wiggle so much, so you radiate more confidence,” or “show a bit more livelihood, you currently sit and act like a robot and make a cold impression.”
FIG 10 schematically depicts an embodiment of a video analyzer Al system 1 which is for instance installed on a cloud server of a social media network, and where a large series of people are broadcasting video streams 510. The video analyzer Al system 1 analyzes the context and body language of the people in the video streams 510, and assigns one or more scores to each video stream 510. Such scores validate a video stream 510 based upon features such as the livelihood of the people, the openness of the people, the attractiveness of the people, how bored the people seem, how defensive or aggressive the people seem, how patient or impatient the people seem, how stupid or smart the people seem, how dull or funny the people seem, how messy or tidy the room is and the styling of the room. The video analyzer Al system 1 for example summarizes multiple scores in a single rating 511 per video. The rating 511 is used, automated or manual by operators, to rank the most interesting video streams 511 first.
In another embodiment a video analyzer Al system 1 analyzes the context and the body language of people in video streams coming from surveillance cameras, and scores the level of potential criminal activity.
In a further embodiment a video analyzer Al system 1 analyzes the body language of students in video streams coming from individual webcams during a online lecture. The scores provide feedback for the lecturer such as “Do I keep my audience interested?” With the help of automated rankings the lecturer can even, when the audience is numerous, give attention to individual students; for instance using a ranking based on body language showing or indicating a lack of understanding the subject.
FIG 11 schematically depicts an embodiment of a video analyzer Al system 1 in a service robot 522 with a video system that is designed to interact with a woman 521 in a supermarket 520. Here, the video analyzer Al system 1 analyzes the body language 525 of the woman 521 to determine if the woman seems irritated or needs help. In the left view, the video analyzer Al system 1 detects, due to her body language 525, that the woman 521 appears to need help. The service robot 522 then approaches the woman 521, and asks if there is anything the service robot 522 can do to help the woman 521 with. Such help can for instance comprise grabbing the product 526 she is reaching for. Alternatively, as depicted in the right view, the service robot 522 senses, due to her body language 525’, that its interaction with the woman 521 causes the woman 521 to get irritated. The service robot 522 then for instance apologizes and moves out of the way of the woman 521.
FIG 12 schematically depicts an embodiment of a video analyzer Al system 1 in a video chat service application and installed on a call center system of a call center 530 servicing a customer services department of an insurance company. The video analyzer Al system 1 provides analysis of the body language 535 of a customer 531 to help an operator 536 better interpret for instance the faithfulness of a customer 531. The body language 535 of the customer 531 shows “nose touching” and may demonstrate that the customer 531 is being untruthful about what he is saying. As a result of an analysis by the video analyzer Al system 1, in an embodiment a message 532 appears on a display 537 of the call center system, warning the operator 536 and instructing the operator 536 to verify and to drill down more into what the customer 531 is describing.
FIG 13 schematically depicts an embodiment of a video analyzer Al system 1 operationally coupled to a camera system of a self-driving car 540. The video analyzer Al system 1 receives images from the camera system of the self-driving car 540, and detects a man 541 and his body language 545 showing a slightly touch of the upper arm of a woman 542 in front of the self-driving car 540. The video analyzer Al system 1 infers from the body language 545 of the man 541 that the man 541 has positive and encouraging feelings for the woman 542, and that the man 541 and woman 542 likely belong together or have some sort of relationship with each other. The self-driving car 540 is instructed to steer around the man 541 and woman 542, and not initiate a steering movement that would require the man 541 and woman 542 to split up.
FIG 14 schematically depicts an embodiment of a video analyzer Al system 1 in a computer robot 551 playing poker on a video-driven poker platform 550 with 2 other players (554 and 556). For each player (554 and 556), the computer robot 551 builds up a baseline of body language exhibited by that player when holding good cards and when holding poor cards. Once a baseline is built up, the computer robot 551 can estimate the quality of the cards during a game by analyzing the body language. In FIG 16. the computer robot 551 estimates the quality of the cards of the player 554 by comparing the body language 555 of the player 554 with the baseline of body language built up by computer robot 551 for the player 554. Likewise, the computer robot 551 estimates the quality of the cards of the player 556 by comparing the body language 557 of the player 556 with the baseline of body language built up by computer robot 551 for the player 556. The computer robot 551 utilizes the estimations of the quality of the cards of the players (554 and 556) to improve its odds to win the poker game.
FIG 15 schematically depicts an embodiment of a video analyzer Al system 1 as part of a surveillance system and operationally coupled to a camera system of a military vehicle 560 driving on a historical square 561. The video analyzer Al system 1 receives images from the camera system of the military vehicle 560, and analyses the body language of each pedestrian 566, which can be as simple as detecting a different way of moving the arms along the body while walking. The video analyzer Al system 1 detects a woman 564 with body language 565 and infers that the woman may be wearing a bomb vest.
FIG 16 schematically depicts an embodiment of a video analyzer Al system 1 integrated in a video camera 4 as part of a surveillance system in a elderly care center 570. The video analyzer Al system 1, comprising a trained machine learning model 573, can determine a baseline body language of the inhabitants 576 while they feel comfortable and at ease. Upon detecting a significant change in the body language of one or more inhabitants 576, compared to a baseline body language, and indicating a need for assistance, the system alerts a care worker. In FIG 18. the inhabitant 574 shows the body language 575 which may indicate this inhabitant needs medical care. While with traditional human 24/7 monitoring system the privacy of inhabitants of an elderly care center is constantly violated, with a video analyzer Al system 1 the privacy of the inhabitants 576 is mostly preserved and is only be violated when there is a need for it.
FIG 17 schematically depicts an embodiment of a video analyzer Al system 1 integrated in a video camera 4 as part of a surveillance system, here in a fast food restaurant 580. The video analyzer Al system 1, comprising a trained machine learning model 583, can determine when there is a line of impatient clients 586 showing body language by (excessive) signals of impatience, and can warn a manager 588 to have a chat with the impatient clients 586 and offer them a drink to ease the inconvenience. In another embodiment the video analyzer Al system 1 can also determine a baseline body language of the waiting clientele when business is quiet as well as a baseline body language when business is busy, and by comparing an actual waiting line with the baseline body language when business is busy, and by comparing the actual waiting line with the baseline body language when business is quiet, the video analyzer Al system 1 can infer if the actual waiting line needs attention.
For instance, video analyzer Al system 1 records the baseline body language for each identified client 586 in the restaurant 580. The baseline body language recording period starts from identifying a client 586 and lasts for a short period (e.g. 30 seconds). Video analyzer Al system 1 monitors the changes of facial and bodily features for each detected and identified client 586. Video analyzer Al system 1 keeps monitoring the clients 586 while they are waiting and detects a level of irritation by comparing facial and bodily features of their body language to their recorded baseline body language.
It will also be clear that the above description and drawings are included to illustrate some embodiments of the invention, and not to limit the scope of protection. Starting from this disclosure, many more embodiments will be evident to a skilled person. These embodiments are within the scope of protection and the essence of this invention and are obvious combinations of prior art techniques and the disclosure of this patent.

Claims (27)

ConclusiesConclusions 1. Een lichaamstaal systeem voor het vaststellen van een lichaamstaal boodschap van een levend wezen, het systeem omvat een kunstmatig intelligentie (AI) systeem, waarbij het AI systeem een computer programma draait dat:1. A body language system for determining a body language message from a living being, the system comprises an artificial intelligence (AI) system, wherein the AI system runs a computer program that: - ten minste een afbeelding van het levend wezen ontvangt en die lichaamstaal laat zien;- receives at least one image of the living entity and shows body language; het levend wezen in de ten minste ene afbeelding labelt, resulterend in een gelabeld levend wezen;tag the living creature in the at least one image, resulting in a labeled living creature; een basis-lichaamstaal van het gelabelde levend wezen bepaalt;determine a basic body language of the labeled living entity; het AI systeem aanpast, gebruik makend van de basis-lichaamstaal;adjust the AI system using the basic body language; het aangepaste AI systeem toepast op ten minste een afbeelding voor het categoriseren van de lichaamstaal resulterend in een categorie, en het toepassen van de categorie voor het bepalen van de lichaamstaal boodschap.apply the modified AI system to at least one image for categorizing the body language resulting in a category, and applying the category to determine the body language message. 2. Het systeem volgens conclusie 1, waarin het computer programma ten minste een afbeelding van het levend wezen ontvangt, en waarin de lichaamstaal een lichaamstaal houding omvat.The system of claim 1, wherein the computer program receives at least one image of the living being, and wherein the body language comprises a body language attitude. 3. Het systeem volgens conclusie 1 of 2, waarin het computerprogramma afbeeldingen van het levend wezen ontvangt, en waarin de lichaamstaal een lichaamstaal gebaar omvat, in het bijzonder zijn de afbeeldingen deel van een tijdsreeks van afbeeldingen.The system according to claim 1 or 2, wherein the computer program receives images from the living entity, and wherein the body language comprises a body language gesture, in particular, the images are part of a time series of images. 4. Het systeem volgens een van de voorgaande conclusies, waarbij ten minste een afbeelding deel is van een tijdsreeks van afbeeldingen.The system according to any of the preceding claims, wherein at least one image is part of a time series of images. 5. Het systeem volgens een van de voorgaande conclusies, waarbij na het aanpassen het AI systeem wordt toegepast op opgehaalde verdere afbeeldingen van het levend wezen.The system according to any of the preceding claims, wherein after the adaptation the AI system is applied to retrieved further images of the living creature. 6. Het systeem volgens een van de voorgaande conclusies, waarbij het vaststellen van de lichaamstaal boodschap van een levend wezen in hoofdzaak onmiddellijk (“real-time”) gebeurt, in een uitvoeringsvorm functioneel onmiddellijk (“realtime”).The system according to any of the preceding claims, wherein the determination of the body language message of a living being occurs substantially immediately ("real-time"), in an embodiment functionally immediately ("real-time"). 7. Het systeem volgens een van de voorgaande conclusies, waarin de lichaamstaal boodschap is vastgesteld binnen een waarschijnlijkheidsbereik.The system of any one of the preceding claims, wherein the body language message is determined within a probability range. 8. Het systeem volgens een van de voorgaande conclusies, waarin het bepalen van de basis-lichaamstaal het bepalen van lichaamstaal karakteristieken die specifiek zijn voor een groep levende wezens inclusief het geïdentificeerde levend wezen omvat.The system of any one of the preceding claims, wherein determining the basic body language comprises determining body language characteristics that are specific to a group of living things including the identified living being. 9. Het systeem volgens conclusie 8, waarin het bepalen van de basis-lichaamstaal het ontvangen van informatie met betrekking tot het gelabelde levend wezen omvat, de informatie omvattende ten minste een gekozen uit soorten, cultuur, ras, religie, uiterlijk, staat van geestelijke gezondheid, staat van fysieke gezondheid en leeftijdsgroep.The system of claim 8, wherein determining the basic body language comprises receiving information relating to the labeled living entity, the information comprising at least one selected from species, culture, race, religion, appearance, state of spiritual health, state of physical health and age group. 10. Het systeem volgens een van de voorgaande conclusies 8 of 9, waarin het ontvangen van informatie voor de basis-lichaamstaal met betrekking tot het gelabelde levend wezen het categoriseren van de lichaamstaal van de verkregen ten minste ene afbeelding omvat.The system according to any of the preceding claims 8 or 9, wherein receiving information for the basic body language with respect to the labeled living entity comprises categorizing the body language of the obtained at least one image. 11. Het systeem volgens een van de voorgaande conclusies, waarin het vaststellen van het basis-lichaamstaal het vaststellen van de lichaamstaal details die specifiek zijn voor het gelabelde levende wezen omvat.The system of any one of the preceding claims, wherein determining the basic body language comprises determining the body language details specific to the labeled living being. 12. Het systeem volgens conclusie 11, waarin het bepalen van de basis-lichaamstaal die specifiek is voor het gelabelde levend wezen omvat dat het computer programma:The system of claim 11, wherein determining the base body language specific to the labeled living being comprises the computer program: - ten minste een verdere afbeelding van het levend wezen dat lichaamstaal laat zien ontvangt;- receive at least one further image of the living entity showing body language; het leven wezen in de ten minste ene verdere afbeelding etiketteert, resulterend in een gelabeld leven wezen;labels the living being in the at least one further image, resulting in a labeled living being; het AI systeem toepast op de ten minste ene verdere afbeelding voor het categoriseren van de lichaamstaal van het levend wezen wat resulteert in een categorie, en het toepassen van de categorie voor het bepalen van de lichaamstaal boodschap.apply the AI system to the at least one further mapping for categorizing the body language of the living entity resulting in a category, and applying the category for determining the body language message. 13. Het systeem volgens een van de voorgaande conclusies, waarin de ten minste ene afbeelding levend wezens omvat die elk lichaamstaal laten zien, en het computer programma ten minste een deel van de levende wezens in de ten minste ene afbeelding etiketteert, een gemeenschappelijke basis-lichaamstaal van de gelabelde levende wezens vaststelt, waardoor het AI systeem aangepast wordt onder gebruikmaking van de gemeenschappelijke basis-lichaamstaal.The system according to any of the preceding claims, wherein the at least one image comprises living creatures showing each body language, and the computer program labels at least a portion of the living creatures in the at least one image, a common basic determines the body language of the labeled living entities, thereby adjusting the AI system using the common basic body language. 14. Het systeem volgens conclusie 13, verder omvattende het gebruik maken van de gemeenschappelijke basis-lichaamstaal voor het groeperen van de gelabelde levende wezens.The system of claim 13, further comprising using the common basic body language for grouping the labeled living things. 15. Het systeem volgens een van de voorgaande conclusies, waarin het computer programma verder een context van het gelabelde levend wezen vaststelt.The system of any one of the preceding claims, wherein the computer program further determines a context of the labeled living entity. 16. Het systeem volgens conclusie 15, waarin de context een omgeving is in welke een levend wezen is gepositioneerd, in het bijzonder is de omgeving ingesteld door ten minste een gekozen uit een locatie, een evenement, een object, een ander levend wezen, weersomstandigheden, lichtomstandigheden, wegomstandigheden, en een combinatie daarvan.The system of claim 15, wherein the context is an environment in which a living creature is positioned, in particular the environment is set by at least one selected from a location, an event, an object, another living creature, weather conditions , light conditions, road conditions, and a combination thereof. 17. Het systeem volgens conclusie 15 of 16, waarin het computer programma de context haalt uit de ten minste een afbeelding.The system of claim 15 or 16, wherein the computer program extracts the context from the at least one image. 18. Een gebarentaal transformatie systeem omvattende het lichaamstaal systeem volgens een van de voorgaande conclusies, het gebarentaal transformatie systeem voor het transformeren van gebarentaal in een computer verwerkbaar gebarentaal boodschap, het lichaamstaal transformatie systeem omvattende een AI systeem dat een gebarentaal computerprogramma draait voor het ontvangen van ten minste een afbeelding van het levend wezen dat gebarentaal toont en het toepassen van het AI systeem voor het transformeren van de gebarentaal in de gebarentaal boodschap, het computer programma van het lichaamstaal systeem voor het vaststellen van een lichaamstaal boodschap van de ten minste ene afbeelding voor het valideren van de lichaamstaal boodschap.A sign language transformation system comprising the body language system according to any of the preceding claims, the sign language transformation system for transforming sign language into a computer processable sign language message, the body language transformation system comprising an AI system running a sign language computer program for receiving at least one image of the living entity showing sign language and applying the AI system to transform the sign language into the sign language message, the computer program of the body language system for determining a body language message of the at least one image for validating the body language message. 19. Het gebarentaal transformatie systeem volgens conclusie 18, waarin het valideren het vergelijken van de gebarentaal boodschap met de lichaamstaal boodschap omvat en een machinaal verwerkbaar resultaat afgeeft, indicerend ofwel dat de lichaamstaal boodschap de gebarentaal benadrukt, ofwel dat de lichaamstaal boodschap de gebarentaal niet benadrukt, of dat de lichaamstaal boodschap verschilt van de gebarentaal boodschap.The sign language transformation system according to claim 18, wherein the validating comprises comparing the sign language message with the body language message and issuing a machine processable result, indicating either that the body language message emphasizes the sign language, or that the body language message does not emphasize the sign language , or that the body language message differs from the sign language message. 20. Het gebarentaal transformatie systeem volgens conclusie 18, waarin:The sign language transformation system of claim 18, wherein: het gebarentaal computerprogramma de gebarentaal boodschap vaststelt binnen een waarschijnlijkheidsgebied;the sign language computer program determines the sign language message within a probability range; het gebarentaal computer programma van het lichaamstaal systeem de lichaamstaal boodschap vaststelt binnen een verder waarschijnlijkheidsgebied, en waarbij het valideren het vergelijken van de waarschijnlijkheidsgebied met het verdere waarschijnlijkheidsgebied omvat.the sign language computer program of the body language system determines the body language message within a further probability region, and wherein validating comprises comparing the probability region with the further probability region. 21. Een opdrachtsgebaartaal transformatie systeem omvattende het lichaamstaal systeem volgens een van de voorgaande conclusies, het opdrachtsgebaartaal transformatie systeem voor het transformeren van opdrachtsgebaartaal naar een opdrachtsgebaartaal boodschap, waarbij het opdrachtsgebaartaal transformatie systeem een AI systeem omvat dat een opdrachtsgebarentaal computer programma draait voor het ontvangen van ten minste een afbeelding van het levende wezen dat opdrachtsgebaartaal laat zien en het toepassen van het AI systeem voor het transformeren van de opdrachtsgebaartaal in de opdrachtsgebaartaal boodschap, het computer programma van het lichaamstaal systeem voor het bepalen van een lichaamstaal boodschap uit de ten minste een afbeelding voor het valideren van de opdrachtsgebaartaal boodschap.A command sign language transformation system comprising the body language system according to any of the preceding claims, the command sign language transformation system for transforming command sign language into a command sign language message, wherein the command sign language transformation system comprises an AI system running a command sign language computer program for receiving at least an image of the living entity showing command sign language and applying the AI system to transform the command sign language into the command sign language message, the computer program of the body language system for determining a body language message from the at least one image for validating the order sign language message. 22. Het opdrachtsgebaartaal transformatie systeem volgens conclusie 21, waarin het valideren het vergelijken van de opdrachtsgebaartaal boodschap met de lichaamstaal boodschap omvat en het uitgeven van een machine verwerkbaar resultaat wijzend op ofwel dat de lichaamstaal boodschap de opdrachtgebaartaal bevestigt, ofwel dat de lichaamstaal boodschap de opdrachtsgebaartaal niet bevestigt, of de lichaamstaal boodschap verschillend is van de opdrachtsgebaartaal boodschap.The command sign language transformation system of claim 21, wherein the validating comprises comparing the command sign language message with the body language message and issuing a machine processable result indicating either that the body language message confirms the command sign language, or that the body language message confirms the command sign language does not confirm whether the body language message is different from the command sign language message. 23. Het opdrachtsgebaartaal transformatie systeem volgens conclusie 21, waarin:The command sign language transformation system according to claim 21, wherein: het opdrachtsgebaartaal computerprogramma de opdrachtsgebarentaai boodschap bepaalt binnen een waarschijnlijkheidsgebied;the command sign language computer program determines the command sign language message within a probability range; het opdrachtsgebaartaal computer programma van het lichaamstaal systeem bepaalt de lichaamstaal boodschap binnen een verder waarschijnlijkheidsgebied, en waarin het valideren het vergelijken van het waarschijnlijkheidsgebied met het verdere waarschijnlijkheidsgebied omvat.the command sign language computer program of the body language system determines the body language message within a further probability region, and wherein validating comprises comparing the probability region with the further probability region. 24. Een gesproken-taal transformatie systeem omvat het lichaamstaal systeem volgens een van de voorgaande conclusies, het gesproken-taal transformatie systeem voor het transformeren van gesproken-taal naar een gesproken-taal boodschap, het gesproken-taal transformatie systeem omvattende een AI systeem dat een gesproken-taal computer programma draait voor het ophalen van een gesprokentaal opname van het gelabelde levende wezen en het toepassen van het AI systeem voor het transformeren van de gesproken-taal opname in de gesproken-taal boodschap, waarbij ten minste een afbeelding tijds-gesynchroniseerd is met de gesproken-taal opname, en het computer programma van het lichaamstaal systeem een lichaamstaal boodschap bepaalt van de ten minste ene afbeelding voor het valideren van de gesproken-taal boodschap.A spoken language transformation system comprises the body language system according to any of the preceding claims, the spoken language transformation system for transforming spoken language into a spoken language message, the spoken language transformation system comprising an AI system which a spoken-language computer program runs for retrieving a spoken-language recording of the labeled living being and applying the AI system to transform the spoken-language recording into the spoken-language message, wherein at least one image is time-synchronized is with the spoken language recording, and the computer program of the body language system determines a body language message from the at least one image to validate the spoken language message. 25. Het gesproken-taal transformatie systeem volgens conclusie 24, waarin het valideren het vergelijken van de gesproken-taal boodschap met de lichaamstaal boodschap omvat en een machine-verwerkbaar uitvoert dat aangeeft ofwel de lichaamstaal boodschap bevestigt de gesproken taal, ofwel de lichaamstaal boodschap bevestigt de gesproken taal boodschap niet, ofwel de lichaamstaal boodschap is anders dan de gesproken-taal boodschap.The spoken language transformation system according to claim 24, wherein the validating comprises comparing the spoken language message with the body language message and executing a machine processable indicating either the body language message confirms the spoken language or the body language message confirms the spoken language message is not, or the body language message is different from the spoken language message. 26. Het gesproken-taal transformatie systeem volgens conclusie 24, waarbij:The spoken language transformation system of claim 24, wherein: het gesproken-taal computerprogramma de gesproken-taal boodschap bepaalt binnen een waarschijnlijkheidsgebied;the spoken language computer program determines the spoken language message within a probability range; het gesproken-taal computer programma van het lichaamstaal systeem bepaalt de lichaamstaal boodschap binnen een verder waarschijnlijkheidsgebied, en waarbij het valideren het vergelijken van het waarschijnlijkheidsgebied met het verdere waarschijnlijkheidsgebieden omvat.the spoken language computer program of the body language system determines the body language message within a further probability region, and wherein validating comprises comparing the probability region with the further probability regions. 27. Een computer programma product voor het controleren van een AI systeem, het computer programma product wanneer werkend op een data processor:27. A computer program product for controlling an AI system, the computer program product when operating on a data processor: ontvangt ten minste een afbeelding van het levend wezen welke lichaamstaal laat zien;receives at least one image of the living entity showing body language; labelt het leven wezen in de ten minste ene afbeelding, resulterend in een gelabeld leven wezen;labels the living creature in the at least one image, resulting in a labeled living creature; - bepaalt een basis-lichaamstaal van het gelabelde levend wezen;- determines a basic body language of the labeled living entity; past een AI systeem aan, gebruik makend van de basis-lichaamstaal;adjusts an AI system, using the basic body language; past het aangepaste AI systeem toe op ten minste een van de ten minste ene afbeelding voor het categoriseren van de lichaamstaal, resulterend in een categorie, en past de categorie toe voor het bepalen van de lichaamstaal boodschap.applies the modified AI system to at least one of the at least one image for categorizing the body language, resulting in a category, and applies the category for determining the body language message. -o-o-o-o-o--o-o-o-o-o- FIG 1A.FIG 1A. FIG 1B.FIG 1B.
NL2020989A 2018-05-25 2018-05-25 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots. NL2020989B1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
NL2020989A NL2020989B1 (en) 2018-05-25 2018-05-25 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots.
US16/956,485 US11443557B2 (en) 2018-05-25 2019-05-24 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots
PCT/NL2019/050301 WO2019226051A1 (en) 2018-05-25 2019-05-24 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots
EP19743015.0A EP3596656B1 (en) 2018-05-25 2019-05-24 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots
US17/931,268 US11908245B2 (en) 2018-05-25 2022-09-12 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots
US18/444,418 US20240193994A1 (en) 2018-05-25 2024-02-16 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
NL2020989A NL2020989B1 (en) 2018-05-25 2018-05-25 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots.

Publications (1)

Publication Number Publication Date
NL2020989B1 true NL2020989B1 (en) 2019-12-04

Family

ID=63684382

Family Applications (1)

Application Number Title Priority Date Filing Date
NL2020989A NL2020989B1 (en) 2018-05-25 2018-05-25 Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots.

Country Status (1)

Country Link
NL (1) NL2020989B1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11134217B1 (en) 2021-01-11 2021-09-28 Surendra Goel System that provides video conferencing with accent modification and multiple video overlaying

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150120465A1 (en) * 2013-10-29 2015-04-30 At&T Intellectual Property I, L.P. Detecting Body Language Via Bone Conduction
US20170124407A1 (en) * 2015-11-04 2017-05-04 Ford Global Technologies, Llc Predicting vehicle movements based on driver body language
US20170319123A1 (en) * 2016-05-06 2017-11-09 The Board Of Trustees Of The Leland Stanford Junior University Systems and Methods for Using Mobile and Wearable Video Capture and Feedback Plat-Forms for Therapy of Mental Disorders

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150120465A1 (en) * 2013-10-29 2015-04-30 At&T Intellectual Property I, L.P. Detecting Body Language Via Bone Conduction
US20170124407A1 (en) * 2015-11-04 2017-05-04 Ford Global Technologies, Llc Predicting vehicle movements based on driver body language
US20170319123A1 (en) * 2016-05-06 2017-11-09 The Board Of Trustees Of The Leland Stanford Junior University Systems and Methods for Using Mobile and Wearable Video Capture and Feedback Plat-Forms for Therapy of Mental Disorders

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
REHRL T ET AL: "Multiple Parallel Vision-Based Recognition in a Real-Time Framework for Human-Robot-Interaction Scenarios", ADVANCES IN COMPUTER-HUMAN INTERACTIONS, 2010. ACHI '10. THIRD INTERNATIONAL CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, 10 February 2010 (2010-02-10), pages 50 - 55, XP031648116, ISBN: 978-1-4244-5693-2 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11134217B1 (en) 2021-01-11 2021-09-28 Surendra Goel System that provides video conferencing with accent modification and multiple video overlaying

Similar Documents

Publication Publication Date Title
US11908245B2 (en) Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots
US11182597B2 (en) Systems and methods for evaluating individual, group, and crowd emotion engagement and attention
Junior et al. First impressions: A survey on vision-based apparent personality trait analysis
Asteriadis et al. Estimation of behavioral user state based on eye gaze and head pose—application in an e-learning environment
US9875445B2 (en) Dynamic hybrid models for multimodal analysis
Kächele et al. Inferring depression and affect from application dependent meta knowledge
Al Osman et al. Multimodal affect recognition: Current approaches and challenges
Avola et al. Deep temporal analysis for non-acted body affect recognition
Palmero et al. Chalearn LAP challenges on self-reported personality recognition and non-verbal behavior forecasting during social dyadic interactions: Dataset, design, and results
Ramakrishnan et al. Toward automated classroom observation: Multimodal machine learning to estimate class positive climate and negative climate
Wang et al. Unlocking the emotional world of visual media: An overview of the science, research, and impact of understanding emotion
Randhavane et al. The liar's walk: Detecting deception with gait and gesture
Salah et al. Challenges of human behavior understanding
Sun et al. Personality assessment based on multimodal attention network learning with category-based mean square error
Alshammari et al. Robotics utilization in automatic vision-based assessment systems from artificial intelligence perspective: A systematic review
Rodrigues et al. Classification of facial expressions under partial occlusion for VR games
Yang et al. Analysis of interaction attitudes using data-driven hand gesture phrases
Kapcak et al. Estimating romantic, social, and sexual attraction by quantifying bodily coordination using wearable sensors
Rozaliev et al. Methods and Models for Identifying Human Emotions by Recognition Gestures and Motion
Akhtar et al. Visual nonverbal behavior analysis: The path forward
Gladys et al. Survey on multimodal approaches to emotion recognition
NL2020989B1 (en) Monitoring and analyzing body language with machine learning, using artificial intelligence systems for improving interaction between humans, and humans and robots.
Rozaliev et al. Recognizing and analyzing emotional expressions in movements
Mai et al. Video-based emotion recognition in the wild for online education systems
Bigand et al. Animating virtual signers: The issue of gestural anonymization