WO2019098835A1 - A computer-controlled method of and apparatus and computer program product for supporting visual clearance of physical content - Google Patents

A computer-controlled method of and apparatus and computer program product for supporting visual clearance of physical content Download PDF

Info

Publication number
WO2019098835A1
WO2019098835A1 PCT/NL2018/050770 NL2018050770W WO2019098835A1 WO 2019098835 A1 WO2019098835 A1 WO 2019098835A1 NL 2018050770 W NL2018050770 W NL 2018050770W WO 2019098835 A1 WO2019098835 A1 WO 2019098835A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
data
computer
eye tracking
observer
Prior art date
Application number
PCT/NL2018/050770
Other languages
French (fr)
Inventor
Peter-Paul VAN MAANEN
Anne-Marie BROUWER
Jonathan Alrik ZWAAN
Original Assignee
Joa Scanning Technology B.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Joa Scanning Technology B.V. filed Critical Joa Scanning Technology B.V.
Publication of WO2019098835A1 publication Critical patent/WO2019098835A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • G06Q10/06395Quality analysis or management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • G06Q10/06398Performance of employee with respect to a job function
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/08Logistics, e.g. warehousing, loading or distribution; Inventory or stock management
    • G06Q10/083Shipping
    • G06Q10/0832Special goods or special handling procedures, e.g. handling of hazardous or fragile goods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/26Government or public services
    • G06Q50/265Personal security, identity or safety

Landscapes

  • Business, Economics & Management (AREA)
  • Human Resources & Organizations (AREA)
  • Engineering & Computer Science (AREA)
  • Economics (AREA)
  • Strategic Management (AREA)
  • Development Economics (AREA)
  • Educational Administration (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Tourism & Hospitality (AREA)
  • Physics & Mathematics (AREA)
  • Marketing (AREA)
  • General Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Game Theory and Decision Science (AREA)
  • Computer Security & Cryptography (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Primary Health Care (AREA)
  • Eye Examination Apparatus (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

A computer-implemented and computer-controlled method of, and an apparatus and computer program product for clearance of physical content to be inspected, such as container content. The computer (12) collects eye tracking data (19) of a human observer (20) performing the inspection by viewing (21 ) the content or an image (23) of the content. From real-time processing of the collected eye tracking data, a clearance profile is determined based on expert eye tracking data (16; 18) of visual content inspection, such as container content inspection. The content is cleared by the computer (12) if the determined clearance profile meets a set clearance profile. The eye tracking data (19) may be enhanced by observer performance data (22) collected during the visual inspection.

Description

Title
A computer-controlled method of and apparatus and computer program product for supporting visual clearance of physical content.
Technical Field
The present disclosure relates to visual content inspection, and more particularly, to visual content inspection by a computer based on a human observer viewing the content or an image of the content to be inspected, generated by an image generator.
Background
Inspection of content, such as container content for security checks, for example the inspection or screening of content of suitcases, hand bags or hand luggage and parcels for transportation by airplanes, trains and busses, and inspection at entrance or access to events, social gatherings, meetings, buildings and or other security checks, for example, is nowadays widely applied an in most cases mandatory by legal requirements.
This inspection is primarily performed by Transportation Safety Officers (TSOs) and trained special security personnel or agents. It is the task of these officers and agents to identify threat items or potential threat items in containers that often comprise a variety of clutteredly packed items. Besides special detection equipment, such as metal detectors, radiation detectors and explosive detectors, for example, the identification is performed by visual inspection of two-dimensional, 2D, or three-dimensional, 3D, images of the content of the containers generated by an X- ray screening or other scanning technology.
Container cargo transport by planes, by trucks over land and by ships over sea, for example, for border security and for custom officials to check whether the contents match the cargo description, and whether illegal cargo can be found, is also subjected to visual inspection by an official observer from an image of the content of a container to be inspected. By viewing an image of the content, the observer must determine whether a container is believed free from threats or whether a container matches the cargo description, for example, as result of which the container is declared“cleared” by the observer and may pass. Whenever there is a potential threat or suspicion of illegal cargo, for example, the container is subjected to further search, which may involve a time and human resource consuming physical inspection involving the unloading of the container, for example. In case the visual inspection leads to a serious suspicion or threat, the container is even to be handled with extra care and/or inspected under strict safety requirements.
In practice, if there is an increased risk or if threats are highly prevalent, inspecting officers or agents will more likely make potential or serious threat decisions, such that“clear” responses are slower, because the observer will apply an increased reliability standard in viewing the container content, to avoid as much as possible making a mistake. At airport and access security checks, this may lead to congestions in the security handling and can lead to frustrations with the persons to be cleared and the clearing officials which, eventually, may have a negative influence on the overall quality of the inspection.
Hand baggage or luggage screening for safety checks, for example, is a repetitive visual search task that often has a very low probability of encountering a threat, but may have high consequences if a serious threat is missed. Since threats are of low prevalence, a“cleared” response becomes the more frequent decision. Observers may tend to abandon a search in less than the average time required to find a target under such circumstances.
Safety standards applicable to baggage or luggage inspection at airports, for example, where officers or agents have to find a target in the midst of distractors, impose strict operating times on observers, such as shifts of twenty minutes, for example. These operating times are determined, with a high safety margin, from the time during which an average trained human observer can stay focussed. It will be appreciated that, based on these safety requirements, for operating baggage or luggage inspection during the opening hours of an airport, a large team of well-trained security or screening personnel is required.
Other applications for clearing physical content by visual inspection involve forgery control, such a forged documents like passports, but also paintings and the like, for which a human observer visually inspects the document or painting, but also images of the human or animal body made by security body scan equipment and X-ray equipment, Computed (Axial) Tomography, CT or CAT, scans and Magnetic Resonance Imaging, MRI, in health care applications, for assessing an injury, such as fractures or dislocations, or a disease, for example bone degeneration, infections or tumours, lung and chest problems. In practice, experts like radiologists visually review the pictures and create a report with their findings to aid in diagnosis. In the case of border security control, for example, a customs official checks the identity of a person by viewing the content of a passport, or other ID document, and the person itself.
With these types of visual inspection, the same problems may arise as outlined above with respect to container content inspection.
Accordingly, there is need for improving visual inspection of physical content when viewing the content or viewing or monitoring a displayed image of the content by a human observer, for at least one of decreasing the time in which an inspection can be completed by an observer, for increasing consistency and reliability in the decision making process by an observer, and for establishing observer operating and rest times for reliable inspection of the content.
Summary
It is an object of the present disclosure to provide a computer- implemented method for visual clearance of physical content to be inspected. It is a further object of the present disclosure to implement such a method in an apparatus for visual clearance of physical content to be inspected, and to provide a container content screening arrangement comprising or arranged for operating with such an apparatus.
Accordingly, in a first aspect of the present disclosure, there is provided a computer-implemented and computer-controlled method of clearing physical content to be inspected, such as container content, wherein the computer connects to an eye tracking device arranged for providing eye tracking data of a human observer inspecting the content by viewing the content or an image of the content, the computer performing the steps of:
collecting eye tracking data of the eye tracking device while the observer viewing the content or the image of the content, real-time processing of the collected eye tracking data for determining a clearance profile based on human expert eye tracking data of visual content inspection, such as container content inspection, and
clearing the content if the determined clearance profile meets a set clearance profile.
Manual clearance of physical content from directly viewing the content itself or from viewing images of the content, requires inspecting officers or agents that are highly trained and experienced in object recognition.
An image of container content for visual inspection purposes, such as an X-ray image of the container content, for example, generally comprises so-called contoured objects. That is, an object in the container, and the container itself, are depicted by their outer or circumferential contour. Separate parts or items inside an object are also displayed by their contour and, optionally, enhanced by a respective colouring. Hence, the objects inside the container to be inspected, for the greater part, have to be recognized by the observer from their shape or contour and, if available, colour and colour differences. Face recognition, in the case of border security control, for example, requires observers trained in recognizing features of a human face.
In general, a decision whether to clear a particular physical content is, at present, made in the brain of the observer, from the information and stimuli gained at viewing the content or an image of the content, in correlation with the observer's object and feature knowledge and experience. In this brain process, if the observer does not perceives a dangerous or forbidden object, for example, or a deviation or an abnormality in a document, such a special security element or feature, or in the resemblance of a human object and a picture at an ID document, such that there is no suspicion concerning the viewed content, the observer clears same, i.e. let the content pass.
Eye tracking technology offers the possibility of capturing visual behaviour in real-time when monitoring or viewing objects and images. In practice, eye tracking technology is proposed for controlling workflows involving a specified sequence of tasks, or for investigating whether all objects in a structured cluster of objects have been viewed at, for example. Eye tracking technology is also proposed for training purposes, for detecting reading or viewing patterns that may lead to poor comprehension, for example. Other applications of eye tracking technology involve monitoring of alertness or distraction and situational awareness of humans while performing a task, such as airplane pilots, truck and car drivers, machine operators, etc.
Different from applying eye tracking technology in monitoring the performance of a predefined or well specified and structured task, the method according to the present disclosure explicitly makes use of information that can be extracted from the viewing as such of a human observer inspecting unstructured, random or cluttered content, using eye tracking technology.
According to the present disclosure, clearance of visually inspected content is determined, by the computer, from the visual behaviour of the observer during the inspection, expressed in a clearance profile, by real-time processing of collected eye tracking data of the observer and validating the processed data against expert eye tracking data of visual content inspection. When the determined clearance profile meets a set clearance profile, the content may be automatically cleared by the computer. It has been found that such an automated decision-making process significantly speeds-up content inspection compared to a human decision-making process.
Instead of massively deploying object or shape recognition technology, requiring high capacity, high resolution and high speed data processing algorithms and equipment, and optionally enhanced by chemical and material sensing technology, for supporting or even replacing a human observer, the present disclosure advantageously makes use of the nowadays readil available and accurate eye tracking technology, producing an amount of data at a rate that allows for real-time processing, not requiring excessive computing equipment, and such that content can be cleared by the computing equipment, even when the human observer in his brain still processes the content information perceived.
With the present disclosure, the human observer is not replaced and remains crucial. The information as to the decision to be made by the computer is obtained from the visual inspection by the human observer. Visual inspection tasks in the context of the present disclosure are best performed by human observers, which can effectively perceive anomalies within complex scenes, such as content comprised of clutteredly packed items, and that can be readily detected from eye tracking data of the human observer.
Hence, the disclosed method, so to say, is based on the best of two worlds, data available from the best performers of visual content inspection tasks, i.e. the expert data, and eye tracking data that enable real-time processing while the content inspection is carried out.
it will be appreciated that the term real-time processing in the present description and the claims refers to the processing or the execution of data in a short time period after collecting same, providing near-instantaneous output. Real-time data processing is also called stream processing, because of the continuous stream of input data required to timely yield output for the purpose of a process that is momentarily carried out. This in contrast to a batch data processing system, that collects data and then processes all the data in bulk at a later point in time, such that the processing result will become available after all the data have been collected.
The method according to the present disclosure can be seamlessly- integrated in visual content inspection tasks, either existing or future tasks and systems.
Hand baggage or luggage applied at airport safety checks, for example, in the majority of cases does not contain suspicious objects or including a threat, hence the majority of these checks is quickly cleared by the computer, before the human observer, thereby effectively reducing the time in which an inspection is completed.
On the other hand, in case of a threat or suspicion, the particular content is also quickly discriminated as such, i.e. automatically declared not cleared, by the computer and can be subjected to inspection by another observer and/or a special treatment, like physical inspection of a not cleared container, or inspection using special equipment and the like.
The computer-implemented method according to the present disclosure further provides for an objective inspection, based on the expertise of the human observer, not being prone to whether suspicious content is encountered in a minority or a majority of the inspection tasks, or other external or subjective influences like an increased state of security, for example. The latter can be effectively dealt with in the present disclosure by applying an adaptable clearance profile, i.e. adapting a set clearance profile, or applying a plurality of clearance profiles, i.e. setting a respective clearance profile, that has to be met for clearance of the content.
In an embodiment of the method according to the present disclosure, the collected eye tracking data at least comprise gaze point data, wherein the clearance profile is determined, by the computer, from at least one of eye fixations, heat maps, areas of interest, time spent, fixation sequences, time-to-first-fixation, saccades, smooth pursuit movements, vergence movements, and vestibule-ocular movements, identified by the computer in the collected eye tracking data.
Gaze points constitute the basic unit of measure in eye tracking technology. One gaze point equals one raw sample captured by an eye tracking device. The viewing time represented by a gaze point depends on the measurement rate of the eye tracking device. Accordingly, several eye tracking data parameters may be processed from the gaze points.
If a series of gaze points happens to be close in time and range, the resulting gaze cluster denotes a fixation, a period in which the eyes are locked toward a specific object. A fixation correlates with informative regions of the content under inspection.
Heat maps are static or dynamic aggregations of gaze points and fixations revealing the distribution of visual attention. Heat maps may also serve as an excellent method to visualize which elements of the stimulus were able to draw attention, using a particular colouring scheme, for example.
Areas of Interest, AOIs, are user-defined sub-regions of a viewed content. Extracting metrics for separate AOIs provide for evaluating the performance of two or more specific areas in a same image, or document, for example. For example, to compare conditions, or different features within the same content.
Time spent quantifies the amount of time that observers spent looking at an AOI. As observers have to blend out other stimuli in the visual periphery that could be equally interesting, the amount of time spent often indicates motivation and conscious attention, prolonged visual attention at a certain region of the content clearly points to a high level of interest, while shorter times indicate that other areas of the content or in the environment might be more attractive.
Fixation sequences are generated based on fixation position and timing information. This is dependent on where observers look and how much time they spend, and provides insight into the order of attention, indicating where observers look first, second, third etc. Fixation sequences also reflect brightness, hue, saturation etc. in a displayed image of the content or environmental content, that are likely to catch attention.
Time-to-First-Fixation, TTFF, indicates the amount of time it takes an observer to look at a specific area of interest from stimulus onset. TTFF can indicate both bottom-up stimulus driven searches, a risk bearing object catching immediate attention, for example, as well as top-down attention driven searches, wherein observers actively decide to search for certain elements or areas in the content.
Saccades are rapid eye movements between fixations. Analysis metrics include fixation or gaze durations, saccadic velocities, saccadic amplitudes, and various transition-based parameters between fixations and/or regions of interest. Hence, saccades represent jumps of the eyes while viewing the content. Saccades can be voluntarily, but also occur reflexively, even when the eyes are fixated.
Smooth pursuit movements are much slower tracking movements of the eye, under voluntarily control. Although smooth pursuit movements are generally made while tracking a moving object or stimulus, highly trained observers can make a smooth pursuit movement in the absence of a moving target. The present disclosure is further not limited to static content, but applies as well to moving physical content.
Vergence movements align the two eyes of the observer to targets at different distances from the observer. Vergence movements are disjunctive or disconjugate, different from conjugate eye movements in which the two eyes move in the same direction.
Vestibule-ocular eye movements stabilize the eyes relative to the external world, thus compensating for head movements.
The clearance profile, in an embodiment, may comprise an observation pattern determined, by the computer, from one or more of the above eye tracking data parameters identified in the collected eye tracking data and patterns of such eye tracking parameters known from or identified in the expert eye tracking data.
In another embodiment, the clearance profile is determined, by the computer, from anomalies in the above-mentioned processed eye tracking data and parameters identified in the collected eye tracking data compared to the expert eye tracking data, and/or unknown patterns identified in the collected and processed eye tracking data.
It will be appreciated that a clearance profile may be content dependent, and that several clearance profiles for similar content may be determined and set for providing a clearance.
In particular when an observer performs the visual inspection by viewing an image of the content, the clearance profile, in an embodiment of the present disclosure, is further determined, by the computer, from at least one of image quality analysis, image dimension analysis and feature extraction analysis of the image of the content.
Low resolution images, less coloured images, low brightness and low contrast are factors that may alter the viewing behaviour of an observer, for example an abnormal increase of fixation sequences, compared to images of excellent quality, i.e. high resolution, high brightness and contrast, for example. Accordingly, false alarms, i.e. false non-clearance, can be avoided by taking the image quality into account when determining a clearance profile.
Content with a plurality or very specific Areas of Interest, AOIs, for example, such as special security features that have to be inspected in the content, may require determining the clearance profile by establishing or resolving same, by the computer, for respective parts of the content or respective locations in the image. The eye tracking data obtained and/or the clearance profile obtained may also be normalized against a reference image or reference content, such to rule out human tendencies or preferences for particular parts of the content to be inspected, not particularly contributing to the inspection task.
In a further advantageous embodiment of the present disclosure, the computer further performs the step of providing annotations based on real-time processing of the collected eye tracking. In particular, the annotating may comprise annotation of a displayed image, such as highlighting of locations at the displayed image.
Notifications may be provided to the observer while the content is not yet cleared. As the clearance profile is established in real-time, it has been found that deviations in the processed eye tracking data parameters, deviations in the clearance profile itself or deviations obtained while validating same against a set clearance profile, may be used to notify the observer of potential threats, suspicious objects, or otherwise areas of interest, even before the observer him/herself has concluded that a respective area or part of the physical content needs more attention in viewing.
Notification may be performed by highlighting the image, such as but not limited to hatching, arrowing, encircling, filtering, contrast/brightness changes, filtering out of surrounding image content, colouring, etc. and by providing written instructions, for example. By this type of quick feedback, the viewing behaviour of the observer is effectively supported and a reduced inspection time is achieved, compared to inspection without such support. Notifications may also include audio and/or tactile and/or visual feedback to an observer, such as spoken words through an earpiece or the like and/or vibrations by a vibration mechanism and/or a warning light or lights, for example.
Notifications may also be provided in case of not yet cleared content, to facilitate physical inspection or another treatment of suspicious content or a so- called second opinion visual inspection, for example. This, to enhance and speed-up as much as possible the screening and clearance of suspicious content. That is, for example, an annotated image of the content to be inspected indicating the suspicious content may be displayed to an officer or other person entrusted with the physical inspection of the not cleared content. It will be appreciated that the computer may be arranged for randomly selecting content for physical inspection, for example for keeping the attention level of an inspector at a required level.
In a yet further embodiment, clearance of the content is calculated from eye tracking data collected during a set time window starting from presentation of the content or an image of the content to the observer.
By setting a time window, anomalies in the viewing behaviour of the observer, for example caused by events or distractions in the environment of the observer, fatigue or other lack of alertness, which may have a negative impact on the quality of the inspection, can be efficiently detected. In case of an observer viewing an image of the content to be inspected, the time window may be automatically set by the computer from at least one of image quality analysis, image dimension analysis and feature extraction analysis of the image of the content, and the expert eye tracking data, such to take the into account the quality of an image and/or diversity or variety thereof.
It will be appreciated that a clearance or a non-clearance by the computer, does not necessarily have to imply that the decision made by the computer should be accepted. The human observer eventually may decide whether to clear, reject, or to elect the content for a further search or investigating, for example by physical inspection of the content of a container, or using additional equipment and tools when inspecting a document, for example.
Accordingly, in an embodiment of the present disclosure the clearance of the content is further based on receiving, by the computer, of an input of an observer viewing the content or an image of the content to be inspected. The input may be any type of input signal provided by hand, foot, voice, movement of the head, facial expression, or detected from the movement of the eyes of the observer, for example.
In this embodiment, the final decision to clear the content is made dependent on both the decision of the observer and the decision made by the computer. That is, effectively, the decision whether to clear the content is based on data available from a trained observer and data of experts of visual content inspection tasks, i.e. the expert data used by the computer, resulting in a decision-making process with improved reliability.
In both cases, that is a clearance decision automatically made by the computer and a clearance decision by both the computer and the observer, the method according to the present disclosure significantly increases the consistency and reliability in the decision making process by an observer inspecting physical content to be cleared. In this context, and for statistical purposes, for example, in a further embodiment of the present disclosure, the computer calculates a figure of merit from the determined clearance profile and the set profile, wherein the content is cleared if the figure of merit is within a set range.
The eye tracking data obtained and the processed clearance profiles may serve for improving the method, i.e. an algorithm loaded in the computer for processing the eye tracking data and determining clearance profiles, thereby facilitating machine learning.
In addition to the collection of eye tracking data, the present disclosure also proposes collecting of observer performance data at least one of physiological data, including electroencephalography data, mental data, including at least one of facial expression data, oral expression data and body movements data of the observer while performing the viewing, i.e. the visual inspection, and to determine, by the computer, the clearance profile from real-time processing of both the collected eye tracking data and the observer performance data.
By associating such observer performance data and the eye tracking data parameters processed from the eye tracking, the quality of the inspection in the sense of an increased number of computer clearances is further improved. The observer performance data may also be used in calculating the figure of merit.
Among others for establishing observer operating and rest times for reliable inspection of the content, in a further embodiment of the method according to the present disclosure, an inspection quality parameter is determined, by the computer, from cumulative real-time processing of collected data, the inspection quality parameter providing an indicator of the observer performance when viewing the content or an image of the content, wherein if the calculated inspection quality parameter does not meet a set inspection quality level representative of a set inspection quality, the computer providing at least one of feedback information to the observer.
The inspection quality parameter is based, in an embodiment, on one or more of eye pupil size/dilation, ocular vergence, blinks and distance to the object or image, indicative for body movements.
An increase in pupil size is referred to as pupil dilation, and a decrease in size is called pupil constriction. Pupil size primarily responds to changes in light (ambient light) or stimulus material (image stimulus). Other properties causing changes in pupil size are emotional arousal (referring to the amount of emotional engagement) and cognitive workload (which refers to how mentally taxing a stimulus is). Pupillary responses may be used as a measure for emotional arousal.
As explained above, the extraction of vergence, i.e. whether left and right eyes move together or apart from each other, is a natural consequence of focusing of the eye near and far. Divergence often happens when the human mind drifts away, when losing focus or concentration. It can be picked up instantly by measuring inter-pupil distance.
Eye tracking data can also provide essential information on cognitive workload by monitoring blinks. Cognitively demanding tasks can be associated with delays in blinks, the so-called attentional blink. However, many other insights can be derived from blinks. A very low frequency of blinks, for example, is usually associated with higher levels of concentration. A rather high frequency is indicative of drowsiness and lower levels of focus and concentration.
Along with pupil size, eye tracking data may also comprise a measurement of the distance to the content or image and the relative position of the observer. Leaning forwards or backwards in front of a remote content is tracked directly and can reflect approach-avoidance behavior.
The feedback provided if the calculated inspection quality parameter does not meet a set inspection quality level representative of a set inspection quality, can be any of visual, audio or tactile feedback, for example. The reliability of a clearance is further enhanced, in an embodiment of the present disclosure, if both the determined clearance profile meets the set clearance profile and the determined inspection quality parameter meets a set inspection quality parameter representative of a set inspection quality.
In addition to the above-mentioned embodiment, the inspection quality parameter, in a further embodiment, the inspection quality parameter is based on the collecting of at least one of physiological data, including electroencephalography data, mental data, including at least one of facial expression data, oral expression data and body movements data of the observer while performing the visual clearance.
Another embodiment according to the present disclosure provides that the clearance profile is determined, by the computer, based on collecting and real time processing of eye tracking data, whether or not in connection with at least one of physiological data, including electroencephalography data, mental data, including at least one of facial expression data, oral expression data and body movements data, or other performance data of a plurality of human observers simultaneously viewing the content or an image of the content to be inspected.
It has been envisaged that the reliability of the clearance decision by the computer can be further enhanced by collecting and real-time processing of data collected from several observers simultaneously inspecting the same content, such as two or three observers. The observers may view a same image or different images of the same content to be inspected. For example images taken at a different viewing angle.
The data collected from the individual observers using individual eye tracking equipment and sensors, may be individually processed, in accordance with the method disclosed above, while the decision to clear the content is based on the individual processing results, for example based on a majority decision and/or taking into account the quality of the individual viewings, the experience of the observers, etc.
Alternatively, the individually collected eye tracking data and/or performance data may be integrally processed by the computer, resulting in a decision whether or not to clear the content. The observer or observers may inspect the image of the content at a position remote from where the image of the content is obtained and the number of observers that view an image simultaneously may change over time.
In a second aspect, the present disclosure provides an apparatus for clearing physical content to be inspected, such as container content, the apparatus comprises a computer and data input/output equipment, the computer being communicatively connected or connectable to a database, the data input equipment comprising at least one eye tracking device arranged for providing eye tracking data of an observer inspecting the content by viewing the content or an image of the content, the computer being arranged for performing the computer-implemented steps of the method disclosed above.
In a further embodiment, the data input/output equipment of the apparatus further comprises at least one sensor for collecting at least one of physiological data, including electroencephalography data, and mental data, including at least one of facial expression data, oral expression data and body movements data of an observer while viewing the content or image.
In another embodiment, the data input/output equipment may comprise a display or screen or monitor, for displaying an annotated image of the content to an officer or other person entrusted with the inspection of content that is not cleared and requires further physical inspection.
The present disclosure, in particular, provides an apparatus or arrangement for container content screening or scanning, comprising a housing having an entrance for receiving containers, such as luggage, baggage or parcels, the content of which having to be inspected by at least one human observer, an exit at which containers that are inspected leaving the apparatus, a transport device for transporting the containers from the entrance to the exit, an image generator, arranged for generating an image of the container content inside the housing for viewing by at last one observer, at least one control console operatively arranged for operating the transport device, and an apparatus or arranged for communicatively connecting to an apparatus according to the second aspect of the present disclosure.
In an embodiment, the arrangement may comprise a display or monitor at or remote of the exit at which an annotated image of the inspected content is displayed for further physical inspection of content that has not been cleared or for random inspection, for example. In a third aspect there is provided a computer program product downloadable from a communication network and/or stored on a computer-readable and/or processor-executable medium, the computer program product comprising program code instructions to cause a computer to carry out the computer-implemented steps of the method according to the present disclosure.
The above-mentioned and other features and advantages of the present disclosure are illustrated in the following description with reference to the enclosed drawings which are provided by way of illustration only and which are not limitative to the present disclosure.
Brief Description of the Drawings
Fig. 1 shows, in a schematic and illustrative manner, an embodiment of an apparatus for clearing physical content.
Fig. 2 shows, in a schematic and illustrative manner, another embodiment of an apparatus for clearing physical content.
Figs. 3, 4 and 5 show typical examples of images for visual inspection of physical content, such as screening of a handbag at an airport security check, a truck for transporting cargo, and a scan of human lungs.
Figs. 6 and 7 show flow chart diagrams of embodiments of the method according to the present disclosure.
Fig. 8 shows, in a schematic and illustrative manner, based on Fig. 1 , an embodiment of an apparatus for clearing physical content operating with a plurality of observers.
Figs. 9, 10 and 1 1 show initially obtained results from a pilot test applying the present method, plotted as Receiver Operating Characteristic, ROC, curves.
Fig. 12 shows, in a schematic and illustrative manner, a luggage or baggage inspection apparatus or arrangement, such as typically used at airport security checks, implementing the teachings of the present disclosure. Detailed Description
It is noted that in the description of the figures, same reference numerals refer to the same or similar components performing a same or an essentially similar function.
In general, two types of devices for eye tracking are distinguished, the so-called screen-based eye trackers and eye tracking glasses.
Screen-based type eye trackers, or desktop or stationary eye trackers comprise a monitor or other electronic display device at which an image to be inspected is displayed, such as an image of content to be inspected. At or near the screen eye tracking capturing equipment is positioned, such as one or more digital cameras, for remote eye tracking of a user or an observer that views the image displayed at the screen.
Eye tracking glasses type eye trackers are devices that are positioned close to the eyes of the observer, usually head-mounted or mounted onto an eyeglass frame. Eye tracking glasses may be used for inspecting content not necessarily displayed on a screen, such as for inspecting hard copy documents, for example.
Although eye tracking glasses provide observers to move freely when performing a task, compared to screen-based eye trackers providing limited freedom for the observer to move, eye tracking glasses may shift in front of the eyes during data capturing, in particular when they are not mounted correctly.
Eye-tracking glasses technology may be combined with so-called smart glasses or virtual reality glasses or video headsets, by which an image of content to be inspected is displayed in front of the eyes of an observer. Like screen-based eye trackers, smart glasses eye trackers allow for providing visual feedback or notifications to the user or observer during the inspection, such as by highlighting the image in the form of hatching, arrowing, encircling, filtering, contrast/brightness changes, filtering out of surrounding image content, etc. to support the inspection task and providing written instructions, for example. In case the inspection task is not sufficiently performed, the display of the image may be blocked, for example.
Figure 1 schematically illustrates an embodiment of an apparatus or assembly 10 for providing computer-implemented clearance of physical content to be inspected, using screen-based eye tracking technology. The apparatus 10 essentially comprises a computer or processor 12 to which an output device connects 36, such as a screen or monitor 1 1 , and an input device connects 35, such as a digital camera or eye tracking capturing module 13. The screen or monitor 1 1 optionally may comprise an audio output device 14, such as a loudspeaker and/or earpiece, and an audio input device 15, such as microphone, connecting 36 to the computer or processor 12.
The computer or processor 12 may access at least one database, such as an internal or local data base and/or a local and working memory 16 and/or an external or remote database 18 communicatively connected or connectable to the computer or processor 12 via a data communication network 17, such as the internet. In use, a user or observer 20 views 21 an image 23 of content to be inspected, which image 23 is displayed at the screen or monitor or display 1 1 , while the eyes of the human observer 20 are tracked 19 by the digital camera or eye tracking capturing module 13.
Reference numeral 22 schematically denotes data input/output equipment of the apparatus 10 comprised of at least one sensor for collecting observer performance data, comprising at least one of physiological data, including electroencephalography data of the observer, while performing the visual clearance. The data input/output equipment 22 connects to the computer 12 by a wired link 34.
Eye tracking data and sensor data collected may be stored in the database 16, 18 for use by the computer 12.
Figure 2 schematically illustrates an apparatus or assembly 24 for supporting visual clearance of physical content to be inspected, comprising eye tracking glasses technology. In the embodiment shown, the human user or observer 20 wears an eye glasses device 25 comprised of a frame 28 at which a digital camera or other eye tracking capturing module 27 is fixed.
In use, the user or observer 20 views 26 a document 30 while eye tracking data 29 captured by the module 27 are transmitted by a wired or wireless transmission link 31 to a computer or processor 12, as explained above with reference to Figure 1. The frame 28, optionally, may comprise a loudspeaker 32 and microphone 33 audio headset, communicatively connected to the computer or processor 12 via the transmission link 31 , for example.
In the embodiment shown, data input/output equipment 22 such as one or more sensor(s) connect by a wireless link 37 to the computer 12. Those skilled in the art will appreciate that in both the embodiments of Figure 1 and Figure 2 the data input/output equipment 22 may connect to the computer 12 by either one of a wired 34 or wireless data communication link 37.
In the case of a virtual reality glasses or video headsets, the frame 28 comprises a screen or display (not shown) at which an image of the physical content to be inspected is displayed.
Eye tracking devices as such are readily known and described in the prior art, such that a further description thereof may be omitted here.
Figure 3 shows a typical example of an X-ray image, available from Wikimedia Commons, obtained with a device for visual inspection of container content such as suitcases, hand bags or hand luggage and parcels for transportation by airplanes, trains and busses, and inspection at entrance or access to events, social gatherings, meetings, buildings and or other security or screening checks, for example.
As can be seen from the image, objects 41 - 44 in the container or bag 40, and the container or bag 40 itself, are depicted by their outer or circumferential contour. Separate parts or items inside an object, such as the electronic equipment 43 are also displayed by their contour. In practice, these objects may be enhanced by a respective different colouring. The objects 41 - 44 inside the container 40 to be inspected, which are clutteredly packed, i.e. in an unstructured manner, have to be recognized by Transportation Safety Officers, TSOs, and special security personnel or agents by their shape or contour and, if available, colour and colour differences, to identify threat items or potential threat items.
Figure 4 shows a typical example of an X-ray image, available from Evergreen Shipping Agency (Europe) GmbH Netherlands branch, obtained of a truck 45 and a trailer 46 containing cargo 47, 48 that is normally not visible from the outside, because the cargo 47, 48 is surrounded by a cover. From the image, well trained border security and custom officials may check whether the contents match the cargo description, and whether illegal cargo can be found, for example in randomly positioned hollow spaces in the chassis 49 of the trailer and the truck or between the cargo items 47, 48.
Figure 5 shows a typical X-ray scan, available from Wikimedia Commons, of the human chest and lungs, from which a trained physician or radiologists by visual inspection of the image 50 has to detect physical content like a tumour or abscess 51 , for example. In the context of the present disclosure, the term physical content is used to identify real physical objects in the most common meaning thereof, different from virtual or artificial objects.
In all these inspection examples, by viewing an image of the content, the observer must determine whether the content is believed free from strange, unusual objects or deviations from a normal or healthy state, or matches a particular description, or the like, as result of which the physical content is declared“cleared” by the observer and my pass. The term 'clearance' in the present disclosure is to be interpreted as any state of the content that matches a predetermined condition.
Although the examples in Figures 3 - 5 show two-dimensional, 2D, images, in practice three-dimensional, 3D, images may be provided for inspection.
In accordance with the present disclosure, eye tracking data of an observer that is performing a visual content inspection are captured by a computer- controlled arrangement or apparatus 10, 24 as illustrated in Figures 1 and 2, respectively, comprising an eye tracking device such as a screen-based eye tracking device or video headset, in the case of viewing images of the content to be inspected, or an eye glasses based eye tracking device when either one of an image or a real physical object is to be inspected, the latter for example in the case of inspecting original paintings, documents and passport control.
An example of the method according to the present disclosure is illustrated by the flow chart diagram 60 shown in Figure 6. In the flow chart, the steps proceed from the top to the bottom of the drawing. Other flow directions are indicated by a respective arrow. The illustrated method is implemented and controlled by the computer 12, running a data processing algorithm, suitably programmed by program code instructions that may be stored in the database 16 or loaded therein as a computer program product from a computer readable medium, such as, but not limited from a memory stick, hard-disk, solid-state drive, or other non-transitory medium, or as data signal, downloaded from a data communication network, for example.
In a first step, content to be inspected is made available for inspection by a human observer, i.e. block 61 "Present content". While the observer views the content or an image of the content, eye tracking data are obtained, i.e. block 62 "Capture eye tracking data", which are real-time processed by the computer 12 for determining a clearance profile, i.e. block 63 "Determining clearance profile". It will be appreciated that the clearance profile is determined by the computer 12 from a particular amount of captured eye tracking data, for example during a time period of one, two, three or four seconds, in correlation with previously established expert eye tracking data.
The amount of expert eye tracking data and the processing algorithm used in determining the clearance profile are content particular, such as a processing algorithm and expert data pertaining to inspection of unstructured or clutteredly packed container content for safety purposes, as illustrated in Figure 3, or a processing algorithm and expert data pertaining to medical content inspection, as illustrated by Figure 5, for face recognition or for forgery content inspection, or a processing algorithm and expert data relating to cargo inspection, as illustrated by Figure 4, for example.
In decision block 64, i.e. "Meets set profile?", the processing algorithm determines whether the clearance profile established in block 63 meets a set or pre-set clearance profile. In the negative, i.e. decision "No" of block 64, if the clearance profile determined and/or the processed eye tracking data conclude to a suspicious content, such as container content forming a threat, i.e. decision block 65 "Suspicious?" answer "Yes", the content may be directly and automatically declared not-cleared by the computer 12, i.e. block 69 "Content not-cleared". If, dependent on the particularities of the determined clearance profile, the computer does not yet directly declare the content not-cleared, i.e. decision "No" of block 65, the computer may decide whether it is appropriate to provide support to the observer, i.e. decision block 66 "Notification?".
In the affirmative, i.e. decision "Yes" of block 66, the viewing of the observer is supported by one or several notifications or annotations, i.e. block 67, "Notification support". In the negative, i.e. decision "No" of block 66, the capturing of eye tracking data may continue.
In the embodiment illustrated by the flow chart 60, a time window is introduced, such that clearance of the content is calculated from eye tracking data collected during a set time window starting from presentation of the content in block 61. In decision block 68, i.e. "Time window lapsed?", it is determined whether the time window set by the computer is lapsed. In the negative, i.e. decision "No" of block 68, the processed eye tracking data accumulates and hence the determination of the clearance profile. If the time window is lapsed, i.e. decision "Yes" of block 68, and the content is not already cleared, the content is declared not-cleared, i.e. block 69.
In the case of content that is not-cleared, further measures may be required, such as a physical inspection of container content, or inspection by a further observer, for example. In all other cases, the content is automatically cleared by the computer 12, merely from processing the captured eye tracking data in relation to the relevant expert eye tracking data, i.e. block 70 "Content cleared".
The automatic computer-controlled clearance of physical content from the captured eye tracking data processing in correlation with expert data, allows for an improved, i.e. shortened, inspection completion time up till clearance of the content. This is, in particular, advantageous in the case of content inspection for security or screening purposes, such as at airport safety and security inspection desks and entrance desks at events, for example, where large quantities of content have to be inspected with a risk of waiting queues building up.
It will be appreciated that also in the case of an automatic not-cleared decision by the computer, i.e. block 69 in Figure 6, the inspection process may be advanced, as the computer, from the captured eye tracking data and the expert data, may already quickly after start of the viewing decide to reject the content, such that same can be taken apart and presented for further physical inspection, whether or not enhanced by notifications for an inspector provided by the computer.
Even in the case of notification support, i.e. block 67 in Figure 6, the inspection completion time till computer clearance of the content can be reduced compared to clearance of the content by the observer without computer control. This, because the notification or annotation can be provided quickly after start of viewing the content. By this fast feedback, the viewing of the content by the observer may be supported such that, in the case of none-suspicious content, a clearance or a none- clearance may be provided, based on the expert data available.
Notification or annotation support 67 may comprise one or more of highlighting the image 23 at the display 1 1 , for example, such as but not limited to hatching, arrowing, encircling, filtering, contrast/brightness changes, filtering out of surrounding image content, colouring, etc. and by providing written instructions, for example. By way of example, in Figure 3 object 55 is annotated by encircling thereof by a dashed circle 56. This annotation may be shown for a short time, for example, just long enough to control the viewing of the observer towards this object 55, without impeding the view thereof or the view at surrounding or neighbouring objects.
In the case of a viewing the object itself, such as illustrated in Figure 2, audible notification support may be provided, through the loudspeaker 32, for example, and/or by a light or laser pointer 38, connected 39 and controlled by the computer 12, for example. To facilitate further measures in case of a non-clearance, support may also be provided in the form of the notifications and/or annotations as disclosed above. That is, notifications may also include tactile and/or visual feedback to an observer. To this end, the data input/output equipment 22 may comprise a vibration mechanism and/or a warning light or lights, for example.
Those skilled in the art will appreciate that certain steps and decisions as shown in the flow chart diagram may be interchanged or differently positioned in the processing flow. For example, decision block 65 may also be positioned between the blocks 63 and 64.
The time window mentioned above with reference to block 68, may be a fixed time window or dynamically set or adapted by the computer 12. This time window provides an effective means for taking the viewing behaviour of the observer into account. The level of experience of the observer, the size of an image to be inspected and the image quality, determined from a feature extraction analysis of the content, for example, may have an influence on the way the image is viewed. A bad quality, for example, may prolong the viewing time before a decision can be made, or may involve eye movements that are differently valuated compared to a bright and full contrasted image, for example, although the content is such that same may be cleared.
As mentioned before, an observer may get distracted form viewing by events in his or her environment, by becoming fatigue or other reasons that lead to a lack of alertness which, eventually, may have a negative impact on the quality of the inspection.
Reference is made to the flow chart diagram 80 shown in Figure 7, illustrating a further embodiment of the present disclosure, implemented and controlled by the computer 12, running a data processing algorithm taking performance data of the observer into account. In block 81 "Present content", the content to be inspected is presented to the human observer 20 (see Figures 1 and 2 and block 61). In addition to eye tracking data of the observer, in block 82 "Capture eye tracking data and performance data", observer performance data are collected, provided by input/output equipment 22 connected to the computer 12, such as a sensor or sensors arranged for collecting at least one of physiological data and electroencephalography data, but also facial expression data of the observer 20 captured by the module 13 or 27, or a separate camera module, such as a video camera module (not shown) and, for example, oral expression data of the observer 20 registered by the microphone 15, 33 and body movements of the observer 20 while performing the visual clearance.
From the cumulative real time processing of the captured eye tracking data and performance data, both the clearance profile and an inspection quality parameter are calculated, i.e. block 83 "Determine clearance profile and inspection quality parameter". The inspection quality parameter, among others, is representative for the mental and/or physical state of the observer 20 and provides a reliable indicator of the performance of the observer 20 when viewing content 30 or an image 23 of the content at a display or screen or monitor 1 1.
The algorithm determines in decision block 84 "Meets set profiles/parameters?" whether the clearance profile and/or the inspection quality parameter determined in block 83 meet a set or pre-set clearance profile and quality inspection parameter. In the negative, i.e. decision "No" of block 84, if the clearance profile determined and/or the processed eye tracking data and/or the inspection quality parameter conclude to a suspicious result, either suspicious content and/or a questionable performance by the observer, i.e. decision block 85 "Suspicious?" answer "Yes", the content may be directly determined, by the computer 12, as being not-cleared and a notification may be provided indicating a possible reason or pointing to an object that requires further treatment/inspection, for example, i.e. block 89 "Content not-cleared/notification". If, dependent on the particularities of the determined clearance profile/inspection quality parameter, the computer does not yet directly declare the content not-cleared, i.e. decision "No" of block 85, the computer may decide whether it is appropriate to provide support to the observer, i.e. decision block 86 "Notification?".
In the affirmative, i.e. decision "Yes" of block 86, the viewing of the observer is supported by one or several notifications or annotations and/or performance feedback, i.e. block 87, "Notification support/performance feedback". In the negative, i.e. decision "No" of block 86, the capturing of eye tracking data may continue, dependent on whether a time window has lapsed, i.e. decision block 88 "Time window lapsed?". In the same manner as explained above with reference to Figure 6.
If the determined inspection quality parameter does not meet a set inspection quality level representative of a required minimum inspection quality, for example, the computer 12 may provide audible, tactile or visible feedback or information to the observer 20 about his/her performance by any of the output equipment disclosed and available with the apparatus 10, 24. Eventually, if the performance gives rise thereto, viewing of the content by the observer 20 may be prohibited, for example.
In a further embodiment, for benchmarking and for use of the captured data and clearance results for improving the algorithms used, i.e. a self-learning facility of the present disclosure, the computer calculates a figure of merit from the determined clearance profile/inspection quality parameters and said set profile/set inspection quality parameters, as shown in Figure 7 by the dashed block 97 "Figure of merit".
The figure of merit may also be used for determining whether to clear the content. In that case an additional decision step may be implemented in the flow chart of Figure 7 between the blocks 90 and 91 , or the figure of merit may be calculated between the blocks 83 and 84 and may be used in the decision by block 84, for example.
The clearance decision in blocks 70, 90 and/or the non-clearance decision in blocks 69, 89 may be further based on receiving, by the computer 12, of an input of an observer viewing the content or an image of the content to be inspected. The input may be any type of input signal provided by the input equipment 22, via a touch screen type display 1 1 , a key board or console connecting to the computer 22, etc.
In this case, the final decision to clear the content is made dependent on both the decision of the observer 20 and the decision made by the computer 12. That is, effectively, the decision whether to clear the content is based on data available from a trained observer 20 and data of experts of visual content inspection tasks, i.e. the expert data used by the computer 12, improving the reliability of the decision.
The reliability of the computer decision whether or not to clear the content is further increased, in an embodiment of a setting 120 as illustrated in Figure 8, wherein a plurality of human observers 121 , 122, 123 simultaneously viewing an image 23 of the content to be inspected. With reference to Figure 1 , each observer 121 , 122, 123 is provided with a screen or monitor 12, and an input device such as a digital camera or eye tracking capturing module 13, connecting by a wired or wireless link for data transfer, such as data bus 35, 36 to a computer or processor 125, which is common to all the observers 121 , 122, 123. The computer or processor 125 may access at least one database, such as an internal or local data base or local and working memory 126 and/or an external or remote database 18 communicatively connected or connectable to the computer or processor 125 via a data communication network 17, such as the internet. In use, a user or observer 121 , 122, 123 views 21 an image 23 of content to be inspected, which image 23 is displayed at his/her screen or monitor or display 1 1 , while the eyes of the observers 121 , 122, 123 are individually tracked 19 by his/her digital camera or eye tracking capturing module 13.
Each observer 121 , 122, 123 may comprise data input/output equipment 22 for individually collecting observer performance data such as comprising at least one of physiological data, including electroencephalography data of the observer, while performing the viewing. The data input/output equipment 22 connect to the computer 125 by a wired or wireless link for data transfer, such as a data bus 34. The individually collected eye tracking data and sensor data may be stored in the database 126, 18 for use by the computer 125.
It will be appreciated that the observers 121 , 122, 123 may also be provided with an apparatus 24 as shown in Figure 2, or a mix of the apparatuses 10 and 24, for example. The number of observers is, of course, not limited to three, and the observers may inspect the image of the content at a position remote from where the image of the content is generated, while the number of observers that view an image simultaneously may change over time.
The computer 125 is arranged for real-time processing of the data collected from the individual observers 121 , 122, 123 while viewing. The data may be individually processed, in accordance with the method disclosed above, while the decision to clear the content is based on the individual processing results, for example based on a majority decision and/or taking into account the quality of the individual viewings, the experience of the observers, etc. as disclosed above. Alternatively, the individually collected eye tracking data and/or performance data may be integrally processed by the computer 12, resulting in a common decision whether or not to clear the content. In a pilot verification of the present method, a validation test has been performed involving six experienced and six less experienced human observers, that had to view 200 images of cluttered container content of hand luggage bags or containers, typically presented at entrance security check points, such as at an airport. Of these 200 images, 100 images where completely free of threat or suspicious items, and 100 images contained at least one threat item or potential threat or suspicious item. The test was performed using collected eye-tracking data only and results of the present method were compared against the decision of the human observer, i.e. which results comprised of the threat items or potential threats or suspicious items as indicated by the human observers.
The viewing time of the individual observers has been represented by gaze point data, resulting in gaze point clusters, denoting a fixation of the viewing at an informative region of the content under inspection in a particular time window of, for example, one or two, three, four and up to 7 - 10 seconds, a period in which the eyes are locked towards a specific object. Based on the gaze clusters, the computer made a decision whether the content was to be cleared or not, i.e. had to presented for further inspection.
The initially obtained results are illustrated in so-called Receiver Operating Characteristic, ROC, curves, shown in Figures 9, 10 and 1 1 , in which the True Positive Rate, TPR, indicated at the vertical axis, is plotted against the False Positive Rate, FPR, indicated at the horizontal axis, for various decision threshold settings. An ROC curve is a commonly used way to visualize the performance of a binary classifier, meaning a classifier with two possible output classes.
In the curves of Figures 9, 10 and 11 , the TPR defines how often the computer predicts positive, i.e. a possible threat, i.e. meaning content that should be physically inspected, when the actual classification is positive, i.e. the number of true positives compared to the number of all positives, according to the results of the observers. The FPR defines when the actual classification is negative, i.e. no threat, meaning content to be cleared or passed, how often the computer incorrectly predicts positive, i.e. the number of false positives compared to the number of all negatives, according to the results of the observers.
The top-right position of the decision threshold line refers to a threshold setting of 0, and the bottom-left position refers to a threshold setting of 1. Accuracy is measured by the Area Under the ROC Curve, AUC. An area of 1 represents a perfect test and an area of 0.5 represents a worthless test, i.e. with a performance of a random classifier.
Figure 9 illustrates the ROC curve of the decisions made by the computer based on all gaze data of the human observers obtained in a viewing time window of 7 seconds. As can be seen from this curve, a 94% hit rate is achieved, i.e. when the observers are also of the opinion that a threat exists, against an only 3% false alarm rate, i.e. the operators are also of the opinion that there is no threat. With a very good AUC of 0.9796.
Figure 10 illustrates the ROC curve of the decisions made by the computer after a viewing window of 2 seconds. The hit rate is already at 77% and the false alarm rate is only at 2%, with AUC = 0.8317. Figure 1 1 illustrates the ROC curve within 1 second, achieving already a hit rate of 70% and only a 2% false alarm rate, with AUC = 0.7564.
Hence, from this pilot experiment, already after 1 second of viewing by a trained human observer, with the method presented, the computer can already find 70% of the possible threats, according to the result of a trained observer, while 98% of the items containing no threat are not falsely classified as a threat. Taken into account that a trained observer makes a decision after about 7 - 10 seconds, the pilot results demonstrate the ability of the present method for speeding-up and improving a content screening process.
Figure 12 shows a typical application of the present screening method in a luggage or baggage inspection or screening apparatus or arrangement 100. The apparatus or arrangement 100 comprises at one side an entrance 101 for receiving luggage or parcels or containers 1 12, the content of which has to be inspected and at another opposite side an exit 102 at which luggage or containers 1 13 that are inspected leave the apparatus or arrangement 100. The luggage or containers are typically transported through the apparatus or arrangement 100 by a transport belt 103. The belt 103 is automatically stopped when a bag or other container has entered the apparatus or arrangement 100 for inspection.
At this stage an image, typical an X-ray image, is generated from the content of the container inside the apparatus or arrangement 100 by an image generator 107, positioned inside the apparatus 100. At an operator side of the apparatus or arrangement 100, a screen or monitor or display 104 and a control console 105 is positioned, for use by an operator or observer performing the inspection. For capturing eye tracking data of the observer, an eye tracking module 108 is provided in the vicinity of the display 104. It will be appreciated that the display 104, control console 105, and eye tracking module 108 may be arranged remote from the apparatus or arrangement 100. In that case, the apparatus or arrangement 100 is provided with data transfer equipment for communicatively exchanging image data and the like.
The apparatus or arrangement 100, inclusive the belt 103, image generator 107 and display 104 are operated under control of a computer or processor 106, for example positioned inside the apparatus or arrangement 100. The setup of the computer 106 and display 104 is equivalent to the equipment shown in Figure 1 , i.e. computer 12, data bases 16, 18, data communication network 17, eye tracking module 13, and the input/output devices 14, 15 and 22, although the latter are not explicitly shown in Figure 8 as same may be omitted in the simplest embodiment of the present disclosure. It will be appreciated that the setup 24 of Figure 2 may also be used.
The apparatus or arrangement 100, in practice, may form part of a security or screening system, involving further screening devices and/or further observers, such as described with reference to Figure 8.
For facilitating physical inspection of content that is not cleared as a pass, a further screen or monitor or display 109 and a control console 1 10, such as keyboard, is provided at or remote of the exit 102, communicatively connecting 1 11 to the computer 106 for receiving an annotated image of the content selected for physical inspection.
For real-time processing of the captured eye tracking data of the observer and/or the observer performance data, several processing algorithms known in practice may be used, for example based on Markov chains, determining one or a plurality of eye fixations, heat maps, areas of interest, time spent, fixation sequences, time-to-first-fixation, saccades, smooth pursuit movements, vergence movements, and vestibule-ocular movements, as elucidated in the Summary part above.
The clearance profile, in an embodiment, comprises an observation pattern determined, by the computer, from the collected and processed eye tracking data and known patterns identified from the expert eye tracking data, whether or not enhanced by collected observer performance data. In another embodiment of the present disclosure, the clearance profile is determined, by the computer, from anomalies identified in the collected eye tracking data compared to the expert eye tracking data and/or unknown patterns identified in the collected and processed eye tracking data, whether or not enhanced by collected observer performance data.
In an embodiment, the clearance profile is determined, by the computer, from a function or model comprising a plurality of expert function or model parameters derived from a collection of observer data gathered from well trained, experienced observers, i.e. experts, while performing a specific visual content inspection task. The data gathered at least comprises eye tracking data and, optionally, observer performance data as elucidated above. Such a function may be established by data analysis experts and/or processor or processing machine controlled using neural network technology, for example.
From the captured eye tracking data and, optionally, collected observer performance data while performing the visual inspection task, function or model parameters in accordance with a task specific function or model are determined by the computer and correlated with the expert function or model parameters. Thereby determining a clearance profile for the purpose of computer-controlled clearance of the inspected content in accordance with the present disclosure.
Algorithms for calculating an inspection quality parameter from collected performance data of the observer may be developed and applied in a like manner as explained above with respect to the clearance profile, for example.
Other variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing the claimed invention, from a study of the drawings, the disclosure, and the appended claims. In the claims, the word "comprising" does not exclude other elements or steps, and the indefinite article "a" or "an" does not exclude a plurality. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measured cannot be used to advantage. Any reference signs in the claims should not be construed as limiting the scope thereof.

Claims

Claims
1. A computer-implemented and computer-controlled method of clearing physical content to be inspected, such as container content, wherein said computer connects to an eye tracking device arranged for providing eye tracking data of an observer inspecting said content by viewing said content or an image of said content, said computer performing the steps of:
collecting eye tracking data of said eye tracking device while said observer viewing said content or said image,
real-time processing of said collected eye tracking data for determining a clearance profile based on expert eye tracking data of visual content inspection, such as container content inspection, and
clearing said content if said determined clearance profile meets a set clearance profile.
2. The method according to claim 1 , wherein said collected eye tracking data at least comprise gaze point data, wherein said clearance profile is determined, by said computer, from at least one of eye fixations, heat maps, areas of interest, time spent, fixation sequences, time-to-first-fixation, saccades, smooth pursuit movements, vergence movements, and vestibule-ocular movements, identified by said computer in said collected eye tracking data.
3. The method according to any of the previous claims, wherein said observer is viewing an image of said content, and said clearance profile is further determined, by said computer, from at least one of image quality analysis, image dimension analysis and feature extraction analysis of said image.
4. The method according to any of the previous claims, wherein said clearance profile is further determined, by said computer, from collecting and real-time processing of at least one of physiological data, including electroencephalography data, mental data, including at least one of facial expression data, oral expression data and body movements data of said observer while viewing.
5. The method according to any of the previous claims, wherein said computer further performs the step of annotating, based on said real-time processing of said collected data, in particular wherein said annotating comprises highlighting of locations in an image of said content.
6. The method according to any of the previous claims, wherein clearance of said content is based on said data collected by said computer during a set time window starting from presentation of said content or said image of said content to said observer.
7. The method according to any of the previous claims, wherein said clearance of said content is further based on receiving, by said computer, of an input of an observer viewing said content or said image of said content.
8. The method according to any of the previous claims, wherein said clearance profile comprises an observation pattern determined, by said computer, from one or more eye tracking parameters identified in said collected eye tracking data and patterns of such eye tracking parameters known from or identified in said expert eye tracking data.
9. The method according to any of the previous claims, wherein said clearance profile is determined, by said computer, from anomalies at least identified in said collected eye tracking data compared to said expert eye tracking data.
10. The method according to any of the previous claims, wherein said computer calculates a figure of merit from said determined clearance profile and said set clearance profile, wherein said content is cleared if said figure of merit is within a set range, in particular wherein an inspection quality parameter is determined, by said computer, from cumulative real-time processing of collected data, said inspection quality parameter providing an indicator of said observer performance when viewing, wherein if said calculated inspection quality parameter does not meet a set inspection quality level representative of a set inspection quality, said computer providing at least one of feedback information to said observer, and wherein said content is cleared if both said determined clearance profile meets said set clearance profile and said determined inspection quality parameter meets a set inspection quality parameter representative of a set inspection quality.
1 1. The method according to any of the previous claims, wherein said clearance profile is determined, by said computer, based on collecting and real-time processing of eye tracking data, whether or not in connection with at least one of physiological data, including electroencephalography data, mental data, including at least one of facial expression data, oral expression data and body movements data, of a plurality of observers simultaneously viewing said content or an image of said content to be inspected.
12. An apparatus for clearing physical content to be inspected, such as container content, said apparatus comprises a computer and data input/output equipment, said computer being communicatively connected or connectable to a database, said data input equipment comprising at least one eye tracking device arranged for providing eye tracking data of an observer inspecting said content by viewing said content or an image of said content, said computer being arranged for performing said computer-implemented steps of any of the claims 1 - 1 1.
13. The apparatus according to claim 12, wherein said data input/output equipment further comprises at least one sensor for collecting at least one of physiological data, including electroencephalography data, and mental data, including at least one of facial expression data, oral expression data and body movements data of an observer while viewing.
14. A container content screening arrangement, comprising a housing having an entrance for receiving containers, such as luggage, baggage or parcels, the content of which having to be inspected by at least one observer, an exit at which containers that are inspected leaving the apparatus, a transport device for transporting said containers from said entrance to said exit, an image generator, arranged for generating an image of said container content inside said housing for viewing by at least one observer, and an apparatus or arranged for communicatively connecting to an apparatus according to claim 12 or 13.
15. A computer program product downloadable from a communication network and/or stored on a computer-readable and/or processor-executable medium, said computer program product comprising program code instructions to cause a computer to carry out said computer-implemented steps of any of the claims 1 - 1 1.
PCT/NL2018/050770 2017-11-16 2018-11-16 A computer-controlled method of and apparatus and computer program product for supporting visual clearance of physical content WO2019098835A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
NL2019927A NL2019927B1 (en) 2017-11-16 2017-11-16 A computer controlled method of and apparatus and computer program product for supporting visual clearance of physical content.
NL2019927 2017-11-16

Publications (1)

Publication Number Publication Date
WO2019098835A1 true WO2019098835A1 (en) 2019-05-23

Family

ID=60766110

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/NL2018/050770 WO2019098835A1 (en) 2017-11-16 2018-11-16 A computer-controlled method of and apparatus and computer program product for supporting visual clearance of physical content

Country Status (2)

Country Link
NL (2) NL2019927B1 (en)
WO (1) WO2019098835A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220207453A1 (en) * 2020-12-28 2022-06-30 Yokogawa Electric Corporation Apparatus, method, and recording medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050105768A1 (en) * 2001-09-19 2005-05-19 Guang-Zhong Yang Manipulation of image data
WO2013150494A2 (en) * 2012-04-05 2013-10-10 Sr Labs S.R.L. Method and system for improving the visual exploration of an image during a target search
US8696121B1 (en) * 2006-01-30 2014-04-15 Evan K. Fram Viewing verification systems

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050105768A1 (en) * 2001-09-19 2005-05-19 Guang-Zhong Yang Manipulation of image data
US8696121B1 (en) * 2006-01-30 2014-04-15 Evan K. Fram Viewing verification systems
WO2013150494A2 (en) * 2012-04-05 2013-10-10 Sr Labs S.R.L. Method and system for improving the visual exploration of an image during a target search

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220207453A1 (en) * 2020-12-28 2022-06-30 Yokogawa Electric Corporation Apparatus, method, and recording medium

Also Published As

Publication number Publication date
NL2022016B1 (en) 2021-05-31
NL2019927B1 (en) 2019-05-22
NL2022016A (en) 2019-05-20

Similar Documents

Publication Publication Date Title
Peterson et al. Looking just below the eyes is optimal across face recognition tasks
US8226574B2 (en) Impaired subject detection system
CN111343927B (en) Cognitive dysfunction diagnostic device and cognitive dysfunction diagnostic program recording medium
Marchi et al. Automated pose estimation captures key aspects of General Movements at eight to 17 weeks from conventional videos
Joubert et al. Rapid visual categorization of natural scene contexts with equalized amplitude spectrum and increasing phase noise
US20220308664A1 (en) System and methods for evaluating images and other subjects
Inoue et al. Automated microaneurysm detection method based on eigenvalue analysis using hessian matrix in retinal fundus images
CN107133584A (en) Implicit intention assessment sorting technique based on eye-tracking
Long et al. A familiar-size Stroop effect in the absence of basic-level recognition
Graves et al. The role of the human operator in image-based airport security technologies
Golan et al. Association and dissociation between detection and discrimination of objects of expertise: Evidence from visual search
CN109620266A (en) The detection method and system of individual anxiety level
CN106572795A (en) Eye condition determination system
NL2022016B1 (en) A computer-controlled method of and apparatus and computer program product for supporting visual clearance of physical content.
US20190294888A1 (en) Method and apparatus for image processing
US9830830B2 (en) Stimulus recognition training and detection methods
Botzer et al. Effects of cues on target search behavior.
Hosseinkhani et al. Significance of Bottom-up Attributes in Video Saliency Detection Without Cognitive Bias
Frame et al. Impact of video content and resolution on the cognitive dynamics of surveillance decision‐making
Cazzato et al. Non-intrusive and calibration free visual exploration analysis in children with autism spectrum disorder
Vaughan Soldier-in-the-loop target acquisition performance prediction through 2001: integration of perceptual and cognitive models
Hosseinkhani et al. Investigating into Saliency Priority of Bottom-up Attributes in 2D Videos Without Cognitive Bias
EP3757950B1 (en) Method and system for selecting parameters of a design or security element of banknotes based on neuroanalysis
Hale et al. Using physiological measures to discriminate signal detection outcome during imagery analysis
Mardell Assisting search and rescue through visual attention

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18839893

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18839893

Country of ref document: EP

Kind code of ref document: A1