EP2347387A1 - Systèmes et procédés pour une analyse de scène depuis de multiples perspectives - Google Patents

Systèmes et procédés pour une analyse de scène depuis de multiples perspectives

Info

Publication number
EP2347387A1
EP2347387A1 EP09740784A EP09740784A EP2347387A1 EP 2347387 A1 EP2347387 A1 EP 2347387A1 EP 09740784 A EP09740784 A EP 09740784A EP 09740784 A EP09740784 A EP 09740784A EP 2347387 A1 EP2347387 A1 EP 2347387A1
Authority
EP
European Patent Office
Prior art keywords
scene
objects
images
visual
visual attention
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP09740784A
Other languages
German (de)
English (en)
Inventor
Brian J. Stankiewicz
Brian E. Brooks
Brian L. Linzie
Nathan J. Anderson
Michael Kelly Canavan
Glenn E. Casner
Timothy J. Gardner
David K. Misemer
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
3M Innovative Properties Co
Original Assignee
3M Innovative Properties Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 3M Innovative Properties Co filed Critical 3M Innovative Properties Co
Publication of EP2347387A1 publication Critical patent/EP2347387A1/fr
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/32Determination of transform parameters for the alignment of images, i.e. image registration using correlation-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/37Determination of transform parameters for the alignment of images, i.e. image registration using transform domain methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/97Determining parameters from multiple pictures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Definitions

  • the human visual system is a capacity limited system in that it can only process a relatively small number of objects at any given time. This is true, despite the fact that there are many objects that may be visible at any given time. From the array of objects visible to a human, that human's visual system will only attend to, or processes, one (or very few) objects at any given time. When a human looks at an image or a scene, his visual system will shift attention (and mental processing) from one object to another.
  • Systems and methods for using visual attention modeling techniques to evaluate a scene from multiple perspectives are taken from a scene (or otherwise generated). Objects within the scene are identified in each of the images, then the scenes are analyzed using a visual attention model. The objects predicted by the visual attention model are tracked, such that predictions can be made about the saliency of objects from a plurality of vantage points in an area.
  • multi-perspective scene analysis may be combined with the techniques and systems related to optimizing a scene. For example, permissible changes to objects may be defined, then multiple images, taken from different vantage points, may be taken or generated, then analyzed. These images may include changes made to the objects.
  • robustness may also be used as a factor in multi-perspective scene analysis. For example, a 3D scene may be analyzed and said to be robust (or a robustness value calculated), or an object within a 3D scene may be determined to be robust.
  • a computer-implemented method comprising: receiving a plurality of images taken from within a 3D scene; receiving input identifying at least some of the objects appearing in the plurality of images; analyzing the plurality of images with a visual attention model; and, based on the analysis, determining identified objects that will tend to receive visual attention in the plurality of images.
  • a number of images could be generated from different perspectives of a common set of objects.
  • a computer-implemented method comprising: defining at least one visual goal for a 3D scene, the scene comprised of a plurality of objects, the visual goal specifying at least one object in the scene which is desired to be predicted to be attended to by a visual attention model, and additionally specifying the perspective from which the at least one object should be attended from; receiving a plurality of images from a plurality of perspectives within the 3D scene; receiving input defining allowable changes to at least some objects within the scene; assigning a cost estimate to at least some of the allowable changes; automatically modifying some of the plurality of objects consistent with the defined allowable changes so as to produce modified images; evaluating the modified images with a visual attention model; and, determining, based on the evaluation, at least some of the modified scenes that achieve the at least one visual goal for the scene.
  • a computer system comprising: a processor and memory; and, a multi-perspective scene analysis module, operative to at least:(l) receive a plurality of images taken from within a 3D scene; (2) identify at least some objects appearing in the images; (3) apply a visual attention module the images, the visual attention module operative to predict objects within the image that will tend to receive visual attention; and, (4) determine which of the identified objects are predicted by the visual attention module.
  • FIG. 1 is a flowchart illustrating high-level functioning of a representative visual attention module.
  • FIG. 2 is a sketch of a representative scene.
  • FIG. 3 A is an artist's rendering of output from a visual attention module.
  • FIG. 3B is an artist's rendering of output from a visual attention module.
  • FIG. 4 is a graph showing properties of two example objects.
  • FIG. 5 shows how a billboard object can be set against different background scenes.
  • FIG. 6 is a diagram of functional modules in a system for evaluating the robustness of a scene, or objects within a scene.
  • FIG. 7 is an illustration showing example scene modifications.
  • FIG. 8 shows an alternative embodiment of the visual attention robustness assessment system, in which the visual attention robustness assessment system additionally includes web server module.
  • FIG. 9 is a high-level flowchart illustrating one embodiment of the process the visual attention robustness assessment system may employ to evaluate the robustness of an object within a scene or a scene itself.
  • FIG. 10 is a flowchart showing the high level process of scene optimization.
  • FIG. 11 is a graph illustrating two different reward structures as a function of the serial position of the attentional saccade to the object of interest.
  • FIG. 12 is a diagram of functional modules in a system for scene optimization.
  • FIG. 13A and 13B are an artist's rendering of two respective scenes.
  • FIG. 14A and 14B are an artist's rendering of two respective scenes.
  • FIG. 15 is a high level flowchart showing multi-perspective scene analysis.
  • FIG. 16 is a diagram of a conference room.
  • FIG. 17A and 17B show different perspectives of the conference room.
  • FIG. 18 is a graph illustrating three different example analyses done to various objects in the conference room.
  • FIG. 19 is a diagram of a multi-perspective scene analysis system.
  • Visual attention models determine the extent to which regions within a scene differ with respect to the likelihood that they will attract visual attention or attract eye movements.
  • a scene is any form of input (for example, graphical image) that is amenable to evaluation by a visual attention model, and may, for instance, be a digital photograph, a virtual 3D scene, a web page, a document, or a video.
  • a visual attention model's disposition to an object refers to how a model of visual attention characterizes the relative saliency of an object within a scene. For example, some visual attention models will superimpose a trace line around predicted objects. Other visual attention models will generate heat maps that may be superimposed over the image or viewed separately from the image.
  • the disposition of an object may be considered as "selected” (when traced) or "not selected” by the model.
  • the disposition of an object is the degree to which the algorithm has selected the object (or not selected the object).
  • the disposition of the object may be the saliency number itself.
  • FIG. 1 shows how a prediction is made for visual attention by assessment of "bottom-up" features such as color, motion, luminance, edges, etc. which serve as building blocks of the visual representations mediating some aspects of human vision.
  • a scene in the form of a digital photograph, is provided to a computer-implemented version of the Itti and Koch model (step 10).
  • a feature extraction process analyzes the digital photograph for colors, intensity, orientations, or other scene cues, such as motion, junctions, terminators, stereo disparity, and shape from shading (step 11).
  • the feature extraction process yields a plurality of feature maps (step 12), which are combined to produce a saliency map (step 13).
  • the saliency data is provided to a user as a rendering of the original digital photograph with the "brightest" objects being to which the model has predicted visual attention will be next allocated.
  • This predicted object is identified to be visually salient (step 14) in a "winner-take-all" type algorithm (step 15), and the process then repeats until a plurality of objects are identified by the model.
  • FIG. 2 is an artist's rendering of a scene 201 that could be provided to a visual attention model such as Itti and Koch. It is a simplified scene included here for illustrative purpose only; in practice the scenes are often actual digital photographs, or videos, and are much more complex.
  • FIG. 2 includes a number of objects within the scene, such as the star 202, flower 203, face 204, star 205, arrow 206, and cup 207.
  • FIG. 3A is an artist's rendering representing how output from the Itti and Koch model could be represented.
  • the highlighted (and in this representative illustration, encircled) objects are those that the model predicts to be visually salient.
  • star 202 is in this figure within highlight border 208; flower 203 is within border 209; face 204 is within border 210; star 205 is within border 211; arrow 206 is within border 212; and cup 207 is within border 213.
  • the model in this instance has determined six objects that are, relative to other objects, more visually salient.
  • This particular model also predicts how attention will move among the objects determined to be above some visual saliency threshold.
  • visual attention pathway 301, 302, 303, 304, and 305 show a predicted visual attention pathway.
  • FIG. 3B is a second artist's rendering showing a further manner in which output from the Itti and Koch model is sometimes represented.
  • FIG. 3B includes the sequence of predicted visual attention.
  • star 202 is labeled "1" (attention sequence number 214)
  • flower 203 is labeled "2" (attention sequence number 215) and so forth.
  • FIG. 3 A and FIG. 3B are just one manner in which a visual attention prediction may be conveyed to a user; different models represent this information (or some subset of it) in different ways. For example, not every model determines a predicted attention sequence, though such an attention sequence could be arrived at by determining the object with the highest level of visual saliency, then eliminating that object and finding the next highest, etc.
  • Itti and Koch 's model is representative of a "bottom-up" visual attention model, in that it makes its predictions based on analysis of the particulars of the scene.
  • Other bottom-up visual salience models are described in these references: Gao, Mahadevan, and Vesconcelos (2008).
  • top down models of visual attention there is another class of models referred to as "top down" models of visual attention.
  • these models start with a scene and either an explicit task (for example, avoiding obstacles and collecting objects) or prior knowledge of the world that will influence where attention will be allocated during a specific search task (for example, chairs tend to be on the floor and not on the ceiling). This knowledge (both task- and scene-based) is used in conjunction with the bottom-up features to direct attention to objects within the observed scene.
  • Some exemplary top-down models are described in Rothkopf, C.A., Ballard, D.H. & Hayhoe, M. M.
  • Torralba A., Contextual Modulation of Target Saliency, Adv. in Neural Information Processing Systems 14 (NIPS), (2001) MIT Press, 2001.
  • Torralba' s model of visual attention has prior knowledge about the features that comprise a particular type of object and information about the absolute and relative locations of these objects within the scene. This prior knowledge provides "top-down" influences on searching for specific targets within a scene.
  • Robustness refers to the sensitivity of predictions output from a visual attention model to either of, or some combination of:
  • object refers to either a thing or area or region within a scene, as the case may be, or the regions within a scene that are analyzed by a visual attention model.
  • object is used interchangeably with "area” or "region” as the case may be.
  • Robustness evaluation is a method for measuring the effect of IE changes on where attention will be allocated within a scene.
  • Robustness of an object within a scene is the degree to which a visual attention model's predicted disposition of a particular object changes, or does not change, despite IE changes.
  • the robustness of a group of objects is the degree to which a visual attention model's disposition to more than one object (a collection of objects) changes or does not change with IE changes.
  • Robustness of a scene is a measure of the degree to which the visual attention model's disposition of objects in the scene will change, or not change, despite IE changes. For example, if a visual attention model predicts that attention or fixations will be made to the same four objects within a scene, and these four objects tend to remain predicted by the visual attention model despite IE changes, the scene tends to be more robust than a if the four objects changed following IE changes.
  • FIG. 6 is a diagram showing exemplary functional modules in visual attention robustness assessment system 403, which is a system for evaluating the robustness of a scene, or the robustness of objects within a scene, or the robustness of objects across multiple scenes. Of course, these methods may also be practiced manually.
  • Visual attention robustness assessment system is, in the embodiment shown in FIG. 6, within computer system 408.
  • Computer system 408 may be any general purpose or application-specific computer or device. It may be a stand-alone laptop computer, or a plurality of networked computers. Further, computer system 408 may be a handheld computer, digital camera, or a tablet PC, or even a cellular telephone.
  • Computer system 408, in one embodiment, has various functional modules (not shown in FIG. 6) that comprise an operating system. Such an operating system facilitates the visual attention robustness assessment system's access to the computer system's resources.
  • Computer system 408 has a processor and memory, and various traditional input/output interfaces.
  • Visual attention module 403 is any embodiment of any visual attention model or combination of models. As mentioned earlier, there are different types of visual attention models, but they all, to some extent, predict an object or area within a scene to which visual attention will tend to be allocated. Visual attention module 403 is shown in FIG. 6 as part of visual attention robustness assessment system 402, but visual attention module 403 in another embodiment operates as a stand-alone computer process or even as a service provided over any type of computer network (such as the World Wide Web) at a remote computer.
  • any type of computer network such as the World Wide Web
  • VAM modification module 404 is the visual attention module modification module.
  • VAM modification module 404 modifies aspects of the visual attention module's parameters or architecture. This modification can by accomplished in many ways depending on the implementation of the visual attention module 403.
  • visual attention module 403 may itself support function calls that modify aspects of how the visual attention module works.
  • visual attention module 403 may support a function call that modifies weighting given to certain aspects of the scene (brightness, for example).
  • various switches could be employed to change variables that are within the visual attention module.
  • the visual attention module 403 is embodied in a script or programming code, the IE modification module could modify the script or programming code itself.
  • the entire visual attention model is replaced by another visual attention model.
  • the VAM modification module 404 modifies the visual attention module 403 's underlying visual attention model (or the application of such a model to a scene) is discussed further below, but one example might be to modify the weights associated with the plurality of feature maps that are used to generate the saliency map. This could be accomplished through a function call that would modify these values based upon a distribution of acceptable values.
  • Scene modification module 405 modifies the scene that will be provided to the visual attention module 403 for analysis.
  • the scene changes are the changes that are associated with the external environment that simulate the dynamic nature of the environment.
  • Such external variability may include, for example, moving objects within the scene (for example, the placement of pedestrians), changes in shadows due to changes in lighting direction, or changes in atmospheric conditions (for example, dust in the air).
  • This variability can be generated in a number of ways. One method is to place a camera within the scene and capture an image at different times. This would capture the natural variability of a real scene. Another method is to capture a single image of the scene and specify possible changes that could be made to the individual elements and objects within that scene. An illustration of such a technique is shown in FIG. 7.
  • Scene 801 is a representation of an original scene.
  • Scene 802 shows the objects from scene 801 replaced with scene variability elements, such as shadows and other objects (for example, cars or birds - any object that can be placed within the scene).
  • Scene 803, 804, 805, and 806 show different combinations of the original scene 801 with scene variability elements.
  • scene variability elements such as shadows and other objects (for example, cars or birds - any object that can be placed within the scene).
  • scene 803, 804, 805, and 806 show different combinations of the original scene 801 with scene variability elements.
  • Scene modification module 405 adds external variability to a scene, which will in turn result in multiple versions of the scene, or scene instances.
  • the collection of scene instances captures the variability within the scene.
  • each scene instance is submitted to visual attention module 403 to produce a prediction as to where visual attention will be allocated within each scene instance (that is, to which objects).
  • Information pertaining to each run is saved in database 407, then robustness assessment module 409 evaluates the data to generate statistics based upon objects to which the model predicted attention would be allocated, across these different instances (and as was described in further detail earlier).
  • Graphical user interface module 406 facilitates interaction with user 401.
  • Graphical user interface module 406 may, for example, call on operating system resources (from computer system 408) to build a graphical user interface to solicit input from user 401.
  • This input in one embodiment includes the location of the scene, as well as other operating parameters for the visual attention robustness assessment system.
  • This input in one embodiment, would specify regions and/or locations within the scene that user 401 is interested in evaluating.
  • user 401 could specify what types of variability is to be considered by the visual attention module 403. This may include general or specific internal variability, or general or specific external variability. For example, one specific type of external variability might be changing the image due to variations in lighting direction. As light changes, the shadows produced will change.
  • Graphical user interface module 406 also facilitates getting input, possibly from the user, regarding from where images of the scene should be acquired. Possible locations include, for example, a database, or a flat file.
  • Robustness assessment module 409 controls the interaction of the other modules in order to evaluate the robustness of objects within a scene or the robustness of the scene itself.
  • the robustness assessment module 409 invokes the visual attention module 403 as necessary, as well as the VAM modification module 404 and the scene modification module 405.
  • the manner in which the robustness assessment module 409 invokes the various modules may be modified by input from user 401, provided to the robustness assessment module 409 via, for example, graphical user interface module 406.
  • Robustness assessment module 409 also evaluates data provided by other modules and generates reports as necessary.
  • Database 407 handles the data storage needs of the visual attention robustness assessment system 402. Among other things, database 407 may hold images of the scene.
  • Database 407 may be any computer memory. It may be random access memory, or a flat file, or one or more database management systems (DBMS) executing on one or more database servers.
  • the database management systems may be a relational (RDBMS), hierarchical (HDBMS), multidimensional (MDBMS), object oriented (ODBMS or OODBMS) or object relational (ORDBMS) database management system.
  • Database 407 for example, could be a single relational database such as SQL Server from Microsoft Corporation.
  • User 401 is any user of the visual attention robustness assessment system.
  • the visual attention robustness assessment system 402 is easy enough to use that a person unfamiliar with visual saliency theory could use the system to evaluate the robustness of an object, region, or scene.
  • User 401 may be a consultant hired by, or an employee who works for a commercial entity to evaluate positioning of signs and non-sign objects within their environment.
  • User 401 may also be a content designer interested in evaluating the design and placement of a digital object (such as advertisement) on a web page in which the visual properties of the page may change.
  • FIG. 8 shows an alternative embodiment of the visual attention robustness assessment system, in which the visual attention robustness assessment system additionally includes web server module 501.
  • Web server module 501 is shown, for convenience, as part of visual attention robustness assessment system 402.
  • Web server module may, however, be embodied as a software module running on computer system 408, in separate memory space. Or web server module 501 may be on a separate computer system coupled, via network, to the visual attention robustness assessment system 402.
  • Web server module 501 provides an interface by which user 401 may communicate with visual attention robustness assessment system 402 via client computer 503 and via network 502.
  • web module 501 executes web server software, such as Internet Information Server from Microsoft Corporation, of Redmond, Washington.
  • Web server module 501 provides a mechanism for interacting with remote user 401 through the use of, for example, Active Server Pages, web pages written in hypertext markup language (HTML) or dynamic HTML, Active X modules, Lotus scripts, Java scripts, Java Applets, Distributed Component Object Modules (DCOM) and the like.
  • Active Server Pages web pages written in hypertext markup language (HTML) or dynamic HTML, Active X modules, Lotus scripts, Java scripts, Java Applets, Distributed Component Object Modules (DCOM) and the like.
  • HTML hypertext markup language
  • DCOM Distributed Component Object Modules
  • server side software modules executing within an operating environment provided by computer system 408, the functional modules that comprise the visual attention robustness assessment system 402 could readily be implemented as
  • client-side software modules executing on computing devices, such as client computer 503, as used by user 401.
  • Visual attention robustness assessment system 402 could, for example, be implemented as Active X modules executed by a web browser executing on client computer 503.
  • Network 502 may be any kind of network, public or private. In one embodiment network 502 is the Internet.
  • FIG. 9 is a high-level flowchart illustrating one embodiment of the process the visual attention robustness assessment system 402 may employ to evaluate the robustness of an object within a scene or a scene itself.
  • the visual attention module is invoked and provided scene input (step 601).
  • Predicted objects from within the scene are then received from the visual attention module (step 602).
  • some indication of the locations/objects/regions that the model predicts to have high relative saliency is saved to a database (step 603).
  • the database in one embodiment, stores information about the objects that reached a threshold saliency value. In another embodiment, the database stores a matrix of values, one for each object identified in the image, and stores, for example, the number of times that the object reached a threshold saliency value.
  • a check is made to see if the process is complete (step 605). In one embodiment, this check might determine if the current iteration exceeds the number of iterations initially set by user 401. In another embodiment, the number of iterations may be determined by the algorithm. In yet another embodiment, the number of iterations may be determined by properties of the scene or statistics concerning the robustness measure. For example, if one were attempting to determine which of two advertisements would be better at a particular site, one could run the algorithm until there was a statistically reliable effect between the two advertisement's robustness values.
  • IE changes will be introduced into either the visual attention module's visual attention model, or the scene input.
  • IE changes may be divided into two categories: structured and random. Random variability is variability that is uncorrelated. For example, random variability in a scene might include random variations in the individual pixel colors and/or intensities. In such case, the pixel variations are uncorrelated.
  • structured variability has correlation between the elements being modified. For example, a scene modified by simulating the movement of an object within the scene, or adding or removing an object within a scene would constitute structured scene variability. In such case, the changes in the pixel modifications are correlated. Random internal variations might include random variations in the attention model used by the visual attention module. Structured variability, on the other hand, might be programmatic biases for attention in one part of the attention map over another. The types methods of generating variability are summarized in Table 1. The sub-processes that introduce IE changes are further detailed below.
  • step 605 analysis is done on how, and the degree to which, the objects predicted by the visual attention module changed in successive iterations. This analysis is further described below.
  • output indicative of robustness is generated (step 607).
  • this output is provided to a user via graphical user interface module 406.
  • the output may also be provided in other forms to, for example, other programs or calling functions.
  • One example of modifying the visual attention model is to programmatically introduce a bias toward a particular feature within the scene, or a particular area within the scene.
  • Such variability to the visual attention model will have an effect on where attention will be allocated and will simulate variability between observers and/or variability between human observers experiencing a scene from one time to another.
  • a saliency "map" is generated by combining the input from the different visual feature maps. For example, in some instantiations there are three feature maps: a first that is sensitive to color, another to orientation and a third associated with luminance.
  • the input to the saliency map is a weighted combination of these three maps.
  • these three maps have equal weights into the saliency map indicating that there is no bias for one type of feature over another and can be represented as a weighting vector (for example, [1 1 1] for an equal weight of the three feature maps).
  • One method for generating variability in the observer's biases is to modify these weights simulating a viewer that might be biased toward one feature over another. For example, one could bias the model toward brighter items in the environment by setting the weights to [0.5 0.5 2].
  • One method of generating internal variability (or simulating viewer variability) is to generate a collection of models that have different weighting values. This could be done by randomly setting the values for each model.
  • Another approach is to randomly modify the model's internal representation of the scene.
  • the neurons in the human visual system are somewhat noisy in that their activation will vary even with the presentation of the same image. Simulating this variability can be done by perturbing the internal representation of the values associated with the individual feature maps (for example, color, shape, orientation, brightness) a visual attention model develops in response to a scene.
  • Another method is to perturb the values in the saliency map that is used to identify which regions of the image are most salient. For example, one could perturb the internal representation following each fixation calculation, or perturb the values independently each time a scene is presented.
  • the scene is changed by creating a new image using a graphical editor (e.g., Photoshop), 3D representation of a scene (for example, virtual reality model), or video (again, all collectively called "scene” herein), based on the previously existing scene in which the properties of the scene are varied.
  • a graphical editor e.g., Photoshop
  • 3D representation of a scene for example, virtual reality model
  • video video
  • varied properties could include simulated (or real) changes in lighting, or newly added or removed objects (real or virtual), or varying the pixel colors randomly.
  • One method for quantifying the degree of external variability is to vary the number of objects manipulated in the scene. For example, returning to the hotel lobby example with pedestrians, one may vary the number of pedestrians in the scene. Few pedestrians would translate to a small amount of variability while a large number of pedestrians would translate to large variability.
  • Another method for measuring variability is to measure the pixel variation for all of the images that are generated.
  • One example measure of pixel variability is to measure the average variance of each pixel for the set of images that are generated. Once the variability of each pixel is calculated, a single number could be produced by computing the mean variance of all of the pixels. Again, as the variability of the within the image increases, this value will also increase.
  • One method for generating changes is to make changes to an image by digitally modifying the image using commercially available image editing software such as that marketed by Adobe Systems Inc. of San Jose under the trade name "Photoshop.” Using this method one can modify the scene by digitally placing objects, removing objects, or simulating lighting changes. These changes would then be translated into a collection of images that would be read and analyzed by the model. These changes can also be applied to images automatically by superimposing layers onto images algorithmically.
  • Another method for generating external structural variability is to generate virtual representations of the scene. Using this approach one can easily modify the locations of objects, object properties, and lighting conditions. From these virtual representations, images can be generated in which the objects, their properties, their poses, and lighting conditions have been changed.
  • approaches to changing the scene could be said to be “synthetic" - that is, they are changes to representations of the scene, rather than to the scene itself.
  • Another approach to modifying the scene is to modify the actual scene. For example, one could take a series of time-delayed images (for example, a picture every 300 seconds for a defined period). This series of pictures could then be used for successive iterations of the visual attention robustness assessment system. Such analysis (multi-image across time) would provide a measure of the external structure variability. Furthermore, given such variability (people and objects moving, along with changes in lighting direction due to changes in, for example, the sun's position) one could analyze where attention will typically be allocated within the environment.
  • Another method for evaluating robustness is to measure the variability in the model's internal representation of the scene.
  • the Itti & Koch model uses an internal "saliency map" representation to determine the locations that visual attention will be allocated.
  • Another method for generating a robustness metric is when the IE changes can be parameterized. This would be the case, for example, where one adds random luminance noise to the scene (thus changing the amount of noise by modifying the luminance energy of the noise relative to the luminance energy in the image). Or one could increase the variability of the connection weights of the visual attention model and characterize the degree of change as a measure of entropy.
  • Another method for characterizing the robustness of an object, group of objects, or scene is to determine the amount of noise that these objects can tolerate while retaining their visual saliency hierarchy. The method for this metric is to measure the likelihood that an object will be attended as a function of the amount of variability generated.
  • Those objects and/or scenes that are highly robust will tolerate large amounts of noise before attention begins to be attracted to the noise elements and not to the objects/regions.
  • the resulting metric would be the level of noise that the object could tolerate before attention was drawn away from the object and toward the noise.
  • FIG. 4 is a graph showing the probability of attending to an object versus the degree of variability for two objects, the first object 451 having a low noise tolerance (low robustness value) and a second object 452 having high noise tolerance (high robustness value).
  • the graph shows the effect of variability on whether attention will be allocated to object 451 and object 452, each with different respective robustness indexes.
  • adding a small amount of variability significantly affects the probability that the model will predict that attention will be allocated to Object 451.
  • FIG. 4 by the drop in probability with only a small amount of variability.
  • Object 452 is unaffected by small amounts of variability. This is illustrated by a "Drop-off point occurring with a much greater level of variability.
  • Such quantification may allow for subsequent analysis such as determining the degree (or the robustness index value) of IE changes that begin to reduce the degree of saliency of a particular object within a scene.
  • Table 2 provides exemplary data in which a target object has a high and low robustness value crossed with scenes that have low and high robustness values.
  • a normalized likelihood value called a Robustness Value below
  • the left column indicates the percentage of time that the object appeared in the first three saccades when there are IE changes.
  • One method for calculating a robustness value for the scene is to calculate the average robustness value for the top K objects (number of attention selections). At the bottom Table 2 is the average robustness value for the top 3 objects.
  • Table 2 categorizes the locations where attention is allocated, by objects. But as mentioned before, the term object is loosely defined as a region or area in an image (or scene). The methods and systems described herein are not limited, however, to an object- based approach - other, similar approaches would also work. For example, a grid could of equal size regions could be defined over the scene, or regions of the scene defined based upon properties of the human visual system (for example, the size of the fovea of the viewer).
  • cross scene robustness is a measure of how the robustness of an object (or a group of objects) changes across different scenes.
  • a robustness metric for an object as discussed above, separate from the scene, one can evaluate the robustness of a particular object in different scenes. For instance, consider the situation in which an advertiser is deciding between two different pieces of advertisement content that will be placed on three different billboard signs that are in three different scenes.
  • FIG. 5 provides an illustration of an example of this scenario.
  • Scene 703 includes billboard 700.
  • Billboard 700 repeats in scene 704 and 705.
  • billboard 706 is seen in scenes 707, 708, and 709.
  • a single object for example, advertisement billboard
  • a set of scenes from a collection of scenes
  • objects to specific scenes based upon a robustness value. For example, a customer might have three different advertisements (the content that is placed on a billboard) that they have developed and twenty different billboards that these signs could be placed upon. By generating robustness values for each of the three advertisements placed in the twenty scenes, the customer could then choose to use the advertisement that has the highest robustness value for each individual scene. Additionally, one could assign the best advertisements to the ten scenes with the highest robustness value.
  • a robustness metric could be used to improve choices made with regard to billboards and advertisements
  • the benefits of a robustness metric are not limited to this domain, and could be used in any domain in which one has a collection of contexts (scenes) and a collection of visual stimuli (objects) to select between, such as digital signs within a store or hotel, static signs, product signs, product packaging configuration, or web sites.
  • sequence target robustness is to compute the likelihood that the target order (that is, Object-A before Object-B) would occur if one were to randomly select the objects.
  • sequence target robustness value would be the probability that the target sequence was achieved divided by the likelihood that it occurred by chance. The principles behind this metric are similar to the principles behind the robustness value described above.
  • the objective is to have attention drawn to Object-A before Object-B, as would be the case for example where a company is advertising a dinner special for a restaurant on a billboard near the restaurant itself.
  • the company may be considering two discreet advertisements the dinner special.
  • the goal is to get people who are passing by to first attend to the dinner special advertisement followed by attending to the restaurant itself.
  • IE changes applying IE changes, one can calculate how often the billboard receives visual attention before the restaurant.
  • Tables 3 and 4 provide possible likelihoods showing the relative ordering of Object-A and Object-B along with whether they received attention at all (Object-# Not Att.). From these tables we see that Object-A before Object-B occurred 65% of the time with Advertisement Content- 1 but only 40% with advertisement Content-2. Thus, if this sequence was important the customer might be inclined to choose Advertisement Content- 1. Advertisement Content- 1
  • This aspect of visibility duration can play an important role when one considers the situation where there are multiple objects within a scene. Under some conditions one may want to distribute the attention capturing elements such that the objects are salient enough to be processed (or capture attention) but not so salient that they distract from other objects within the scene. (Note: this concept also relates to scene optimization, which concerns modifying a scene in a way that achieves particular goals. Scene optimization is discussed further below.) Given this, one would want to consider the visibility duration when analyzing the robustness of an object, a collection of objects or a scene.
  • an object that has a short visibility duration on a path one may want to increase the saliency elements (for example, motion, brightness, color contrast, etc.) to increase the likelihood that a particular object will be processed during that short visibility duration.
  • the saliency elements for example, motion, brightness, color contrast, etc.
  • another object that is visible along that same path, but has a longer visibility duration one may be able to take advantage of the fact that there are more opportunities to process that object and thus one can allocate fewer (or weaker) saliency elements to that object.
  • the visibility duration would be factored into the robustness analysis by considering whether the object received visual attention anywhere along the path. Thus particular objects that have longer visibility durations than others would have more attentional fixations associated with them, and if attention was allocated to the object anywhere along that sequence then it would be considered a "hit.” Therefore, objects with longer visibility durations would have more samples and would have a higher likelihood of being processed and thus may require a lower level of saliency to be processed. An object with shorter visibility duration would have fewer samples and thus would be less likely to be attended and thus, may require higher level of saliency elements to be detected during that shorter sequence.
  • visual attention modeling related concepts that, in some embodiments, may benefit from the robustness-related methods and systems described earlier, but do not necessarily require assessment of robustness.
  • One such related concept is that of scene optimization.
  • visual attention models exist that may predict where a subject will allocate his or her visual attention within a scene.
  • such models do not provide any mechanism for identifying how a scene can be modified to achieve a specific visual goal. Because a human's visual system does not actually process the entire viewable area of a scene, but instead only processes those regions in which attention is drawn, it is desirable in many real-world situations not only to get people to 'view' specific objects within a scene, but to have them 'attend' to specific objects.
  • a visual goal refers to the desired manner in which a subject will attend to objects within a scene.
  • a visual goal could be simply a desire that particular objects are attended to (that is, the collection of objects within a scene that one determines as important (from a visual attention perspective) in addition to the collection of objects that are deemed unimportant or even detrimental.
  • it could be a desire to have particular objects attended to in a particular sequence or at a particular time, or it could be a desire to have particular objects attended to from a particular viewing point, but not necessarily from others.
  • the remainder of this section discusses ways in which one could utilize a computational model of visual attention to optimize a scene in order to achieve a visual goal.
  • scene optimization starts by assigning an explicit cost/reward structure on the objects within the scene (step 1205).
  • the cost/reward structure defines which objects within the scene are high value objects, low value objects and even objects that the user views as distracting or detrimental to the visual goal.
  • the user will place "attention utility" values on to each of the objects that are deemed to be part of the visual goal (positive rewards) or are detrimental (negative costs). Or, the user can place priorities specifying which elements are "more" valuable to the visual goal than others.
  • step 1210 a number of possible changes to the scene are defined (step 1210). These could be simple changes such as lighting, color of objects, positioning of objects, etc. or more complex design changes such as where within a lobby a reception area should be built. Of course, ideally, the time to evaluate something as fundamental as positioning of the reception area is best done before building a hotel lobby, so it is expected one utility of the scene optimization methods described herein will be for evaluating synthetic, or partially synthetic, scenes and design / layout choices within such scenes.
  • the actual real-world "cost" (or estimate) is then associated with each possible scene change (step 1215). For example, in the case of a hotel lobby, where the goal is for patrons to attend to a particular sign, changing the color of the sign may be relatively inexpensive (could be assigned an estimated change value of $200), while changing the color of the granite floor would be expensive (could be assigned an estimated change value of $20,000). Assigning real-world cost estimates makes it possible to associate a price figure with a plurality of changes.
  • the method provides the scene configuration that maximizes the attentional utility while minimizing the cost associated with the object feature configurations (step 1220).
  • real-world costs are just an example of how relative costs of particular changes could be co-associated - other systems, such as point systems with higher points correlating with higher costs for particular changes, and lower points being less expensive changes, could just as easily be used.
  • a graphical editor with the ability to track and manipulate discrete objects may be useful in defining possible changes to a scene. For example, a sign in a scene that is being viewed within such a graphical editor could be identified (right-clicked with a mouse, for example), and the user would be able to select the changeable properties of that object. These changeable properties might include color, lighting, positioning within layers (for example, the object could be put in front of or behind some other object), contrast, shadow, size, etc.
  • the user may also have the ability to define the allowed scope of change or other relevant parameters. For example, with respect to color, the only colors a client or user may find acceptable for a particular wall that is within a scene may be shades of tan. Thus the color attribute is defined to only be varied within the specified spectral range. Similarly, if the attribute is size, there may be an obvious limit to the size a particular object may grow to (or be reduced to) before the size of the object is not acceptable to the client or user. As mentioned earlier, the real-world cost estimate is associated with each possible change.
  • a user may define the cost of exemplary points within the range and the supporting computer system (later described) will extrapolate a best-fit curve for the example points.
  • the user may be presented with a number of possible best-fit algorithms and thus be asked to select which one to use. For example, the user may just want to define, with respect to the size of a sign, that the smallest allowable sign costs $100, and the cost of the intermediate signs increase linearly (based on size) between these two cost points.
  • the next step is to evaluate the benefit of possible changes to the scene and attempt to find the scene configurations that achieve the visual goal defined in step 1201. This is done, in one embodiment, by algorithmically modifying properties of the scene to maximize the expected reward, or the scene configuration that minimizes costs while maximizing rewards, or the scene configuration that simply meets the defined goals for the least cost.
  • the expected reward is calculated as the likelihood that a particular object will receive attention and the reward/cost for attending to those objects.
  • the optimization routine In addition to the reward/cost for attending to the location in the scene, the optimization routine also factors in the cost/reward for making specific types of changes to the scene (for example, changing the color and/or position of an object within the scene).
  • One method for calculating the likelihood of an object receiving attention may be determined using the robustness analysis described in the previous section.
  • Equation 1 provides one example of how to formulate a utility function using a computational model of visual attention.
  • F is the set of changeable attributes for the objects within the scene. For example, it would include the color, texture, or position for each of the objects within the scene.
  • A is the set of attentional fixations that the model predicts given this feature configuration F.
  • R(Oxy) is the reward (both positive and negative) for attending to the object at the position (xy) with the feature set/
  • a(xy) is the likelihood that the model predicts that attention will be allocated to location xy.
  • a(xy) can be a binary value (0 or 1 indicating whether attention will or won't be allocated to the location), while for other models and approaches, this might be a likelihood that attention will be allocated to this location (0...1). Either way, this part of the equation specifies the reward for the allocation of attention for the current object feature set.
  • the other part of the function specifies the cost for using feature set/for object o.
  • R(Of) specifies the reward/cost for using feature set/on object o.
  • the cost for a particular feature might be 0.0 (for example, this might be the case for the current feature set for object o).
  • the rewards for all features would be equivalent (or most easily 0.0).
  • the allowable feature configurations in some situations (for example, changing the color of the carpet in a scene versus moving a vase) there will be constraints on the allowable feature configurations. In such situations, one could specify these values in terms of costs/rewards on the object (o) and the features (J), or they could simply refrain from defining the non-allowable feature configurations as an allowable change.
  • Equation 1 the expected reward function above (Equation 1) may be expanded to optimize the feature configuration such that the expected reward is order dependent. This can be done by taking advantage of a variable reward structure as a function of saccade number. Note that in Equation 1 the reward for attending to an object is indexed by the individual attentional saccade (Ra). By specifying a different reward based upon the saccade 's sequential position (a) one can generate a method by which the scene is optimized by the predicted saccade order.
  • Fig. 11 is a graph illustrating two reward structures. One reward structure is represented by invariant with saccade position (line 1225) and the second is dependent upon saccade position (line 1230).
  • the saccade dependent reward structure specifies that the expected reward for attending to this particular object is very high if it occurs early, but it declines as attention is allocated later in the sequence.
  • This type of reward structure might be associated with "high value" objects, such as pedestrians in a construction zone scene.
  • a hotel owner may have two outdoor signs. One advertises a special in their restaurant and the second displaying the name and logo of their hotel. The hotel owner decides that it is important that the advertising special sign should be seen before the hotel's sign. Given this visual goal the analysis would place a very high reward on when the "restaurant special" sign is attended to before the hotel name.
  • the sequence could be a relative sequence (before versus after) or could be an absolute sequence (no reward for attending to objects A and B unless Object-A occurs as the first object that receives visual attention and Object-B is the second object that receives visual attention).
  • the sequence could be a relative sequence (before versus after) or could be an absolute sequence (no reward for attending to objects A and B unless Object-A occurs as the first object that receives visual attention and Object-B is the second object that receives visual attention).
  • the optimization routine may be set up to provide the best configuration for a scene given the reward structure, but one may find that the resulting scene "recommendation" not appropriate. For example, one might wonder how the attentional sequence changes if one were to modify the color of one object to make it more, or less, conspicuous. Making one object more conspicuous can have non-obvious, non-linear effects on the predicted sequence. This is due to the fact that attention allocation is by definition a zero-sum game: allocating attention to one object will necessarily mean that attention is allocated away from another object.
  • Robustness can also be used in scene optimization. For example, one could optimize a scene not only to the properties of that image, but to the scene as it might appear at different times of day, or to different viewers. That is, the optimization routine recommends scenes that are robust to the variability that the scene may experience.
  • one method for capturing the variability of observers is to vary the parameters of the visual attention model (for example, the weights of the individual feature maps for calculating saliency) and running these models on the same image.
  • visibility duration refers to the period of time that an object and/or scene will be viewed. Visibility durations, which may be specified in terms of time will typically be translated into the number of predicted attentional fixations (or salient regions) that will be considered in the optimization routine. The visibility duration would be used to limit the set of fixations that would be used for the analysis of the various scenes. More specifically, it will be used in the visual goal analysis.
  • the visibility duration may have an effect when considering multiple visual goals.
  • one may have a document or content that has six visual goals with different rewards — say the reward values are 1,2,3,4,5,6. If the content is displayed on a digital sign for 5 seconds, and people make about 2 fixations per second, this translates to a visibility duration of 10 fixations. Given that the visibility duration is 10 fixations, the visibility duration is long enough to capture all of the visual goals. Under this condition, the model would make the saliency of the six items relatively equal (assuming no other objects in the scene and an equal cost for making changes). The reason the optimization routine with visibility duration will make the visual goals roughly equal, is that the model is more likely to attend to all of the targets under this condition.
  • Fig. 12 is a high-level diagram of scene optimization system 1255. Like -named components in this figure are similar in functionality and capability to those modules discussed earlier.
  • Scene optimization system 1255 includes database graphical user interface 1245, which receives input from user 401, via computer 503, over network 502, to define one or more visual goals for a scene, as well as possible changes to the scene. These possible changes to objects in the scene are stored in database 1235.
  • Scene optimization module 1260 iterates explores the universe of possible scenes that meet the visual goal, and invokes scene modification module 1240 to modify scene input in ways consistent with the possible changes defined by user 401. This produces a modified scene, which is provided to visual attention module 403, which provides output relevant to visual attention, which is stored in database 1235.
  • Scene optimization system 1255 may include web server module 1250 if user 401 is remote (and as shown in Fig. 12).
  • Fig. 13A is an artist's rendering of a hotel lobby scene including a reception desk having digital sign 1290 behind the desk.
  • Fig. 13A The scene has been analyzed by a visual attention model which predicted the two objects that draw the most attention in Fig. 13A are the digital sign 1290 and advertisement card 1285.
  • Fig. 13B is the same reception area in the same hotel lobby, but digital sign 1290 is not among the several objects identified by the visual attention model. If the visual goal is to have patrons attend to the digital sign 1290 and only the scene shown in Fig. 13A is analyzed, there will be a false sense of security that the visual goal is being consistently met. Thus there is a need to have an effective way to measure and evaluate scene content within a 3D scene.
  • 3D scene evaluation (which herein we refer to as multi-perspective scene analysis) is not the same as merely extending 2D visual attention analysis to such things as movie sequences.
  • Such an approach may provide data from many perspectives, but ultimately has limitations similar to that of the single image approach, in that it may provide one with an ability to analyze the allocation of attention for a particular movement through space, but it may not provide the ability to consider other paths and/or movements through the space.
  • a visual attention analysis of static scenes nor a plurality of static scenes in succession (videos) effectively deals with the fact that the images are derived from a 3D setting.
  • Fig. 15 is a flowchart showing the multi-perspective visual attention modeling process.
  • site planning consists of determining which locations and objects within the 3D scene one wants to analyze. In practice, this may mean acquiring or developing a floor plan of the 3D scene that will be analyzed, and determining locations in the floor plan that will be representative of the observing perspectives of users.
  • site planning may mean acquiring or developing a floor plan of the 3D scene that will be analyzed, and determining locations in the floor plan that will be representative of the observing perspectives of users.
  • a less rigorous approach could be used, in which the user simply goes to the site and makes decisions about which locations are of interest and takes photographs from those locations. Additionally, the user might record the position and orientation of where the image was taken, which could be useful for reporting purposes.
  • video technology in which frames are captured from the video or video sequences could also be used.
  • frames images from video
  • images from digital cameras, or example
  • the user can also use view interpolation techniques to generate views that are between two different images.
  • the images are not limited to being generated from a real environment, but they can also be generated from synthetic (virtual) environments.
  • synthetic environments in both cases, one must record or pre- specify the locations in the environment where the images are taken from, and the perspective of the camera (which way the camera is pointed).
  • a simple approach is to specify locations, then have each successive image be generated from a perspective that advances 90 degrees around the north, east, south, west axis (as is done in an example below).
  • the camera could instead be tracked using GPS tracking technology possibly in combination with optical tracking technology.
  • an instrument or instruments would be attached to the camera such that every time an image is taken, the system would record the three positional dimensions (X,Y,Z) and the three orientation dimensions (yaw, pitch and roll) to provide an explicit representation of the viewpoint from which the image is captured. These six values would be stored in memory and associated with the image captured at that time. Of course, cameral position and orientation could be manually recorded.
  • the multiple images representing a set of views that an observer may experience as the observer interacts and navigates through a 3D environment (step 1330). In one embodiment, this is accomplished by taking multiple photographs taken from multiple positions and orientations within the 3D scene (again, such as the lobby of a hotel). In another embodiment, a video is made of the 3D scene, with shots taken from multiple representative areas an observer might be expected to be at. In yet another embodiment, a virtual 3D model is used and views are generated by moving a virtual camera through the virtual 3D model. No matter how generated, the result is a plurality of 2D images from various positions within the 3D scene. Data representative of the location within the 3D environment and the camera orientation are also collected, if not pre-specified. This data will allow one to evaluate a scene from many perspectives along with evaluating many different paths that an observer might take through a 3D space.
  • the regions of the two-dimensional image that correspond to the objects of interest are selected and identified (step 1335). This can be accomplished using a number of different methods which might include (but are not limited to) methods that automatically extract these objects, manual selection of regions, or even hybrid approaches that use both automated and manual tagging and labeling techniques.
  • An illustration of graphical results of one method (manual) for accomplishing this process is shown Fig. 14A and Fig. 14B.
  • the objects such as mural 1310, digital sign 1315, and vase 1320 have all been identified by a manual selection process.
  • the images including the objects of interest are processed using a visual attention model to produce visual attention data (step 1340).
  • a visual attention model is that described by Itti and Koch (2001), but any visual attention model could be used.
  • the two-dimensional locations where the model predicts that visual attention will be allocated when an observer is at each viewpoint are then recorded in, for example, a database. These locations are then compared to the regions that have been tagged and labeled for each viewpoint to determine which objects within the visible area the model predicts will receive attention.
  • data concerning which objects will be attended from which locations is generated and stored. This data may then be used to, for example, determine the likelihood that a particular object will be viewed (at all) by a potential observer as she traverses the 3D scene; the percentage of potential observers that will in fact observe particular objects within the 3D scene; the likelihood of an object being attended to for a sub-set of particular views (may be useful for analyzing information about particular travel paths, such as entering a lobby versus exiting the lobby), or the likelihood that an object will be attended when that object is viewable (some objects may need to inconspicuous from most of the possible viewpoints within the environment, but for the viewpoints that the object is viewable, one may want to have a high degree of certainty that attention will be drawn to it), or the viewpoints within a 3D scene that an object is visible (or that the object will be attended to).
  • This 3D visual attention modeling may be combined with the systems and methods discussed above with respect to scene optimization.
  • the visual goals that could be defined with respect to 3D visual attention modeling may be different.
  • the visual goals may be defined such that 90% of potential observers do in fact observe a particular digital sign at some point while traversing a hotel lobby.
  • a hotel that has three different forms of advertising content that it wants its customers to view and attend to during their visit.
  • a first content is advertising a special on room upgrades;
  • a second content is advertising a special on room service;
  • a third content is advertising tickets for a play that is being held in the ballroom of the hotel.
  • the room upgrade is relevant when the customer is checking into the hotel but is not relevant at other times.
  • Room service is relevant when the customer is going to their room, but not necessarily when the customer is leaving the hotel.
  • the play tickets are relevant to customers at almost any time.
  • the 3D visual attention modeling techniques combined with the scene optimization techniques both described above.
  • the hotel would also want signs that are not relevant to be inconspicuous (for example, room upgrade when waiting for an elevator).
  • the conference room included a table 1345, a yellow sign 1350, green basket 1360, telephone 1355, purple sign 1365, as well as other typical things one would expect to find in a conference room (chairs, waste basket, screen).
  • Representative points from which an observer might be expected to view the room were determined manually, yielding eight representative observing locations (observing locations 1366, 1377, 1378, 1379, 1380, 1381, 1382, and 1383).
  • the observing locations were spaced approximately 4 feet apart throughout the unencumbered regions (no furniture) of the room.
  • Four images were taken, using a digital camera, from each of the eight observing locations, to yield 32 images.
  • the arrows extending outward from the observing locations indicate the general direction the digital camera was aimed for each picture - about a 90-degree orientation difference for each picture at each observing location.
  • the location and orientation for each of the 32 different images was recorded.
  • Fig. 17a and 17B shows an artist's rendering of two of the 32 images, where polygons encircle objects of interest, such as green basket 1360 and yellow sign 1350.
  • "Tagging” simply refers to naming the regions that comprise objects (such as "yellow sign"). Note that the polygons encircling the objects are representative of the tagging software's purpose; the underlying images are not actually modified with the polygon; the identified regions, once specified by the user, are not signified in any way on the original image.
  • the images were submitted to a visual attention model to collect the locations in the image where the model predicts visual attention will be allocated.
  • the computer recorded, for each image, the x,y coordinates where the model predicted that visual attention would be allocated.
  • the computer also calculated for each of these attention fixations whether it fell within a region of the image that was tagged and labeled by the user.
  • the computer also recorded each image that included a tagged region that was not predicted to receive attention (a "miss"). All data was stored in a database, which was then used to generate a series of summaries concerning the conspicuity of the objects within the scene.
  • Fig. 18 shows graph 1395 illustrating three different example analyses that were done for the 12 objects 1400 tagged and labeled within the conference room.
  • the first analysis is the likelihood that the object is visible or viewable (p(Visible)). This is the ratio of the images in which the object is in the image divided by the total number of images. p(Visible) is, then, a metric giving some indication of how well placed the object of interest is within the setting.
  • the metric that is determined is the likelihood that the object was attended given that it was visible (p(Attended
  • the metric calculated was the likelihood that a particular object will be attended at all (p(Attended)), which is calculated by taking the number of images in which the model predicted attention allocation to an object at least once, then dividing that value by the total number of images.
  • Multi-perspective scene analysis represents the fact that a target object can be viewed from many different distances. For example, consider the billboard example described above. As one is driving down a long flat road, there are many opportunities to attend to, and thus process the billboard. By contrast, for another billboard, there may be a hill or a group of trees that are occluding the sign until the very last minute. By taking into account these different viewpoints, one can more accurately analyze the likelihood that an object will receive visual attention from the different perspectives that it can be viewed. Without multi-perspectives, and using only a single view, one may incorrectly predict that an object will or will not receive visual attention given the number of perspectives that it could be attended.
  • Fig. 18 shows a subset of possible evaluations that one could carry out on data resulting from the analysis described above. There are a number of conclusions one could draw from the data.
  • the PurpleSign object is visible (white bar on graph) less often than the Screen object (that is, it was in fewer of the images taken within the environment).
  • the black bars for these two objects we see that when the PurpleSign is visible (that is, it is present in the image) it is always attended (p(Attended
  • Visible) l .0), but when the screen is visible, is never attended. This suggests that although the PurpleSign is located at a place where it won't be seen very often, when it is in view the model predicts that attention will be allocated to it.
  • the PurpleSign object is, then, generating the types of properties that were discussed as desirable earlier in the context of the hotel lobby. That is, the sign is not visible from many locations (it is inconspicuous) but when people are in a location where it is visible (by the check-in counter), they will almost always attend to that object. This is illustrated by the fact that the probability that the PurpleSign object is visible (white bar) is about 15%. But the probability that attention will be captured by the PurpleSign when it is visible (black bar) is 1.0.
  • Fig. 19 is a block diagram showing high-level functional modules that comprise a multi- perspective scene analysis system 1415, which in various embodiments may perform the multi-perspective visual attention modeling process described with respect to Fig. 15. It resides, in one embodiment, in a computer system 408, which includes a number of other functional modules (such as an operating system), and hardware, such as a memory or processor (neither of which are represented in Fig. 19). Though shown in Fig. 19 as a single computer, in practice various portion of the functionality could be spread among several or many computers in a networked configuration. Multi-perspective scene analysis system 1415 includes visual attention module 403 (discussed above).
  • Multi-perspective scene analysis module 1425 which invokes, as needed, visual attention module 403 to do the analysis of interest (discussed earlier with respect to step 1340 of Fig. 15), and receive results from the analysis (including which defined objects were identified by the visual attention module in which images), and stores these results, or summaries of the results, in database 1410.
  • Database 1410 is any data storage device or system, such as a computer flat file, computer memory, or a database.
  • Multi-perspective scene analysis system 1415 also includes graphical user interface module 1420, which facilitates the input of multiple images (acquired in step 1330 in Fig. 15), then, in this embodiment, facilitates the identification and tagging (step 1335 in Fig. 15) of objects of interest within the images.
  • User 1405 is any person or other computer system interested in interacting with the multi- perspective scene analysis system.
  • user 1405 is a consultant hired by a company to analyze and recommend configuration changes to a 3D scene owned or controlled by the company.
  • Visual foraging is a situation in which observer is not looking for anything in particular, but simply looking around the environment for something that is interesting.
  • this visual foraging the person will attend to different pieces of information by moving his eyes and when his eyes reach the edge of their rotation axis, the person will move his head. Typically he will move his head and eyes such that the item of interest is at the center of fixation. Current state-of-the-art does not simulate this type of re-centering action on the images.
  • the center of the image always remains fixed. This is true even when the fixation point is at the edge of the image (or visual field). Without re-centering, the current state-of-the-art approach will only be able to fixate to the edge of the image but will never fixate beyond that point. Humans, by contrast, will attend to the edge of their visual field and rotate their head so their eyes are re-aligned to the center of gaze. This would allow the visual system to make another fixation in the same direction. With a single image, one cannot make a fixation in the same direction due to the fact that there is no more information off of the edge of the image.
  • the systems and methods described herein can, in some embodiments, use multiple views of a scene or a single panoramic view to simulate the re-centering of the eyes during visual foraging. This could be done as follows:
  • the images are generated by using a 360-degree panoramic camera or multiple single images rotated around the viewing axis (vertical or otherwise). Multiple images may be taken in which the views "overlap" one another. The orientations of the views would also be assigned to each view.
  • starting view is given to the visual attention model.
  • the view can be determined by the typical viewing orientation that someone starts their visual foraging (for example, in a store line, it might be looking "forward" toward the cashier).
  • a “slice” of the panoramic view could be used centered on the "starting" view.
  • the image that is centered closest on the starting location would be used.
  • the "starting view” is analyzed with a visual attention model.
  • the initial fixation is predicted by the model.
  • the orientation of this fixation would be calculated (this can be done using trigonometry). If a panoramic view is used, a new “slice” of the panoramic view would be made centered on this new fixation. If multiple images are used, the image that is centered most closely to this new fixation would be used.
  • Example 1 Robustness Calculation Using External Variation Background: A hotel owner wants to install two digital signs in the hotel lobby. She wants them noticed by customers and can put them in any of 3 potential locations, resulting in 3 possible configurations of two digital signs (that is, signs at locations 1-2, 1- 3, or 2-3). Robustness is calculated to recommend the best sign locations to be noticed.
  • Modify the image to include simulated digital signs Generate three modified images, each simulating two digital signs such that all combinations of two signs in the three potential locations are produced. Use a standard digital photograph manipulation program such as PhotoshopTM (available from Adobe Co., San Jose, CA). Each simulated digital sign is properly scaled and has simulated content, such as a graphic of the hotel logo. Store the image regions, as defined by pixel addresses, associated with each of the three digital sign locations in a file on the computer.
  • the output will include the predicted salient regions of the modified images. Each salient region is compared to the digital sign pixel addresses stored in Step 2. If the salient region falls within or overlaps with the stored pixel addresses, then the predicted attention goes to the desired location. Each of the three modified images shows the digital signs to be in the top ten fixations, thus confirming that any of the three locations is a good candidate.
  • Step 4 Capture multiple images of the same scene, either using a series of still photos or using a video and sampling images from the video stream. Images are taken every 5 minutes over a 16 hour period, thus capturing the scene external variability resulting from a variety of lighting conditions and pedestrian movements. The goal is to have the sign locations robust to these types of variability (lighting and pedestrian movements). Load these images into the computer and modify them with simulated digital signs as described in Step 2.
  • Each modified image from step 4 is analyzed by the attention model and compared to the stored pixel addresses as described in Step 3.
  • the series of modified images associated with sign locations 1 and 2 showed that predicted fixations went to both digital sign locations in 20% of the images. Similarly, locations 1 and 3 had 35% of fixations going to both sign locations, while locations 2 and 3 had 85% of fixations going to both sign locations. Having signs installed at locations 2 and 3 resulted in the most robust configuration, providing the best solution for the hotel. Recommend this solution to the hotel owner.
  • Modify the image to include simulated digital signs Generate three modified images, each simulating two digital signs such that all combinations of two signs in the three potential locations are produced. Use a standard digital photograph manipulation program such as PhotoshopTM (available from Adobe Co., San Jose, CA). Each simulated digital sign is properly scaled and has simulated content, such as a graphic of the hotel logo. Store the image regions, as defined by pixel addresses, associated with each of the three digital sign locations in a file on the computer. 3. Run the modified images through the attention model. The output will include the predicted salient regions of the modified images. Each salient region is compared to the digital sign pixel addresses stored in Step 2. If the salient region falls within or overlaps with the stored pixel addresses, then the predicted attention goes to the desired location. Each of the three modified images shows the digital signs to be in the top ten fixations, thus confirming that any of the three locations is a good candidate.
  • PhotoshopTM available from Adobe Co., San Jose, CA
  • Each simulated digital sign is properly scaled and has simulated content, such as
  • Step 1 Specify the number of model variations to utilize in analyzing the modified images (for example, 100 model variations).
  • Each visual attention model has three different feature maps (color, orientation, and luminance); the saliency map is computed as a weighted combination of each of these maps.
  • the basic visual attention model sets the weighting parameters for each map as equal (1, 1, 1).
  • To produce 100 model variations randomly set the weighting vector for each model. This is completed by an algorithm that randomly sets each weight and normalizes the sum of the weights to 3 (3 * (RandWeights/sum(RandWeights)) .
  • Her specific visual goals are for customers to notice four target objects: a first and a second digital sign, a static graphic sign advertising the hotel restaurant, and the staff behind the check-in counter. 1.
  • rewards are given for changes that draw attention to the target objects, and real-world costs are associated with permissible changes.
  • An estimated cost in dollars relating to labor and supply costs is assigned to the potential changes being considered: moving a painting currently located behind the check-in counter: $100, changing the lighting behind the restaurant sign: $2500, and redesigning content displayed on the two digital signs: $250 dollars each.
  • Reward values assigned for achieving the visual goals are as follows: drawing attention to the two digital signs: $500 each, drawing attention to the restaurant sign: $250, and drawing attention to the staff behind the check-in counter: $150.
  • Example 3 Continuing from Example 3, recommended changes have been made. The hotel owner would like to understand the visual saliency of each target object as viewed from multiple perspectives in the hotel lobby.
  • Each of the 16 images from Step 1 are analyzed using the attention model. For each image, it is determined by the visual attention model which target objects are visible and which target objects are predicted to draw attention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Graphics (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Multimedia (AREA)
  • Geometry (AREA)
  • Image Analysis (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

L’invention concerne des systèmes et des procédés permettant d’utiliser des techniques de modélisation d’attention visuelle pour évaluer une scène depuis de multiples perspectives.
EP09740784A 2008-10-03 2009-10-01 Systèmes et procédés pour une analyse de scène depuis de multiples perspectives Withdrawn EP2347387A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10267008P 2008-10-03 2008-10-03
PCT/US2009/059255 WO2010039976A1 (fr) 2008-10-03 2009-10-01 Systèmes et procédés pour une analyse de scène depuis de multiples perspectives

Publications (1)

Publication Number Publication Date
EP2347387A1 true EP2347387A1 (fr) 2011-07-27

Family

ID=41510890

Family Applications (1)

Application Number Title Priority Date Filing Date
EP09740784A Withdrawn EP2347387A1 (fr) 2008-10-03 2009-10-01 Systèmes et procédés pour une analyse de scène depuis de multiples perspectives

Country Status (7)

Country Link
EP (1) EP2347387A1 (fr)
JP (1) JP5667061B2 (fr)
KR (1) KR20110081240A (fr)
CN (1) CN102227748A (fr)
AU (1) AU2009298438B2 (fr)
CA (1) CA2739273A1 (fr)
WO (1) WO2010039976A1 (fr)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2591458A2 (fr) * 2010-07-07 2013-05-15 3M Innovative Properties Company Systèmes et procédés d'amélioration de modèles d'attention visuelle
WO2013086137A1 (fr) 2011-12-06 2013-06-13 1-800 Contacts, Inc. Systèmes et procédés pour obtenir une mesure d'écart pupillaire à l'aide d'un dispositif informatique mobile
US9483853B2 (en) 2012-05-23 2016-11-01 Glasses.Com Inc. Systems and methods to display rendered images
US9286715B2 (en) 2012-05-23 2016-03-15 Glasses.Com Inc. Systems and methods for adjusting a virtual try-on
US20130314401A1 (en) 2012-05-23 2013-11-28 1-800 Contacts, Inc. Systems and methods for generating a 3-d model of a user for a virtual try-on product
CN103226695A (zh) * 2013-04-02 2013-07-31 北京林业大学 基于选择性视觉注意机制的马尾松多尺度智能识别模型
JP2015046089A (ja) 2013-08-29 2015-03-12 ソニー株式会社 情報処理装置および情報処理方法
CN106354251B (zh) * 2016-08-17 2019-04-02 深圳前海小橙网科技有限公司 一种虚拟场景与真实场景融合的模型系统与方法
CN108322648B (zh) * 2018-02-02 2020-06-30 Oppo广东移动通信有限公司 图像处理方法和装置、电子设备、计算机可读存储介质
JP7295132B2 (ja) * 2018-03-12 2023-06-20 ライン プラス コーポレーション ニューラルネットワークを利用して状況を認知する拡張現実提供装置、提供方法、およびコンピュータプログラム
CN110097539B (zh) * 2019-04-19 2020-09-11 贝壳找房(北京)科技有限公司 一种在虚拟三维模型中截取图片的方法及装置
RU2725682C1 (ru) * 2019-04-29 2020-07-03 Кэнон Кабусики Кайся Устройство обработки информации, способ обработки информации и носитель данных
WO2021009819A1 (fr) * 2019-07-12 2021-01-21 日本電信電話株式会社 Dispositif de commande d'affichage, procédé de commande d'affichage et programme de commande d'affichage
CN114092783A (zh) * 2020-08-06 2022-02-25 清华大学 一种基于注意力机制连续视角的危险品检测方法
CN112802303A (zh) * 2021-02-09 2021-05-14 成都视海芯图微电子有限公司 一种基于3d视觉的实时风险预测方法及系统
WO2023004734A1 (fr) * 2021-07-30 2023-02-02 Pricewaterhousecoopers Llp Dispositif de prédiction de regard

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2370438A (en) * 2000-12-22 2002-06-26 Hewlett Packard Co Automated image cropping using selected compositional rules.

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO2010039976A1 *

Also Published As

Publication number Publication date
JP2012504830A (ja) 2012-02-23
AU2009298438B2 (en) 2013-10-24
WO2010039976A1 (fr) 2010-04-08
AU2009298438A1 (en) 2010-04-08
KR20110081240A (ko) 2011-07-13
CA2739273A1 (fr) 2010-04-08
CN102227748A (zh) 2011-10-26
JP5667061B2 (ja) 2015-02-12

Similar Documents

Publication Publication Date Title
US8577156B2 (en) Systems and methods for multi-perspective scene analysis
AU2009298428B2 (en) Systems and methods for optimizing a scene
AU2009298416B8 (en) Systems and methods for evaluating robustness
AU2009298438B2 (en) Systems and methods for multi-perspective scene analysis
US10515163B2 (en) Systems and methods for improving visual attention models
JP6495254B2 (ja) 視覚的注意モデリングの結果の提示

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20110426

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK SM TR

AX Request for extension of the european patent

Extension state: AL BA RS

DAX Request for extension of the european patent (deleted)
17Q First examination report despatched

Effective date: 20120227

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20160503