WO2010089989A1 - 情報表示装置及び情報表示方法 - Google Patents
情報表示装置及び情報表示方法 Download PDFInfo
- Publication number
- WO2010089989A1 WO2010089989A1 PCT/JP2010/000595 JP2010000595W WO2010089989A1 WO 2010089989 A1 WO2010089989 A1 WO 2010089989A1 JP 2010000595 W JP2010000595 W JP 2010000595W WO 2010089989 A1 WO2010089989 A1 WO 2010089989A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- information
- notification information
- display
- screen
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42201—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] biosensors, e.g. heat sensor for presence detection, EEG sensors or any limb activity sensors worn by the user
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/4223—Cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/442—Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
- H04N21/44213—Monitoring of end-user related data
- H04N21/44218—Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4882—Data services, e.g. news ticker for displaying messages, e.g. warnings, reminders
Definitions
- the present invention relates to an information display device that displays notification information, which is information for notifying a user, on a screen.
- TVs are not only used for viewing broadcast content, but also open to new uses such as simultaneous use of multiple information and lists of large amounts of information. It's getting on.
- a TV provided with a wall of a living room as a display.
- various information closely related to daily life can be presented at an appropriate timing.
- the user can operate a plurality of devices using one remote controller. Further, the user can also check the video of the network camera on the TV screen. In addition to these devices, if white goods such as washing machines or microwave ovens are connected to the home network, the user can check the status of each device on the TV. That is, a plurality of devices connected to the network cooperate with each other and present notification information from each device to a display device such as a TV, so that the user can acquire information on various devices while staying in front of the TV. .
- Patent Document 1 As a method for presenting such notification information to the user, a method for controlling the timing of presenting the notification information to the user has been proposed (for example, see Patent Document 1).
- the notification information is presented to the user in consideration of a policy for determining an appropriate timing for presenting the notification information and the user's situation such as the user's cost for interruption.
- Patent Document 2 a method of presenting information to the user in consideration of the size of the user's effective visual field has been proposed (see, for example, Patent Document 2).
- the size of the image is adjusted according to the display position of the image displayed on the screen and the distance to the center of the visual field. As a result, it is possible to prevent the user from recognizing the image between the central portion of the visual field and the peripheral portion of the visual field.
- an object of the present invention is to solve the above-described conventional problems and to provide an information display device capable of presenting notification information to a user without giving a sudden impression to the user.
- an information display device is an information display device that displays notification information, which is information for notifying a user, on a screen, and the physical state of the user
- a user state detection unit that detects a user state
- a concentration degree estimation unit that estimates a degree of concentration indicating a concentration degree of the user with respect to the screen, based on the user state detected by the user state detection unit
- An application control unit that determines the initial display position of the notification information based on the concentration degree estimated by the concentration degree estimation unit so that the initial display position is located outside the effective visual field region that is visually recognized by the user;
- the notification information is displayed at the initial display position determined by the application control unit, and the display position and display mode of the notification information after display are reduced. And a screen displaying part also be changed one.
- the initial display position of the notification information it is possible to determine the initial display position of the notification information so that the initial display position is located outside the effective visual field area, so that a sudden impression is given to the user when the notification information is initially displayed. Can be suppressed. Furthermore, by changing the display position or display mode of the notification information after display, the notification information can be noticed casually to the user, and the notification information can be presented to the user without giving a sudden impression.
- the user state detection unit detects the position of the user's gazing point on a plane including the screen as the user state
- the application control unit has the degree of concentration estimated by the concentration degree estimation unit. It is preferable that the initial display position is determined such that the smaller the value is, the farther from the position determined by the position of the gazing point detected by the user state detection unit.
- the initial display position can be determined so that the smaller the degree of concentration is, the farther from the position determined by the position of the gazing point, and the initial display position is easily located so that the initial display position is located outside the effective visual field area Can be determined.
- the application control unit further determines a movement speed so that the movement speed increases as the concentration degree estimated by the concentration degree estimation unit increases, and the screen drawing unit is determined by the application control unit. It is preferable to change the display position of the notification information by moving the display position of the notification information at the moving speed.
- the moving speed of the display position of the notification information can be determined according to the degree of concentration, so that the notification information can be presented to the user without giving a more abrupt impression.
- the user state detection unit detects a position of the user's point of interest on a plane including the screen as the user state, and the screen drawing unit is detected within a predetermined period by the user state detection unit. It is preferable to change the display position of the notification information by moving the display position of the notification information toward a position that represents the position of a plurality of gazing points.
- the display position of the notification information can be moved toward a position representing a plurality of gazing points detected within a predetermined period, so that the notification information is presented to the user without giving a more sudden impression. be able to.
- the screen drawing unit changes the display position of the notification information by moving the display position of the notification information toward a predetermined position in the display area of the content displayed on the screen.
- the display position of the notification information can be moved toward the position in the content display area, so that the notification information can be presented to the user without giving a more abrupt impression.
- the screen drawing unit moves the display position of the notification information toward a position outside the display area of the content displayed on the screen and near the boundary of the display area of the content. It is preferable to change the display position of the notification information.
- the display position of the notification information can be moved to a position outside the content display area and near the boundary of the content display area, so notification without giving a more sudden impression Information can be presented to the user.
- the application control unit further determines the size of the display area such that the larger the degree of concentration estimated by the concentration degree estimation unit, the larger the display region, and the screen drawing unit includes the application control unit.
- the notification information is displayed at the initial display position determined by the above, it is preferable to display the notification information in the display area having the determined size.
- This configuration allows the notification information to be displayed in a size corresponding to the degree of concentration, so that the notification information can be presented to the user without giving a more abrupt impression.
- the user status detecting unit further includes a relevance estimation unit that estimates a relevance level indicating a relevance level between the content displayed on the screen and the notification information.
- the position of the viewpoint is detected as the user state, and the application control unit detects from the position determined by the position of the gazing point detected by the user state detection unit as the degree of association estimated by the degree of association estimation unit is small. It is preferable to determine the initial display position so that the initial display position becomes far.
- the initial display position of the notification information can be determined based on the degree of association between the notification information and the content, the notification information can be presented to the user without giving a more abrupt impression.
- the application control unit further determines a moving speed so that the moving speed increases as the relevance estimated by the relevance estimating unit increases, and the screen drawing unit notifies at the determined moving speed. It is preferable to change the display position of the notification information by moving the display position of the information.
- the moving speed of the notification information can be determined based on the degree of association between the notification information and the content, the notification information can be presented to the user without giving a more abrupt impression.
- an importance level / emergency level acquisition unit that acquires an importance level indicating the importance level of the notification information or an urgent level indicating the urgency level of the notification information is provided, and the application control unit includes the importance level / emergency level. Determining the initial display position such that the smaller the importance or urgency acquired by the degree acquisition unit is, the farther the initial display position is from the position determined by the position of the gazing point detected by the user state detection unit. Is preferred.
- the initial display position of the notification information can be determined based on the importance or urgency of the notification information, so that notification information with a high importance or urgency can be presented to the user earlier.
- the application control unit further determines a moving speed so that the moving speed increases as the importance level or the urgency level acquired by the importance / urgent level acquisition unit increases, and the screen drawing unit determines It is preferable to change the display position of the notification information by moving the display position of the notification information at the moved speed.
- the moving speed of the notification information can be determined based on the importance or urgency of the notification information, so that notification information with a high importance or urgency can be presented to the user earlier.
- the user state detection unit detects a position of a user's gazing point on a plane including the screen as the user state, and the concentration degree estimation unit is detected by the user state detection unit within a predetermined period.
- the degree of concentration is estimated based on a gaze point distribution including a plurality of gaze points.
- This configuration makes it possible to estimate the user concentration with high accuracy.
- the user state detection unit detects a position of the user's gazing point on a plane including the screen as the user state
- the concentration degree estimation unit detects the gazing point detected by the user state detection unit. It is preferable to estimate the degree of concentration based on the amount of movement.
- This configuration makes it possible to estimate the user concentration with high accuracy.
- the user state detection unit detects a user's face orientation as the user state
- the concentration degree estimation unit is a face orientation distribution composed of a plurality of face orientations detected within a predetermined period by the user state detection unit. It is preferable to estimate the degree of concentration based on the above.
- This configuration makes it possible to estimate the user concentration with high accuracy.
- the user state detection unit detects a user posture as the user state, and the concentration degree estimation unit estimates the concentration degree based on the posture detected by the user state detection unit. .
- This configuration makes it possible to estimate the user concentration with high accuracy.
- the user status database further includes a user information database that stores the degree of concentration and effective visual field area information indicating the size of the effective visual field area in association with each other.
- the application control unit obtains effective visual field region information corresponding to the degree of concentration estimated by the concentration degree estimation unit by referring to the user information database.
- the initial display position is determined outside the effective visual field area estimated using the acquired effective visual field area information and the gazing point detected by the user state detection unit.
- the initial display position of the notification information can be easily set so that the initial display position is located outside the effective visual field area. Can be determined.
- the application control unit further determines whether a distance between the display position of the notification information and the position of the user's point of interest is less than a threshold value while the display position of the notification information is changed by the screen drawing unit. If it is determined whether or not the distance is less than the threshold value, it is preferable to update the effective visual field area information stored in the user information database using the display position.
- This configuration can improve the accuracy of the effective visual field area information stored in the user information database.
- a user identification unit that identifies a user in front of the screen
- the user information database associates, for each user, a concentration level and effective visual field area information indicating the size of the effective visual field area.
- the application control unit stores the effective visual field area information corresponding to the user identified by the user identification unit.
- This configuration makes it possible to determine the initial display position with high accuracy so that the initial display position is located outside the effective visual field region.
- An information display method is an information display method for displaying notification information, which is information for notifying a user, on a screen, and detects a user state indicating a physical state of the user.
- a user state detection step a concentration degree estimation step for estimating a concentration degree indicating a user's concentration degree on the screen based on the user state detected in the user state detection step, and an area visually recognized by the user.
- An application control step for determining an initial display position of the notification information based on the concentration degree estimated in the concentration degree estimation step, and an application control step so that the initial display position is located outside the effective visual field region;
- the notification information is displayed at the initial display position, and the display position and table of the notification information after display are displayed.
- the present invention can also be realized as a program for causing a computer to execute such an information display method.
- a program can be distributed via a computer-readable recording medium such as CD-ROM (Compact Disc Only Memory) or a transmission medium such as the Internet.
- the initial display position of the notification information is determined so that the initial display position is located outside the effective visual field area of the user. Can do. Therefore, the information display device can suppress giving a sudden impression to the user when the notification information is initially displayed. Furthermore, the information display device can casually notice the notification information by changing the display position or display mode of the notification information after the display, and presents the notification information to the user without giving a sudden impression. can do.
- FIG. 1 is a diagram showing an external appearance of the information display device according to Embodiment 1 of the present invention.
- FIG. 2 is a block diagram showing a functional configuration of the information display apparatus according to Embodiment 1 of the present invention.
- FIG. 3 is a flowchart showing the operation of the information display device in Embodiment 1 of the present invention.
- FIG. 4 is a diagram for explaining the operation of the information display apparatus according to Embodiment 1 of the present invention.
- FIG. 5 is a block diagram showing a functional configuration of the information display apparatus according to Embodiment 2 of the present invention.
- FIG. 6 is a diagram showing an example of a user information database according to Embodiment 2 of the present invention.
- FIG. 1 is a diagram showing an external appearance of the information display device according to Embodiment 1 of the present invention.
- FIG. 2 is a block diagram showing a functional configuration of the information display apparatus according to Embodiment 1 of the present invention.
- FIG. 3 is a flowchar
- FIG. 7 is a flowchart showing the operation of the information display apparatus in Embodiment 2 of the present invention.
- FIG. 8 is a diagram for explaining an example of an interface between the appearance of the information display device and related devices in Embodiment 3 of the present invention.
- FIG. 9 is a block diagram showing a functional configuration of the information display apparatus according to Embodiment 3 of the present invention.
- FIG. 10A is a diagram for describing a user position calculation method by a user state detection unit according to Embodiment 3 of the present invention.
- FIG. 10B is a diagram for explaining a user position calculation method by the user state detection unit according to Embodiment 3 of the present invention.
- FIG. 10A is a diagram for describing a user position calculation method by a user state detection unit according to Embodiment 3 of the present invention.
- FIG. 10B is a diagram for explaining a user position calculation method by the user state detection unit according to Embodiment 3 of the present invention.
- FIG. 10A is
- FIG. 11 is a flowchart showing a flow of processing relating to the detection of the line-of-sight direction in the third embodiment of the present invention.
- FIG. 12 is a diagram for explaining the process of detecting the face orientation of the user in the third embodiment of the present invention.
- FIG. 13 is a diagram for describing the reference direction reference plane.
- FIG. 14 is a diagram for explaining the detection of the center of the black eye.
- FIG. 15 is a diagram for explaining the detection of the center of the black eye.
- FIG. 16A is a diagram showing an example of a user information database according to Embodiment 3 of the present invention.
- FIG. 16B is a diagram showing an example of a user information database according to Embodiment 3 of the present invention.
- FIG. 16C is a diagram showing an example of a user information database according to Embodiment 3 of the present invention.
- FIG. 17A is a diagram showing an example of notification information in Embodiment 3 of the present invention.
- FIG. 17B is a diagram showing an example of notification information in Embodiment 3 of the present invention.
- FIG. 17C is a diagram showing an example of notification information in Embodiment 3 of the present invention.
- FIG. 18A is a diagram for describing a use scene of the information display device in Embodiment 3 of the present invention.
- FIG. 18B is a diagram for describing a use scene of the information display device in Embodiment 3 of the present invention.
- FIG. 18A is a diagram for describing a use scene of the information display device in Embodiment 3 of the present invention.
- FIG. 18B is a diagram for describing a use scene of the information display device in Embodiment 3 of the present invention.
- FIG. 18A is a diagram for
- FIG. 19 is a flowchart showing the flow of processing of the information display device in Embodiment 3 of the present invention.
- FIG. 20 is a diagram for explaining an operation example of the information display apparatus according to Embodiment 3 of the present invention.
- FIG. 21 is a diagram for explaining an operation example of the information display apparatus according to Embodiment 3 of the present invention.
- FIG. 22A is a block diagram showing a functional configuration of the information display device in Modification 1 of Embodiment 3 of the present invention.
- FIG. 22B is a flowchart showing a processing flow of the information display device in the first modification of the third embodiment of the present invention.
- FIG. 23A is a diagram for explaining an operation example of the information display device in the first modification of the third embodiment of the present invention.
- FIG. 20 is a diagram for explaining an operation example of the information display apparatus according to Embodiment 3 of the present invention.
- FIG. 21 is a diagram for explaining an operation example of the information display apparatus according to Embodiment 3 of
- FIG. 23B is a diagram for explaining an operation example of the information display device in the first modification of the third embodiment of the present invention.
- FIG. 24A is a diagram for explaining an overview of a display area control method based on a user position in the second modification of the third embodiment of the present invention.
- FIG. 24B is a diagram for describing an outline of a display area control method based on the user position in the second modification of the third embodiment of the present invention.
- FIG. 24C is a diagram for describing an outline of a display area control method based on the user position in the second modification of the third embodiment of the present invention.
- FIG. 25 is a diagram for explaining an operation example of the information display device in the second modification of the third embodiment of the present invention.
- FIG. 26A is a diagram for explaining an operation example of the information display device in the third modification of the third embodiment of the present invention.
- FIG. 26B is a diagram for explaining an operation example of the information display device in the third modification of the third embodiment of the present invention.
- FIG. 27A is a diagram for explaining an operation example of the information display device in the fourth modification of the third embodiment of the present invention.
- FIG. 27B is a diagram for explaining an operation example of the information display device in the fourth modification of the third embodiment of the present invention.
- FIG. 27C is a diagram for describing an operation example of the information display device in the fourth modification of the third embodiment of the present invention.
- FIG. 28A is a diagram for explaining an operation example of the information display device in the fifth modification of the third embodiment of the present invention.
- FIG. 28B is a diagram for explaining an operation example of the information display device in the fifth modification of the third embodiment of the present invention.
- FIG. 29A is a diagram for explaining an operation example of the information display device in the sixth modification of the third embodiment of the present invention.
- FIG. 29B is a diagram for explaining an operation example of the information display device in the sixth modification of the third embodiment of the present invention.
- FIG. 30A is a diagram for explaining an operation example of the information display device in the sixth modification of the third embodiment of the present invention.
- FIG. 30B is a diagram for describing an operation example of the information display device in the sixth modification of the third embodiment of the present invention.
- FIG. 31A is a diagram for explaining an operation example of the information display device in the seventh modification of the third embodiment of the present invention.
- FIG. 31B is a diagram for explaining an operation example of the information display device in the seventh modification of the third embodiment of the present invention.
- FIG. 31C is a diagram for describing an operation example of the information display device in the seventh modification of the third embodiment of the present invention.
- FIG. 32 is a diagram for explaining an operation example of the information display device in the modification 8 of the third embodiment of the present invention.
- FIG. 1 is a diagram showing an external appearance of the information display device according to Embodiment 1 of the present invention.
- FIG. 2 is a block diagram showing a functional configuration of the information display apparatus according to Embodiment 1 of the present invention.
- the information display device 10 in the present embodiment is a device that displays notification information on the screen, and displays the notification information on the screen so that the notification information is initially displayed outside the effective visual field area of the user.
- the information display device 10 is preferably applied to a large screen display.
- the information display device 10 includes a user state detection unit 11, a concentration degree estimation unit 12, an application control unit 13, and a screen drawing unit 14.
- the user state detection unit 11 detects a user state indicating the physical state of the user. Specifically, the user state detection unit 11 detects, for example, the position of the user's point of gaze on a plane including the screen as the user state, and holds the detected user state. A method for detecting the position of the user's point of interest will be described in detail in Embodiment 3.
- the user state detection unit 11 may detect the user's face orientation or the user's posture as the user state. In this case, the user state detection unit 11 detects the user's face orientation using, for example, an image of the user's face taken by the camera. In addition, for example, the user state detection unit 11 detects the user's posture using an image of the user taken by the camera or a pressure sensor installed on the floor surface in front of the screen.
- the concentration level estimation unit 12 estimates a concentration level indicating a user's concentration level on the screen based on the detected user state.
- the concentration level estimation unit 12 estimates the concentration level based on a gaze point distribution including a plurality of gaze points detected by the user state detection unit 11 within a predetermined period. For example, the degree-of-concentration estimation unit 12 estimates that the degree of concentration is smaller the greater the variance of the gaze point distribution.
- the predetermined period is, for example, a period from the most recent time at which the gazing point is detected to a time that is a certain time back.
- the concentration level estimation unit 12 may estimate the concentration level based on the movement amount of the gazing point detected by the user state detection unit 11. In this case, for example, the concentration degree estimation unit 12 calculates the movement amount of the gazing point from the positions of the plurality of gazing points detected within the predetermined period by the user state detection unit 11, and the calculated movement amount of the gazing point is large. It is estimated that the degree of concentration is small.
- the concentration degree estimation unit 12 may estimate the concentration degree based on a face orientation distribution including a plurality of face orientations detected by the user state detection unit 11 within a predetermined period. In this case, for example, the concentration degree estimation unit 12 estimates that the degree of concentration is smaller as the variance of values indicating a plurality of face orientations detected by the user state detection unit 11 within a predetermined period is larger.
- the concentration degree estimation unit 12 may estimate the concentration degree based on the user posture detected by the user state detection unit 11.
- the concentration degree estimation unit 12 refers to a database that stores the concentration degree corresponding to the posture (standing position, sitting position, or the like), and thereby the concentration degree corresponding to the detected posture of the user. Is estimated.
- the application control unit 13 determines the initial display position of the notification information based on the estimated degree of concentration so that the initial display position is located outside the effective visual field area that is an area visually recognized by the user. Specifically, the application control unit 13 determines the initial display position so that the smaller the estimated concentration is, the farther from the position determined by the position of the detected gazing point. Note that the application control unit 13 may determine the initial display position such that the smaller the estimated concentration is, for example, the farther from the center position of the display area of the content displayed on the screen or the center position of the screen. Good.
- the effective visual field region is a region where the user can recognize the displayed image and the like relatively clearly, and is a region whose size changes according to the degree of concentration of the user.
- the effective visual field region is a circular or elliptical region centered on the center position of the gazing point distribution, and is estimated to be a region that increases in size as the degree of user concentration decreases. If the notification information is suddenly displayed within this effective visual field region, the user feels awkward and feels bothersome.
- the position determined by the position of the gazing point is, for example, the position of the gazing point itself, or the center of gravity position or the center position of the gazing point distribution including a plurality of gazing points detected within a predetermined period.
- the notification information is information for notifying the user.
- the notification information is, for example, text information or image information indicating a state of a device connected to the information display device 10 via a network, or text information or image information related to display content. More specifically, the notification information is, for example, an icon indicating the appearance of the microwave oven indicating that the heating process by the microwave oven has been completed.
- the screen drawing unit 14 displays the notification information on a screen such as a plasma display panel (PDP: Plasma Display Panel) or a liquid crystal panel (Liquid Crystal Panel).
- a plasma display panel PDP: Plasma Display Panel
- a liquid crystal panel Liquid Crystal Panel
- the screen drawing unit 14 first displays the notification information at the determined initial display position.
- the screen drawing unit 14 changes at least one of the display position and display mode of the notification information after display.
- the screen drawing unit 14 changes the display position of the notification information, for example, by moving an image indicating the notification information to the target position.
- the target position is, for example, a position representing a plurality of gazing points detected within a predetermined period, and is typically the center position of the gazing point distribution.
- the target position is a position within the display area of the display content, and may typically be the center position of the display area of the display content. Further, the target position may be a position outside the display area of the display content and a position near the boundary of the display area of the display content.
- the screen drawing unit 14 changes the display mode of the notification information by changing the color or definition of the image indicating the notification information, the size of the display area of the notification information, or the like. Specifically, the screen drawing unit 14 gradually increases the display area of the notification information. Further, the screen drawing unit 14 may gradually increase the sharpness of the image indicating the notification information. The screen drawing unit 14 may gradually change the color of the image indicating the notification information to a color with high saturation.
- the screen drawing unit 14 may change both the display position and the display mode of the notification information.
- FIG. 3 is a flowchart showing the operation of the information display apparatus according to Embodiment 1 of the present invention.
- the user state detection unit 11 detects a user state indicating the physical state of the user (S102). Subsequently, the concentration level estimation unit 12 estimates the concentration level indicating the concentration level of the user with respect to the screen based on the detected user state (S104).
- the application control unit 13 determines the initial display position of the notification information based on the estimated concentration so that the initial display position is located outside the effective visual field region (S106).
- the screen drawing unit 14 displays the notification information at the determined initial display position (S108). Then, the screen drawing unit 14 changes at least one of the display position and display mode of the notification information after display (S110), and ends the process.
- FIG. 4 is a diagram for explaining the operation of the information display apparatus according to Embodiment 1 of the present invention. Specifically, FIG. 4 is a diagram illustrating a change with time of the notification information displayed on the screen.
- the peripheral visual field region is a region surrounding the central vision, which is a region where the user can recognize the object with high resolution, and is a region where the user can recognize the movement or change of the object.
- the outer edge of the peripheral visual field region is a position where the viewing angle of the user is about 180 to 210 degrees.
- the effective visual field area is an area in which the user can recognize the object relatively clearly in the peripheral visual field area.
- the size of the effective visual field region varies depending on the psychological factors of the user, and becomes smaller as the user's degree of concentration increases.
- the outer edge of the effective viewing area is a position where the viewing angle of the user is about 4 to 20 degrees.
- the information display device 10 first determines the initial display position based on the degree of concentration so that the initial display position is located in the screen area outside the effective visual field area in the peripheral visual field area. Then, the information display device 10 displays the notification information at the determined initial display position. Thereby, the information display apparatus 10 can suppress giving a sudden impression to the user by displaying the notification information.
- the information display device 10 changes at least one of the display position and the display mode of the notification information displayed on the screen area outside the effective visual field area among the peripheral visual field areas. For example, as illustrated in FIG. 4, the information display device 10 moves the notification information displayed outside the effective visual field region in a direction toward the center position (central view) of the effective visual field region, and also displays the size of the image indicated by the notification information. Increase the size gradually. Thereby, the information display apparatus 10 can make the user notice the notification information displayed in the peripheral visual field region casually.
- the information display device 10 can determine the initial display position of the notification information so that the initial display position is located outside the effective visual field area of the user. Therefore, the information display device 10 can suppress giving a sudden impression to the user when the notification information is initially displayed. Further, the information display device 10 can casually notice the notification information by changing the display position or display mode of the notification information after display, and can notify the user of the notification information without giving a sudden impression. Can be presented.
- the information display device 10 can determine the initial display position so that the smaller the degree of concentration, the farther from the position determined by the position of the gazing point, and the initial display position can be easily located outside the effective visual field region. Can be determined.
- the information display device 10 can estimate the user's concentration degree with high accuracy based on the user state such as the distribution of the gazing point, the moving amount of the gazing point, the user's face orientation, or the user's posture.
- the information display device 20 in the present embodiment is mainly different from the information display device 10 in the first embodiment in that the initial display position is determined with reference to the user information database 23.
- FIG. 5 is a block diagram showing a functional configuration of the information display apparatus according to Embodiment 2 of the present invention. 5, the same components as those in FIG. 2 are denoted by the same reference numerals, and the description thereof is omitted.
- the application control unit 21 determines the initial display position of the notification information based on the estimated concentration level so that the initial display position is located outside the effective visual field area that is an area visually recognized by the user.
- the application control unit 21 refers to the user information database 23 to acquire effective visual field area information corresponding to the estimated concentration level. Then, the application control unit 21 determines an initial display position outside the effective visual field area estimated using the acquired effective visual field area information and the detected gazing point.
- the application control unit 21 indicates the distance indicated by the effective visual field area information from the detected central position of the gazing point distribution. A position separated by a distance obtained by adding a certain distance to is determined as the initial display position. Further, for example, when the effective viewing area information indicates the viewing angle of the user, the application control unit 21 calculates a straight line connecting the user position and the center position of the gaze point distribution and a straight line connecting the user position and the initial display position. The initial display position is determined so that the formed angle is equal to or greater than the viewing angle.
- the application control unit 21 determines whether or not the distance between the display position of the notification information and the position of the user's gazing point is less than the threshold while the display position of the notification information is changed by the screen drawing unit 22. Determine.
- the threshold value is an upper limit value of the distance estimated that the user's attention is directed to the notification information, and is a value given in advance by experience or experiment.
- the application control unit 21 updates the effective visual field area information stored in the user information database 23 using the display position.
- the application control unit 21 determines that the distance between the display position of the notification information and the gazing point of the user is less than the threshold value. When the determination is made, the distance between the display position of the notification information and the center position of the gaze point distribution is calculated. Then, the application control unit 21 updates the distance indicated by the effective visual field area information to the calculated distance.
- the application control unit 21 connects the straight line connecting the display position of the notification information and the position of the user's eyes, the center position of the gazing point distribution, and the position of the user's eyes. Calculate the angle made with the straight line. Then, the application control unit 21 updates the viewing angle indicated by the effective viewing area information to the calculated angle.
- the screen drawing unit 22 displays the notification information at the determined initial display position. And the screen drawing part 22 changes the display position of the notification information after a display.
- the user information database 23 stores the degree of concentration and the effective visual field area information in association with each other as shown in FIG.
- FIG. 6 is a diagram showing an example of a user information database according to Embodiment 2 of the present invention.
- the effective visual field area information indicates the size of the effective visual field area, and in FIG. 6, the distance from the center position of the gaze point distribution.
- the degree of concentration is “0.8”
- a position separated by “0.5” m or more from the center position of the gazing point distribution is located outside the effective visual field region. Indicates.
- FIG. 7 is a flowchart showing the operation of the information display apparatus according to Embodiment 2 of the present invention. In FIG. 7, the same processes as those in FIG.
- the application control unit 21 refers to the user information database 23 to acquire effective visual field area information corresponding to the estimated concentration level (S202). Subsequently, the application control unit 21 estimates the effective visual field area using the acquired effective visual field area information and the center position of the gazing point distribution, and determines an initial display position outside the estimated effective visual field area (S204). ).
- the screen drawing unit 22 displays the notification information at the determined initial display position (S206). Furthermore, the screen drawing unit 22 changes the display position of the notification information after display (S208). Next, the user state detection unit 11 newly detects the user's gaze point (S210).
- the application control unit 21 determines whether or not the distance between the current notification information display position and the gaze point detected in step S209 is equal to or less than a threshold value (S212).
- a threshold value (S212)
- the application control unit 21 uses the current display position to obtain the effective visual field area information stored in the user information database 23. Update (S214) and end the process.
- the process returns to Step S208.
- the information display device 20 refers to the user information database 23 and acquires the effective visual field region information corresponding to the degree of concentration so that the initial display position is outside the effective visual field region.
- the initial display position can be easily determined so as to be positioned.
- the information display device 20 updates the effective visual field area information using the display position, thereby storing the information in the user information database 23.
- the accuracy of stored effective visual field area information can be increased.
- the application control part 21 in this Embodiment updated the user information database 23, it is not necessary to update the user information database 23 necessarily. Even when the application control unit 21 does not update the user information database 23, the information display device 20 refers to the user information database 23 so that the initial display position is located outside the effective visual field region. The display position can be determined.
- the application control unit 21 determines whether the distance between the display position of the notification information and the user's gaze point is less than the threshold. It may be determined whether or not the state where the distance from the gazing point is less than the threshold value continues for a predetermined time or more. Thereby, the fall of the determination precision by the misdetection of a gaze point etc. can be suppressed.
- the information display device 30 in the present embodiment is suitable for being applied to a large screen display viewed by one or more users, and performs display control of notification information based on the viewing status of the user with respect to display content.
- FIG. 8 is a diagram for explaining an example of an interface between the appearance of the information display device and related devices in Embodiment 3 of the present invention.
- the information display device 30 acquires content and image information from one or a plurality of user detection cameras 102 in addition to the antenna 101 for receiving a broadcast program. Further, the user detection camera 102 may be laid on the wall or ceiling where the screen is installed, instead of being laid on the information display device 30 as shown in FIG. In addition, the user detection camera 102 may be laid on both the information display device 30 and a wall or a ceiling.
- the information display device 30 includes a notification source 106 such as a mobile phone 103, a network camera 104, and a household appliance group 105 (for example, a refrigerator, a washing machine, a microwave oven, an air conditioner, and a lighting device) and a wireless network or a wired network. Are connected through.
- the information display device 30 is also connected to the Internet via the router / hub 107.
- FIG. 9 is a block diagram showing a functional configuration of the information display apparatus according to Embodiment 3 of the present invention.
- the information display device 30 includes a user identification unit 31, a user state detection unit 32, a concentration degree estimation unit 33, a user information database 34, a relevance degree estimation unit 35, and an application control unit 36. And a screen drawing unit 37 and a screen 38.
- the user detection camera 102 is a camera including an image sensor such as a CCD (Charge Coupled Device) or a CMOS (Complementary Metal Oxide Semiconductor) installed around the screen 38.
- an image sensor such as a CCD (Charge Coupled Device) or a CMOS (Complementary Metal Oxide Semiconductor) installed around the screen 38.
- the user detection camera 102 images a user existing in front of the screen 38.
- the user identification unit 31 identifies a user by performing face region extraction from an image photographed by the user detection camera 102 and then comparing the face image registered in advance with the extracted face image. And the user identification part 31 outputs the user identification information for identifying the specified user.
- the user state detection unit 32 detects the position of the user's gaze point on the screen 38.
- the user state detection unit 32 detects the user position and the line-of-sight direction, and detects the position of the gazing point based on the detection result.
- detection of the user position, detection of the gaze direction, and detection of the position of the gazing point will be described in order.
- the user state detection unit 32 extracts an area in which a user is captured (hereinafter simply referred to as “user area”) in each of a plurality of images captured by the user detection camera 102. Thereafter, the user state detection unit 32 calculates the relative position between the user and the screen 38 (hereinafter simply referred to as “user position”) from the correspondence relationship between the user areas in the plurality of images using parallax based on stereo vision.
- 10A and 10B are diagrams for explaining a user position calculation method by the user state detection unit according to Embodiment 3 of the present invention.
- two user detection cameras 102 are installed in parallel with respect to the screen 38 as a set, separated from each other by a distance B. Then, the user state detection unit 32 extracts a user area in an image captured by each user detection camera 102. Then, the user state detection unit 32 calculates the distance D from the screen 38 based on the position shift of the corresponding user area between the images.
- the user state detection unit 32 stores, for example, an image captured by each camera in the absence of a user in advance, and is captured when the user appears in the shooting range (user detectable region).
- the user area is extracted by obtaining the difference between the image and the stored image.
- the user state detection unit 32 can also extract a user face area obtained by face image detection and face image matching as a user area.
- FIG. 10B is a diagram showing the principle of distance measurement by stereo vision used when determining the distance D between the user and the camera installation surface (screen 38) based on the positional relationship between corresponding user areas on two images.
- the image of the user whose position is to be measured is projected onto the imaging surfaces of the image sensors of the two user detection cameras 102 as shown in FIG. 10B.
- the user state detection unit 32 uses the focal length f of the camera and the distance B between the optical axes of the camera.
- a distance D between the screen 38 and the screen 38 can be calculated as shown in Equation (1).
- the user state detection unit 32 can obtain the user position in the direction parallel to the screen 38 based on the position of the user area in the image and the distance D calculated by the equation (1).
- the user state detection unit 32 detects and outputs the relative position of the user with respect to the screen 38 in this way.
- the user state detection unit 32 does not necessarily need to detect the user position based on stereo parallax.
- the user state detection unit 32 may detect the user position from distance information obtained by the principle of light wave distance measurement (Time of Flight).
- the user detection camera 102 includes a distance image sensor that outputs distance information using the principle of light wave distance, and at least one of the user detection cameras 102 may be installed.
- the user state detection unit 32 may detect the user position using a pressure value obtained from a floor pressure sensor installed on the floor surface in front of the screen 38. In this case, it is not necessary to install the user detection camera 102 for detecting the user position.
- the user state detection unit 32 detects the user's line-of-sight direction as described below.
- FIG. 11 is a flowchart showing a flow of processing relating to the detection of the line-of-sight direction in the third embodiment of the present invention.
- the user state detection unit 32 performs a result of the process (S510) for detecting the face orientation of the user and a process (S530) for detecting a relative gaze direction that is a relative gaze direction with respect to the face orientation. Based on the result, the line-of-sight direction is detected (S550).
- the user state detection unit 32 detects a face area from a user image that is photographed by the user detection camera 102 and exists in front of the screen 38 (S512). Next, the user state detection unit 32 applies the face part feature point area corresponding to each reference face direction to the detected face area, and cuts out an area image of each face part feature point (S514).
- the user state detection unit 32 calculates the degree of correlation between the clipped region image and the template image stored in advance (S516). Subsequently, the user state detection unit 32 obtains a weighted sum obtained by weighting and adding the angles indicated by the respective reference face orientations according to the calculated degree of correlation, and the user face orientation corresponding to the detected face region (S518).
- the user state detection unit 32 detects the user's face orientation by executing the processing of steps S512 to S518.
- the user state detection unit 32 detects the three-dimensional positions of the left and right eyes of the user using the images taken by the user detection camera 102 (S532). Subsequently, the user state detection unit 32 detects the three-dimensional position of the center of the left and right eyes of the user using the image captured by the user detection camera 102 (S534). Then, the user state detection unit 32 detects the relative line-of-sight direction using the line-of-sight direction reference plane obtained from the three-dimensional positions of the left and right eye heads and the three-dimensional position of the center of the left and right eyes (S536).
- the user state detection unit 32 detects the relative line-of-sight direction by executing the processing of steps S532 to S536.
- the user state detection unit 32 detects the user's line-of-sight direction using the user's face direction and the relative line-of-sight direction detected as described above.
- FIG. 12 is a diagram for explaining processing for detecting the face orientation of the user in the third embodiment of the present invention.
- the user state detection unit 32 obtains a facial part feature point area from a facial part area DB storing a facial part feature point area corresponding to each reference face direction. read out. Subsequently, as shown in FIG. 12B, the user state detection unit 32 applies the face part feature point area to the face area of the photographed image for each reference face direction, and the face part feature point area image. For each reference face orientation.
- the user state detection unit 32 calculates the correlation between the clipped region image and the template image held in the face part region template DB for each reference face direction.
- the user state detection unit 32 calculates a weight for each reference face direction according to the high degree of correlation indicated by the calculated degree of correlation. For example, the user state detection unit 32 calculates, as a weight, the ratio of the correlation degree of each reference face direction to the sum of the correlation degrees of the reference face direction.
- the user state detection unit 32 calculates the sum of values obtained by multiplying the calculated angle by the angle indicated by the reference face direction, and detects the calculation result as the user's face direction. To do.
- the weight for the reference face orientation +20 degrees is “0.85”
- the weight for 0 degrees is “0.14”
- the weight for ⁇ 20 degrees is “0.01”.
- the user state detection unit 32 calculates the degree of correlation for the area image of the facial part feature points, but it is not always necessary to target the area image of the facial part feature points.
- the user state detection unit 32 may calculate the degree of correlation for the entire face area image.
- a method for detecting the face orientation there is a method of detecting facial part feature points such as eyes, nose and mouth from the face image and calculating the facial orientation from the positional relationship of the facial part feature points.
- a method of calculating the face orientation from the positional relationship of the facial part feature points rotate and enlarge the 3D model of the facial part feature points prepared in advance so as to best match the facial part feature points obtained from one camera.
- the position of the facial part feature point positions in the left and right cameras is calculated using the principle of stereo vision based on the images taken by the two cameras.
- the user state detection unit 32 first detects the gaze direction reference plane, then detects the three-dimensional position of the center of the black eye, and finally detects the relative gaze direction.
- FIG. 13 is a diagram for explaining the reference direction reference plane.
- the user state detection unit 32 detects the gaze direction reference plane by detecting the three-dimensional positions of the left and right eye heads (end points on the nose side of the left eye and the right eye).
- the gaze direction reference plane is a plane that serves as a reference when detecting the relative gaze direction, and is the same as the left-right symmetric plane of the face as shown in FIG. It should be noted that the position of the head of the eye is less fluctuated due to facial expressions and less erroneously detected than the position of other facial parts such as the corner of the eye, the corner of the mouth or the eyebrows. Therefore, in the present embodiment, the user state detection unit 32 detects the gaze direction reference plane, which is a left-right symmetric plane of the face, using the three-dimensional position of the eye.
- the user state detection unit 32 detects the right and left eye regions using a face detection module and a face component detection module in each of two images taken simultaneously by the two user detection cameras 102. To do. Then, the user state detection unit 32 detects a three-dimensional position of each of the right and left eyes using a positional shift (parallax) between the images of the detected eye area. Furthermore, as shown in FIG. 13, the user state detection unit 32 detects a vertical bisector of a line segment with the detected three-dimensional positions of the left and right eyes as end points, as a gaze direction reference plane.
- 14 and 15 are diagrams for explaining detection of the center of the black eye.
- the person visually recognizes the object by the light from the object reaching the retina through the pupil, being converted into an electric signal, and the electric signal being transmitted to the brain. Therefore, the line-of-sight direction can be detected using the position of the pupil.
- the Japanese iris is black or brown, it is difficult to discriminate between the pupil and the iris by image processing.
- the center of the pupil and the center of the black eye substantially coincide. Therefore, in the present embodiment, the user state detection unit 32 detects the center of the black eye when detecting the relative line-of-sight direction.
- the user state detection unit 32 first detects the positions of the corners of the eyes and the eyes from the captured image. And the user state detection part 32 detects an area
- the user state detection unit 32 sets a black eye detection filter composed of a first area and a second area as shown in FIG. 15 at an arbitrary position in the black eye area. Then, the user state detection unit 32 searches for the position of the black eye detection filter that maximizes the inter-region variance between the luminance of the pixels in the first region and the luminance of the pixels in the second region, and the search result indicates The position is detected as the center of the black eye. Similarly to the above, the user state detection unit 32 detects the three-dimensional position of the center of the black eye using the shift of the position of the center of the black eye in the two images taken at the same time.
- the user state detection unit 32 detects the relative gaze direction using the detected gaze direction reference plane and the three-dimensional position of the center of the black eye. It is known that there is almost no individual difference in the diameter of an eyeball of an adult. For example, in the case of a Japanese, it is about 24 mm. Therefore, if the position of the center of the black eye when the direction of the reference (for example, the front) is known is known, it can be converted and calculated in the line-of-sight direction by obtaining the displacement from there to the current center position of the black eye.
- the direction of the reference for example, the front
- this method utilizes the fact that when facing the front, the midpoint of the line segment connecting the centers of the left and right black eyes exists on the center of the face, that is, on the line-of-sight direction reference plane. That is, the user state detection unit 32 detects the relative line-of-sight direction by calculating the distance between the midpoint of the line segment connecting the left and right black eye centers and the line-of-sight direction reference plane.
- the user state detection unit 32 uses the eyeball radius R and the distance d between the midpoint of the line segment connecting the left and right black eye centers and the gaze direction reference plane as shown in Expression (2).
- the rotation angle ⁇ in the left-right direction with respect to the face direction is detected as the relative line-of-sight direction.
- the user state detection unit 32 detects the relative line-of-sight direction using the line-of-sight reference plane and the three-dimensional position of the center of the black eye. And the user state detection part 32 detects a user's gaze direction using the detected face direction and relative gaze direction.
- the user state detection unit 32 detects the position of the user's gazing point on the plane including the screen, using the user position detected as described above and the user's line-of-sight direction. Specifically, the user state detection unit 32 detects the position of the user's gazing point by calculating the intersection of a straight line extending in the line-of-sight direction from the user position and a plane including the screen.
- the user state detection unit 32 detects the position of the user's gazing point as the user state, and further detects the user's face direction detected when detecting the position of the gazing point as the user state. Good. Furthermore, the user state detection unit 32 may detect the user's posture as the user state.
- the concentration level estimation unit 33 uses the user state detected by the user state detection unit 32 to estimate the concentration level for each user who is the viewing target of the display content identified by the user identification unit 31. Specifically, the concentration degree estimation unit 33 may calculate the user's concentration degree based on, for example, the distribution of the user's face orientation every predetermined time. In addition, the concentration degree estimation unit 33 may calculate the user's concentration degree based on the distribution of the user's gaze point for each predetermined time. Further, the concentration degree estimation unit 33 may calculate the user's concentration degree based on the user's posture.
- the user information database 34 stores various types of information shown in FIGS. 16A to 16C.
- FIGS. 16A to 16C are diagrams showing an example of a user information database according to Embodiment 3 of the present invention.
- the user information database 34 stores basic attribute information, body characteristic information, and cognitive characteristic information shown in FIGS. 16A to 16C. Specifically, as shown in FIG. 16A, the user information database 34 stores basic attribute information such as name, sex, age, date of birth, and relationship in association with an ID for identifying the user. ing.
- the user information database 34 is associated with an ID for identifying a user, and height and eye height in a standing position, height and eye height in a sitting position, dominant hand, It stores body feature information such as body shape and viewing ability according to the user's posture, such as dominant eye, visual acuity, and hearing ability.
- the user information database 34 stores cognitive characteristic information for each user. Specifically, the user information database 34 stores cognitive characteristic information that associates time, concentration, and effective visual field area information for each user. In the cognitive characteristic information shown in FIG. 16C, a viewing angle is stored as effective viewing area information.
- the user information database 34 further displays the contents of the display content (terrestrial broadcast 5ch drama, photo browsing application, etc.), and the positional relationship of people around (“HG003 (0.4, 0.6)”, etc. ) May be stored in association with the degree of concentration.
- “HG003 (0.4, 0.6)” indicates that the user whose ID is HG003 is located at a position separated by 0.4 m in the x coordinate direction and 0.6 m in the y coordinate direction.
- the notification source 106 outputs the notification information to the information display device 30.
- Examples of the notification source 106 include a home appliance group 105 connected to a home network such as a refrigerator, a washing machine, and a microwave oven, a network camera 104, and a mobile phone 103 as shown in FIG.
- FIGS. 17A to 17C are diagrams illustrating an example of notification information according to Embodiment 3 of the present invention.
- a notification icon as shown in FIG. 17A
- a notification text as shown in FIG. 17B
- image information or text information such as a thumbnail image / video as shown in FIG.
- notification information is not limited to this example.
- Various types of information can be used as notification information, such as notification of device status or work progress, incoming e-mail, or schedule notification.
- the relevance level estimation unit 35 calculates a relevance level r indicating the relevance level between the notification information and the display content. A method for calculating the relevance will be described later.
- the application control unit 36 uses the user identification information output from the user identification unit 31, the user state output from the user state detection unit 32, and the user concentration level output from the concentration level estimation unit 33 as input information. Do. In addition to the input information, the application control unit 36 performs display control using the input information output from the user information database 34, the notification source 106, and the relevance estimation unit 35.
- the application control unit 36 When updating the drawing content of the screen, the application control unit 36 outputs update information of the drawing content to the screen drawing unit 37.
- the screen drawing unit 37 presents the drawing content on the screen 38.
- 18A and 18B are diagrams for explaining a use scene of the information display device according to Embodiment 3 of the present invention.
- the information display device 30 displays an icon indicating notification information output from the device.
- the area S is displayed at the initial display position on the screen. This initial display position is a position away from the first target position by a distance d1, and is a position away from the second target position by a distance d2.
- the gaze point distribution center 41 is the center position of the gaze point distribution composed of a plurality of gaze points detected every predetermined time.
- the current gazing point 42 is the position of the user's gazing point detected most recently.
- the information display device 30 gives the user a more sudden impression than necessary by gradually approaching the icon from the initial display position toward the display area of the display content viewed by the user at the speed v. Display notification information so that there is no.
- the target position to which the icon approaches consists of a first target position and a second target position.
- the information display device 30 first moves the icon from the initial display position toward the first target position. And even if an icon arrives at a 1st target position, if a user does not point a gaze at the icon continuously for a predetermined time or more, information display device 30 will aim an icon from the 1st target position to the 2nd target position. To approach.
- the first target position is a position away from the boundary of the main content display area by a predetermined distance ⁇ d. That is, the first target position is a position outside the main content display area and in the vicinity of the boundary of the main content display area.
- the predetermined distance ⁇ d is preferably half the width of the icon indicating the notification information.
- the second target position is a predetermined position in the content display area or a position representing a plurality of gazing points detected within a predetermined period. Specifically, the second target position is the center of the display area of the display content as shown in FIG. 18A or the gaze point distribution center 41 of the user as shown in FIG. 18B.
- the second target position does not necessarily need to be the center of the display area of the display content or the gaze distribution center 41 of the user.
- the second target position may be the center of the image displayed in a part of the display area of the display content.
- the second target position may be the center of gravity of the gaze point distribution.
- FIG. 19 is a flowchart showing the flow of processing of the information display device in Embodiment 3 of the present invention.
- FIG. 20 is a figure for demonstrating the operation example of the information display apparatus in Embodiment 3 of this invention.
- FIG. 20 there are a user A and a user B in front of the information display device 30, and a news program is displayed as display content.
- the viewing user of this display content is the user A, and the user B is the viewing user.
- FIG. 20A when the user A is watching a news program, as shown in FIG. 20B, an icon indicating notification information from the microwave is displayed at the initial display position. .
- the icon gradually approaches the first target position.
- FIG. 20D when the user's line of sight is poured into the icon, for example, detailed information “cooking completed” linked to the icon is displayed.
- the user identification unit 31 performs user identification by collating with body feature information stored in the user information database 34 registered in advance (S301). Then, the user state detection unit 32 detects the user position for each identified user (S302). Further, the user state detection unit 32 detects the face direction and the line-of-sight direction for each identified user (S303). Furthermore, the user state detection unit 32 detects and holds the current gazing point 42 based on the user position and the line-of-sight direction.
- the user state detection unit 32 determines the viewing user of the display content (S304). For example, the user state detection unit 32 may determine a user who is within a predetermined distance from the display content as a viewing user. In addition, it is preferable that the user state detection part 32 discriminate
- the concentration degree estimation unit 33 calculates the center position (gaze point distribution center 41) of the gaze point distribution composed of a plurality of gaze points detected every predetermined time (S305). Then, the concentration degree estimation unit 33 calculates the degree of concentration c according to the following equation (3) using the calculated variance ⁇ of the distance between the gazing point distribution center 41 and the position of each gazing point (S306).
- the degree of concentration c means that the larger the value is, the more concentrated the user is on the display content.
- the concentration degree estimation unit 33 sets the allowable notification strength Int according to the following equation (4).
- the permissible notification strength indicates the strength at which notification information should be noticed by the user.
- the allowable notification intensity is high, the information display device 30 needs to perform “noticeable information notification”.
- the allowable notification intensity is low, the information display device 30 performs “subtle information notification” so that the user can notice the notification information without giving a sudden impression to the notification information. There is a need.
- n is the gain
- the allowable notification intensity Int is smaller when the degree of concentration c is smaller, that is, when the user is not concentrated on the display content.
- the relevance estimation unit 35 calculates a relevance r between the notification information and the display content (S308).
- the relevance r is a numerical value between 0 and 1.
- the relevance estimation unit 35 calculates the relevance to be high if the notification information is related to the program content, and relates to the relevance if the notification information is not related to the program content.
- the degree is calculated to be low.
- the relevance may be set as a binary value such that 0 is low and 1 is high.
- the relevance level estimation unit 35 weakens the allowable notification strength according to the value of the relevance level r (S310). If the degree of relevance r is equal to or greater than a preset threshold value (No in S309), the degree of relevance estimation unit 35 increases the allowable notification strength according to the value of the degree of relevance r (S311).
- the application control unit 36 determines the display parameter of the notification information using the allowable notification strength Int (S312).
- the display parameter here is information indicating the initial display position, the size, and the approach method to the first target position or the second target position of the notification information.
- the application control unit 36 determines display parameters so that the user does not have a sudden impression of the notification information.
- the distance di between the initial display position of the notification information and the target position is calculated according to the following equation (5).
- gd is a gain
- d0 is a predetermined constant value
- the moving speed v of the notification information and the display area S of the notification information are calculated according to the following equations (6) and (7).
- gv and gS are gains, and v0 and S0 are predetermined constant values.
- the application control unit 36 determines the initial display position so that the smaller the estimated concentration c or relevance r is, the farther from the position determined by the position of the detected gaze point. In addition, the application control unit 36 determines the moving speed v so that the moving speed v increases as the estimated concentration degree c or relevance degree r increases. In addition, the application control unit 36 determines the size of the display area (display area S) such that the larger the estimated concentration c or relevance r is, the larger the display area is.
- the screen drawing unit 37 displays the notification information on the screen 38 according to the display parameters (S313). Then, as shown in FIG. 20 (d), when attention is directed to the notification information displayed by the user, that is, when the line of sight is continuously directed to the notification information for a predetermined time or longer, the screen drawing unit 37 displays the notification information. Detailed information is displayed on the screen 38. That is, the application control unit 36 determines whether or not the state where the distance between the gazing point and the display position of the notification information is less than the threshold continues for a predetermined time or more. Then, the screen drawing unit 37 changes the content of the notification information when it is determined that the notification is continued. That is, the notification information is initially displayed in the form of an icon, and is displayed in a form in which predetermined information linked to the icon can be confirmed after the user's attention is directed.
- FIG. 21 is a diagram for explaining an operation example of the information display device according to the third embodiment of the present invention. Specifically, FIG. 21 is a diagram for explaining a method for updating the user information database.
- the screen drawing unit 37 moves the display position of the notification information toward the gazing point distribution center 41 of the user A.
- the application control unit 36 determines the distance d between the current gazing point 42 and the gazing point distribution center 41 when the user's current gazing point 42 moves to the display position of the notification information. Is calculated.
- the application control unit 36 estimates the inside of the circle with the radius d centering on the gazing point distribution center 41 as the effective visual field region, and updates the effective visual field region information stored in the user information database 34.
- the application control part 36 calculates the viewing angle k according to following Formula (11).
- the application control unit 36 updates the user information database 34 shown in FIG. 16C.
- the information display device 30 in the present embodiment can achieve the same effects as the information display device 10 or 20 of the first embodiment or the second embodiment.
- the information display device 30 can move the display position of the notification information toward the first target position or the second target position, the notification information can be presented to the user without giving a more sudden impression. it can.
- the information display device 30 can identify the user, the initial display position can be determined with higher accuracy by using the user information database 34 in which information is stored for each user.
- the information display device 30 further determines the display parameter in consideration of the importance u indicating the importance of the notification information or the urgency u indicating the urgency of the notification information.
- FIG. 22A is a block diagram showing a functional configuration of the information display device in Modification 1 of Embodiment 3 of the present invention.
- the information display device 30 in the present modification further includes an importance / emergency level acquisition unit 39 in addition to the components included in the information display device in the third embodiment.
- the importance / emergency level acquisition unit 39 acquires the importance level indicating the importance level of the notification information or the urgency level indicating the urgency level of the notification information. Specifically, the importance / emergency level acquisition unit 39 acquires the importance level or the urgency level of the notification information from the notification source 106 that outputs the notification information, for example. Further, for example, the importance / emergency level acquisition unit 39 reads the importance level or the urgency level stored in association with the notification source or the type of the notification information that has output the notification information, so that the importance level of the notification information or The urgency level may be acquired.
- FIG. 22B is a flowchart showing a process flow of the information display apparatus in the first modification of the third embodiment of the present invention.
- FIG. 22B includes a step of acquiring the importance or urgency of the notification information and adjusting the allowable notification intensity according to the acquired importance in the flowchart of FIG.
- the information display device 30 can perform notification control based on the importance or urgency of the notification information. That is, when the importance or urgency of the notification information is high, the information display device 30 can increase the permissible notification intensity so that the notification information is noticed in preference to the main content.
- the application control unit 36 determines the initial display position so that the smaller the acquired importance u or urgency u is, the farther from the position determined by the detected gaze point. Further, the application control unit 36 determines the moving speed v so that the moving speed v increases as the acquired importance u or urgency u increases. In addition, the application control unit 36 determines the size of the display area (display area S) so that the larger the acquired importance u or urgency u is, the larger the display area is.
- FIG. 23A and FIG. 23B are diagrams for explaining an operation example of the information display device in the first modification of the third embodiment of the present invention.
- the notification information is a work completion notification from the washing machine
- the urgency of the notification information is not high, so the information display device 30 weakens the allowable notification strength.
- the distance between the initial display position of the notification information and the target position is set longer than a predetermined value, and “subtle information notification” becomes possible.
- the notification information is a notification that informs the customer
- the urgency of the notification information is high, so the information display device 30 increases the allowable notification strength.
- the distance between the initial display position of the notification information and the target position is set shorter than a predetermined value, and “noticeable information notification” becomes possible.
- 24A to 24C are diagrams for explaining an outline of a display area control method based on the user position in the second modification of the third embodiment of the present invention.
- the application control unit 36 determines the display position on the screen of the main content to be presented to the user based on the user position detected by the user state detection unit 32, for example. As a result, as shown in FIGS. 24A and 24B, the information display device 30 can continue to present the main content at a position that can be easily viewed by the user even when the user moves.
- the application control unit 36 can display information at a position on the screen close to the position of the user.
- the application control unit 36 displays the information in a size that is easy for the user to view by reducing or enlarging the display size of the information. Can do.
- the application control unit 36 changes the information displayed as the main content to more detailed information. It is preferable.
- the application control unit 36 displays the main content at a height that is easy for each user to see according to the height position of each user's face obtained by referring to the user information database 34. be able to.
- FIG. 25 is a diagram for explaining an operation example of the information display device in the second modification of the third embodiment of the present invention.
- FIG. 25 shows an example in which the display position of the main content follows the user position when the user passes in front of the screen.
- the application control unit 36 moves the target position of the notification information as the user moves. That is, when the display position of the main content moves, the application control unit 36 moves the target position of the notification information following the movement of the display position of the main content.
- the information display device 30 according to the present modification can make the user aware that the notification information is displayed even when the display position of the main content is changed.
- Modification 3 of Embodiment 3 Next, Modification 3 of Embodiment 3 will be described. In this modification, a case where a plurality of viewing users are determined will be described.
- FIG. 26A and FIG. 26B are diagrams for explaining an operation example of the information display device in the third modification of the third embodiment of the present invention.
- 26A and 26B show a case where a plurality of viewing users are in front of the screen 38.
- FIG. 26A and 26B show a case where a plurality of viewing users are in front of the screen 38.
- FIG. 26A is a diagram illustrating a state in which display contents associated with each of the user A and the user B are displayed on the screen.
- FIG. 26B is a diagram showing a state where display content common to user A and user B is displayed on the screen.
- the information display device 30 applies the process shown in FIG. 19 or 22B to each of the user A and the user B.
- the information display device 30 selects a user having a higher degree of concentration among the users A and B, and calculates an allowable notification intensity based on the degree of concentration of the selected user.
- the information display device 30 may calculate the allowable notification strength based on the average value of the user A concentration level and the user B concentration level.
- the information display device 30 is configured to display the initial notification information so that the information display device 30 is located outside the effective visual field region according to the degree of concentration of the viewing user even when there are a plurality of viewing users.
- the display position can be determined.
- Modification 4 of Embodiment 3 Next, Modification 4 of Embodiment 3 will be described. In the present modification, a description will be given of how the information display device 30 changes the display method of the notification information according to the size of the area where the main content is not displayed.
- FIGS. 27A to 27C are diagrams for explaining an operation example of the information display device in the fourth modification of the third embodiment of the present invention.
- FIG. 27A to FIG. 27C the width of the area that can be used for displaying the notification information is represented by w1 and w2, and the distance between the initial display position of the notification information and the first target position located near the boundary of the display area of the main content is shown.
- d1 the width of the area that can be used for displaying the notification information
- FIG. 27A shows “width w1 ⁇ distance d1, width w2> distance d1”
- FIG. 27B shows “width w1 ⁇ distance d1, width w2> distance d1”
- FIG. 27C shows “width w1, The state of width w2 ⁇ distance d1 ”is shown.
- the information display device 30 may display the notification information in either or both of the area A and the area B. However, when the relationship of “width w1 ⁇ distance d1, width w2> distance d1” is satisfied as shown in FIG. 27B, the information display device 30 displays the notification information in the region B having a width longer than the distance d1. . Further, as shown in FIG. 27C, when “width w1, width w2 ⁇ distance d1”, the widths of both the area A and the area B are shorter than the distance d1, and therefore the information display device 30 has the longer area B. Display notification information on.
- the information display device 30 changes the display mode of the notification information after display in accordance with the allowable notification intensity.
- FIG. 28A and FIG. 28B are diagrams for explaining an operation example of the information display device in the fifth modification of the third embodiment of the present invention.
- the information display device 30 When the allowable notification intensity is higher than the threshold, the information display device 30 approaches the target position while keeping the size of the image indicated by the notification information constant as shown in FIG. 28A. On the other hand, when the allowable notification intensity is lower than the threshold, the information display device 30 moves the notification information closer to the target position while gradually increasing the notification information, as shown in FIG. 28B.
- the information display device 30 can change the display mode of the notification information after display based on the degree of user concentration, the degree of association between the notification information and the display content, or the importance or urgency of the notification information. Since it can be changed, a more casual impression can be given to the user.
- Modification 6 of Embodiment 3 Next, Modification 6 of Embodiment 3 will be described. In this modification, an operation after the notification information display position is moved to the target position will be described.
- FIG. 29A, FIG. 29B, FIG. 30A and FIG. 30B are diagrams for explaining an operation example of the information display device in the sixth modification of the third embodiment of the present invention.
- the screen drawing part 37 changes at least one of the display position and display mode of notification information so that a user may become hard to notice notification information.
- the screen drawing unit 37 is in a state where the display position of the notification information is within a predetermined range from the target position when the importance or urgency of the notification information is not high. If the user's attention does not increase beyond a predetermined level even after a predetermined time elapses, the display size of the notification information is gradually reduced and hidden. Further, for example, as illustrated in FIG. 29B, the screen drawing unit 37 may gradually move the notification information away from the target position so that the notification information moves away from the effective view area.
- the screen drawing unit 37 keeps the user's eyes on the notification information continuously for a predetermined time or more even after a certain time has elapsed since the notification information reached the target position. If not, the notification information is moved away from the target position by a predetermined distance.
- the notification information gradually moves away from the effective visual field area of the user and moves to a position where it is difficult to draw attention, so that the user's attention to the display content can be prevented more than necessary.
- Modification 7 of Embodiment 3 Next, Modification 7 of Embodiment 3 will be described. In this modification, a case will be described in which the information display device 30 presents a plurality of pieces of notification information to the user at the same time.
- FIG. 31A to FIG. 31C are diagrams for explaining an operation example of the information display device in the modified example 7 of the third embodiment of the present invention.
- the present invention can be applied to a case where a plurality of pieces of notification information are simultaneously presented to the user as shown in FIGS. 31A to 31C.
- the information display device 30 causes a plurality of pieces of notification information to approach the target position from one of the left and right sides of the display area of the main content.
- the information display device 30 causes a plurality of pieces of notification information to approach the target position from both the left and right sides of the display area of the main content.
- the information display device 30 determines display parameters such as the initial display position and speed of the notification information in accordance with the degree of association between each piece of notification information and display content.
- Modification 8 of Embodiment 3 Next, Modification 8 of Embodiment 3 will be described. In this modification, a case where main content is displayed in the entire area of the screen will be described.
- FIG. 32 is a diagram for explaining an operation example of the information display device in the modified example 8 of the third embodiment of the present invention.
- the main content has been displayed in a partial area of the screen has been described so far, but the main content may be displayed on the entire screen as shown in FIG.
- the present invention is applicable.
- FIG. 32A a description will be given by taking as an example main content in which a plurality of objects are displayed on the entire screen.
- FIG. 32 (a) shows a situation where a fish is swimming on the screen or a wallpaper image is displayed on the entire screen. Therefore, as illustrated in FIG. 32B, the information display device 30 displays the notification information at the initial display position, and causes the displayed notification information to approach the target position. Then, after detecting that the user's attention is directed to the notification information as shown in FIG. 32 (c), the information display device 30 is associated with the notification information in advance as shown in FIG. 32 (d). Display secondary content on the screen.
- the information display device 30 displays a menu screen that asks the user whether to display the sub-contents associated with the notification information on the screen 38 after detecting that the user's attention is directed to the notification information. Also good. In that case, the information display device displays the sub-contents on the displayed menu screen as shown in FIG. 32 (d) when the user explicitly designates the sub-contents to be displayed by the remote controller or the gesture. 38 may be displayed.
- the first target position is set to a position separated by a predetermined distance ⁇ d from the boundary of the display area of the main content as shown in FIGS. 18A and 18B.
- the boundary of the display area of the main content becomes a frame of the screen 38 and is set at a position outside the screen. Therefore, when the main content is displayed on the entire screen, the first target position is a position away from the user's gaze point distribution center 41 or the center of the main content display area (the center of the screen 38) by a predetermined distance. It is preferable to set to.
- the information display device has been described based on the embodiments and the modifications thereof, but the present invention is not limited to these embodiments or the modifications thereof. Unless it deviates from the gist of the present invention, various modifications conceived by those skilled in the art are applied to the present embodiment or the modified examples thereof, or a form constructed by combining the components in the different embodiments or modified examples thereof. It is included within the scope of the present invention.
- the information display device includes a screen such as a plasma display panel or a liquid crystal panel, but it is not always necessary to include a screen.
- the information display device may be a projector that projects content onto a projection surface such as a screen or a wall surface.
- Each of the above devices is specifically a computer system including a microprocessor, a ROM, a RAM, a hard disk unit, a display unit, a keyboard, a mouse, and the like.
- a computer program is stored in the RAM or the hard disk unit.
- Each device achieves its function by the microprocessor operating according to the computer program.
- the computer program is configured by combining a plurality of instruction codes indicating instructions for the computer in order to achieve a predetermined function.
- Each device is not limited to a computer system including a microprocessor, a ROM, a RAM, a hard disk unit, a display unit, a keyboard, a mouse, and the like, but may be a computer system including a part of them.
- a part or all of the constituent elements constituting each of the above devices may be constituted by one system LSI (Large Scale Integration).
- the system LSI is a super multifunctional LSI manufactured by integrating a plurality of components on one chip, and specifically, a computer system including a microprocessor, a ROM, a RAM, and the like. .
- a computer program is stored in the RAM.
- the system LSI achieves its functions by the microprocessor operating according to the computer program.
- system LSI may be called IC, LSI, super LSI, or ultra LSI depending on the degree of integration.
- method of circuit integration is not limited to LSI's, and implementation using dedicated circuitry or general purpose processors is also possible.
- An FPGA Field Programmable Gate Array
- a reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
- a part or all of the constituent elements constituting each of the above devices may be constituted by an IC card or a single module that can be attached to and detached from each device.
- the IC card or the module is a computer system including a microprocessor, a ROM, a RAM, and the like.
- the IC card or the module may include the super multifunctional LSI described above.
- the IC card or the module achieves its function by the microprocessor operating according to the computer program. This IC card or this module may have tamper resistance.
- the present invention may be a method in which an operation of a characteristic component included in the information display device described above is used as a step. Further, the present invention may be a computer program that realizes these methods by a computer, or may be a digital signal composed of the computer program.
- the present invention also provides a computer-readable recording medium such as a flexible disk, hard disk, CD-ROM, MO, DVD, DVD-ROM, DVD-RAM, BD (Blu-ray Disc). (Registered trademark)) or a semiconductor memory. Further, the present invention may be the computer program or the digital signal recorded on these recording media.
- a computer-readable recording medium such as a flexible disk, hard disk, CD-ROM, MO, DVD, DVD-ROM, DVD-RAM, BD (Blu-ray Disc). (Registered trademark)) or a semiconductor memory.
- the present invention may be the computer program or the digital signal recorded on these recording media.
- the computer program or the digital signal may be transmitted via an electric communication line, a wireless or wired communication line, a network represented by the Internet, a data broadcast, or the like.
- the present invention may also be a computer system including a microprocessor and a memory.
- the memory may store the computer program, and the microprocessor may operate according to the computer program.
- the program or the digital signal is recorded on the recording medium and transferred, or the program or the digital signal is transferred via the network or the like, and is executed by another independent computer system. It is good.
- the information display device initially displays the notification information outside the effective visual field area of the user, the notification information can be noticed without giving a sudden impression to the user.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Analytical Chemistry (AREA)
- Chemical & Material Sciences (AREA)
- Neurosurgery (AREA)
- Biophysics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Databases & Information Systems (AREA)
- User Interface Of Digital Computer (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
(実施の形態1)
図1は、本発明の実施の形態1における情報表示装置の外観を示す図である。また、図2は、本発明の実施の形態1における情報表示装置の機能構成を示すブロック図である。
次に、本発明の実施の形態2について、実施の形態1と異なる点を中心に図面を参照しながら説明する。本実施の形態における情報表示装置20は、ユーザ情報データベース23を参照して初期表示位置を決定する点が、実施の形態1における情報表示装置10と主として異なる。
本実施の形態における情報表示装置30は、1以上のユーザにより視聴される大画面ディスプレイに適用されるのに好適であり、表示コンテンツに対するユーザの視聴状況に基づいて通知情報の表示制御を行う。
次に、実施の形態3の変形例1について説明する。本変形例において、情報表示装置30は、さらに、通知情報の重要度合いを示す重要度u、又は通知情報の緊急度合いを示す緊急度uを考慮して、表示パラメータを決定する。
次に、実施の形態3の変形例2について説明する。本変形例では、主コンテンツの表示位置あるいは大きさなどがユーザの動きに応じて変更される場合について説明する。
次に、実施の形態3の変形例3について説明する。本変形例では、複数の視聴ユーザが判別された場合について説明する。
次に、実施の形態3の変形例4について説明する。本変形例では、情報表示装置30が、主コンテンツが表示されていない領域の大きさに応じて、通知情報の表示方法を変更することについて説明する。
次に、実施の形態3の変形例5について説明する。本変形例では、情報表示装置30は、許容通知強度の高さに応じて、表示後の通知情報の表示態様を変化させる。
次に、実施の形態3の変形例6について説明する。本変形例では、通知情報の表示位置をターゲット位置に移動させた後の動作について説明する。
次に、実施の形態3の変形例7について説明する。本変形例では、情報表示装置30が同時に複数の通知情報をユーザに提示する場合について説明する。
次に、実施の形態3の変形例8について説明する。本変形例では、主コンテンツが画面の全領域に表示されている場合について説明する。
11、32 ユーザ状態検出部
12、33 集中度推定部
13、21、36 アプリケーション制御部
14、22、37 画面描画部
23、34 ユーザ情報データベース
31 ユーザ識別部
35 関連度推定部
38 画面
39 重要度/緊急度取得部
41 注視点分布中心
42 現在注視点
101 アンテナ
102 ユーザ検出カメラ
103 携帯電話
104 ネットワークカメラ
105 家電機器群
106 通知ソース
107 ルータ/ハブ
Claims (20)
- ユーザに通知するための情報である通知情報を画面に表示する情報表示装置であって、
前記ユーザの物理的な状態を示すユーザ状態を検出するユーザ状態検出部と、
前記ユーザ状態検出部により検出されたユーザ状態に基づいて、前記画面に対する前記ユーザの集中度合いを示す集中度を推定する集中度推定部と、
前記ユーザによって視認される領域である有効視野領域の外部に初期表示位置が位置するように、前記集中度推定部により推定された集中度に基づいて通知情報の初期表示位置を決定するアプリケーション制御部と、
前記アプリケーション制御部により決定された初期表示位置に通知情報を表示し、かつ、表示後の通知情報の表示位置及び表示態様の少なくとも一方を変化させる画面描画部とを備える
情報表示装置。 - 前記ユーザ状態検出部は、前記画面を含む平面上における前記ユーザの注視点の位置を、前記ユーザ状態として検出し、
前記アプリケーション制御部は、前記集中度推定部により推定された集中度が小さいほど、前記ユーザ状態検出部により検出された注視点の位置により定まる位置から遠くなるように、前記初期表示位置を決定する
請求項1に記載の情報表示装置。 - 前記アプリケーション制御部は、さらに、前記集中度推定部により推定された集中度が大きいほど移動速度が大きくなるように移動速度を決定し、
前記画面描画部は、前記アプリケーション制御部により決定された移動速度で通知情報の表示位置を移動させることにより、通知情報の表示位置を変化させる
請求項1に記載の情報表示装置。 - 前記ユーザ状態検出部は、前記画面を含む平面上におけるユーザの注視点の位置を、前記ユーザ状態として検出し、
前記画面描画部は、前記ユーザ状態検出部により所定期間内に検出された複数の注視点の位置を代表する位置に向けて通知情報の表示位置を移動させることにより、通知情報の表示位置を変化させる
請求項1に記載の情報表示装置。 - 前記画面描画部は、前記画面に表示されたコンテンツの表示領域内の所定位置に向けて通知情報の表示位置を移動させることにより、通知情報の表示位置を変化させる
請求項1に記載の情報表示装置。 - 前記画面描画部は、前記画面に表示されたコンテンツの表示領域外の位置であって、かつ前記コンテンツの表示領域の境界近傍の位置に向けて、通知情報の表示位置を移動させることにより、通知情報の表示位置を変化させる
請求項1に記載の情報表示装置。 - 前記アプリケーション制御部は、さらに、前記集中度推定部により推定された集中度が大きいほど表示領域が大きくなるように表示領域の大きさを決定し、
前記画面描画部は、前記アプリケーション制御部により決定された初期表示位置に通知情報を表示する際に、決定された大きさの表示領域に通知情報を表示する
請求項1に記載の情報表示装置。 - さらに、
前記画面に表示されたコンテンツと前記通知情報との関連度合いを示す関連度を推定する関連度推定部を備え、
前記ユーザ状態検出部は、前記画面を含む平面上におけるユーザの注視点の位置を、前記ユーザ状態として検出し、
前記アプリケーション制御部は、前記関連度推定部により推定された関連度が小さいほど、前記ユーザ状態検出部により検出された注視点の位置により定まる位置から初期表示位置が遠くなるように、前記初期表示位置を決定する
請求項1に記載の情報表示装置。 - 前記アプリケーション制御部は、さらに、前記関連度推定部により推定された関連度が大きいほど移動速度が大きくなるように移動速度を決定し、
前記画面描画部は、決定された移動速度で通知情報の表示位置を移動させることにより、通知情報の表示位置を変化させる
請求項8に記載の情報表示装置。 - さらに、
前記通知情報の重要度合いを示す重要度、又は前記通知情報の緊急度合いを示す緊急度を取得する重要度/緊急度取得部を備え、
前記アプリケーション制御部は、前記重要度/緊急度取得部により取得された重要度又は緊急度が小さいほど、前記ユーザ状態検出部により検出された注視点の位置により定まる位置から初期表示位置が遠くなるように、前記初期表示位置を決定する
請求項1に記載の情報表示装置。 - 前記アプリケーション制御部は、さらに、前記重要度/緊急度取得部により取得された重要度又は緊急度が大きいほど移動速度が大きくなるように移動速度を決定し、
前記画面描画部は、決定された移動速度で通知情報の表示位置を移動させることにより、通知情報の表示位置を変化させる
請求項10に記載の情報表示装置。 - 前記ユーザ状態検出部は、前記画面を含む平面上におけるユーザの注視点の位置を、前記ユーザ状態として検出し、
前記集中度推定部は、前記ユーザ状態検出部により所定期間内に検出された複数の注視点からなる注視点分布に基づいて、前記集中度を推定する
請求項1に記載の情報表示装置。 - 前記ユーザ状態検出部は、前記画面を含む平面上におけるユーザの注視点の位置を、前記ユーザ状態として検出し、
前記集中度推定部は、前記ユーザ状態検出部により検出された注視点の移動量に基づいて、前記集中度を推定する
請求項1に記載の情報表示装置。 - 前記ユーザ状態検出部は、ユーザの顔向きを前記ユーザ状態として検出し、
前記集中度推定部は、前記ユーザ状態検出部により所定期間内に検出された複数の顔向きからなる顔向き分布に基づいて、前記集中度を推定する
請求項1に記載の情報表示装置。 - 前記ユーザ状態検出部は、ユーザの姿勢を前記ユーザ状態として検出し、
前記集中度推定部は、前記ユーザ状態検出部により検出された姿勢に基づいて、前記集中度を推定する
請求項1に記載の情報表示装置。 - さらに、
集中度と有効視野領域の大きさを示す有効視野領域情報とを対応付けて格納しているユーザ情報データベースとを備え、
前記ユーザ状態検出部は、前記画面を含む平面上におけるユーザの注視点の位置を、前記ユーザ状態として検出し、
前記アプリケーション制御部は、前記ユーザ情報データベースを参照することにより、前記集中度推定部により推定された集中度に対応する有効視野領域情報を取得し、取得した有効視野領域情報と、前記ユーザ状態検出部により検出された注視点とを用いて推定される有効視野領域の外部に前記初期表示位置を決定する
請求項1に記載の情報表示装置。 - 前記アプリケーション制御部は、さらに、前記画面描画部によって通知情報の表示位置が変化させられている間に、通知情報の表示位置とユーザの注視点の位置との距離が閾値未満であるか否かを判定し、距離が閾値未満であると判定された場合に、当該表示位置を用いて、前記ユーザ情報データベースに格納された有効視野領域情報を更新する
請求項16に記載の情報表示装置。 - さらに、
前記画面の前方にいるユーザを識別するユーザ識別部を備え、
前記ユーザ情報データベースは、ユーザごとに、集中度と有効視野領域の大きさを示す有効視野領域情報とを対応付けて格納し、
前記アプリケーション制御部は、前記ユーザ識別部により識別されたユーザに対応する有効視野領域情報を取得する
請求項17に記載の情報表示装置。 - ユーザに通知するための情報である通知情報を画面に表示する情報表示方法であって、
ユーザの物理的な状態を示すユーザ状態を検出するユーザ状態検出ステップと、
前記ユーザ状態検出ステップにおいて検出されたユーザ状態に基づいて、前記画面に対するユーザの集中度合いを示す集中度を推定する集中度推定ステップと、
前記ユーザによって視認される領域である有効視野領域の外部に初期表示位置が位置するように、前記集中度推定ステップにおいて推定された集中度に基づいて通知情報の初期表示位置を決定するアプリケーション制御ステップと、
前記アプリケーション制御ステップにおいて決定された初期表示位置に通知情報を表示し、かつ、表示後の通知情報の表示位置及び表示態様の少なくとも一方を変化させる画面描画部とを含む
情報表示方法。 - 請求項19に記載の情報表示方法をコンピュータに実行させるためのプログラム。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP10738331.7A EP2395420B1 (en) | 2009-02-05 | 2010-02-02 | Information display device and information display method |
US13/143,861 US20110267374A1 (en) | 2009-02-05 | 2010-02-02 | Information display apparatus and information display method |
JP2010549388A JP5286371B2 (ja) | 2009-02-05 | 2010-02-02 | 情報表示装置及び情報表示方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2009-024662 | 2009-02-05 | ||
JP2009024662 | 2009-02-05 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010089989A1 true WO2010089989A1 (ja) | 2010-08-12 |
Family
ID=42541900
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/000595 WO2010089989A1 (ja) | 2009-02-05 | 2010-02-02 | 情報表示装置及び情報表示方法 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20110267374A1 (ja) |
EP (1) | EP2395420B1 (ja) |
JP (1) | JP5286371B2 (ja) |
WO (1) | WO2010089989A1 (ja) |
Cited By (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012114557A (ja) * | 2010-11-22 | 2012-06-14 | Nec Saitama Ltd | 表示装置、表示制御方法、プログラム及び電子機器 |
JP2013073529A (ja) * | 2011-09-28 | 2013-04-22 | Kyocera Corp | 装置、方法、及びプログラム |
WO2013057878A1 (ja) * | 2011-10-19 | 2013-04-25 | パナソニック株式会社 | 表示制御装置、集積回路、表示制御方法およびプログラム |
WO2015029554A1 (ja) * | 2013-09-02 | 2015-03-05 | ソニー株式会社 | 情報処理装置、情報処理方法、及びプログラム |
JP5686169B1 (ja) * | 2013-09-30 | 2015-03-18 | 沖電気工業株式会社 | 表示制御装置、表示制御方法およびプログラム |
JP2015514254A (ja) * | 2012-04-27 | 2015-05-18 | ヒューレット−パッカード デベロップメント カンパニー エル.ピー.Hewlett‐Packard Development Company, L.P. | ユーザからのオーディオ入力 |
JP2015109104A (ja) * | 2015-01-23 | 2015-06-11 | 沖電気工業株式会社 | 表示制御装置、表示制御方法およびプログラム |
JP2015132786A (ja) * | 2014-01-16 | 2015-07-23 | コニカミノルタ株式会社 | 眼鏡型表示装置 |
JP2015228214A (ja) * | 2014-05-30 | 2015-12-17 | イマージョン コーポレーションImmersion Corporation | 触覚通知マネージャ |
JP2016012377A (ja) * | 2015-10-22 | 2016-01-21 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
JP2016091263A (ja) * | 2014-11-04 | 2016-05-23 | 三菱電機インフォメーションシステムズ株式会社 | 制御装置及び制御プログラム |
CN106125921A (zh) * | 2011-02-09 | 2016-11-16 | 苹果公司 | 3d映射环境中的凝视检测 |
JP2016212174A (ja) * | 2015-05-01 | 2016-12-15 | パラマウントベッド株式会社 | 画像表示制御装置、画像表示システム及びプログラム |
JP2017054201A (ja) * | 2015-09-07 | 2017-03-16 | 株式会社ソニー・インタラクティブエンタテインメント | 情報処理装置および画像生成方法 |
US10152953B2 (en) | 2011-06-03 | 2018-12-11 | Sony Corporation | Information processing apparatus and information processing method |
WO2019235135A1 (ja) * | 2018-06-07 | 2019-12-12 | ソニー株式会社 | タスク対応情報の表示位置を変更する情報処理装置 |
JP2020149336A (ja) * | 2019-03-13 | 2020-09-17 | 株式会社リコー | 情報処理装置、表示制御方法、及びプログラム |
JPWO2020189030A1 (ja) * | 2019-03-20 | 2020-09-24 | ||
JP2021026204A (ja) * | 2019-08-07 | 2021-02-22 | 伊藤組土建株式会社 | 個人宛情報表示装置、個人宛情報表示システムおよび個人宛情報表示プログラム |
US11960647B2 (en) | 2020-04-24 | 2024-04-16 | Sharp Nec Display Solutions, Ltd. | Content display device, content display method, and storage medium using gazing point identification based on line-of-sight direction detection |
JP7519253B2 (ja) | 2020-10-02 | 2024-07-19 | 東芝ライフスタイル株式会社 | 家電システム |
Families Citing this family (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6055421B2 (ja) | 2011-12-28 | 2016-12-27 | サターン ライセンシング エルエルシーSaturn Licensing LLC | 表示装置、表示制御方法、携帯端末装置、およびプログラム |
WO2013099632A1 (ja) | 2011-12-28 | 2013-07-04 | ソニー株式会社 | 表示装置、表示制御方法、およびプログラム |
JP6225029B2 (ja) * | 2011-12-28 | 2017-11-01 | サターン ライセンシング エルエルシーSaturn Licensing LLC | 表示装置、表示制御方法、およびプログラム |
US20130205314A1 (en) * | 2012-02-07 | 2013-08-08 | Arun Ramaswamy | Methods and apparatus to select media based on engagement levels |
KR101922589B1 (ko) | 2012-02-15 | 2018-11-27 | 삼성전자주식회사 | 디스플레이장치 및 그 시선추적방법 |
CN104246682B (zh) | 2012-03-26 | 2017-08-25 | 苹果公司 | 增强的虚拟触摸板和触摸屏 |
JP2013219556A (ja) * | 2012-04-09 | 2013-10-24 | Olympus Imaging Corp | 撮像装置 |
KR20130134785A (ko) * | 2012-05-31 | 2013-12-10 | 엘지전자 주식회사 | 유저 입력 리스폰스 출력 방법 및 이를 위한 홈 디바이스 |
US9310611B2 (en) * | 2012-09-18 | 2016-04-12 | Qualcomm Incorporated | Methods and systems for making the use of head-mounted displays less obvious to non-users |
US9400551B2 (en) * | 2012-09-28 | 2016-07-26 | Nokia Technologies Oy | Presentation of a notification based on a user's susceptibility and desired intrusiveness |
KR102111457B1 (ko) * | 2013-05-15 | 2020-05-15 | 엘지전자 주식회사 | 이동 단말기 및 이동 단말기의 제어 방법 |
JP6318470B2 (ja) * | 2013-05-15 | 2018-05-09 | ソニー株式会社 | 表示制御装置、表示制御方法および記録媒体 |
US9661230B2 (en) * | 2013-07-05 | 2017-05-23 | Lg Electronics Inc. | Image display apparatus and method of operating the image display apparatus |
JP2015087824A (ja) * | 2013-10-28 | 2015-05-07 | オムロン株式会社 | 画面操作装置および画面操作方法 |
KR102163850B1 (ko) * | 2014-01-29 | 2020-10-12 | 삼성전자 주식회사 | 디스플레이장치 및 그 제어방법 |
DE102014005759A1 (de) * | 2014-04-17 | 2015-10-22 | Audi Ag | Displaysteuerung, Anzeigevorrichtung, Fahrzeug und Anzeigeverfahren zum Darstellen von Bildinformation |
US20150348513A1 (en) * | 2014-05-27 | 2015-12-03 | Lenovo (Singapore) Pte. Ltd. | Gaze based notification placement |
JP6990962B2 (ja) * | 2014-06-09 | 2022-01-12 | 雅人 桑原 | 情報処理装置 |
US20160034121A1 (en) * | 2014-07-30 | 2016-02-04 | Wal-Mart Stores, Inc. | Method and Apparatus for Automatically Displaying Multiple Presentations for Multiple Users |
US9922403B2 (en) * | 2015-01-09 | 2018-03-20 | Panasonic Intellectual Property Management Co., Ltd. | Display control apparatus, projection apparatus, display control method, and non-transitory computer readable medium |
US20160342327A1 (en) * | 2015-05-22 | 2016-11-24 | Lg Electronics Inc. | Watch-type mobile terminal and method of controlling therefor |
KR101697286B1 (ko) * | 2015-11-09 | 2017-01-18 | 경북대학교 산학협력단 | 사용자 스타일링을 위한 증강현실 제공 장치 및 방법 |
JP6841241B2 (ja) * | 2016-02-08 | 2021-03-10 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
US10515121B1 (en) * | 2016-04-12 | 2019-12-24 | Tableau Software, Inc. | Systems and methods of using natural language processing for visual analysis of a data set |
GB2564610A (en) | 2016-04-15 | 2019-01-16 | Walmart Apollo Llc | Systems and methods for providing content-based product recommendations |
CA3020974A1 (en) | 2016-04-15 | 2017-10-19 | Walmart Apollo, Llc | Partiality vector refinement systems and methods through sample probing |
EP3451135A4 (en) * | 2016-04-26 | 2019-04-24 | Sony Corporation | INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING PROCESS AND PROGRAM |
KR20190020779A (ko) * | 2016-06-23 | 2019-03-04 | 가부시키가이샤 가이아 시스템 솔루션 | 인게이지먼트값 처리 시스템 및 인게이지먼트값 처리 장치 |
US10373464B2 (en) | 2016-07-07 | 2019-08-06 | Walmart Apollo, Llc | Apparatus and method for updating partiality vectors based on monitoring of person and his or her home |
JP6851745B2 (ja) * | 2016-08-04 | 2021-03-31 | 富士通株式会社 | 画像制御方法、装置、及びプログラム |
CN106650661A (zh) * | 2016-12-21 | 2017-05-10 | 奇酷互联网络科技(深圳)有限公司 | 一种终端使用状态的检测方法和装置 |
US11082383B2 (en) | 2018-03-13 | 2021-08-03 | ROVl GUIDES, INC. | Systems and methods for displaying a notification at an area on a display screen that is within a line of sight of a subset of audience members to whom the notification pertains |
GB2576904B (en) * | 2018-09-06 | 2021-10-20 | Sony Interactive Entertainment Inc | Content modification system and method |
GB2576910B (en) * | 2018-09-06 | 2021-10-20 | Sony Interactive Entertainment Inc | User profile generating system and method |
WO2020076300A1 (en) | 2018-10-09 | 2020-04-16 | Hewlett-Packard Development Company, L.P. | Selecting a display with machine learning |
GB2596541B (en) * | 2020-06-30 | 2023-09-13 | Sony Interactive Entertainment Inc | Video processing |
CN112506345B (zh) * | 2020-12-10 | 2024-04-16 | 北京达佳互联信息技术有限公司 | 一种页面显示方法、装置、电子设备及存储介质 |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000250677A (ja) * | 1999-03-02 | 2000-09-14 | Toshiba Corp | マルチモーダルインターフェース装置及びマルチモーダルインターフェース方法 |
JP2002229918A (ja) * | 2000-12-22 | 2002-08-16 | Xerox Corp | 電子掲示板システム |
JP2004220179A (ja) * | 2003-01-10 | 2004-08-05 | Toshiba Corp | 画像生成装置、画像生成方法、及び画像生成プログラム |
JP2005157134A (ja) * | 2003-11-27 | 2005-06-16 | Nippon Telegr & Teleph Corp <Ntt> | 情報出力方法及び装置及びプログラム及び情報出力プログラムを格納したコンピュータ読み取り可能な記憶媒体 |
JP2005227522A (ja) * | 2004-02-13 | 2005-08-25 | Hitachi Ltd | 情報表示装置 |
JP2006178842A (ja) * | 2004-12-24 | 2006-07-06 | Matsushita Electric Ind Co Ltd | 情報提示装置 |
JP2007133305A (ja) * | 2005-11-14 | 2007-05-31 | Nippon Telegr & Teleph Corp <Ntt> | 情報表示制御装置および情報表示制御方法 |
WO2008012716A2 (en) * | 2006-07-28 | 2008-01-31 | Koninklijke Philips Electronics N. V. | Private screens self distributing along the shop window |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3058004B2 (ja) * | 1994-03-23 | 2000-07-04 | キヤノン株式会社 | 目視制御画像表示装置 |
JPH09293147A (ja) * | 1996-04-24 | 1997-11-11 | Canon Inc | 視線検出を用いた画像処理装置 |
US7743340B2 (en) * | 2000-03-16 | 2010-06-22 | Microsoft Corporation | Positioning and rendering notification heralds based on user's focus of attention and activity |
US20020044152A1 (en) * | 2000-10-16 | 2002-04-18 | Abbott Kenneth H. | Dynamic integration of computer generated and real world images |
US20020113872A1 (en) * | 2001-02-16 | 2002-08-22 | Naoto Kinjo | Information transmitting system |
US8292433B2 (en) * | 2003-03-21 | 2012-10-23 | Queen's University At Kingston | Method and apparatus for communication between humans and devices |
JP2005315802A (ja) * | 2004-04-30 | 2005-11-10 | Olympus Corp | ユーザ支援装置 |
US8232962B2 (en) * | 2004-06-21 | 2012-07-31 | Trading Technologies International, Inc. | System and method for display management based on user attention inputs |
US9710961B2 (en) * | 2009-09-17 | 2017-07-18 | Nokia Technologies Oy | Method and apparatus for providing contextual rendering of a map |
-
2010
- 2010-02-02 US US13/143,861 patent/US20110267374A1/en not_active Abandoned
- 2010-02-02 WO PCT/JP2010/000595 patent/WO2010089989A1/ja active Application Filing
- 2010-02-02 JP JP2010549388A patent/JP5286371B2/ja active Active
- 2010-02-02 EP EP10738331.7A patent/EP2395420B1/en not_active Not-in-force
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000250677A (ja) * | 1999-03-02 | 2000-09-14 | Toshiba Corp | マルチモーダルインターフェース装置及びマルチモーダルインターフェース方法 |
JP2002229918A (ja) * | 2000-12-22 | 2002-08-16 | Xerox Corp | 電子掲示板システム |
JP2004220179A (ja) * | 2003-01-10 | 2004-08-05 | Toshiba Corp | 画像生成装置、画像生成方法、及び画像生成プログラム |
JP2005157134A (ja) * | 2003-11-27 | 2005-06-16 | Nippon Telegr & Teleph Corp <Ntt> | 情報出力方法及び装置及びプログラム及び情報出力プログラムを格納したコンピュータ読み取り可能な記憶媒体 |
JP2005227522A (ja) * | 2004-02-13 | 2005-08-25 | Hitachi Ltd | 情報表示装置 |
JP2006178842A (ja) * | 2004-12-24 | 2006-07-06 | Matsushita Electric Ind Co Ltd | 情報提示装置 |
JP2007133305A (ja) * | 2005-11-14 | 2007-05-31 | Nippon Telegr & Teleph Corp <Ntt> | 情報表示制御装置および情報表示制御方法 |
WO2008012716A2 (en) * | 2006-07-28 | 2008-01-31 | Koninklijke Philips Electronics N. V. | Private screens self distributing along the shop window |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012114557A (ja) * | 2010-11-22 | 2012-06-14 | Nec Saitama Ltd | 表示装置、表示制御方法、プログラム及び電子機器 |
CN106125921B (zh) * | 2011-02-09 | 2019-01-15 | 苹果公司 | 3d映射环境中的凝视检测 |
EP2672880A4 (en) * | 2011-02-09 | 2017-09-13 | Apple Inc. | Gaze detection in a 3d mapping environment |
CN106125921A (zh) * | 2011-02-09 | 2016-11-16 | 苹果公司 | 3d映射环境中的凝视检测 |
US10152953B2 (en) | 2011-06-03 | 2018-12-11 | Sony Corporation | Information processing apparatus and information processing method |
US10176787B2 (en) | 2011-06-03 | 2019-01-08 | Sony Corporation | Information processing apparatus and information processing method for occlusion avoidance in tabletop displays |
JP2013073529A (ja) * | 2011-09-28 | 2013-04-22 | Kyocera Corp | 装置、方法、及びプログラム |
WO2013057878A1 (ja) * | 2011-10-19 | 2013-04-25 | パナソニック株式会社 | 表示制御装置、集積回路、表示制御方法およびプログラム |
JPWO2013057878A1 (ja) * | 2011-10-19 | 2015-04-02 | パナソニック インテレクチュアル プロパティ コーポレーション オブアメリカPanasonic Intellectual Property Corporation of America | 表示制御装置、集積回路、表示制御方法およびプログラム |
JP2015514254A (ja) * | 2012-04-27 | 2015-05-18 | ヒューレット−パッカード デベロップメント カンパニー エル.ピー.Hewlett‐Packard Development Company, L.P. | ユーザからのオーディオ入力 |
US9626150B2 (en) | 2012-04-27 | 2017-04-18 | Hewlett-Packard Development Company, L.P. | Audio input from user |
US10379610B2 (en) | 2013-09-02 | 2019-08-13 | Sony Corporation | Information processing device and information processing method |
WO2015029554A1 (ja) * | 2013-09-02 | 2015-03-05 | ソニー株式会社 | 情報処理装置、情報処理方法、及びプログラム |
JPWO2015029554A1 (ja) * | 2013-09-02 | 2017-03-02 | ソニー株式会社 | 情報処理装置、情報処理方法、及びプログラム |
JP5686169B1 (ja) * | 2013-09-30 | 2015-03-18 | 沖電気工業株式会社 | 表示制御装置、表示制御方法およびプログラム |
JP2015132786A (ja) * | 2014-01-16 | 2015-07-23 | コニカミノルタ株式会社 | 眼鏡型表示装置 |
US10365878B2 (en) | 2014-05-30 | 2019-07-30 | Immersion Corporation | Haptic notification manager |
JP2015228214A (ja) * | 2014-05-30 | 2015-12-17 | イマージョン コーポレーションImmersion Corporation | 触覚通知マネージャ |
JP2016091263A (ja) * | 2014-11-04 | 2016-05-23 | 三菱電機インフォメーションシステムズ株式会社 | 制御装置及び制御プログラム |
JP2015109104A (ja) * | 2015-01-23 | 2015-06-11 | 沖電気工業株式会社 | 表示制御装置、表示制御方法およびプログラム |
JP2016212174A (ja) * | 2015-05-01 | 2016-12-15 | パラマウントベッド株式会社 | 画像表示制御装置、画像表示システム及びプログラム |
US10627628B2 (en) | 2015-09-07 | 2020-04-21 | Sony Interactive Entertainment Inc. | Information processing apparatus and image generating method |
JP2017054201A (ja) * | 2015-09-07 | 2017-03-16 | 株式会社ソニー・インタラクティブエンタテインメント | 情報処理装置および画像生成方法 |
JP2016012377A (ja) * | 2015-10-22 | 2016-01-21 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
WO2019235135A1 (ja) * | 2018-06-07 | 2019-12-12 | ソニー株式会社 | タスク対応情報の表示位置を変更する情報処理装置 |
JP2020149336A (ja) * | 2019-03-13 | 2020-09-17 | 株式会社リコー | 情報処理装置、表示制御方法、及びプログラム |
JP7215254B2 (ja) | 2019-03-13 | 2023-01-31 | 株式会社リコー | 情報処理装置、表示制御方法、及びプログラム |
JP2023033559A (ja) * | 2019-03-13 | 2023-03-10 | 株式会社リコー | 情報処理装置、表示制御方法、及びプログラム |
JPWO2020189030A1 (ja) * | 2019-03-20 | 2020-09-24 | ||
WO2020189030A1 (ja) * | 2019-03-20 | 2020-09-24 | 株式会社Nttドコモ | 情報生成装置、及び制御システム |
JP7291779B2 (ja) | 2019-03-20 | 2023-06-15 | 株式会社Nttドコモ | 情報生成装置、及び制御システム |
JP2021026204A (ja) * | 2019-08-07 | 2021-02-22 | 伊藤組土建株式会社 | 個人宛情報表示装置、個人宛情報表示システムおよび個人宛情報表示プログラム |
US11960647B2 (en) | 2020-04-24 | 2024-04-16 | Sharp Nec Display Solutions, Ltd. | Content display device, content display method, and storage medium using gazing point identification based on line-of-sight direction detection |
JP7519253B2 (ja) | 2020-10-02 | 2024-07-19 | 東芝ライフスタイル株式会社 | 家電システム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2010089989A1 (ja) | 2012-08-09 |
EP2395420A1 (en) | 2011-12-14 |
US20110267374A1 (en) | 2011-11-03 |
JP5286371B2 (ja) | 2013-09-11 |
EP2395420A4 (en) | 2014-12-03 |
EP2395420B1 (en) | 2018-07-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5286371B2 (ja) | 情報表示装置及び情報表示方法 | |
JP5602155B2 (ja) | ユーザインタフェース装置および入力方法 | |
JP5869558B2 (ja) | 表示制御装置、集積回路、表示制御方法およびプログラム | |
JP5518713B2 (ja) | 情報表示装置及び情報表示方法 | |
CN111886564B (zh) | 信息处理装置、信息处理方法和程序 | |
JP6886117B2 (ja) | 1つの表示装置に表示された画像の画質の制御方法 | |
JP5260643B2 (ja) | ユーザインターフェース装置、ユーザインターフェース方法、及び記録媒体 | |
WO2013018267A1 (ja) | 提示制御装置、及び提示制御方法 | |
WO2017113668A1 (zh) | 一种根据眼部动作对终端进行控制的方法及装置 | |
US20120133754A1 (en) | Gaze tracking system and method for controlling internet protocol tv at a distance | |
JP2007220004A (ja) | テレビおよび認証装置 | |
JP2009087026A (ja) | 映像表示装置 | |
JPWO2018025458A1 (ja) | 情報処理装置、情報処理方法、およびプログラム | |
JP2011166305A (ja) | 画像処理装置および撮像装置 | |
JP6241107B2 (ja) | 視線方向検出装置および視線方向検出方法 | |
CN107430841B (zh) | 信息处理设备、信息处理方法、程序以及图像显示系统 | |
CN114740966A (zh) | 多模态图像显示控制方法、系统及计算机设备 | |
KR101889463B1 (ko) | 사용자 단말의 화면 표시 제어 방법 및 이를 이용한 사용자 단말 | |
JP2016192096A (ja) | 物体認識選択装置、物体認識選択方法及びプログラム | |
JP2018181339A (ja) | システム、情報処理方法及びプログラム | |
JP2013061988A (ja) | 表示制御装置、表示制御方法および表示制御プログラム | |
US20230308613A1 (en) | Display device and method for controlling display device | |
WO2022091589A1 (ja) | 情報処理装置、情報処理方法、およびプログラム | |
CN118764596A (zh) | 投影方法、装置、设备及其存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10738331 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2010549388 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13143861 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010738331 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |