EP3588247B1 - Informationsverarbeitungsvorrichtung, informationsverarbeitungsverfahren und programm - Google Patents

Informationsverarbeitungsvorrichtung, informationsverarbeitungsverfahren und programm Download PDF

Info

Publication number
EP3588247B1
EP3588247B1 EP17897499.4A EP17897499A EP3588247B1 EP 3588247 B1 EP3588247 B1 EP 3588247B1 EP 17897499 A EP17897499 A EP 17897499A EP 3588247 B1 EP3588247 B1 EP 3588247B1
Authority
EP
European Patent Office
Prior art keywords
irradiation
information processing
processing apparatus
detection unit
person
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
EP17897499.4A
Other languages
English (en)
French (fr)
Other versions
EP3588247A1 (de
EP3588247A4 (de
Inventor
Yuichiro Saito
Shinji Matsuda
Tomoki Hamajima
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Group Corp
Original Assignee
Sony Group Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Group Corp filed Critical Sony Group Corp
Publication of EP3588247A1 publication Critical patent/EP3588247A1/de
Publication of EP3588247A4 publication Critical patent/EP3588247A4/de
Application granted granted Critical
Publication of EP3588247B1 publication Critical patent/EP3588247B1/de
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B21/00Projectors or projection-type viewers; Accessories therefor
    • G03B21/14Details
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B17/00Details of cameras or camera bodies; Accessories therefor
    • G03B17/48Details of cameras or camera bodies; Accessories therefor adapted for combination with other photographic or optical apparatus
    • G03B17/54Details of cameras or camera bodies; Accessories therefor adapted for combination with other photographic or optical apparatus with projector
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/36Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/30Transforming light or analogous information into electric information
    • H04N5/33Transforming infrared radiation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3179Video signal processing therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3191Testing thereof
    • H04N9/3194Testing thereof including sensor feedback
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/04Changes in size, position or resolution of an image
    • G09G2340/0464Positioning
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2354/00Aspects of interface with display user

Definitions

  • the present disclosure relates to an information processing apparatus, an information processing method, and a program.
  • Patent Document 1 discloses a projection apparatus including an irradiation unit that can irradiate various image information.
  • Patent Document 1 Japanese Patent Application Laid-Open No. 2003-044839
  • Such a projection apparatus has been required to have a technology that allows a user to view the image information more comfortably.
  • Prior Art Includes: EP 2744206A1 ; US 2007/001111A1 ; US 6460999B1 and US 2011/304833 .
  • EP 2744206 A1 discloses a laser projector comprising a front-facing projection component and a rear-facing directional microphone configured to detect a person and to control an emission of light based on a detected sound.
  • the person detection unit is controlled on the basis of the irradiation direction so that an operation unintended by a user is less likely to be performed.
  • the user can thus view the image information more comfortably.
  • the above effect is not necessarily restrictive, but any of the effects illustrated in the present specification or other effects that can be perceived from the present specification may be achieved together with the above effect or in place of the above effect.
  • the information processing apparatus 10 includes an irradiation unit 11, a person detection unit 12, a position detection unit 13, an attitude detection unit 14, a distance detection unit 15, an operation unit 16, a communication unit 17, and a control unit 18.
  • the information processing apparatus 10 is a projection apparatus, for example.
  • the information processing apparatus 10 includes the configuration of hardware such as a central processing unit (CPU), a read only memory (ROM), a random access memory (RAM), a non-volatile memory, an irradiation device, a person detection device, a position detection device, an attitude detection device, a distance detection device, an operation button, and a communication device.
  • Information necessary for the operation of the information processing apparatus 10 such as a program is recorded in the ROM.
  • the CPU reads and executes the program recorded in the ROM.
  • the irradiation unit 11, the person detection unit 12, the position detection unit 13, the attitude detection unit 14, the operation unit 16, the communication unit 17, and the control unit 18 are implemented.
  • the irradiation unit 11 includes the irradiation device, for example, and irradiates various image information within an irradiation area 11A as illustrated in Fig. 2 .
  • the irradiation unit 11 irradiates the image information on a horizontal surface 100.
  • a surface where the irradiation area 11A and the horizontal surface 100 intersect is an irradiation surface 11B.
  • the image information is displayed on the irradiation surface 11B.
  • the horizontal surface 100 includes, for example, a table top of a desk or the like, a floor, or the like.
  • An optical axis of the irradiation unit 11 (that is, a central axis of the irradiation area 11A) is slightly tiled toward an outer wall surface of the information processing apparatus 10 from a direction perpendicular to the outer wall surface.
  • the direction of the optical axis that is, an irradiation direction
  • the irradiation direction includes a downward direction illustrated in Fig. 2 or an upward direction illustrated in Fig. 3 .
  • the downward direction is a direction below a horizontal direction
  • the upward direction is a direction above the horizontal direction. In a case where the irradiation direction is the upward direction as illustrated in Fig.
  • the irradiation unit 11 irradiates the image information on a vertical surface 200, for example.
  • the information processing apparatus 10 is placed on a horizontal surface 300.
  • a surface where the irradiation area 11A and the vertical surface 200 intersect is the irradiation surface 11B.
  • the image information is displayed on the irradiation surface 11B.
  • the vertical surface 200 includes, for example, a wall surface, a side surface of various objects, or the like.
  • a user (a person U illustrated in Fig. 2 ) can display the image information on the horizontal surface 100 by placing the information processing apparatus 10 on the horizontal surface 100 and setting the irradiation direction in the downward direction.
  • the user can display the image information on the vertical surface 200 by placing the information processing apparatus 10 near the vertical surface 200 and setting the irradiation direction in the upward direction.
  • the information processing apparatus 10 can thus display the image information at various positions desired by the user. It is needless to say that the angle between the optical axis of the irradiation unit 11 and the outer wall surface of the information processing apparatus 10 and the types of the irradiation direction are not limited to the above.
  • the surface on which the image information is to be irradiated is not limited to the horizontal surface 100 and the vertical surface 200.
  • the person detection unit 12 includes the person detection device, for example, and detects a person.
  • the person detection device includes, for example, a human sensor, a camera, and the like.
  • the human sensor detects a person by detecting heat of the person and/or a movement of the heat.
  • the camera detects a person by imaging the person.
  • the information processing apparatus 10 need only include either the human sensor or the camera, but may include both.
  • the person detection unit 12 forms a detection area 12A in a direction opposite to the irradiation direction (that is, the optical axis of the irradiation unit 11) in which the irradiation unit 11 irradiates the image information.
  • a person to be detected is, for example, the user of the information processing apparatus 10.
  • the person detection unit 12 detects a person within the detection area 12A.
  • An angle between an optical axis of the detection area 12A (that is, a central axis of the detection area 12A) and the optical axis of the irradiation unit 11 is larger than 90°
  • the person detection unit 12 detects a person
  • the person detection unit generates person detection information to that effect and outputs the information to the control unit 18.
  • the control unit 18 acquires the person detection information
  • the control unit causes the irradiation unit 11 to irradiate the image information.
  • the person detection unit 12 is controlled on the basis of the irradiation direction of the irradiation unit 11, whereby an operation unintended by the user is less likely to be performed.
  • the person detection unit 12 of the present embodiment forms the detection area 12A in the direction opposite to the irradiation direction in which the irradiation unit 11 irradiates the image information, the detection area 12A may be formed in a direction different from the irradiation direction (irradiation surface).
  • the position detection unit 13 includes the position detection device, for example, and detects an object present within the irradiation area 11A of the image information.
  • the position detection device includes an infrared sensor, for example.
  • the position detection unit 13 is driven in conjunction with the control unit 18 and the irradiation unit 11, and detects an object present within a detection area 13A.
  • the detection area 13A covers substantially an entire area of the irradiation surface 11B.
  • a target to be detected by the position detection unit 13 is a hand (more specifically, a finger) of a person, for example.
  • the user can perform an input operation (for example, a tap operation, a drag operation, a pinch-in/out operation, and the like) on the image information displayed on the irradiation surface 11B.
  • the position detection unit 13 detects such an input operation and outputs position detection information regarding a result of the detection to the control unit 18.
  • the control unit 18 performs processing based on the position detection information such as processing of changing the image information.
  • the detection area 13A of the position detection unit 13 is of course not limited to the above example but may cover, for example, only a part of the irradiation surface 11B.
  • the attitude detection unit 14 includes the attitude detection device, for example, and detects an attitude of the information processing apparatus 10.
  • the attitude detection device includes a gyro sensor, an acceleration sensor, or the like, for example.
  • the attitude detection unit 14 outputs attitude detection information regarding a result of detection to the control unit 18.
  • the control unit 18 determines the irradiation direction on the basis of the attitude detection information. Then, the control unit 18 then controls the person detection unit 12 on the basis of the irradiation direction. Details of the control will be described later.
  • the distance detection unit 15 includes the distance detection device, for example, and measures a distance between the information processing apparatus 10 and the irradiation surface 11B.
  • the distance detection device includes a range sensor, for example.
  • the range sensor measures the distance between the information processing apparatus 10 and the irradiation surface 11B using a laser, ultrasound, or the like.
  • the distance detection unit 15 measures a distance from a surface of the information processing apparatus 10 facing the irradiation surface 11B to the irradiation surface 11B, the surface being one of the outer wall surfaces of the information processing apparatus.
  • This distance corresponds to the length of a perpendicular drawn from the surface facing the irradiation surface 11B to a plane including the irradiation surface 11B (which is the vertical surface 200 in the example of Fig. 3 .
  • the distance detection unit 15 outputs distance detection information regarding the distance detected to the control unit 18.
  • the operation unit 16 includes the operation button and the like, for example, and allows an input operation by the user.
  • the operation button here includes a power button, for example.
  • the operation unit outputs input operation information regarding the input operation to the control unit 18.
  • the control unit 18 performs processing corresponding to the input operation. For example, the control unit 18 activates the information processing apparatus 10 in a case where the user presses the power button. Moreover, the control unit 18 causes the irradiation unit 11 to stop irradiation of the image information in a case where the user presses the power button, that is, performs an irradiation stop operation, while the information processing apparatus 10 is activated.
  • the control unit 18 thus causes the information processing apparatus 10 to shift to a sleep mode. Note that such a state is also referred to as a standby mode, a suspend mode, or the like.
  • the control unit 18 causes the irradiation unit 11 to restart irradiation of the image information in a case where the user presses the power button, that is, performs an irradiation restart operation, during the sleep mode.
  • the communication unit 17 includes the communication device, for example, and communicates with another information processing apparatus or the like. As a result, the communication unit 17 acquires various kinds of information such as the image information from the other information processing apparatuses or the like.
  • the control unit 18 includes a CPU or the like, for example, and controls each component of the information processing apparatus 10. Furthermore, the control unit 18 also controls the person detection unit 12 on the basis of the irradiation direction. Furthermore, the control unit 18 performs each processing described above. Details of the processing will be described later.
  • step S10 the attitude detection unit 14 detects an attitude of the information processing apparatus 10 (for example, in which direction the placement of the apparatus allows for irradiation).
  • the attitude detection unit 14 outputs attitude detection information regarding a result of detection to the control unit 18.
  • step S11 the control unit 18 determines the irradiation direction on the basis of the attitude detection information.
  • the control unit 18 determines whether the irradiation direction is the downward direction illustrated in Fig. 2 or the upward direction illustrated in Fig. 3 .
  • step S12 the control unit 18 determines whether or not the irradiation direction is the downward direction (a first direction).
  • the control unit 18 proceeds to step S13 if the irradiation direction is the downward direction (the first direction), that is, the irradiation direction is horizontal to the ground, and the irradiation surface is on a desk or a floor.
  • the control unit 18 proceeds to step S14 if the irradiation direction is the upward direction (a second direction different from the first direction), that is, the irradiation direction is perpendicular to the ground, and the irradiation surface is on a wall.
  • step S13 the control unit 18 stops the operation of the person detection unit 12.
  • the detection area 12A of the person detection unit 12 is upward as illustrated in Fig. 2 .
  • a ceiling light or the like possibly enters the detection area 12A of the person detection unit 12.
  • Such a light is also a source of heat. Therefore, in a case where the person detection unit 12 includes the human sensor, the light may be detected as a person by mistake.
  • the person detection unit 12 outputs person detection information to the control unit 18 so that the control unit 18 causes the irradiation unit 11 to irradiate the image information.
  • the irradiation unit 11 thus possibly irradiates the image information even in a case where the user does not intend to view the image information. Moreover, since the user does not intend to view the image information, an object (for example, a glass or the like) may be put on the irradiation surface 11B. In this case, the position detection unit 13 possibly detects the presence of the object and outputs position detection information to the control unit 18. The control unit 18 in this case possibly performs processing unintended by the user. For example, in a case where an icon is displayed at the position of the object, the control unit 18 possibly determines by mistake that the user has selected the icon. As a result, the information processing apparatus 10 may perform an operation unintended by the user. Thus, in the present embodiment, the control unit 18 stops the operation of the person detection unit 12 in the case where the irradiation direction is the downward direction.
  • the user switches the presence or absence of irradiation by pressing the power button.
  • the user turns on the power button in a case where the irradiation unit 11 does not perform irradiation of the image information (that is, during the sleep mode).
  • the control unit 18 thus causes the irradiation unit 11 to restart irradiation of the image information.
  • the user presses the power button again.
  • the control unit 18 thus causes the irradiation unit 11 to stop the irradiation of the image information.
  • the information processing apparatus 10 shifts to the sleep mode.
  • the information processing apparatus 10 can thus perform processing in accordance with the intention of the user, whereby the user can view the image information more comfortably.
  • the control unit 18 thereafter ends the present basic processing.
  • control unit 18 determines whether or not the power button has been pressed.
  • the control unit 18 proceeds to step S15 if the power button has been pressed, or ends the present basic processing if the power button has not been pressed.
  • step S15 the control unit 18 determines whether or not the irradiation unit 11 is irradiating the image information.
  • the control unit 18 proceeds to step S16 if the irradiation unit 11 is irradiating the image information.
  • the control unit 18 proceeds to step S20 if the irradiation unit 11 is not irradiating the image information, that is, in the sleep mode.
  • step S16 the control unit 18 causes the irradiation unit 11 to stop irradiation of the image information.
  • the control unit 18 causes the information processing apparatus 10 to shift to the sleep mode.
  • the pressing of the power button in step S14 is the irradiation stop operation.
  • step S17 the control unit 18 stops the operation of the person detection unit 12.
  • step S18 the control unit 18 determines whether or not a predetermined irradiation restart condition is satisfied. The control unit 18 proceeds to step S19 if the irradiation restart condition is satisfied, or returns to step S17 if the irradiation restart condition is not satisfied. In other words, the control unit 18 waits until the irradiation restart condition is satisfied.
  • the irradiation restart condition is a condition that a predetermined time has elapsed since the irradiation stop operation.
  • step S19 the control unit 18 restarts the operation of the person detection unit 12.
  • the control unit 18 thereafter causes the irradiation unit 11 to restart irradiation of the image information in a case where the person detection unit 12 detects a person (for example, the user).
  • the control unit 18 thereafter ends the present basic processing.
  • the control unit 18 stops the operation of the person detection unit 12 until the predetermined irradiation restart condition is satisfied.
  • the irradiation direction is the upward direction illustrated in Fig. 3 .
  • the detection area 12A of the person detection unit 12 is thus oriented in a direction away from the vertical surface 200. Therefore, the person detection unit 12 can detect a person facing the vertical surface 200.
  • the person detection unit 12 outputs person detection information to the control unit 18. Then, the control unit 18 then causes the irradiation unit 11 to restart irradiation of the image information on the basis of the person detection information. Therefore, the information processing apparatus 10 possibly releases the sleep mode immediately after shifting to the sleep mode.
  • the above operation may not be intended by the user.
  • the user often presses the power button not with the intention to cause such an operation but in order to stop the irradiation by the irradiation unit 11 for a while.
  • Various cases are assumed for such a situation to occur such as a case where the user has another business to take care of. In this case, it is assumed that the user leaves the information processing apparatus 10 as illustrated in Fig. 4 .
  • the control unit 18 stops the operation of the person detection unit 12 until the predetermined irradiation restart condition is satisfied. More specifically, the control unit 18 stops the operation of the person detection unit 12 until a predetermined time elapses after the irradiation stop operation is performed.
  • This predetermined time is an estimated time until the user goes out of the detection area 12A of the person detection unit 12 as illustrated in Fig. 5 . Therefore, the predetermined time may be set on the basis of, for example, the size of the detection area 12A, the speed of travel (for example, an average walking speed) of the user, or the like. The predetermined time may be 10 seconds, for example.
  • the control unit 18 can thus restart the operation of the person detection unit 12 after the user goes out of the detection area 12A of the person detection unit 12. In this case, since the user is outside the detection area 12A, the sleep mode is not released even if the operation of the person detection unit 12 is restarted. The user thereafter takes care of the other business and enters the detection area 12A. In response to this, the person detection unit 12 outputs person detection information to the control unit 18. The control unit 18 causes the irradiation unit 11 to restart irradiation of the image information. The information processing apparatus 10 can therefore perform processing in accordance with the intention of the user.
  • the control unit 18 may cause the irradiation unit 11 to restart the irradiation in a case where the camera detects the face of the user.
  • the irradiation unit 11 may restart the irradiation after it is acknowledged that the user has started viewing the image information.
  • the camera and the human sensor may be used in combination.
  • the control unit 18 first restarts the operation of only the human sensor in a case where the operation of the person detection unit 12 is restarted. Then, the control unit 18 also restarts the operation of the camera in a case where the human sensor detects a person.
  • the control unit 18 may then cause the irradiation unit 11 to restart the irradiation in a case where the camera detects the face of the user.
  • step S20 the control unit 18 causes the irradiation unit 11 to restart the irradiation of the image information.
  • the pressing of the power button in step S14 is the irradiation start operation.
  • the control unit 18 thereafter ends the present basic processing.
  • the control unit 18 sets the aforementioned predetermined time on the basis of the distance between the information processing apparatus 10 and the irradiation surface 11B.
  • Information regarding the distance is given from the distance detection unit 15.
  • the following effects can be expected as a result. That is, the user in some cases wishes to view the image information at a distance from the information processing apparatus 10 because the irradiation surface 11B is large. In this case, the user is assumed to leave the information processing apparatus 10 after temporarily setting the information processing apparatus 10 to the sleep mode. Then, the distance the user moves away from the information processing apparatus 10 to an extent is assumed to be longer as the irradiation surface 11B is larger.
  • the control unit 18 sets the aforementioned predetermined time on the basis of the distance between the information processing apparatus 10 and the irradiation surface 11B. Specifically, the control unit 18 sets the predetermined time to a longer time as the distance between the information processing apparatus 10 and the irradiation surface 11B is longer. As a result, the control unit 18 can restart the operation of the person detection unit 12 after the user has reached a desired position. The information processing apparatus 10 can therefore perform processing in accordance with the intention of the user.
  • the control unit 18 may adjust the sensitivity of the person detection unit 12 on the basis of the distance between the information processing apparatus 10 and the irradiation surface 11B. As a result, the person detection unit 12 can detect the user more reliably even if the user is away from the information processing apparatus 10.
  • the control unit 18 may cause the irradiation unit 11 to restart the irradiation in a case where the camera detects the face of the user.
  • the camera and the human sensor may be used in combination. The specific processing is as described above.
  • the person detection unit 12 is the camera. That is, the person detection unit 12 can detect the face of a person. Then, in step S17, the control unit 18 causes the person detection unit 12 to continue its operation. However, even if the person detection unit 12 outputs person detection information, the control unit 18 does not cause the irradiation unit 11 to restart the irradiation. Thereafter, in step S18, the control unit 18 determines whether or not an irradiation restart condition is satisfied.
  • the irradiation restart condition is that the person detection unit 12 stops detecting the face of a person and then detects the face of a person again.
  • the user who sets the information processing apparatus 10 to the sleep mode is assumed to turn his back on the information processing apparatus 10 and leave the information processing apparatus 10.
  • the person detection unit 12 no longer detects the face of a person, that is, the face of the user.
  • the person detection unit 12 again detects the face of a person, that is, the face of the user.
  • the control unit 18 proceeds to step S19.
  • the control unit 18 causes the irradiation unit 11 to restart irradiation of the image information.
  • the information processing apparatus 10 can perform processing in accordance with the intention of the user.
  • the person detection unit 12 detects the face of the user so that more accurate processing is possible.
  • the person detection unit 12 is the camera. That is, the person detection unit 12 can detect a line of sight of a person. Then, in step S17, the control unit 18 causes the person detection unit 12 to continue its operation. However, even if the person detection unit 12 outputs person detection information, the control unit 18 does not cause the irradiation unit 11 to restart the irradiation. Thereafter, in step S18, the control unit 18 determines whether or not an irradiation restart condition is satisfied.
  • the irradiation restart condition is that a line of sight of a person detected by the person detection unit 12 moves outside the irradiation area 11A (more specifically, the irradiation surface 11B) of the image information and then moves back to the irradiation area 11A (more specifically, the irradiation surface 11B) of the image information.
  • the user who sets the information processing apparatus 10 to the sleep mode is assumed to move his line of sight away from the irradiation surface 11B.
  • the person detection unit 12 detects that the line of sight of a person, that is, the line of sight of the user, has moved away from the irradiation surface 11B.
  • the user thereafter moves his line of sight back to the irradiation surface 11B in a case where the user wishes to view the image information.
  • the person detection unit 12 detects that the line of sight of a person, that is, the line of sight of the user, has moved back to the irradiation surface 11B.
  • the control unit 18 proceeds to step S19.
  • step S19 the control unit 18 causes the irradiation unit 11 to restart irradiation of the image information.
  • the information processing apparatus 10 can perform processing in accordance with the intention of the user.
  • the person detection unit 12 detects the line of sight of the user so that more accurate processing is possible.
  • the control unit 18 controls the person detection unit 12 on the basis of the irradiation direction.
  • the control unit 18 is less likely to perform an operation unintended by the user.
  • the control unit 18 can perform processing more in line with the intention of the user. The user can thus view the image information more comfortably.
  • the irradiation direction includes at least one of the downward direction or the upward direction so that processing in line with the intention of the user can be performed in a case where the irradiation direction corresponds with one of these directions.
  • control unit 18 causes the irradiation unit 11 to irradiate the image information in a case where the person detection unit 12 detects a person or where a predetermined irradiation start operation is performed, and stops the operation of the person detection unit 12 in a case where the irradiation direction is the downward direction.
  • the control unit 18 can perform processing more in line with the intention of the user.
  • control unit 18 drives the position detection unit 13 in conjunction with the irradiation unit 11. Moreover, the control unit 18 then performs the aforementioned processing to be able to prevent misdetection by the position detection unit 13. As a result, the control unit 18 can perform processing more in line with the intention of the user.
  • the person detection unit 12 is in some cases the human sensor that detects a person by detecting heat.
  • the control unit 18 can perform processing more in line with the intention of the user.
  • control unit 18 causes the irradiation unit 11 to irradiate the image information in a case where the person detection unit 12 detects a person or where a predetermined irradiation start operation is performed. Moreover, the control unit 18 also causes the irradiation unit 11 to stop the irradiation of the image information in a case where a predetermined irradiation stop operation is performed. Furthermore, in a case where the irradiation direction is the upward direction, the control unit 18 stops the operation of the person detection unit 12 until a predetermined irradiation restart condition is satisfied after the irradiation stop operation is performed. As a result, the control unit 18 can perform processing more in line with the intention of the user.
  • the irradiation restart condition may be a lapse of a predetermined time since the irradiation stop operation is performed.
  • the control unit 18 can perform processing more in line with the intention of the user.
  • control unit 18 sets the predetermined time on the basis of the distance between the information processing apparatus 10 and the irradiation surface 11B. As a result, the control unit 18 can perform processing more in line with the intention of the user.
  • the person detection unit 12 in some cases is the human sensor that detects a person by detecting heat and is sensitivity adjustable.
  • the control unit 18 adjusts the sensitivity of the person detection unit 12 on the basis of the distance between the information processing apparatus 10 and the irradiation surface 11B. As a result, the control unit 18 can perform processing more in line with the intention of the user.
  • the person detection unit 12 is in some cases the camera that can detect the face of a person.
  • the irradiation restart condition may be that the person detection unit 12 stops detecting the face of a person and then detects the face of a person again.
  • the control unit 18 can perform processing more in line with the intention of the user.
  • the person detection unit 12 is in some cases the camera that can detect the line of sight of a person.
  • the irradiation restart condition may be that the line of sight of a person detected by the person detection unit 12 moves outside the irradiation surface 11B and thereafter moves back to the irradiation surface 11B.
  • the control unit 18 can perform processing more in line with the intention of the user.
  • the control unit 18 may be performed by another information processing apparatus.
  • the control unit 18 in the other information processing apparatus acquires necessary information (for example, person detection information, position detection information, attitude detection information, distance detection information, input operation information, and the like) from the information processing apparatus 10 and performs the aforementioned processing on the information processing apparatus 10.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computer Hardware Design (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • User Interface Of Digital Computer (AREA)
  • Control Or Security For Electrophotography (AREA)
  • Circuit Arrangement For Electric Light Sources In General (AREA)
  • Controls And Circuits For Display Device (AREA)
  • Image Analysis (AREA)

Claims (12)

  1. Informationsverarbeitungsvorrichtung (10), aufweisend eine Steuereinheit (18), die eine Personenerkennungseinheit (12) auf Basis einer Strahlungsrichtung steuert, in welche eine Strahlungseinheit (11), die zum Ausstrahlen von Bildinformationen imstande ist, die Bildinformationen ausstrahlt, wobei die Personenerkennungseinheit (12) gesteuert wird, um einen Erkennungsbereich (12A) in einer Richtung entgegengesetzt zur Strahlungsrichtung zu bilden, wobei ein Winkel zwischen einer optischen Achse des Erkennungsbereichs (12A) und einer optischen Achse der Strahlungseinheit (11) besteht, der größer als 90° ist; wobei die Informationsverarbeitungsvorrichtung (10) dazu ausgebildet ist, auf einer horizontalen Oberfläche angeordnet zu werden, und die Strahlungsrichtung mindestens eine von einer Abwärtsrichtung, die eine Richtung unter einer horizontalen Richtung ist, oder einer Aufwärtsrichtung, die eine Richtung über der horizontalen Richtung ist, in Abhängigkeit von der Installationseinstellung der Informationsverarbeitungsvorrichtung (10) aufweist;
    eine Einstellungserkennungseinheit (14), die zum Erkennen der Installationseinstellung der Informationsverarbeitungsvorrichtung (10) ausgebildet ist;
    wobei die Steuereinheit (18) die Strahlungseinheit (11) zum Bestimmen der Installationseinstellung basierend auf der Ausgabe von der Einstellungserkennungseinheit (14) und zum Ausstrahlen der Bildinformationen veranlasst, falls die Personenerkennungseinheit (12) eine Person erkennt oder wenn ein vorbestimmter Strahlungsstartvorgang durchgeführt wird, und einen Betrieb der Personenerkennungseinheit (12) stoppt, falls die Strahlungsrichtung die Abwärtsrichtung ist.
  2. Informationsverarbeitungsvorrichtung (10) nach Anspruch 1, wobei die Steuereinheit (18) eine Positionserkennungseinheit (13) in Verbindung mit der Strahlungseinheit (11) ansteuert, wobei die Positionserkennungseinheit (13) eine Position eines Objekts innerhalb eines Strahlungsbereichs der Bildinformationen erkennt und wobei die Steuereinheit (18) die Bildinformationen basierend auf der erkannten Position des Objekts ändert.
  3. Informationsverarbeitungsvorrichtung (10) nach Anspruch 1, wobei die Personenerkennungseinheit (12) eine Person durch Erkennen von Wärme erkennt.
  4. Informationsverarbeitungsvorrichtung (10) nach Anspruch 1, wobei die Steuereinheit (18) die Strahlungseinheit (11) zum Ausstrahlen der Bildinformationen veranlasst, falls die Personenerkennungseinheit (12) eine Person erkennt oder wenn ein vorbestimmter Strahlungsstartvorgang durchgeführt wird, die Strahlungseinheit (11) zum Stoppen des Ausstrahlens der Bildungsinformationen veranlasst, falls ein vorbestimmter Strahlungsstoppvorgang durchgeführt wird, und, falls die Strahlungsrichtung die Aufwärtsrichtung ist, den Betrieb der Personenerkennungseinheit (12) stoppt, bis eine vorbestimmte Strahlungsneustartbedingung nach der Durchführung des Strahlungsstoppvorgangs erfüllt wird.
  5. Informationsverarbeitungsvorrichtung (10) nach Anspruch 1, wobei die Strahlungsneustartbedingung eine Spanne einer vorbestimmten Zeit seit der Durchführung des Strahlungsstoppvorgangs ist.
  6. Informationsverarbeitungsvorrichtung (10) nach Anspruch 5, wobei die Steuereinheit (18) die vorbestimmte Zeit auf Basis eines Abstands zwischen der Informationsverarbeitungsvorrichtung (10) und einer Strahlungsfläche einstellt, auf die die Bildinformationen gestrahlt werden, wobei der Abstand basierend auf Informationen bestimmt wird, die durch eine Abstandserkennungseinheit (15) bereitgestellt werden.
  7. Informationsverarbeitungsvorrichtung (10) nach Anspruch 6, wobei die Personenerkennungseinheit (12) eine Person durch Erkennen von Wärme erkennt und ihre Empfindlichkeit angepasst werden kann, und
    die Steuereinheit (18) die Empfindlichkeit der Personenerkennungseinheit (12) auf Basis des Abstands zwischen der Informationsverarbeitungsvorrichtung (10) und der Strahlungsfläche (11B) anpasst, auf die die Bildinformationen gestrahlt werden.
  8. Informationsverarbeitungsvorrichtung (10) nach Anspruch 4, wobei die Personenerkennungseinheit (12) ein Gesicht einer Person erkennen kann, und
    die Strahlungsneustartbedingung ist, dass die Personenerkennungseinheit (12) Erkennen eines Gesichts einer Person stoppt und dann ein Gesicht einer Person erneut erkennt.
  9. Informationsverarbeitungsvorrichtung nach Anspruch 4, wobei die Personenerkennungseinheit eine Blicklinie einer Person erkennen kann, und
    die Strahlungsneustartbedingung ist, dass eine Blicklinie einer Person, die von der Personenerkennungseinheit erkannt wird, sich aus dem Strahlungsbereich der Bildinformationen bewegt und sich dann wieder in den Strahlungsbereich der Bildinformationen zurückbewegt.
  10. Informationsverarbeitungsvorrichtung (10) nach Anspruch 1, wobei die Bildverarbeitungsvorrichtung (10) die Strahlungseinheit (11) und die Personenerkennungseinheit (12) aufweist.
  11. Informationsverarbeitungsverfahren, aufweisend, dass eine Steuereinheit (18) einer Informationsverarbeitungsvorrichtung (10) eine Personenerkennungseinheit (12) auf Basis einer Strahlungsrichtung steuert, in welche eine Strahlungseinheit (11), die zum Ausstrahlen von Bildinformationen imstande ist, die Bildinformationen ausstrahlt, wobei die Personenerkennungseinheit (12) gesteuert wird, um einen Erkennungsbereich in einer Richtung entgegengesetzt zur Strahlungsrichtung zu bilden; wobei ein Winkel zwischen einer optischen Achse des Erkennungsbereichs und einer optischen Achse der Strahlungseinheit (11) besteht, der größer als 90° ist;
    wobei die Informationsverarbeitungsvorrichtung (10) dazu ausgebildet ist, auf einer horizontalen Oberfläche angeordnet zu werden, und die Strahlungsrichtung mindestens eine von einer Abwärtsrichtung, die eine Richtung unter einer horizontalen Richtung ist, oder einer Aufwärtsrichtung, die eine Richtung über der horizontalen Richtung ist, in Abhängigkeit von der Installationseinstellung der Informationsverarbeitungsvorrichtung (10) aufweist;
    wobei das Verfahren ferner ein Erkennen der Installationseinstellung der Informationsverarbeitungsvorrichtung (10) umfasst und die Steuereinheit die Installationseinstellung der Informationsverarbeitungsvorrichtung (10) bestimmt und die Strahlungseinheit (18) zum Ausstrahlen der Bildinformationen veranlasst, falls die Personenerkennungseinheit eine Person erkennt oder wenn ein vorbestimmter Strahlungsstartvorgang durchgeführt wird, und einen Betrieb der Personenerkennungseinheit stoppt, falls die Strahlungsrichtung die Abwärtsrichtung ist.
  12. Programm, das eine Informationsvorrichtung (10) nach einem der Ansprüche 1 bis 10 zum Implementieren eines Steuerverfahrens nach Anspruch 11 veranlasst.
EP17897499.4A 2017-02-24 2017-12-22 Informationsverarbeitungsvorrichtung, informationsverarbeitungsverfahren und programm Active EP3588247B1 (de)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2017033594 2017-02-24
PCT/JP2017/046255 WO2018154954A1 (ja) 2017-02-24 2017-12-22 情報処理装置、情報処理方法、及びプログラム

Publications (3)

Publication Number Publication Date
EP3588247A1 EP3588247A1 (de) 2020-01-01
EP3588247A4 EP3588247A4 (de) 2020-01-08
EP3588247B1 true EP3588247B1 (de) 2023-07-12

Family

ID=63252604

Family Applications (1)

Application Number Title Priority Date Filing Date
EP17897499.4A Active EP3588247B1 (de) 2017-02-24 2017-12-22 Informationsverarbeitungsvorrichtung, informationsverarbeitungsverfahren und programm

Country Status (5)

Country Link
US (1) US10976648B2 (de)
EP (1) EP3588247B1 (de)
JP (1) JP6790223B2 (de)
CN (1) CN110168474A (de)
WO (1) WO2018154954A1 (de)

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3086406B2 (ja) * 1995-10-04 2000-09-11 オプテックス株式会社 受動型赤外線式人体検知装置
DE19640404A1 (de) * 1996-09-30 1998-04-09 Ldt Gmbh & Co Vorrichtung zur Darstellung von Bildern
JP4269425B2 (ja) * 1999-09-03 2009-05-27 ソニー株式会社 投影装置およびその安全装置
JP4620909B2 (ja) 2001-07-26 2011-01-26 日東光学株式会社 画像入出力装置
WO2006113848A2 (en) * 2005-04-19 2006-10-26 Virtek Vision International, Inc. Method and apparatus for protecting personnel using laser projection systems
JP2008287142A (ja) * 2007-05-21 2008-11-27 Brother Ind Ltd 画像投影装置
JP5277703B2 (ja) * 2008-04-21 2013-08-28 株式会社リコー 電子機器
JP5832069B2 (ja) 2009-02-25 2015-12-16 京セラ株式会社 携帯電子機器
JP5254906B2 (ja) * 2009-08-27 2013-08-07 京セラ株式会社 電子機器
JP5560721B2 (ja) * 2010-01-12 2014-07-30 セイコーエプソン株式会社 画像処理装置、画像表示システム、及び画像処理方法
JP5740822B2 (ja) * 2010-03-04 2015-07-01 ソニー株式会社 情報処理装置、情報処理方法およびプログラム
JP5648298B2 (ja) * 2010-03-15 2015-01-07 ソニー株式会社 情報処理装置、情報処理方法、およびプログラム
JP2013065061A (ja) * 2011-09-15 2013-04-11 Funai Electric Co Ltd プロジェクタ
JP6186869B2 (ja) * 2012-05-23 2017-08-30 株式会社リコー 画像処理方法及び画像処理装置
JP6015218B2 (ja) * 2012-08-06 2016-10-26 富士通株式会社 方位検出装置
JP2014115488A (ja) * 2012-12-11 2014-06-26 Funai Electric Co Ltd 画像表示装置
JP6167529B2 (ja) * 2013-01-16 2017-07-26 株式会社リコー 画像投影装置、画像投影システム、制御方法およびプログラム
JP6075122B2 (ja) * 2013-03-05 2017-02-08 株式会社リコー システム、画像投影装置、情報処理装置、情報処理方法およびプログラム
US10262462B2 (en) * 2014-04-18 2019-04-16 Magic Leap, Inc. Systems and methods for augmented and virtual reality
WO2015045125A1 (ja) * 2013-09-27 2015-04-02 日立マクセル株式会社 映像投射装置
WO2015092905A1 (ja) * 2013-12-19 2015-06-25 日立マクセル株式会社 投写型映像表示装置及び投写型映像表示方法
JP6222830B2 (ja) * 2013-12-27 2017-11-01 マクセルホールディングス株式会社 画像投射装置
WO2016035231A1 (ja) * 2014-09-03 2016-03-10 パナソニックIpマネジメント株式会社 ユーザインターフェース装置およびプロジェクタ装置
JPWO2016129489A1 (ja) * 2015-02-10 2018-01-18 シャープ株式会社 表示装置、表示装置の制御方法、制御プログラム、記録媒体

Also Published As

Publication number Publication date
US20190310538A1 (en) 2019-10-10
EP3588247A1 (de) 2020-01-01
US10976648B2 (en) 2021-04-13
WO2018154954A1 (ja) 2018-08-30
JPWO2018154954A1 (ja) 2019-12-26
JP6790223B2 (ja) 2020-11-25
CN110168474A (zh) 2019-08-23
EP3588247A4 (de) 2020-01-08

Similar Documents

Publication Publication Date Title
US10762876B2 (en) Information processing apparatus and control method
US9124686B2 (en) Portable device including automatic scrolling in response to a user's eye position and/or movement
US20160054794A1 (en) Eye-control reminding method, eye-control image display method and display system
BR112016026613B1 (pt) Sistema e método para fornecer retroalimentação háptica para auxiliar na captura de imagens
KR102449593B1 (ko) 카메라 장치를 제어하기 위한 방법 및 그 전자 장치
EP3051345B1 (de) Videoprojektionsvorrichtung
EP3208699A1 (de) Steuerungsvorrichtung, steuerungsverfahren und programm
KR101855141B1 (ko) 사용자 디바이스의 옵션 설정 방법 및 장치
JP5974189B2 (ja) 投写型映像表示装置及び投写型映像表示方法
RU2011137460A (ru) Устройство для захвата изображения и способ управления им
JP2011043876A (ja) 画像表示装置
US20120249524A1 (en) Information processing apparatus, display controlling method, and display controlling apparatus
WO2016188188A1 (zh) 投影设备及其控制方法,计算机存储介质
JP2018067840A (ja) 情報処理装置及び画像処理装置
JP2011166332A (ja) 情報処理装置
US20160231918A1 (en) Electronic device and method of providing user interface therefor
JP2017182109A (ja) 表示システム、情報処理装置、プロジェクター及び情報処理方法
EP3076272A3 (de) Inhaltsanzeigesteuerungsverfahren und -system basierend auf der erfassung der reflektion eines lichtstrahls
EP3588247B1 (de) Informationsverarbeitungsvorrichtung, informationsverarbeitungsverfahren und programm
KR102241074B1 (ko) 디스플레이 디바이스 및 그 제어 방법
EP3139258A1 (de) Verfahren und vorrichtung zur steuerung der automatischen drehung eines bildschirms und endgerät
CN111526346A (zh) 投影系统及投影方法
WO2014117675A1 (zh) 一种信息处理方法及电子设备
KR20130138935A (ko) 빔프로젝터 시스템 및 그 제공방법
JP2005190283A (ja) 情報入力装置

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20190904

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

A4 Supplementary search report drawn up and despatched

Effective date: 20191209

RIC1 Information provided on ipc code assigned before grant

Ipc: G09G 5/36 20060101ALI20191203BHEP

Ipc: G06T 7/70 20170101ALI20191203BHEP

Ipc: G06F 3/03 20060101ALI20191203BHEP

Ipc: G06F 3/01 20060101AFI20191203BHEP

Ipc: G06T 7/00 20170101ALI20191203BHEP

Ipc: G03B 21/14 20060101ALI20191203BHEP

Ipc: G09G 5/00 20060101ALI20191203BHEP

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20210409

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: SONY GROUP CORPORATION

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

INTG Intention to grant announced

Effective date: 20230217

RIC1 Information provided on ipc code assigned before grant

Ipc: H04N 9/31 20060101ALI20230203BHEP

Ipc: G06F 3/03 20060101ALI20230203BHEP

Ipc: G06F 3/01 20060101AFI20230203BHEP

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602017071347

Country of ref document: DE

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG9D

REG Reference to a national code

Ref country code: NL

Ref legal event code: MP

Effective date: 20230712

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 1587825

Country of ref document: AT

Kind code of ref document: T

Effective date: 20230712

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20231013

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20231112

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20231113

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20231012

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20231112

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20231013

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20231121

Year of fee payment: 7

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20230712

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT