US10015864B2 - Lighting preference arbitration - Google Patents
Lighting preference arbitration Download PDFInfo
- Publication number
- US10015864B2 US10015864B2 US15/509,631 US201515509631A US10015864B2 US 10015864 B2 US10015864 B2 US 10015864B2 US 201515509631 A US201515509631 A US 201515509631A US 10015864 B2 US10015864 B2 US 10015864B2
- Authority
- US
- United States
- Prior art keywords
- user
- light source
- light
- lighting
- controller
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Images
Classifications
-
- H05B37/0227—
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
- H05B47/175—Controlling the light source by remote control
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
- H05B47/105—Controlling the light source in response to determined parameters
- H05B47/115—Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings
- H05B47/125—Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings by using cameras
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02B—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
- Y02B20/00—Energy efficient lighting technologies, e.g. halogen lamps or gas discharge lamps
- Y02B20/40—Control techniques providing energy savings, e.g. smart controller or presence detection
Definitions
- the present disclosure relates to the arbitration of lighting preferences of users to control light sources.
- LEDs light-emitting diodes
- HID High-intensity discharge
- LED technology i.e. illumination based on semiconductor light sources, such as light-emitting diodes (LEDs)
- LEDs light-emitting diodes
- HID High-intensity discharge
- Lighting systems have been disclosed wherein a person can input his or her lighting preferences, and a central controller can control LEDs or other light sources to implement the person's lighting preferences.
- the inventors have recognized that in known solutions arbitration based on a user's dynamically changing orientation is not considered, meaning that their directional Field of View (FOV) is not considered.
- FOV Field of View
- Embodiments of the present disclosure relate to (1) applying a user's lighting preferences to an environment, based on what the user sees in this environment (i.e. his/her field of view); and (2) applying the user's lighting preferences to the environment based on where other users are in this environment and what these other users see.
- a controller comprising: an output for controlling one or more light source arranged to emit light to illuminate an environment; and an input for receiving a signal output from at least one sensor; wherein the controller is configured to: identify presence of a user and at least one further user in said environment based on the signal output from the at least one sensor; determine lighting preferences of the user and the at least one further user; detect an orientation of the user and the at least one further user based on the signal output from the at least one sensor; estimate a field of view of the user and the at least one further user based on the detected orientation of the user and the at least one further user; determine that at least one light source of the one or more light source is positioned to emit light in both the field of view of said user and the field of view of said at least one further user, arbitrate between the lighting preferences of the user and the at least one further user to determine light settings to be applied to the at least light source, and control said at least one light source based on the determined light settings.
- the controller is configured to perform said arbitration based on a comparison of a proportion of the field of view of the user in which the light emitted from the at least one light source is incident, and a proportion of the field of view of the at least one further user in which the light emitted from the at least one light source is incident; and the determined light settings are based (to a greater extent) on (i) the lighting preferences of the user if the proportion of the field of view of the user is greater than proportion of the field of view of the at least one further user, or (ii) the lighting preferences of the at least one further user if the proportion of the field of view of the at least one further user is greater than proportion of the field of view of the user.
- the controller is configured to perform said arbitration based on a comparison of a location in the field of view of the user at which the light emitted from the at least one light source is incident, and a location in the field of view of the at least one further user at which the light emitted from the at least one light source incident; and the determined light settings are (to a greater extent) based on (i) the lighting preferences of the user if the light emitted from the light source is incident in the field of view of the user in a more central location than in the field of view of the at least one further user, or (ii) the lighting preferences of the least one further user if the light emitted from the light source is incident in the field of view of the least one further user in a more central location than in the field of view of the user.
- the light settings to be applied to the at least one light source are determined based on the lighting preferences of the user and the at least one further user.
- the light settings to be applied to the at least one light source may be determined based on an average of the lighting preferences of the user and the at least one further user.
- the light settings to be applied to the at least one light source may be determined based on the outcome of an arbitration algorithm.
- Such an arbitration algorithm can determine, for example, matching preferences (e.g. there are two users, each of them prefers a dim level of 40%; as such this is the dim level that is set), overlapping preference ranges (e.g.
- the first user prefers a dim level between 30% and 60%
- the second user prefers a dim level between 40% and 80%
- the third user prefers a dim level between 50% and a 100%; as such a dim level is selected within the overlapping range of 50% to 60% —the selection of the dim level within this range can be based on the average dim level preference over all three users) or the algorithm can mediate between divergent preferences (e.g. a first user prefers a dim level between 20%-50% and a second user prefers a dim level between 70% and 100%; the dim level that is set is halfway between the upper bound of the first user and the lower bound of the second user), request user feedback to solve differences between preferences (e.g. request users to adapt their preferences), prioritize user preferences based on ranking (e.g. a user that is in his home location has preferences over a user that is a guest at that location), etc.
- divergent preferences e.g. a first user prefers a dim level between 20%-50% and a second user prefers a dim level
- the controller may further comprise an input for receiving location information of the one or more light source from a memory, wherein the controller is configured to determine that said at least one light source is positioned to emit light in both the field of view of said user and the field of view of said at least one further user based on said location information.
- the controller may be configured to determine said light settings based on properties of the at least one light source.
- the properties of the at least one light source may comprise one or any combination of: a light source type of the at least one light source, an available color range of light emitted from the at least one light source, an available brightness range of light emitted from the at least one light source, and an available beam spread range of light output from the at least one light source.
- the controller may be further configured to control said at least one light source of the one or more light source by transmitting said determined light settings to a lighting controller, the lighting controller configured to control said at least one light source based on received light settings.
- the controller may be configured to: capture biometric information of the user based on the signal output from the at least one sensor, and determine if the captured biometric information corresponds to biometric information stored in a memory coupled to the controller; wherein if the captured biometric information corresponds to biometric information stored in said memory, the controller further configured to retrieve lighting preferences associated with the biometric information stored in said memory to determine lighting preferences of the user;
- the controller further configured to retrieve default lighting preferences stored in said memory to determine lighting preferences of the user.
- the controller may be configured to: capture biometric information of the at least one further user based on the signal output from the at least one sensor, and determine if the captured biometric information corresponds to biometric information stored in a memory coupled to the controller; wherein if the captured biometric information corresponds to biometric information stored in said memory, the controller further configured to retrieve lighting preferences associated with the biometric information stored in said memory to determine lighting preferences of the at least one further user; wherein if the captured biometric information does not correspond to biometric information stored in said memory, the controller further configured to retrieve default lighting preferences stored in said memory to determine lighting preferences of the at least one further user.
- the lighting preferences of the user may comprise one or any combination of the user's preferred lighting brightness, lighting color, spread of light incident in the user's FOV, and light source type; and the lighting preferences of the at least one further user may comprise one or any combination of the at least one further user's preferred lighting brightness, lighting color, spread of light incident in the at least one further user's field of view, and light source type.
- the controller may be configured to execute an image recognition algorithm to detect an orientation of the user's head to detect the orientation of the user and to detect an orientation of the least one further user's head to detect the orientation of the least one further user.
- the controller may be configured to control said at least one light source of the one or more light source to emit light to illuminate said area of the environment based on an average of determined lighting preferences of the user and the at least one further user.
- the controller may be further configured to: determine that an area of the environment is present in only the field of view of said user; and determine that at least one light source of the one or more light source is positioned to emit light to illuminate the area of the environment that is present in only the field of view of said user, and control said at least one light source of the one or more light source to emit light to illuminate said area of the environment that is present in only the field of view of said user based on only the determined lighting preferences of said user.
- the controller may be further configured to: determine that an area of the environment is present in only the field of view of the at least one further user; and determine that at least one light source of the one or more light source is positioned to emit light to illuminate the area of the environment that is present in only the field of view of said at least one further user, and control said at least one light source of the one or more light source to emit light to illuminate said area of the environment that is present in only the field of view of said at least one further user based on only the determined lighting preferences of said at least one further user.
- a lighting system comprising: the controller referred to above, the at least one sensor; and the one or more light source.
- the at least one sensor may comprise one or any combination of a time of flight sensor and a structured light sensor.
- a computer program product for performing sensing using a signal output from at least one sensor to control one or more lighting source ( 106 ) to illuminate an environment
- the computer program product comprising code embodied on a computer-readable medium and configured so as when executed on one or more processing units to perform operations of: identifying presence of a user and at least one further user in said environment based on the signal output from the at least one sensor; determining lighting preferences of the user and the at least one further user; detecting an orientation of the user and the at least one further user based on the signal output from the at least one sensor; estimating a field of view of the user and the at least one further user based on the detected orientation of the user and the at least one further user; determine that at least one light source of the one or more light source is positioned to emit light in both the field of view of said user and the field of view of said at least one further user, arbitrate between the lighting preferences of the user and the at least one further user to determine light settings to be applied to the at least one
- FIG. 1 is a schematic block diagram of a lighting system
- FIG. 2 is a flow chart of a method to control lighting devices of the lighting system
- FIG. 3 a illustrates a forward facing vertical view angle of a user's field of view
- FIG. 3 b illustrates a forward facing horizontal view angle of a user's field of view
- FIG. 4 illustrates the field of views of user in an environment of the lighting system.
- FIG. 1 illustrates a schematic block diagram of a lighting system 100 .
- the lighting system 100 comprises an imaging controller 102 coupled to at least one sensor 104 and a memory 106 .
- the imaging controller 102 comprises an input 105 a for receiving a signal output from the sensor(s) 104 .
- the sensor(s) 104 may comprise a time of flight sensor comprising a time-of-flight sensing element.
- the time-of-flight sensing element is able to sense radiation emitted from an emitter, and this sensing is synchronized with the emission of the radiation from the emitter.
- the emitter may be a dedicated emitter which may be part of the time of flight sensor. In this case the emitted radiation may be radiation other than visible light, e.g.
- infrared, RF or ultrasound in order not to intrude upon or be confused with the visible light in the environment of the lighting system 100 ; or the radiation could be visible light modulated with an identifiable signal to distinguish it from the rest of the light in the environment of the lighting system 100 .
- the time of flight sensor can be used to determine the amount of time between emission from the emitter and reception back at the sensing element, i.e. time-of-flight information.
- the sensing element takes the form of a two-dimensional pixel array, and is able to associate a time-of-flight measurement with a measurement of the radiation captured by some or all of the individual pixels.
- the time-of-flight sensor is operable to capture a depth-aware or three-dimensional image in its SR, including a detected object.
- the time-of-flight sensor may also be referred to as a depth-aware or 3D camera.
- image recognition By applying image recognition to the depth-aware or 3D images captured by the time of flight sensor, it is possible to detect information such as the location of a detected person in the environment of the lighting system 100 , and the direction in which the person is facing i.e. the persons orientation. Details of time-of-flight based image sensing in themselves will be familiar to a person skilled in the art, and are not described in any further detail herein.
- the sensor(s) 104 may comprise a structured light sensor.
- Structured lighting is a well-known technique for use in extraction of 3D information from a scene. The extraction of 3D information is based on the projection of a known lighting pattern on a scene, and capturing of the resulting image with a sensor (camera) of the scene. The technique works on the basis of disparity with a distance (often called the baseline) between the camera and a projector that projects the lighting pattern. As a result of this disparity, the depth of an object (seen from the camera) can be determined.
- the sensor(s) 104 may comprises wearable or other sensors local to the user—such as a smartwatch, glasses etc.
- wearable technology such Google Glasses can be used to determine the orientation of a user's head, and a smartwatch which will usually be worn on a user's wrist, can be used to determine the orientation of the user's head given that a smartwatch will typically have a defined relationship with the orientation of the users head.
- the imaging controller 102 is coupled to a lighting controller 108 .
- the imaging controller 102 may be coupled to the lighting controller 108 via a wired connection over a wired network or a wireless connection over a wireless network.
- the imaging controller 102 may be coupled to the lighting controller 108 via an I 2 C bus, a DMX bus, a DALI bus, Ethernet connection Wi-Fi connection or ZigBee connection.
- the imaging controller 102 is configured to determine light settings based on signals output from the sensor(s) 104 and supply these to the lighting controller 108 via an interface 105 c of the imaging controller 102 .
- the lighting controller 108 is coupled to one or more light source 110 that are operable to emit light to illuminate an environment of the lighting system 100 .
- the environment may be any indoor or outdoor space.
- a light source may be mounted on the ceiling of an indoor space, a wall, floor or ground, or may be disposed elsewhere such as in a free-standing unit.
- the lighting controller 108 In response to receiving light settings from the imaging controller 102 , the lighting controller 108 is configured to transmit appropriate control signals to the light source(s) 110 to apply the light settings to the light source(s) 110 . That is, the lighting controller 108 responds to instructions received from the imaging controller 102 in order to control the light source(s) 110 .
- a light source may comprise any suitable source of light such as e.g. a laser diode, an LED, an incandescent source (e.g., filament lamp, halogen lamp), fluorescent source, and phosphorescent source.
- a laser diode e.g., an LED, an incandescent source (e.g., filament lamp, halogen lamp), fluorescent source, and phosphorescent source.
- the term “LED” should be understood to include any electroluminescent diode or other type of carrier injection/junction-based system that is capable of generating radiation in response to an electric signal.
- the term LED includes, but is not limited to, various semiconductor-based structures that emit light in response to current, light emitting polymers, organic light emitting diodes (OLEDs), electroluminescent strips, and the like.
- FIG. 2 illustrates a process 200 implemented by the imaging controller 102 for controlling the light emitted from the light source(s) 110 .
- the functionality of the imaging controller 102 may be implemented in code (software) stored on a memory comprising one or more storage media, and arranged for execution on a processor comprising on or more processing units.
- the code is configured so as when fetched from the memory and executed on the processor to perform operations in line with embodiments discussed below.
- some or all of the functionality of the imaging controller 102 is implemented in dedicated hardware circuitry, or configurable hardware circuitry like a field-programmable gate array (FPGA).
- FPGA field-programmable gate array
- the imaging controller 102 detects the presence of the first user and the second user in the environment of the lighting system 100 based on the signal received via interface 105 a from the sensor(s) 104 .
- Step 5202 may be implemented by means of the imaging controller 102 executing a facial recognition algorithm to analyses images captured by the sensor(s) 104 to detect one or more facial feature of the first user and the second user. It is common for such facial recognition algorithms to detect the presence of a user based on detecting a user's nose, based on its central position within the face, the fact that it is likely to not be obscured, and its relative ease of detection using both amplitude and depth based approaches.
- the imaging controller 102 determines lighting preferences of the first user and the second user present in the environment of the lighting system 100
- the characteristics of known users may comprise biometric information.
- the biometric information may comprise one or more of facial measurement data (e.g. distances between the eyes, nose and mouth of a consumer) other body measurement data, and iris pattern data.
- the imaging controller 102 is able to determine lighting preferences of the known user at step S 204 by comparing biometric information extracted from images captured by sensor(s) 104 with the stored biometric associated with the known user, and retrieving lighting preferences associated with the stored biometric information which matches the biometric information extracted from the images captured by sensor(s) 104 via the interface 105 b.
- the known user information may be collected as part of a registration process which a user carries out.
- the registration process collects data associated with a user (for example biometric information of the user) and lighting preferences of the user.
- the imaging controller is configured to determine biometric information of a user based on the signal received from the sensor(s) 104 and store the determined biometric information in the memory 106 .
- a user may load their lighting preferences into the memory 106 using an input device e.g. keyboard mouse, touch screen, microphone etc.
- the input device may be coupled to the imaging controller 102 via a wired or wireless connection in which case the imaging controller 102 is configured, in response to receiving a user's lighting preferences to store the received lighting preferences in the memory 106 .
- the input device may be coupled to the memory 106 via a wired or wireless connection in which case lighting preferences may be loaded directly into memory 106 . Following the registration process a user becomes a “known” user.
- Lighting preferences of a known user may include for example, the user's preferred lighting brightness, lighting color, light source type (e.g. preference for incandescent lighting over fluorescent lighting), and spread of light incident in the user's FOV.
- the imaging controller 102 may determine that a user is an “unknown” user based on determining that biometric information of the user extracted from images captured by sensor(s) 104 does not match any of the biometric associated with known users stored in the memory 106 . In these scenarios, the imaging controller 102 is configured to associate an unknown user with predetermined lighting preferences (e.g. default lighting preferences). These predetermined lighting preferences may be stored in memory 106 .
- predetermined lighting preferences e.g. default lighting preferences
- the imaging controller 102 determines the location and properties of each light source of the light source(s) 110 that is controllable by the lighting controller 108 . It will be appreciated that all of the light source(s) 110 may be controllable by the lighting controller 108 , however in other embodiments one or more light source may not be controllable by the lighting controller 108 .
- the properties of a light source may include the light source type, the available range of color of light emitted from the light source, the available range of brightness of light emitted from the light source, the available beam spread range of light output by the light source (defined by a beam angle or beam angle range) etc.
- the information on the location and the properties of the light source(s) 110 may be stored in memory 106 and retrieved by the imaging controller 102 via interface 105 b.
- the imaging controller 102 detects the orientation of the first user and the second user present in the environment of the lighting system 100 by detecting the orientation of each user's head based on the signal received via interface 105 a from the sensor(s) 104 .
- Step S 208 may be implemented by means of the imaging controller 102 executing a facial recognition algorithm and/or utilizing other image recognition techniques, for example by tracking facial features of the users or by analyzing the movement of the users' limb(s).
- the imaging controller 102 estimates a Field of View (FOV) of each user present in the environment of the lighting system 100 .
- a FOV of a user is the area in the environment of the lighting system 100 that is visible to the user.
- a FOV is defined by a forward facing vertical view angle ⁇ v (shown in FIG. 3 a ) and a forward facing horizontal view angle ⁇ h (shown in FIG. 3 b ).
- the imaging controller 102 is preconfigured with values of the angles ⁇ v and ⁇ h. It has been found that humans have an almost 180-degree forward-facing horizontal field of view, whilst the vertical range of the field of view in humans is typically around 135-degrees. Therefore the forward facing vertical view angle ⁇ v may be set to 135 degrees, and the forward facing horizontal view angle ⁇ h may be set to 180 degrees. It will be appreciated that these values of the angles ⁇ v and ⁇ h are merely examples.
- the first user's FOV may overlap with the second user's FOV.
- the imaging controller 102 determines that at least one light source (controllable by the lighting controller 108 ) is positioned to emit light in both the FOVs of the first user and the second user, using the information on the location of the light source(s) 110 in the environment of the lighting system 100 (determined at step S 206 ) and the FOVs estimated at step S 210 .
- the arbitration performed at step S 214 may be implemented in a number of ways.
- the imaging controller 102 is configured to perform the arbitration of step S 214 based on what extent the light from the light source 110 is present in the user's FOV compared with in the further user's FOV. That is, the imaging controller 102 compares a proportion of the FOV of the first user in which the light emitted from the light source 110 is incident, and a proportion of the FOV of the second user in which the light emitted from the light source 110 is incident.
- the light settings to be applied to the light source 110 are based on (i) the lighting preferences of the first user if the proportion of the field of view of the first user is greater than proportion of the field of view of the second user, or (ii) the lighting preferences of the second user if the proportion of the field of view of the second user is greater than proportion of the field of view of the first user. For example if the light emitted by the light source 110 is present in 5% of the first user's FOV and 70% in the second user's FOV then the imaging controller 102 determines light settings to be applied to the light source 110 based on the lighting preferences of the second user.
- the imaging controller 102 is configured to perform the arbitration of step S 214 based where the light effect of the light source 110 is visible in the first user's FOV compared with in the second user's FOV. That is, the imaging controller 102 compares a location in the FOV of the first user at which the light emitted from the light source 110 incident, and location in the FOV of the second user at which the light emitted from the light source incident.
- the light settings to be applied to the light source 110 are based on (i) the lighting preferences of the first user if the light emitted from the light source 110 is incident in the FOV of the first user in a more central location than in the field of view of the second user, or (ii) the lighting preferences of the second user if the light emitted from the light source 110 is incident in the field of view of the second user in a more central location than in the field of view of the first user. For example, if the light effect of the light source 110 is in the center of the first user's FOV but in the periphery of the second user's FOV then the imaging controller 102 determines light settings to be applied to the light source 110 based on the lighting preferences of the first user.
- Reference herein to the central location of a user's FOV is used to refer to an imaginary axis projecting perpendicular from a user's eye line in the direction the user is facing, thus the closer the location of the light effect of the light source 110 that is incident in the field of view of a user is to this axis the more central its location.
- the imaging controller 102 is configured to perform the arbitration of step S 214 based on the lighting preferences of both the user and the second user.
- the arbitration performed at step S 214 may comprise determining an average of the lighting preferences of the first user and the lighting preferences of the second user.
- the first user may prefer the light emitted from the plurality of lighting devices that is incident in his FOV to have a first brightness value
- the second user may prefer the light emitted from the plurality of lighting devices that is incident in his FOV to have a second brightness value
- the imaging controller 102 may determine an average brightness value (a computed light setting) based on the first brightness value and the second brightness value.
- the imaging controller 102 may determine an arbitrated brightness range (a computed light setting) comprising brightness values falling in both the first brightness range and the second brightness range.
- the brightness value can be expressed for example in terms of illuminance (in lux) i.e. in terms of the amount of light emitted from the light source(s) 110 that is incident over a plane of interest in the environment of the lighting system 100 or in terms of luminous flux (in lumens) i.e. the quantity of the energy of the light emitted from the light source(s) 110 (that emit light that is incident in the area of the environment where the FOVs of the first user and the second user overlap) per second in all directions.
- illuminance in lux
- luminous flux in lumens
- a transition profile may be assigned to the users according to the FOV of each user, the light visible in their FOV and the overlap of their FOV.
- the imaging controller 102 controls the light which is present in the shared FOV to achieve a transition between the preferences of each user such that the light present in parts of their FOV which is shared with another user transitions between the preferences of the users.
- the transition profile may be linear or non-linear, and conform to one of a range of functions.
- the imaging controller 102 is configured to determine light settings to be applied to the light source 110 according to the individual lighting preferences of the first user only.
- the imaging controller 102 is configured to determine light settings to be applied to the light source 110 according to the individual lighting preferences of the second user only.
- the spread of light from a light source may be controlled so as to control the light which is present in each user's FOV. This is especially useful in situations where light from a single light source spreads across two users FOV even though the FOV of the two users do not overlap. In such a situation the spread of a lamp may be narrowed so that it has greater influence on a single user's FOV.
- the spread of the light source may be controlled according to how similar the lighting preferences desired by each user are. For example, if the desired lighting preferences are the same, then the spread can be set at its maximum value. If they are different, then the spread of the light can be restricted to a specific user.
- the imaging controller 102 communicates the light settings to the lighting controller 108 via the interface 105 c of the imaging controller 102 .
- the lighting controller 108 is configured to transmit appropriate control signals to the light source(s) 110 to apply the light settings to the light source(s) 110 .
- FIG. 4 illustrates a FOV of a first user 402 and a FOV of a second user 404 .
- the first user 402 is sitting opposite to, and facing, the second user 404 in the environment of the lighting system 100 .
- the imaging controller 102 is configured to determine light settings for light emitted by the light source(s) 110 that is incident in the area 406 of the environment of the lighting system 100 in the FOV of the first user 402 only, according to the individual lighting preferences of the first user 402 .
- the imaging controller 102 is configured to determine light settings for light emitted by the light source(s) 110 that is incident in the area 408 of the environment of the lighting system 100 in the FOV of the second user 404 only, according to the individual lighting preferences of the second user.
- Area 410 of the environment of the lighting system 100 is where the FOV of the first user 402 overlaps with the FOV of the second user 404 .
- the imaging controller 102 identifies each of the light sources that are positioned such that they emit light into the area 410 and arbitrates between the lighting preferences of the first user 402 and the second user 404 in order to determine light settings for each of these identified light sources.
- the steps described above with reference to FIG. 2 may be iterated periodically on a time driven basis, or may be iterated on an event driven basis.
- the steps of FIG. 2 may be iterated based on the image controller detecting a gesture performed by a user in the environment of the lighting system 100 .
- the imaging controller is configured to detect a gesture performed by a user based on the signal received from the sensor(s) 104 .
- a gesture is an action performed explicitly by a human user for the purpose of signaling an intention to control some system or apparatus.
- the term “gesture” as used herein refers to motion of at least one bodily member through the air (or other fluid medium of the space the user is occupying, e.g. water). As such it excludes an action detected purely by detecting the bodily member in question being pressed or held against a button or a control surface such as a touch screen, pressure pad or track pad (whether directly or indirectly via an implement such as a stylus). It also excludes detection purely based on moving a sensing device over a surface while in contact with that surface, e.g. a mouse. Rather, detecting a gesture means detecting the motion of the user by means of a contactless or non-tactile sensing principle.
- a gesture may be an explicit gesture such as a user raising or lowering their hand or a non-explicit gesture such as the user walking, reading, writing or operating a personal computer.
- a user may associate lighting preferences with one or more gesture.
- Information on a user's gestures and associated lighting preferences are stored in memory 106 .
- a user may specify that when the user raises a hand, light sources which emit light in his field of view are to be dimmed up in proportion to a magnitude of the gesture .g. how far is the hand raised.
- a user may specify that when the user raises a hand, light sources which emit light in his field of view are to be dimmed down in proportion to a magnitude of the gesture .g. how far is the hand lowered.
- “proportional” as used herein does not necessarily mean proportional in the strict mathematical sense, i.e. not necessarily a linear relationship between any two quantities such as intensity and the distance a hand is moved (though such a relationship is one possibility).
- the image controller 102 is configured to detect a gesture performed by a user based on the signal received via interface 105 a from the sensor(s) 104 , and retrieve lighting preferences of the user associated with the detected gesture from memory 106 at step S 204 .
- the effect of the requested change on other users in the environment of the lighting system 100 is calculated according to (i) the orientation of other users present in the environment of the lighting system 100 and (ii) the preferences of other users present in the environment of the lighting system 100 .
- the requested lighting change is arbitrated by scaling it according to the effect the requested change in lighting has on other users, based on the orientation and preferences of the users present in the environment of the lighting system 100 .
- a computer program may be stored/distributed on a suitable medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems. Any reference signs in the claims should not be construed as limiting the scope.
Abstract
Description
Claims (15)
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP14183935 | 2014-09-08 | ||
EP14183935 | 2014-09-08 | ||
EP14183935.7 | 2014-09-08 | ||
PCT/EP2015/068069 WO2016037772A1 (en) | 2014-09-08 | 2015-08-05 | Lighting preference arbitration. |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170265277A1 US20170265277A1 (en) | 2017-09-14 |
US10015864B2 true US10015864B2 (en) | 2018-07-03 |
Family
ID=51492223
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/509,631 Expired - Fee Related US10015864B2 (en) | 2014-09-08 | 2015-08-05 | Lighting preference arbitration |
Country Status (6)
Country | Link |
---|---|
US (1) | US10015864B2 (en) |
EP (1) | EP3192330B1 (en) |
JP (1) | JP6736543B2 (en) |
CN (1) | CN106797692B (en) |
RU (1) | RU2698303C2 (en) |
WO (1) | WO2016037772A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170285594A1 (en) * | 2016-03-30 | 2017-10-05 | Lenovo (Singapore) Pte. Ltd. | Systems and methods for control of output from light output apparatus |
Families Citing this family (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11592723B2 (en) | 2009-12-22 | 2023-02-28 | View, Inc. | Automated commissioning of controllers in a window network |
US11054792B2 (en) | 2012-04-13 | 2021-07-06 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
US10964320B2 (en) | 2012-04-13 | 2021-03-30 | View, Inc. | Controlling optically-switchable devices |
US9872367B2 (en) | 2012-07-01 | 2018-01-16 | Cree, Inc. | Handheld device for grouping a plurality of lighting fixtures |
US10721808B2 (en) | 2012-07-01 | 2020-07-21 | Ideal Industries Lighting Llc | Light fixture control |
US9572226B2 (en) | 2012-07-01 | 2017-02-14 | Cree, Inc. | Master/slave arrangement for lighting fixture modules |
US10342102B2 (en) * | 2012-07-01 | 2019-07-02 | Cree, Inc. | Light fixture control |
US10219338B2 (en) | 2012-07-01 | 2019-02-26 | Cree, Inc. | Modular lighting control |
US10154569B2 (en) | 2014-01-06 | 2018-12-11 | Cree, Inc. | Power over ethernet lighting fixture |
US10859983B2 (en) | 2014-03-05 | 2020-12-08 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
US10278250B2 (en) | 2014-05-30 | 2019-04-30 | Cree, Inc. | Lighting fixture providing variable CCT |
CN104284486A (en) * | 2014-09-26 | 2015-01-14 | 生迪光电科技股份有限公司 | Intelligent lighting device and system and intelligent lighting control method |
GB2545171B (en) * | 2015-12-03 | 2020-08-19 | Sony Interactive Entertainment Inc | Light source identification apparatus and method |
CN116893741A (en) * | 2016-04-26 | 2023-10-17 | 唯景公司 | Controlling an optically switchable device |
WO2018008101A1 (en) * | 2016-07-06 | 2018-01-11 | 株式会社オプティム | Image provision system, image provision method, and program |
EP3513630B1 (en) * | 2016-09-16 | 2020-09-02 | Signify Holding B.V. | Illumination control |
WO2018059985A1 (en) * | 2016-09-29 | 2018-04-05 | Philips Lighting Holding B.V. | Depth queue by thermal sensing |
EP3735804B1 (en) | 2018-01-05 | 2021-10-13 | Signify Holding B.V. | A controller for controlling a lighting device and a method thereof |
EP3813488A4 (en) * | 2018-06-19 | 2022-02-23 | Iris Communication Kabushiki Kaisha | Lighting control system, lighting control method, and lighting control program |
WO2020254225A1 (en) | 2019-06-18 | 2020-12-24 | Signify Holding B.V. | System and methods to provide group lighting interaction |
NL2023811B1 (en) * | 2019-09-11 | 2021-05-17 | Microsoft Technology Licensing Llc | Validating and providing accessible electronic messages |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010079388A1 (en) | 2009-01-07 | 2010-07-15 | Koninklijke Philips Electronics N.V. | Intelligent controllable lighting networks and schemata therefore |
US20120206050A1 (en) * | 2002-07-12 | 2012-08-16 | Yechezkal Evan Spero | Detector Controlled Illuminating System |
WO2013111134A1 (en) | 2012-01-24 | 2013-08-01 | Yechezkal Evan Spero | Detector controlled illuminating system |
EP2645823A1 (en) | 2012-03-21 | 2013-10-02 | Osram Sylvania Inc. | Dynamic lighting based on activity type |
WO2014064587A1 (en) * | 2012-10-26 | 2014-05-01 | Koninklijke Philips N.V. | Lighting methods for providing personalized lighting to users positioned proximal to one another |
US9655193B2 (en) * | 2014-01-08 | 2017-05-16 | Philips Lighting Holding B.V. | Lighting unit providing reduced intensity light output based on user proximity and related methods |
US20170231064A1 (en) * | 2014-09-25 | 2017-08-10 | Philips Lighting Holding B.V. | Control of lighting |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08273839A (en) * | 1995-04-04 | 1996-10-18 | Matsushita Electric Ind Co Ltd | Lighting control system |
US7111952B2 (en) * | 2003-03-24 | 2006-09-26 | Lutron Electronics Co., Inc. | System to control daylight and artificial illumination and sun glare in a space |
EP2208397B1 (en) * | 2007-11-06 | 2018-10-03 | Philips Lighting Holding B.V. | Light control system and method for automatically rendering a lighting scene |
US8108055B2 (en) * | 2007-12-28 | 2012-01-31 | Larry Wong | Method, system and apparatus for controlling an electrical device |
WO2010125325A1 (en) * | 2009-04-28 | 2010-11-04 | Dialight Corporation | Method and apparatus for multi-zoned illumination |
CA2766889C (en) * | 2009-06-30 | 2017-08-01 | Koninklijke Philips Electronics N.V. | Systems and methods for managing interaction with controllable lighting networks |
CN102484930B (en) * | 2009-09-21 | 2015-08-19 | 皇家飞利浦电子股份有限公司 | For the method and system in lighting atmosphere market |
WO2012146256A2 (en) * | 2011-04-28 | 2012-11-01 | Lighten Aps | A lighting system and a method for locally changing light conditions |
JP2013054992A (en) * | 2011-09-06 | 2013-03-21 | Sharp Corp | Apparatus control device |
-
2015
- 2015-08-05 WO PCT/EP2015/068069 patent/WO2016037772A1/en active Application Filing
- 2015-08-05 RU RU2017111815A patent/RU2698303C2/en active
- 2015-08-05 EP EP15744618.8A patent/EP3192330B1/en active Active
- 2015-08-05 US US15/509,631 patent/US10015864B2/en not_active Expired - Fee Related
- 2015-08-05 JP JP2017512930A patent/JP6736543B2/en not_active Expired - Fee Related
- 2015-08-05 CN CN201580048293.3A patent/CN106797692B/en not_active Expired - Fee Related
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120206050A1 (en) * | 2002-07-12 | 2012-08-16 | Yechezkal Evan Spero | Detector Controlled Illuminating System |
WO2010079388A1 (en) | 2009-01-07 | 2010-07-15 | Koninklijke Philips Electronics N.V. | Intelligent controllable lighting networks and schemata therefore |
WO2013111134A1 (en) | 2012-01-24 | 2013-08-01 | Yechezkal Evan Spero | Detector controlled illuminating system |
EP2645823A1 (en) | 2012-03-21 | 2013-10-02 | Osram Sylvania Inc. | Dynamic lighting based on activity type |
WO2014064587A1 (en) * | 2012-10-26 | 2014-05-01 | Koninklijke Philips N.V. | Lighting methods for providing personalized lighting to users positioned proximal to one another |
US9655193B2 (en) * | 2014-01-08 | 2017-05-16 | Philips Lighting Holding B.V. | Lighting unit providing reduced intensity light output based on user proximity and related methods |
US20170231064A1 (en) * | 2014-09-25 | 2017-08-10 | Philips Lighting Holding B.V. | Control of lighting |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170285594A1 (en) * | 2016-03-30 | 2017-10-05 | Lenovo (Singapore) Pte. Ltd. | Systems and methods for control of output from light output apparatus |
US10719059B2 (en) * | 2016-03-30 | 2020-07-21 | Lenovo (Singapore) Pte. Ltd. | Systems and methods for control of output from light output apparatus |
Also Published As
Publication number | Publication date |
---|---|
JP2017527087A (en) | 2017-09-14 |
EP3192330A1 (en) | 2017-07-19 |
WO2016037772A1 (en) | 2016-03-17 |
CN106797692A (en) | 2017-05-31 |
JP6736543B2 (en) | 2020-08-05 |
RU2017111815A3 (en) | 2019-03-14 |
EP3192330B1 (en) | 2020-12-02 |
RU2698303C2 (en) | 2019-08-26 |
RU2017111815A (en) | 2018-10-10 |
US20170265277A1 (en) | 2017-09-14 |
CN106797692B (en) | 2019-08-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10015864B2 (en) | Lighting preference arbitration | |
RU2729045C2 (en) | Adaptive lighting system for mirror component and method for controlling adaptive lighting system | |
JP6445025B2 (en) | Gesture control | |
US9301372B2 (en) | Light control method and lighting device using the same | |
US9504126B2 (en) | Coded light detector | |
US9942966B2 (en) | Control of lighting | |
US20150230320A1 (en) | Controlling light source(s) via a portable device | |
EP3198372B1 (en) | Control of lighting | |
US20180054876A1 (en) | Out of plane sensor or emitter for commissioning lighting devices | |
WO2016206991A1 (en) | Gesture based lighting control | |
JP7286159B2 (en) | Depth cues by thermal sensing | |
US20170097685A1 (en) | Method and system for controlling fixtures | |
BR112018068962B1 (en) | ADAPTIVE LIGHTING SYSTEM FOR A MIRROR COMPONENT CONFIGURED TO SHOW A MIRROR IMAGE OF A USER ON A MIRROR SURFACE, MIRROR ARRANGEMENT, VIRTUAL MIRROR ARRANGEMENT, METHOD FOR CONTROLING THE LIGHT SOURCES OF AN ADAPTABLE LIGHTING SYSTEM FOR A COMPONENT OF MIRROR CONFIGURED TO SHOW A MIRROR IMAGE OF A USER ON A MIRROR SURFACE OF THE MIRROR COMPONENT, METHOD FOR CONTROLING A VIRTUAL MIRROR, AND COMPUTER READABLE MEDIA FOR CONTROLING THE LIGHT SOURCES OF AN ADAPTIVE LIGHTING SYSTEM FOR A COMPONENT MIRROR |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KONINKLIJKE PHILIPS N.V., NETHERLANDS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NOLAN, JULIAN CHARLES;LAWRENSON, MATTHEW JOHN;VAN EEUWIJK, ALEXANDER HENRICUS WALTERUS;AND OTHERS;SIGNING DATES FROM 20150824 TO 20150826;REEL/FRAME:043182/0904 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: PHILIPS LIGHTING HOLDING B.V., NETHERLANDS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KONINKLIJKE PHILIPS N.V.;REEL/FRAME:050429/0060 Effective date: 20160201 |
|
AS | Assignment |
Owner name: SIGNIFY HOLDING B.V., NETHERLANDS Free format text: CHANGE OF NAME;ASSIGNOR:PHILIPS LIGHTING HOLDING B.V.;REEL/FRAME:050837/0576 Effective date: 20190201 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20220703 |