WO2019246431A1 - Sytème d'éclairage doté d'un capteur intégré - Google Patents

Sytème d'éclairage doté d'un capteur intégré Download PDF

Info

Publication number
WO2019246431A1
WO2019246431A1 PCT/US2019/038297 US2019038297W WO2019246431A1 WO 2019246431 A1 WO2019246431 A1 WO 2019246431A1 US 2019038297 W US2019038297 W US 2019038297W WO 2019246431 A1 WO2019246431 A1 WO 2019246431A1
Authority
WO
WIPO (PCT)
Prior art keywords
light sources
spectrum
primary light
scene
image data
Prior art date
Application number
PCT/US2019/038297
Other languages
English (en)
Inventor
Wouter Soer
Willem SILLEVIS-SMITT
Original Assignee
Lumileds Llc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US16/015,697 external-priority patent/US10582589B2/en
Application filed by Lumileds Llc filed Critical Lumileds Llc
Priority to JP2020571355A priority Critical patent/JP7036952B2/ja
Priority to EP19733920.3A priority patent/EP3811736A1/fr
Publication of WO2019246431A1 publication Critical patent/WO2019246431A1/fr

Links

Classifications

    • HELECTRICITY
    • H05ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
    • H05BELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
    • H05B45/00Circuit arrangements for operating light-emitting diodes [LED]
    • H05B45/20Controlling the colour of the light
    • H05B45/22Controlling the colour of the light using optical feedback

Definitions

  • Tunable lighting systems may be used to illuminate one or more scenes containing objects and may be adjustable such that the light output by such systems is varied based on user input. Such tunable lighting systems may be adjusted to, for example, increase or decrease the amount and/or type of light that is illuminated onto a scene. Further, such tunable lighting systems may include multiple light sources, such as multiple light bulbs, to illuminate a scene.
  • the following description includes techniques and devices provided for sensing image data from a scene and activating primary light sources based on the image data.
  • Subsets of a plurality of primary light sources may be activated to emit a sensing spectrum of light onto a scene.
  • Image data may be sensed from the scene while the subsets of the plurality of primary light sources are activated.
  • Reflectance information for the scene may be determined based on the combined image data.
  • Spectrum optimization criteria for the primary light sources may be determined based on the reference information and a desired output parameter provided by a user or determined by a controller.
  • the plurality of primary light sources may be activated to emit a lighting spectrum based on the spectrum optimization criteria.
  • FIG. l is a flowchart for activating primary light sources based on spectrum optimization criteria
  • FIG. 2 is an example diagram of a image sensor, controller, and plurality of primary light sources
  • Fig. 3 A is an example chart of chromaticities of five primaries
  • Fig. 3B is an example chart of the spectra of the five primaries of Fig. 3 A;
  • Fig. 4A is an example chart of spectra of the five primaries of Figs. 3 A and 3B with their corresponding TM-30 indices including spectral power plotted against wavelength;
  • Fig. 4B is an example chart of spectra of the five primaries of Figs. 3 A and 3B with their corresponding TM-30 indices including Rg plotted against Rf;
  • Fig. 4C is an example chart of spectra of the five primaries of Figs. 3 A and 3B with their corresponding TM-30 indices including Rf plotted against hue bins;
  • Fig. 4D is an example chart of spectra of the five primaries of Figs. 3A and 3B with their corresponding TM-30 indices including Res plotted against hue bins;
  • Fig. 5 shows an example chart of reference spectra with estimated reference spectra using the five primaries of Figs. 3 A and 3B and a polynomial fit algorithm
  • Fig. 6 shows actual color points of TM-30 CES 1-99 in CAM02-USC and estimated color points using the five primaries of Fig. 3 A and 3B;
  • Fig. 7A shows example color points of several CES in each of the hue bins 1,
  • Fig. 7B shows an example color vector diagram showing three color rending modes for high saturation of the hue bins for high fidelity of Fig. 7A;
  • FIG. 8 shows a flowchart for factory input data provided to an on-board processing unit.
  • Tunable lighting arrays may support applications that benefit from distributed intensity, spatial, and temporal control of light distribution.
  • Primary light sources may be light emitting devices such as LEDs that emit a given color.
  • Tunable lighting array based applications may include, but are not limited to, precise spatial patterning of emitted light from pixel blocks or individual pixels. Depending on the application, emitted light may be spectrally distinct, adaptive over time, and/or environmentally responsive.
  • the light emitting arrays may provide scene based light distribution in various intensity, spatial, or temporal patterns. The emitted light may be based at least in part on received sensor data, as disclosed herein.
  • Associated optics may be distinct at a pixel, pixel block, or device level. Common applications supported by light emitting arrays include architectural and area illumination, professional lighting, retail lighting and/or exhibition lighting, and the like.
  • Use of a tunable light system may provide controlled illumination of portions of a scene for a determined amount of time. This may allow the array to, for example, emphasize certain colors or color properties within a scene, emphasize white backgrounds, emphasize moving objects and/or the like. Architectural and area illumination may also benefit from the lighting disclosed herein.
  • an optimal lighting spectrum may vary based on the illuminated scene.
  • the variation may be a result of the objects, colors, and angles presented in the scene and may also vary based on one or more intended desired output parameters.
  • a tunable lighting system may automatically adjust the lighting spectrum based on the scene to be illuminated and/or one or more desired output parameters.
  • a first subset of a plurality of primary light sources may be activated to emit a first sensing spectrum onto a scene at 110.
  • the sensing spectrum may refer to the light emitted by a subset of the plurality of primary light sources while image data is collected via an image sensor.
  • the sensing spectrum may include light that is not visible to a human viewing the scene.
  • first image data may be sensed from the scene while the first subset of the plurality of primary light sources are activated.
  • a second subset of the plurality of primary light sources may be activated to emit a second sensing spectrum onto a scene.
  • second image data may be sensed from the scene while the second subset of the plurality of primary light sources are activated.
  • the plurality of light sources may be activated in subsets such that a subset is activated, image data is collected while that subset is activated, and then another subset is activated and additional image data is collected. This process may be repeated such that each subset of the plurality of primary light sources corresponds to a primary light source and image data is collected as each primary light source is activated.
  • at least four or five primary light sources may be provided in a lighting system disclosed herein.
  • reference information for the scene may be determined based on combined image data where combined image data is a combination of the first image data and the second image data.
  • This combined image data may be collected by combining image data while different subsets of the plurality of primary light sources are activated. It should be noted that combined image data does not necessarily refer to different image data that is added together as combined image data may simply be the collection of a number of different image data.
  • spectrum optimization criteria for the plurality of primary light sources may be determined based on the reference information and one or more desired output parameters. The desired output parameters may be input by a user or a component or may be determined based on the scene, as further disclosed herein.
  • the plurality of primary light sources may be activated to emit a lighting spectrum based on the spectrum optimization criteria.
  • the lighting spectrum may refer to the light emitted by the plurality of primary light sources based on the spectrum optimization criteria such that the lighting spectrum is visible to a human viewing the scene.
  • FIG. 2 shows an example diagram of a lighting system 200 as disclosed herein.
  • a substrate 210 may be a mount or housing on which the components of the lighting system 200 are attached to or placed on.
  • a plurality of primary light sources 260 may be provided and may emit light as disclosed herein.
  • the plurality of primary light sources 260 may be separately addressable channels such that a first channel may correspond to a first primary light source (e.g., LEDs that emit red light) and a second channel may correspond to a second primary light source (e.g., LEDs that emit royal blue light).
  • a first optic lens 240 may be proximate to the primary light sources 260 such that all or a part of the light emitted by the primary light sources 260 may pass through the first lens 240 and may be shaped or adjusted by the first optic lens 240.
  • first optic lens 240 is shown as one component, it may be a combination of multiple components and multiple components may be configured such that one or a subset of the components are aligned with one or a subset of the plurality of primary light sources.
  • an image sensor 220 may be provided and may be in connection with the substrate 210 or generally within the same housing as the plurality of primary light sources 260. Alternatively, the image sensor 220 may be separate from the plurality of primary light sources 260 and may be provided in a separate housing.
  • a second optic lens 230 may be proximate to the image sensor 220 such that all or a part of the image data sensed or gathered by the image sensor 220 may pass through the second optic lens 230 and may be shaped or adjusted by the second optic lens 230.
  • a controller 250 may be provided and may be in connection with the substrate 210 or generally within the same housing as the plurality of primary light sources 260 and image sensor 220. Alternatively, the controller 250 may be separate from the plurality of primary light sources 260 and/or image sensor 220 and may be provided in a separate housing. The controller 250 may be configured to receive data from the image sensor 220 and/or the plurality of primary light sources 260 and may also provide control or other information to the plurality of primary light sources 260 and/or image sensor 220.
  • a first subset of a plurality of primary light sources may be activated to emit a first sensing spectrum onto a scene.
  • the first subset of the plurality of primary light sources may correspond to a channel that activates one or more light sources that correspond to a primary color (e.g., red).
  • a primary color e.g., red
  • the red light emitting diodes (LEDs) of a plurality of primary light sources 260 of Fig. 2 may be activated.
  • the light sources that correspond to a primary color may be grouped together or, preferably, may be spread out across an array of light sources.
  • primary light sources 260 include a plurality of light sources.
  • the red LEDs may be spread out throughout the light sources 260 such that they can reach various sections of a scene.
  • the first subset of the plurality of light sources may be activated such that their activation is not visible to a human viewing the scene due to, for example, a high frequency, short duration, and/or low amplitude modulation at which the activation occurs.
  • the light from the first subset of primary light sources 260 may emit via the first optic lens 240.
  • the primary light sources 260 may include, for example, primary colors royal blue, cyan, lime, amber, and red. Properties of the primary light sources 260, used in accordance with the subject matter disclosed herein, may be known to the system, and specifically, for example, may be known to the controller 250. As an example, the primary light sources 260 may have chromaticities as shown in Fig. 3 A and wavelength spectra as shown in Fig. 3B. Each of the dots in Fig. 3 A including 310, 311, 312, 313, and 314 may correspond to one of the five primaries of the primary light sources 260, in this example. The dotted line may correspond to a single wavelength.
  • a curved blackbody locus 320 may be followed more closely, for example, in a tunable white system.
  • the color output by the primary light sources 260 of Fig. 2 may have chromaticity corresponding to the area enclosed by the dots in Fig. 3 A including 310, 311, 312, 313, and 314.
  • Fig. 3B shows the spectra 341,
  • Fig. 4A shows a graphical depiction of the spectral power distribution
  • the spectral power distribution 415 corresponds to a color rendering mode that gives the maximum color fidelity within the range of the primaries, specifically the five primaries in this example.
  • Fig. 4B shows a graphical depiction 420 of the gamut index R g and fidelity index R f where the gamut index R g is the TM-30 measure for average relative gamut and the fidelity index R f is the TM-30 measure for average color fidelity.
  • the points 430, 431, 432, 433, and 434 corresponds to the different color rendering modes and the square 435 corresponds to a maximum color fidelity mode.
  • FIG. 4C shows a graphical depiction 440 of R f values as a function of the sixteen hue bins of TM-30.
  • data lines 442, 443, 444, 445, and 446 correspond to the R f values for the corresponding hue bins for the different color rendering modes.
  • Data line 441 corresponds to a maximum color fidelity mode.
  • Fig. 4D shows a graphical depiction 450 of R es values as a function of the sixteen hue bins of TM-30.
  • data lines 451, 452, 453, 454, and 455 correspond to the Res values for the corresponding hue bins for the different color rendering modes.
  • Data line 456 corresponds to a maximum color fidelity mode.
  • first image data may be sensed from the scene while the first subset of the plurality of primary light sources are activated.
  • the first image data may be sensed using an image sensor 220 and the first image data sensed by the image sensor 220 may reach the image sensor 220 via the second optic lens 230.
  • image data may include characteristics about the scene that may enable the controller 250 to approximate the reflectance spectrum for each pixel of the image sensor and/or create a color map of the scene.
  • the image sensor 220 may be a light sensor with spatial resolution such that the image sensor 220 and/or controller 250 may avoid averaging out the different colors present in a scene illuminated by the process described by step 110 of Fig. 1.
  • the controller 250 controls the subsets of a plurality of primary light sources as they are activated to emit sensing spectrums onto a scene, the image sensor does not need to have wavelength-resolving capability in order to obtain information.
  • the controller 250 may utilize the known information about the primary light sources 260, as subsets of the primary light sources 260 emit light onto a scene, in order to obtain color information about the scene.
  • spectral information about the scene may be obtained without using a spectrally selective sensor. It should be noted that because the spectral information, via the image data, is obtained based on the light emitted by the subsets of primary light sources 260, the resolution of the spectral information is limited by the bandwidth of the primary light sources 260. However, it should be noted that such spectral information is sufficient to optimize color rendering by the primary light sources 260 because the primary light sources 260 will have the same limitation in spectral rendering when emitting a lighting spectrum as they have when emitting the sensing spectrum.
  • a second subset of the plurality of primary light sources may be activated to emit a second sensing spectrum onto the scene.
  • the second subset of the plurality of primary light sources may correspond to a channel that activates one or more light sources that correspond to a different color (e.g., royal blue) than the first subset of the plurality of primary light sources.
  • the royal blue light emitting diodes (LEDs) of a plurality of primary light sources 260 of Fig. 2 may be activated.
  • the light sources that correspond to the royal blue color may be grouped together or, preferably, may be spread out across an array of primary light sources 260.
  • primary light sources 260 include a plurality of light sources.
  • the royal blue LEDs may be spread out throughout the primary light sources 260 such that they can reach various sections of a scene.
  • the second subset of the plurality of light sources may be activated such that their activation is not visible to a human viewing the scene due to, for example, a high frequency, short duration, and/or low amplitude modulation at which the activation occurs.
  • the light from the second subset of primary light sources 260 may emit via the first optic lens 240.
  • second image data may be sensed from the scene while the second subset of the plurality of primary light sources 260 of Fig. 2 are activated.
  • the second image data may be sensed using an image sensor 220 and the second image data sensed by the image sensor 220 may reach the image sensor 220 via the second optic lens 230.
  • image data may include characteristics about the scene that may enable the controller 250 to approximate the reflectance spectrum for each pixel of the image sensor and/or create a color map of the scene.
  • the controller 250 which may be factory programmed or user programmable to provide the desired response, as further disclosed herein, may modulate the primary light sources 260 such that the first subset is activated and the image sensor 220 collects first image data and then the second subset is activated and then the image sensor 220 collects second image data.
  • image data may be sensed for additional available primary light sources.
  • four or more and more preferably, five or more primary light sources may be available.
  • third, fourth and fifth image data corresponding to third, fourth and fifth spectrums, respectively may be sensed and provided to a controller such as controller 250 of Fig. 2.
  • reference information for the scene may be determined based on combined image data where combined image data is a combination of the available image data such as the first image data and the second image data.
  • a controller such as controller 250 of Fig. 2, may determine the reference information based on combined image data such as the combination of the first image data and the second image data. Additionally, according to an implementation, the controller may also have sensing spectrum information regarding the primary light sources 260.
  • Figs. 4A-D as described herein, provide example graphical depictions of spectra and corresponding TM-30 indices that the controller may have or have access to and that can be realized with the primary light sources of Figs. 3 A and Fig. 3B.
  • the reference information may correspond to an estimate of an approximate reflectance spectrum for each pixel of the image sensor and, thus, may correspond to a color map of the scene.
  • the color map may be expressed as the relative response of each pixel to each of the subsets of the primary light sources 260 of Fig.
  • Table 1 includes the relative reflected intensities sensed by a single pixel of an image sensor, for four different example reflectance spectra.
  • the relative intensities are sensed for the five primary light sources, as shown in Table 1, such that a relative reflected intensity for a given primary source (i.e., channel) is sensed when that primary source emits light onto the scene.
  • the four example reflectance spectra correspond to four Color Evaluation Samples (CES) as defined in TM-30 and correspond to CES 5
  • CES Color Evaluation Samples
  • Table 1 shows that the relative reflected intensity sensed by the image sensor 220 sensing a maroon portion of a scene while a royal blue primary light source is emitting royal blue light onto that part of the scene is .098. Similarly, as shown in Table 1, the relative reflected intensity sensed by the image sensor 220 sensing a maroon portion of a scene while a red primary light source is emitting red light onto that part of the scene is .5468. Because the color red is closer to the
  • the controller may develop a color map of the scene based on the data gathered via the pixel(s) of the image sensor.
  • a color map may be expressed in a standardized color space such as CIE1931, CIE1976, CAM02-UCS, or the like. Expressing the color map in such a standardized color space may allow more advanced spectral optimization algorithms and/or more intuitive programming of the desired response.
  • the reflectance spectrum of each pixel of an image sensor may be estimated and, subsequently the associated color coordinates for the pixel may be calculated based on the estimated reflectance spectrum.
  • Fig. 5 shows an example implementation including reflectance spectra of CES 5, CES 64, CES 32 and CES 81 colors from TM-30 represented by solid lines 511, 512, 513 and 514 respectively. Dashed lines 521, 522, 523, and 524 show the respective estimated reflectance spectra using the five primary light sources of Figs. 3 A and 3B.
  • the dashed lines 521, 522, 523, and 524 are estimated based on the image data collected by an image sensor.
  • the royal blue primary channel emits a peak wavelength at 450nm shown by 541.
  • Fig. 5 shows that the relative reflected intensity sensed by an image sensor, such as image sensor 220 of Fig 2, senses four different CES color points 531, 532, 533, and 534 while the royal blue primary channel is activated, and emits a peak wavelength at 450nm shown by 541.
  • the four different CES color points in this example correspond to the maroon (CES 5) 531, the teal (CES 64) 532, the mustard (CES 32) 533, and the purple (CES 81) 534.
  • the five wavelengths corresponding to the five primary light sources are shown by 541 for the royal blue, 542 for the cyan, 543 for the lime, 544 for the amber, and 545 for the red.
  • the image sensor may register a relative reflectance intensity of roughly .098 corresponding to the maroon CES 5, as shown by point 531 in Fig.
  • the image sensor 220 may capture four CES color points at the peak wavelength for each of the five primary light sources of Figs. 3 A and 3B, for a total of 20 SPD data points, in this example.
  • SPD data points are obtained at the centroid wavelength of each primary.
  • An approximate reflectance spectrum can subsequently be estimated based on these data points via polynomial fits such as those shown by the dashed lines 521, 522, 523, and 524.
  • Each dashed line 521, 522, 523, and 524 represents a best polynomial fit for a respective CES color based on data points collected at peak wavelengths of the five primary sources, with conditions defined at 380nm and 780nm. It should be noted that other interpolation methods may also be used such as a linear interpolation, spline interpolation, or moving average interpolation.
  • Fig. 6 shows an analysis of the polynomial fit with the five primaries of Fig. 3 A and 3B where data points for all 99 CES colors from TM-30 were calculated by sequentially activating each of the five primaries.
  • Fig. 6 shows a graph 600 where of the 99 CES from TM-30, 58 CES colors are identified by their correct TM-30 hue bin (1-16), and 96 CES colors are identified correctly within plus or minus one hue bin.
  • the circles represent the original CES color point and the corresponding diamond represent the estimated color point as determined by using the five primary light sources.
  • the spectrum optimization criteria for primary light sources may be determined based on the reference information and one or more desired output parameters.
  • the spectrum optimization criteria may be the criteria that the primary light sources are operated based on when emitting a lighting spectrum onto a scene. Accordingly, the spectrum optimization criteria are the criteria that achieve the desired output based on the reference information of the scene.
  • the reference information may be determined based on combined image data as disclosed herein in reference to 150 of Fig. 1.
  • the desired output parameters may be generated via any applicable manner such as based on user input, based the location of a device or component, based on the image data, based on predetermined criteria, or the like.
  • a user may provide input via a wireless signal such as via Bluetooth, WiFi, RFID, infrared, or the like.
  • a user may provide input via a keyboard, mouse, touchpad, haptic response, voice command, or the like.
  • a controller such as controller 250 of Fig. 2, may utilize the reference information and the desired output parameter(s) to generate the spectrum optimization criteria.
  • spectrum optimization criteria may be pre calculated offline based on potential image data and output parameters and may be stored via an applicable technique, such as a lookup table, on a controller or a memory accessible by the controller. Such pre-calculation and storing may reduce the need for complex calculations by an on-board controller that may be limited in its computational ability.
  • Fig. 8 shows an example flowchart diagram of such an implementation.
  • factory input data 810 may be provided to an on-board processing system 820.
  • the factory input data 810 may be provided to a scene color mapping module 821 to generate spectral data points based on for example, image data collected while primary light sources emit a sensing spectrum as well as intensity values and factory input data 810.
  • the factory input data 810 may also be provided to a source spectrum optimization module 822 which may calculate desired indices/spectrum optimization criteria based on: the output from the scene color mapping module 821 and specified response behavior (output parameters) from the user programming module 815.
  • the spectrum optimization module 822 may also set the channel drive currents based on the determined spectrum optimization criteria.
  • the plurality of primary light sources may be activated to emit a lighting spectrum based on the spectrum optimization criteria.
  • the spectrum optimization criteria may be provided to the plurality of primary light sources by the controller either directly or via applicable communication channels, such as a wired or wireless communication channel as further disclosed herein.
  • the spectrum optimization criteria may result in different color rendering modes to be emitted.
  • the desired output parameters may be to maximize the saturation or fidelity of the most contrasting dominant colors in a scene.
  • a controller may utilize the image data to determine the most contrasting dominant colors in a given scene and generate spectrum optimization criteria for the light sources to emit such that the saturation or fidelity of those colors is maximized.
  • the five primary light sources of Figs. 3A-B may be used for color saturation.
  • Fig. 7 A shows chart 710 that includes estimated and actual color points of several CES colors in each of the TM-30 hue bins 1, 5, 9, and 13. As shown in Fig.
  • color saturation may primarily be achieved in either of two directions: along the red to cyan axis (e.g., TM-30 hue bins 1 and 9) shown along the horizontal axis in Fig. 7A or the green-yellow to purple axis (e.g., bins 5 and 13) shown along the vertical axis in Fig. 7A. Accordingly, as a specific example, as shown in Fig.
  • a controller may select spectrum optimization criteria based on one of three color rendering modes where the bins correspond to TM-30 hue bins and 730 corresponds to a perfect TM-30 circle: (1) bin 1 and bin 9 over saturation if mainly red and/or cyan are detected as represented by trace 721, (2) bin 5 and bin 13 oversaturation if mainly green-yellow and/or purple are detected as represented by trace 722, and (3) a high fidelity spectrum if there is no dominant color detected in one of these hue bins as represented by trace 723.
  • a controller may, based on output parameters, select optimization criteria for a spectrum that maximizes oversaturation of the dominant color detected, or a spectrum that maximizes over saturation of all detected colors weighted by their occurrence.
  • slightly oversaturated colors may be subjectively preferred, as may be indicated by the output parameters.
  • over saturation may be quantified by chroma shift such as, for example, the Res indices in TM-30.
  • a typical preferred range for Res may be 0-20%, and a more preferred range may be 5-10%.
  • image data may be compared to a previously recorded image data to determine the colors of a moving or new object of interest such that the spectrum can be optimized for this object.
  • the average reflectance spectrum of the image may be used to optimize the spectrum for the average color.
  • the chromaticity may be kept constant or allowed to change.
  • the output parameters may correspond to targeting a certain chromaticity of the emitted light, based on a given scene as determined based on the image data. For example, a cool white may be desirable when the scene contains cool hues such as blue, cyan and green, whereas a warm white may illuminate yellow, orange and red hues. Such a scheme may enhance color gamut and visual brightness of the scene.
  • a controller may provide spectrum optimization criteria corresponding to three or more primaries.
  • the output parameters may correspond to achieving a desired color point.
  • a controller may utilize the reflected light information within the image to determine the spectrum optimization criteria for the emitted spectrum that is needed to achieve a desired overall color point. For example, in a space where a colored object or wall is illuminated near a white background, the light reflected off the colored object or wall may cause the white background to appear non-white, may be undesirable, as indicated by the output parameters. Accordingly, a controller may generate spectrum optimization criteria such that the primary light sources emit a lighting spectrum that maintains the white background as white.
  • Table 2 shows a summary of example output parameters, example image data, and corresponding example spectrum optimization criteria.
  • the lighting system disclosed herein may include a communication interface that may enable communication to an external component or system.
  • the communication may be facilitated by wired or wireless transmissions and may incorporate any applicable modes including Bluetooth, WiFi, cellular, infrared, or the like.
  • the controller may be external to the lighting system such that image data is provided to the external controller and spectrum optimization criteria is determined and/or provided by the external controller.
  • output criteria may be provided via an external input device (e.g., a mobile phone) and/or may be provided to an external component such as an external controller.
  • the sensing spectrum may be emitted by a first subset of primary light sources while a lighting spectrum is emitted by the remaining or other subset of primary light sources.
  • the first subset may emit the sensing spectrum such that the sensing spectrum is not visible to humans (e.g., at a high frequency).
  • Image data may be collected, as disclosed herein, based on the first subset emitting the sensing spectrum and may subsequently be collected when a second subset emits a sensing spectrum subsequent the first subset switching to emitting a lighting spectrum.
  • ROM read only memory
  • RAM random access memory
  • register cache memory
  • semiconductor memory devices magnetic media such as internal hard disks and removable disks, magneto-optical media, and optical media such as CD-ROM disks, and digital versatile disks (DVDs).

Landscapes

  • Circuit Arrangement For Electric Light Sources In General (AREA)
  • Image Processing (AREA)

Abstract

L'invention concerne des techniques et des dispositifs permettant de détecter des données d'image d'une scène et d'activer des sources de lumière primaires sur la base d'informations détectées à partir de la scène. Des sous-ensembles d'une pluralité de sources de lumière primaires peuvent être activés pour émettre un spectre de détection de lumière sur une scène. Des données d'image combinées peuvent être détectées à partir de la scène tandis que les sous-ensembles de la pluralité de sources de lumière primaires sont activés. Des informations de réflectance de la scène peuvent être déterminées sur la base des données d'image combinées et des spectres de détection combinés. Des critères d'optimisation de spectres des sources de lumière primaires peuvent être déterminés sur la base des informations de référence et d'un paramètre de sortie souhaité fourni par un utilisateur ou déterminé par un dispositif de commande. Les sources de la pluralité de sources de lumière primaires peuvent être activées pour émettre un spectre d'éclairage sur la base des critères d'optimisation de spectres.
PCT/US2019/038297 2018-06-22 2019-06-20 Sytème d'éclairage doté d'un capteur intégré WO2019246431A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2020571355A JP7036952B2 (ja) 2018-06-22 2019-06-20 一体型センサ付き照明システム
EP19733920.3A EP3811736A1 (fr) 2018-06-22 2019-06-20 Sytème d'éclairage doté d'un capteur intégré

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US16/015,697 US10582589B2 (en) 2018-06-22 2018-06-22 Lighting system with integrated sensor
US16/015,697 2018-06-22
EP18186839 2018-08-01
EP18186839.9 2018-08-01

Publications (1)

Publication Number Publication Date
WO2019246431A1 true WO2019246431A1 (fr) 2019-12-26

Family

ID=67070918

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2019/038297 WO2019246431A1 (fr) 2018-06-22 2019-06-20 Sytème d'éclairage doté d'un capteur intégré

Country Status (4)

Country Link
EP (1) EP3811736A1 (fr)
JP (1) JP7036952B2 (fr)
TW (1) TWI728385B (fr)
WO (1) WO2019246431A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7710369B2 (en) * 2004-12-20 2010-05-04 Philips Solid-State Lighting Solutions, Inc. Color management methods and apparatus for lighting devices
EP2473008A2 (fr) * 2011-01-04 2012-07-04 Zumtobel Lighting GmbH Procédé et dispositif d'éclairage pour l'éclairage d'un objet avec plusieurs sources de lumière
US9332612B1 (en) * 2014-02-25 2016-05-03 Lumenetix, Inc. System and method for rapidly generating color models for LED-based lamps

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101313171B (zh) * 2005-11-22 2010-05-19 皇家飞利浦电子股份有限公司 具有多组光源的照明系统
JP2009048989A (ja) * 2007-07-20 2009-03-05 Toshiba Lighting & Technology Corp 照明装置
TWI813693B (zh) * 2012-03-05 2023-09-01 美商恩光碼公司 多頻帶色覺濾光器及使用線性規劃最佳化之方法
BR112015001396A2 (pt) * 2012-07-27 2017-07-04 Koninklijke Philips Nv disposição de iluminação configurada para iluminar um objeto, método para controlar uma disposição de iluminação, e, produto de programa de computador
JP6388637B2 (ja) * 2013-04-09 2018-09-12 フィリップス ライティング ホールディング ビー ヴィ ターゲット物体の視覚的外観を変えるためのアレンジメント
US9704441B2 (en) * 2015-02-04 2017-07-11 Snaptrack, Inc. System and method to adjust displayed primary colors based on illumination
DE202016008521U1 (de) * 2015-06-08 2018-05-17 Opple Lighting Co. Ltd. Beleuchtungsvorrichtung und Steuersystem dafür
TWI561797B (en) * 2015-12-02 2016-12-11 Chromatic detector of led light source

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7710369B2 (en) * 2004-12-20 2010-05-04 Philips Solid-State Lighting Solutions, Inc. Color management methods and apparatus for lighting devices
EP2473008A2 (fr) * 2011-01-04 2012-07-04 Zumtobel Lighting GmbH Procédé et dispositif d'éclairage pour l'éclairage d'un objet avec plusieurs sources de lumière
US9332612B1 (en) * 2014-02-25 2016-05-03 Lumenetix, Inc. System and method for rapidly generating color models for LED-based lamps

Also Published As

Publication number Publication date
TW202001465A (zh) 2020-01-01
EP3811736A1 (fr) 2021-04-28
JP7036952B2 (ja) 2022-03-15
JP2021527931A (ja) 2021-10-14
TWI728385B (zh) 2021-05-21

Similar Documents

Publication Publication Date Title
US11382195B2 (en) Lighting system with integrated sensor
KR101011662B1 (ko) Led 기반 발광을 위한 컬러 제어
JP5689476B2 (ja) 対象物を強調するための照明システム
US8981672B2 (en) Color control of dynamic lighting
US8115779B2 (en) Color control of white LED lamps
US8390643B2 (en) Dynamic gamut control
JP2012506614A (ja) 複数の光源タイプを用いた色生成
JP5345271B2 (ja) 画像表示装置
KR20130141920A (ko) 색역 변환 시스템 및 방법
US9294742B2 (en) Spectral improvement of digital camera color images
KR20110099306A (ko) 조명 기구의 성능을 최대화하는 방법
WO2020226970A1 (fr) Sélection de paramètres dans une application d'ajustement de couleur
US11076461B2 (en) User control modality for LED color tuning
WO2019246431A1 (fr) Sytème d'éclairage doté d'un capteur intégré
US9980338B2 (en) Illumination system
WO2020236525A1 (fr) Modalité de commande d'utilisateur pour le réglage de couleur de del
CN113689510B (zh) 一种光源性能表征方法、装置及计算机可读存储介质
JP5096674B2 (ja) 表示装置
JP2022091243A (ja) 色温度可変型照明装置
TW201122707A (en) Method and apparatus of projector color enhancement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19733920

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020571355

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2019733920

Country of ref document: EP

Effective date: 20210122