EP4222731A1 - System and method for a multi-primary wide gamut color system - Google Patents

System and method for a multi-primary wide gamut color system

Info

Publication number
EP4222731A1
EP4222731A1 EP21876192.2A EP21876192A EP4222731A1 EP 4222731 A1 EP4222731 A1 EP 4222731A1 EP 21876192 A EP21876192 A EP 21876192A EP 4222731 A1 EP4222731 A1 EP 4222731A1
Authority
EP
European Patent Office
Prior art keywords
primary
color
approximately
image data
white
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
EP21876192.2A
Other languages
German (de)
French (fr)
Inventor
Gary B. Mandle
James M. DeFilippis
Mitchell J. Bogdanowicz
Corey P. Carbonara
Michael F. Korpi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Baylor University
Original Assignee
Baylor University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US17/060,917 external-priority patent/US11030934B2/en
Priority claimed from US17/076,383 external-priority patent/US11069279B2/en
Priority claimed from US17/082,741 external-priority patent/US11069280B2/en
Priority claimed from US17/209,959 external-priority patent/US11373575B2/en
Application filed by Baylor University filed Critical Baylor University
Publication of EP4222731A1 publication Critical patent/EP4222731A1/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/02Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B5/00Optical elements other than lenses
    • G02B5/20Filters
    • G02B5/201Filters in the form of arrays
    • GPHYSICS
    • G02OPTICS
    • G02FOPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
    • G02F1/00Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
    • G02F1/01Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour 
    • G02F1/13Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour  based on liquid crystals, e.g. single liquid crystal display cells
    • G02F1/133Constructional arrangements; Operation of liquid crystal cells; Circuit arrangements
    • G02F1/1333Constructional arrangements; Manufacturing methods
    • G02F1/1335Structural association of cells with optical devices, e.g. polarisers or reflectors
    • G02F1/1336Illuminating devices
    • G02F1/133614Illuminating devices using photoluminescence, e.g. phosphors illuminated by UV or blue light
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2300/00Aspects of the constitution of display devices
    • G09G2300/04Structural and physical details of display devices
    • G09G2300/0439Pixel structures
    • G09G2300/0452Details of colour pixel setup, e.g. pixel composed of a red, a blue and two green components
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/0666Adjustment of display parameters for control of colour parameters, e.g. colour temperature
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2330/00Aspects of power supply; Aspects of display protection and defect management
    • G09G2330/06Handling electromagnetic interferences [EMI], covering emitted as well as received electromagnetic radiation
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/06Colour space transformation
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2370/00Aspects of data communication
    • G09G2370/14Use of low voltage differential signaling [LVDS] for display data communication
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G3/00Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
    • G09G3/001Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes using specific devices not provided for in groups G09G3/02 - G09G3/36, e.g. using an intermediate record carrier such as a film slide; Projection systems; Display of non-alphanumerical information, solely or in combination with alphanumerical information, e.g. digital display on projected diapositive as background

Definitions

  • U.S. Application No. 17/060,917 is a continuation-in-part of U.S. Application No. 17/009,408, filed September 1, 2020, which is a continuation-in-part of U.S. Application No. 16/887,807, filed May 29, 2020, which is a continuation-in-part of U.S. Application No.
  • U.S. Application No. 17/082,741 is a continuation-in-part of U.S. Application No. 17/009,408, filed September 1, 2020, which is a continuation-in-part of U.S. Application No. 16/887,807, filed May 29, 2020, which is a continuation-in-part of U.S. Application No.
  • U.S. Application No. 17/209,959 is a continuation-in-part of U.S. Application No. 17/082,741, filed October 28, 2020, which is a continuation-in-part of U.S. Application No. 17/009,408, filed September 1, 2020, which is a continuation-in-part of U.S. Application No. 16/887,807, filed May 29, 2020, which is a continuation-in-part of U.S. Application No.
  • the present invention relates to color systems, and more specifically to a wide gamut color system with an increased number of primary colors.
  • U.S. Patent Publication No. 20200144327 for Light emitting diode module and display device by inventors Lee, et al., filed June 27, 2019 and published May 7, 2020, is directed to a light emitting diode module that includes a cell array including first to fourth light emitting diode cells, each cell having a first conductive semiconductor layer, an active layer, and a second conductive semiconductor layer, the cell array having a first surface and a second surface opposite to the first surface; first to fourth light adjusting portions on the second surface of the cell array to respectively correspond to the first to fourth light emitting diode cells, to provide red light, first green light, second green light, and blue light, respectively; light blocking walls between the first to fourth light adjusting portions to isolate the first to fourth light adjusting portions from one another; and an electrode portion on the first surface of the cell array, and electrically connected to the first to fourth light emitting diode cells to selectively drive the first to fourth light emitting diode cells.
  • U.S. Patent No. 10,847,498 for Display device and electronic device by inventors Nakamura, et al., filed April 10, 2020 and issued November 24, 2020, is directed to a display panel that includes a plurality of light-emitting elements.
  • Light emitted from a first lightemitting element has a CIE 1931 chromaticity coordinate x of greater than 0.680 and less than or equal to 0.720 and a CIE 1931 chromaticity coordinate y of greater than or equal to 0.260 and less than or equal to 0.320.
  • Light emitted from a second light-emitting element has a CIE 1931 chromaticity coordinate x of greater than or equal to 0.130 and less than or equal to 0.250 and a CIE 1931 chromaticity coordinate y of greater than 0.710 and less than or equal to 0.810.
  • Light emitted from a third light-emitting element has a CIE 1931 chromaticity coordinate x of greater than or equal to 0.120 and less than or equal to 0.170 and a CIE 1931 chromaticity coordinate y of greater than or equal to 0.020 and less than 0.060.
  • U.S. Patent No. 10,504,437 for Display panel, control method thereof, display device and display system for anti -peeping display by inventors Zhang, et al., filed May 26, 2016 and issued December 10, 2019, is directed to a display panel, a control method thereof, a display device and a display system comprising such a display panel.
  • the display panel includes a plurality of pixel units. Each pixel unit has a plurality of subpixels Each subpixel has a display subpixel and an interference subpixel. Additionally, the interference subpixel and the display subpixel are different in at least one of color and gray scale.
  • the display panel also includes a first control unit configured to control the display subpixel to be switched on during a first period of time in each display period, and to control the interference subpixel to be switched off during the first period of time in each display period and switched on during a second period of time in each display period.
  • 20200128220 for Image processing method and apparatus, electronic device, and computer storage medium by inventors Bao, et al., filed December 19, 2019 and published April 23, 2020 is directed to an image processing method including: obtaining a facial skin tone area in an image to be processed; filtering the image to be processed to obtain a filtered smooth image; obtaining a high-frequency image based on the smooth image and the image to be processed; obtaining a facial skin tone high-frequency image based on the high-frequency image and a facial skin tone mask; and superimposing the high-frequency image and the image to be processed based on the facial skin tone mask and preset first superimposition strength in a luma channel, and superimposing a luma channel signal of the facial skin tone high-frequency image onto a luma channel signal of the image to be processed, to obtain a first image.
  • U.S. Patent Publication No. 20200209678 for Reflective pixel unit, reflective display panel and display apparatus by inventors Hsu, et al., filed April 17, 2019 and published July 2, 2020, is directed to a reflective pixel unit, a reflective display panel and a display apparatus.
  • the reflective pixel unit includes a substrate, a reflective plate on the substrate, and a reflective filter layer on a side of the reflective plate facing away from the substrate.
  • the reflective filter layer is configured such that a surface of the reflective filter layer facing away from the reflective plate receives visible light and reflects a part of light having wavelengths within a specific range in the visible light, and allows another part of the light having wavelengths within the specific range to pass through the reflective filter layer to the reflective plate.
  • the reflective plate is configured to reflect the another part of the light having wavelengths within the specific range passed through the reflective filter layer.
  • U.S. Patent No. 10,222,263 for RGB value calculation device by inventor Yasuyuki Shigezane, filed February 6, 2017 and issued March 5, 2019, is directed to a microcomputer that equally divides the circumference of an RGB circle into 6xn (n is an integer of 1 or more) parts, and calculates an RGB value of each divided color. (255, 0, 0) is stored as a reference RGB value of a reference color in a ROM in the microcomputer.
  • the microcomputer converts the reference RGB value depending on an angular difference of the RGB circle between a designated color whose RGB value is to be found and the reference color, and assumes the converted RGB value as an RGB value of the designated color.
  • U.S. Patent No. 9,373,305 for Semiconductor device, image processing system and program by inventor Hiorfumi Kawaguchi, filed May 29, 2015 and issued June 21, 2016, is directed to an image process device including a display panel operable to provide an input interface for receiving an input of an adjustment value of at least a part of color attributes of each vertex of n axes (n is an integer equal to or greater than 3) serving as adjustment axes in an RGB color space, and an adjustment data generation unit operable to calculate the degree of influence indicative of a following index of each of the n-axis vertices, for each of the n axes, on a basis of distance between each of the n-axis vertices and a target point which is an arbitrary lattice point in the RGB color space, and operable to calculate adjusted coordinates of the target point in the RGB color space.
  • U.S. Publication No. 20130278993 for Color-mixing bi-primary color systems for displays by inventor Heikenfeld, et.al, filed September 1, 2011 and published October 24, 2013, is directed to a display pixel.
  • the pixel includes first and second substrates arranged to define a channel.
  • a fluid is located within the channel and includes a first colorant and a second colorant.
  • the first colorant has a first charge and a color.
  • the second colorant has a second charge that is opposite in polarity to the first charge and a color that is complimentary to the color of the first colorant.
  • a first electrode with a voltage source, is operably coupled to the fluid and configured to moving one or both of the first and second colorants within the fluid and alter at least one spectral property of the pixel.
  • U.S. Patent No. 8,599,226 for Device and method of data conversion for wide gamut displays by inventor Ben-Chorin, et. al, filed February 13, 2012 and issued December 3, 2013, is directed to a method and system for converting color image data from a, for example, three-dimensional color space format to a format usable by an n-primary display, wherein n is greater than or equal to 3.
  • the system may define a two-dimensional sub-space having a plurality of two-dimensional positions, each position representing a set of n primary color values and a third, scaleable coordinate value for generating an n-primary display input signal. Furthermore, the system may receive a three-dimensional color space input signal including out-of range pixel data not reproducible by a three-primary additive display, and may convert the data to side gamut color image pixel data suitable for driving the wide gamut color display.
  • U.S. Patent No. 8,081,835 for Multiprimary color sub-pixel rendering with metameric filtering by inventor Elliot, et. al, filed July 13, 2010 and issued December 20, 2011, is directed to systems and methods of rendering image data to multi primary displays that adjusts image data across metamers as herein disclosed.
  • the metamer filtering may be based upon input image content and may optimize sub-pixel values to improve image rendering accuracy or perception. The optimizations may be made according to many possible desired effects.
  • One embodiment comprises a display system comprising: a display, said display capable of selecting from a set of image data values, said set comprising at least one metamer; an input image data unit; a spatial frequency detection unit, said spatial frequency detection unit extracting a spatial frequency characteristic from said input image data; and a selection unit, said unit selecting image data from said metamer according to said spatial frequency characteristic.
  • U.S. Patent No. 7,916,939 for High brightness wide gamut display by inventor Roth, et. al, filed November 30, 2009 and issued March 29, 2011, is directed to a device to produce a color image, the device including a color filtering arrangement to produce at least four colors, each color produced by a filter on a color filtering mechanism having a relative segment size, wherein the relative segment sizes of at least two of the primary colors differ.
  • U.S. Patent No. 6,769,772 for Six color display apparatus having increased color gamut by inventor Roddy, et. al, filed October 11, 2002 and issued August 3, 2004, is directed to a display system for digital color images using six color light sources or two or more multicolor LED arrays or OLEDs to provide an expanded color gamut. Apparatus uses two or more spatial light modulators, which may be cycled between two or more color light sources or LED arrays to provide a six-color display output. Pairing of modulated colors using relative luminance helps to minimize flicker effects.
  • the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter, wherein the single display device and the image data converter are in network communication, wherein the image data converter further includes a first link component and a second link component
  • the single display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the single display device is based on the set of image data.
  • the present invention provides system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, wherein the set of SDP parameters is modifiable, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color
  • SDP Session Description Protocol
  • the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with a single display device, and wherein the image data converter further includes a first link component and a second link component, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the single display device using the image data converter, transporting the first set of color channel data to the single display device using the first link component, and transporting the second set of color channel data to the single display device using the first link component in parallel with the first link component, wherein the at least one white emitter includes at least three
  • FIG. 1 illustrates one embodiment of a six primary system including a red primary, a green primary, a blue primary, a cyan primary, a magenta primary, and a yellow primary (“6P-B”) compared to ITU-R BT.709-6.
  • FIG. 2 illustrates another embodiment of a six primary system including a red primary, a green primary, a blue primary, a cyan primary, a magenta primary, and a yellow primary (“6P-C”) compared to SMPTE RP431 -2 for a D60 white point.
  • 6P-C yellow primary
  • FIG. 3 illustrates yet another embodiment of a six primary system including a red primary, a green primary, a blue primary, a cyan primary, a magenta primary, and a yellow primary (“6P-C”) compared to SMPTE RP431 -2 for a D65 white point.
  • 6P-C yellow primary
  • FIG. 4 illustrates Super 6Pa compared to 6P-C.
  • FIG. 5 illustrates Super 6Pb compared to Super 6Pa and 6P-C.
  • FIG. 6 illustrates an embodiment of an encode and decode system for a multiprimary color system.
  • FIG. 7 illustrates a sequential method where three color primaries are passed to the transport format as full bit level image data and inserted as normal (“System 2”).
  • FIG. 8 illustrates one embodiment of a system encode and decode process using a dual link method (“System 3”).
  • FIG. 9 illustrates one embodiment of an encoding process using a dual link method.
  • FIG. 10 illustrates one embodiment of a decoding process using a dual link method.
  • FIG. 11 illustrates one embodiment of a six-primary color system encode using a 4:4:4 sampling method.
  • FIG. 12 illustrates one embodiment for a method to package six channels of primary information into the three standard primary channels used in current serial video standards by modifying bit numbers for a 12-bit SDI and a 10-bit SDI.
  • FIG. 13 illustrates a simplified diagram estimating perceived viewer sensation as code values define each hue angle.
  • FIG. 14 illustrates one embodiment for a method of stacking/encoding six-primary color information using a 4:4:4 video system.
  • FIG. 15 illustrates one embodiment for a method of unstacking/decoding six- primary color information using a 4:4:4 video system.
  • FIG. 16 illustrates one embodiment of a 4:4:4 decoder for a six-primary color system.
  • FIG. 17 illustrates one embodiment of an optical filter.
  • FIG. 18 illustrates another embodiment of an optical filter.
  • FIG. 19 illustrates an embodiment of the present invention for sending six primary colors to a standardized transport format.
  • FIG. 20 illustrates one embodiment of a decode process adding a pixel delay to the RGB data for realigning the channels to a common pixel timing.
  • FIG. 21 illustrates one embodiment of an encode process for 4:2:2 video for packaging five channels of information into the standard three-channel designs.
  • FIG. 22 illustrates one embodiment for a non-constant luminance encode for a six- primary color system.
  • FIG. 23 illustrates one embodiment of a packaging process for a six-primary color system.
  • FIG. 24 illustrates a 4:2:2 unstack process for a six-primary color system.
  • FIG. 25 illustrates one embodiment of a process to inversely quantize each individual color and pass the data through an electronic optical function transfer (EOTF) in a non-constant luminance system.
  • EOTF electronic optical function transfer
  • FIG. 26 illustrates one embodiment of a constant luminance encode for a six- primary color system.
  • FIG. 27 illustrates one embodiment of a constant luminance decode for a six- primary color system.
  • FIG. 28 illustrates one example of 4:2:2 non-constant luminance encoding.
  • FIG. 29 illustrates one embodiment of a non-constant luminance decoding system.
  • FIG. 30 illustrates one embodiment of a 4:2:2 constant luminance encoding system.
  • FIG. 31 illustrates one embodiment of a 4:2:2 constant luminance decoding system.
  • FIG. 32 illustrates a raster encoding diagram of sample placements for a six- primary color system.
  • FIG. 33 illustrates one embodiment of the six-primary color unstack process in a 4:2:2 video system.
  • FIG. 34 illustrates one embodiment of mapping input to the six-primary color system unstack process.
  • FIG. 35 illustrates one embodiment of mapping the output of a six-primary color system decoder.
  • FIG. 36 illustrates one embodiment of mapping the RGB decode for a six-primary color system.
  • FIG. 37 illustrates one embodiment of an unstack system for a six-primary color system.
  • FIG. 38 illustrates one embodiment of a legacy RGB decoder for a six-primary, non-constant luminance system.
  • FIG. 39 illustrates one embodiment of a legacy RGB decoder for a six-primary, constant luminance system.
  • FIG. 40 illustrates one embodiment of a six-primary color system with output to a legacy RGB system.
  • FIG. 41 illustrates one embodiment of six-primary color output using a nonconstant luminance decoder.
  • FIG. 42 illustrates one embodiment of a legacy RGB process within a six-primary color system.
  • FIG. 43 illustrates one embodiment of packing six-primary color system image data into an ICjCp (ITP) format.
  • FIG. 44 illustrates one embodiment of a six-primary color system converting RGBCMY image data into XYZ image data for an ITP format.
  • FIG. 45 illustrates one embodiment of six-primary color mapping with SMPTE ST424.
  • FIG. 46 illustrates one embodiment of a six-primary color system readout for a SMPTE ST424 standard.
  • FIG. 47 illustrates a process of 2160p transport over 12G-SDI.
  • FIG. 48 illustrates one embodiment for mapping RGBCMY data to the SMPTE ST2082 standard for a six-primary color system.
  • FIG. 49 illustrates one embodiment for mapping YRGB YCMY CR CB CC CY data to the SMPTE ST2082 standard for a six-primary color system.
  • FIG. 50 illustrates one embodiment for mapping six-primary color system data using the SMPTE ST292 standard.
  • FIG. 51 illustrates one embodiment of the readout for a six-primary color system using the SMPTE ST292 standard.
  • FIG. 52 illustrates modifications to the SMPTE ST352 standards for a six-primary color system.
  • FIG. 53 illustrates modifications to the SMPTE ST2022 standard for a six-primary color system.
  • FIG. 54 illustrates a table of 4:4:4 sampling for a six-primary color system for a 10-bit video system.
  • FIG. 55 illustrates a table of 4:4:4 sampling for a six-primary color system for a 12-bit video system.
  • FIG. 56 illustrates sequence substitutions for 10-bit and 12-bit video in 4:2:2 sampling systems in a Y Cb Cr Cc Cy color space.
  • FIG. 57 illustrates sample placements of six-primary system components for a 4:2:2 sampling system image.
  • FIG. 58 illustrates sequence substitutions for 10-bit and 12-bit video in 4:2:0 sampling systems using a Y Cb Cr Cc Cy color space.
  • FIG. 59 illustrates sample placements of six-primary system components for a 4:2:0 sampling system image.
  • FIG. 60 illustrates modifications to SMPTE ST2110-20 for a 10-bit six-primary color system in 4:4:4 video.
  • FIG. 61 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:4:4 video.
  • FIG. 62 illustrates modifications to SMPTE ST2110-20 for a 10-bit six primary color system in 4:2:2 video.
  • FIG. 63 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:2:0 video.
  • FIG. 64 illustrates an RGB sampling transmission for a 4:4:4 sampling system.
  • FIG. 65 illustrates a RGBCMY sampling transmission for a 4:4:4 sampling system.
  • FIG. 66 illustrates an example of System 2 to RGBCMY 4:4:4 transmission.
  • FIG. 67 illustrates a Y Cb Cr sampling transmission using a 4:2:2 sampling system.
  • FIG. 68 illustrates a Y Cr Cb Cc Cy sampling transmission using a 4:2:2 sampling system.
  • FIG. 69 illustrates an example of a System 2 to Y Cr Cb Cc Cy 4:2:2 Transmission as non-constant luminance.
  • FIG. 70 illustrates a Y Cb Cr sampling transmission using a 4:2:0 sampling system.
  • FIG. 71 illustrates a Y Cr Cb Cc Cy sampling transmission using a 4:2:0 sampling system.
  • FIG. 72 illustrates a dual stack LCD projection system for a six-primary color system.
  • FIG. 73 illustrates one embodiment of a single projector.
  • FIG. 74 illustrates a six-primary color system using a single projector and reciprocal mirrors.
  • FIG. 75 illustrates a dual stack DMD projection system for a six-primary color system.
  • FIG. 76 illustrates one embodiment of a single DMD projector solution.
  • FIG. 77 illustrates one embodiment of a color filter array for a six-primary color system with a white OLED monitor.
  • FIG. 78 illustrates one embodiment of an optical filter array for a six-primary color system with a white OLED monitor.
  • FIG. 79 illustrates one embodiment of a matrix of an LCD drive for a six-primary color system with a backlight illuminated LCD monitor.
  • FIG. 80 illustrates one embodiment of an optical filter array for a six-primary color system with a backlight illuminated LCD monitor.
  • FIG. 81 illustrates an array for a Quantum Dot (QD) display device.
  • FIG. 82 illustrates one embodiment of an array for a six-primary color system for use with a direct emissive assembled display.
  • FIG. 83 illustrates one embodiment of a six-primary color system in an emissive display that does not incorporate color filtered subpixels.
  • FIG. 84 illustrates a graph of one embodiment of a four primary system with respect to CIE 1931.
  • FIG. 85 illustrates a graph of one embodiment of a five primary system with respect to CIE 1931.
  • FIG. 86 illustrates a graph of one embodiment of a six primary system with respect to CIE 1931.
  • FIG. 87 illustrates a graph of one embodiment of a seven primary system with respect to CIE 1931.
  • FIG. 88 illustrates a graph of one embodiment of an eight primary system with respect to CIE 1931.
  • FIG. 89 illustrates a graph of one embodiment of a ten primary system with respect to CIE 1931.
  • FIG. 90 illustrates a graph of one embodiment of a twelve primary system with respect to CIE 1931.
  • FIG. 91 illustrates a graph of another embodiment of a twelve primary system with respect to CIE 1931.
  • FIG. 92 illustrates a graph of a twelve primary system that is backwards compatible with 6P-C with respect to CIE 1931.
  • FIG. 93 shows one embodiment of transportation of twelve individual color channels on a first link (Link A) and a second link (Link B).
  • FIG. 94A shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a first link (Link A).
  • FIG. 94B shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a second link (Link B).
  • FIG. 95 A shows one embodiment of a 4:2:2 Constant Luminance Encode for a first link (Link A).
  • FIG. 95B shows one embodiment of a 4:2:2 Constant Luminance Encode for a second link (Link B).
  • FIG. 96A shows one embodiment of a 4:4:4 Encode for a first link (Link A).
  • FIG. 96B shows one embodiment of a 4:4:4 Encode for a second link (Link B).
  • FIG. 97A shows one embodiment of component mapping into SMPTE 2081-1 for a first link (Link A).
  • FIG. 97B shows one embodiment of component mapping into SMPTE 2081-1 for a second link (Link B).
  • FIG. 98A shows one embodiment of a twelve primary system mapping into SMPTE 2081-1 for a first link (Link A).
  • FIG. 98B shows one embodiment of the twelve primary system mapping into
  • FIG. 99A shows one embodiment of a 4:2:2 Non-Constant Luminance Decode for a first link (Link A).
  • FIG. 99B shows one embodiment of a 4:2:2 Non-Constant Luminance Decode for a second link (Link B).
  • FIG. 100A shows one embodiment of a 4:2:2 Constant Luminance Decode for a first link (Link A).
  • FIG. 100B shows one embodiment of a 4:2:2 Constant Luminance Decode for a second link (Link B).
  • FIG. 101A shows one embodiment of a 4:4:4 Decode for a first link (Link A).
  • FIG. 101B shows one embodiment of a 4:4:4 Decode for a second link (Link B).
  • FIG. 102A illustrates a front view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
  • FIG. 102B illustrates a normal orthogonal view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
  • FIG. 102C illustrates a top view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
  • FIG. 103A illustrates a front view of a three-dimensional plot of DCI-P3 in XYZ space.
  • FIG. 103B illustrates a normal orthogonal view of a three-dimensional plot of DCI-P3 in XYZ space.
  • FIG. 103C illustrates a top view of a three-dimensional plot of DCI-P3 in XYZ space.
  • FIG. 104A illustrates a front view of 6P-C in XYZ space.
  • FIG. 104B illustrates a normal orthogonal view of 6P-C in XYZ space.
  • FIG. 104C illustrates a top view of 6P-C in XYZ space.
  • FIG. 105A illustrates a front view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
  • FIG. 105B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
  • FIG. 105C illustrates a top view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
  • FIG. 106A illustrates a front view of DCI-P3 (red) and 6P-C (green) in XYZ space.
  • FIG. 106B illustrates a normal orthogonal view of DCI-P3 (red) and 6P-C (green) in XYZ space.
  • FIG. 106C illustrates a top view of DCI-P3 (red) and 6P-C (green) in XYZ space.
  • FIG. 107A illustrates a front view of 4P in XYZ space.
  • FIG. 107B illustrates a normal orthogonal view of 4P in XYZ space.
  • FIG. 107C illustrates a top view of 4P in XYZ space.
  • FIG. 108A illustrates a front view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space.
  • FIG. 108B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space.
  • FIG. 108C illustrates a top view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space
  • FIG. 109A illustrates a front view of DCI-P3 (red) and 4P (blue) in XYZ space.
  • FIG. 109B illustrates a normal orthogonal view of DCI-P3 (red) and 4P (blue) in
  • FIG. 109C illustrates a top view of DCI-P3 (red) and 4P (blue) in XYZ space.
  • FIG. 110A illustrates a front view of 4P-N in XYZ space.
  • FIG. HOB illustrates a normal orthogonal view of 4P-N in XYZ space.
  • FIG. 110C illustrates a top view of 4P-N in XYZ space.
  • FIG. 111 A illustrates a front view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
  • FIG. 11 IB illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
  • FIG. 111C illustrates a top view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
  • FIG. 112A illustrates a front view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
  • FIG. 112B illustrates a normal orthogonal view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
  • FIG. 112C illustrates a top view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
  • FIG. 113A illustrates one embodiment of a quadrature method (“System 2A”).
  • FIG. 113B illustrates another embodiment of a quadrature method (“System 2A”).
  • FIG. 113C illustrates yet another embodiment of a quadrature method (“System
  • FIG. 114A illustrates an embodiment of a stereo quadrature method (“System
  • FIG. 114B illustrates another embodiment of a stereo quadrature method (“System 2A”).
  • FIG. 114C illustrates yet another embodiment of a stereo quadrature method
  • FIG. 115 illustrates one embodiment of a Yxy encode with an OETF.
  • FIG. 116 illustrates one embodiment of a Yxy encode without an OETF.
  • FIG. 117 illustrates one embodiment of a Yxy decode with an electro-optical transfer function (EOTF).
  • EOTF electro-optical transfer function
  • FIG. 118 illustrates one embodiment of a Yxy decode without an EOTF.
  • FIG. 119 illustrates one embodiment of a 4:2:2 Yxy encode with an OETF.
  • FIG. 120 illustrates one embodiment of a 4:2:2 Yxy encode without an OETF.
  • FIG. 121 illustrates one embodiment of a 4:4:4 Yxy encode with an OETF.
  • FIG. 122 illustrates one embodiment of a 4:4:4 Yxy encode without an OETF.
  • FIG. 123 illustrates sample placements of Yxy system components for a 4:2:2 pixel mapping.
  • FIG. 124 illustrates sample placements of Yxy system components for a 4:2:0 pixel mapping.
  • FIG. 125 illustrates one embodiment of a SMPTE ST292 Yxy system mapping.
  • FIG. 126 illustrates one embodiment of a SMPTE ST2082 Yxy system mapping.
  • FIG. 127 illustrates one embodiment of Yxy inserted into a CTA 861 stream.
  • FIG. 128 illustrates one embodiment of a Yxy decode with an EOTF.
  • FIG. 129 illustrates one embodiment of a Yxy decode without an EOTF.
  • FIG. 130A illustrates one embodiment of an IPT 4:4:4 encode.
  • FIG. 130B illustrates one embodiment of an IPT 4:4:4 decode.
  • FIG. 131 A illustrates one embodiment of an ICTCP 4:2:2 encode.
  • FIG. 13 IB illustrates one embodiment of an ICTCP 4:2:2 decode.
  • FIG. 132 illustrates the emissive spectra of Xenon lamps and UHPHg lamps.
  • FIG. 133 illustrates one embodiment of the dual-panel display system using a Cyan filter.
  • FIG. 134 illustrates one embodiment of a Vi gamma function.
  • FIG. 135 illustrates a graph of maximum quantizing error using the Vi gamma function.
  • FIG. 136 illustrates one embodiment of a 1/3 gamma function.
  • FIG. 137 is a schematic diagram of an embodiment of the invention illustrating a computer system.
  • the present invention is generally directed to a multi-primary color system.
  • the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter, wherein the single display device and the image data converter are in network communication, wherein the image data converter further includes a first link component and a second link component
  • the single display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the single display device is based on the set of image data.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 493nm, a third primary at approximately 540nm, and a fourth primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 485nm, a third primary at approximately 510nm, a fourth primary at approximately 535nm, and a fifth primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 490nm, a third primary at approximately 506nm, a fourth primary at approximately 520nm, a fifth primary at approximately 545nm, and a sixth primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 508nm, a fifth primary at approximately 520nm, a sixth primary at approximately 540nm, and a seventh primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 500nm, a fifth primary at approximately 51 Inm, a sixth primary at approximately 521nm, a seventh primary at approximately 545nm, and an eighth primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 502nm, a sixth primary at approximately 512nm, a seventh primary at approximately 520nm, an eighth primary at approximately 535nm, a ninth primary at approximately 550nm, and a tenth primary at approximately 660nm.
  • the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 505nm, a seventh primary at approximately 511nm, an eighth primary at approximately 517nm, a ninth primary at approximately 523nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 550nm, and a twelfth primary at approximately 670nm.
  • the at least four primaries include a first primary at approximately 400nm, a second primary at approximately 468nm, a third primary at approximately 484nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 506nm, a seventh primary at approximately 512nm, an eighth primary at approximately 518nm, a ninth primary at approximately 524nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 556nm, and a twelfth primary at approximately 700nm.
  • the at least four primaries include a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary.
  • the set of SDP parameters is modifiable.
  • the mid-Kelvin white emitter is modified to include a green bias.
  • the first set of color channel data is converted by the first link component and the second set of color channel data is converted by the second link component, and wherein the first set of color channel data and the second set of color channel data are combined to form the set of image data for display on the single display device.
  • the system further includes a standardized transport format, wherein the first link component includes a first standardized transport format link and wherein the second link component includes a second standardized transport format link, wherein the standardized transport format is operable to receive the first set of image data and the second set of image data using the first standardized transport format link and the second standardized transport format link, and wherein the first standardized transport format link and the second standardized transport format link are operable to combine the first set of image data and the second set of image data into a combined set of image data.
  • the present invention provides system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, wherein the set of SDP parameters is modifiable, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a midKelvin white emitter, wherein the single display device and the image data converter are in network communication, wherein the image data converter further includes
  • the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with a single display device, and wherein the image data converter further includes a first link component and a second link component, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the single display device using the image data converter, transporting the first set of color channel data to the single display device using the first link component, and transporting the second set of color channel data to the single display device using the first link component in parallel with the first link component, wherein the at least one white emitter includes at least three
  • the at least one white emitter includes a white emitter matching a white point of the primary color system.
  • the present invention relates to color systems. A multitude of color systems are known, but they continue to suffer numerous issues. As imaging technology is moving forward, there has been a significant interest in expanding the range of colors that are replicated on electronic displays. Enhancements to the television system have expanded from the early CCIR 601 standard to ITU-R BT.709-6, to SMPTE RP431-2, and ITU-R BT.2020. Each one has increased the gamut of visible colors by expanding the distance from the reference white point to the position of the Red (R), Green (G), and Blue (B) color primaries (collectively known as “RGB”) in chromaticity space.
  • RGB Red
  • RGB Green
  • B Blue
  • Enhancements in brightness have been accomplished through larger backlights or higher efficiency phosphors. Encoding of higher dynamic ranges is addressed using higher range, more perceptually uniform electro-optical transfer functions to support these enhancements to brightness technology, while wider color gamuts are produced by using narrow bandwidth emissions. Narrower bandwidth emitters result in the viewer experiencing higher color saturation. But there can be a disconnect between how saturation is produced and how it is controlled. What is believed to occur when changing saturation is that increasing color values of a color primary represents an increase to saturation. This is not true, as changing saturation requires the variance of a color primary spectral output as parametric. There are no variable spectrum displays available to date as the technology to do so has not been commercially developed, nor has the new infrastructure required to support this been discussed.
  • VASARI Visual Arts System for Archiving and Retrieval of Images
  • the multiprimary systems of the present invention include at least four primaries.
  • the at least four primaries preferably include at least one red primary, at least one green primary, and/or at least one blue primary.
  • the at least four primaries include a cyan primary, a magenta primary, and/or a yellow primary.
  • the at least four primaries include at least one white primary.
  • the multi-primary system includes six primaries.
  • the six primaries include a red (R) primary, a green (G) primary, a blue (B) primary, a cyan (C) primary, a magenta (M) primary, and a yellow (Y) primary, often referred to as “RGBCMY”.
  • R red
  • G green
  • B blue
  • C cyan
  • M magenta
  • Y yellow
  • 6P-B is a color set that uses the same RGB values that are defined in the ITU-R BT.709-6 television standard. The gamut includes these RGB primary colors and then adds three more color primaries orthogonal to these based on the white point.
  • the white point used in 6P-B is D65 (ISO 11664-2).
  • the red primary has a dominant wavelength of 609nm
  • the yellow primary has a dominant wavelength of 571nm
  • the green primary has a dominant wavelength of 552nm
  • the cyan primary has a dominant wavelength of 491nm
  • the blue primary has a dominant wavelength of 465nm as shown in Table 1.
  • the dominant wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the dominant wavelength is within ⁇ 5% of the value listed in the table below.
  • the dominant wavelength is within ⁇ 2% of the value listed in the table below.
  • FIG. 1 illustrates 6P-B compared to ITU-R BT.709-6.
  • 6P-C is based on the same RGB primaries defined in SMPTE RP431-2 projection recommendation. Each gamut includes these RGB primary colors and then adds three more color primaries orthogonal to these based on the white point.
  • the white point used in 6P-B is D65 (ISO 11664-2).
  • Two versions of 6P-C are used. One is optimized for a D60 white point (SMPTE ST2065-1), and the other is optimized for a D65 white point.
  • the red primary has a dominant wavelength of 615nm
  • the yellow primary has a dominant wavelength of 570nm
  • the green primary has a dominant wavelength of 545nm
  • the cyan primary has a dominant wavelength of 493nm
  • the blue primary has a dominant wavelength of 465nm as shown in Table 2.
  • the dominant wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the dominant wavelength is within ⁇ 5% of the value listed in the table below.
  • the dominant wavelength is within ⁇ 2% of the value listed in the table below.
  • FIG. 2 illustrates 6P-C compared to SMPTE RP431-2 for a D60 white point.
  • the red primary has a dominant wavelength of 615nm
  • the yellow primary has a dominant wavelength of 570nm
  • the green primary has a dominant wavelength of 545nm
  • the cyan primary has a dominant wavelength of 423nm
  • the blue primary has a dominant wavelength of 465nm as shown in Table 3.
  • the dominant wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the dominant wavelength is within ⁇ 5% of the value listed in the table below.
  • the dominant wavelength is within ⁇ 2% of the value listed in the table below.
  • FIG. 3 illustrates 6P-C compared to SMPTE RP431-2 for a D65 white point.
  • ITU-R BT.2020 One of the advantages of ITU-R BT.2020 is that it can include all of the Pointer colors and that increasing primary saturation in a six-color primary design could also do this.
  • Pointer is described in “The Gamut of Real Surface Colors, M.R. Pointer”, Published in Colour Research and Application Volume #5, Issue #3 (1980), which is incorporated herein by reference in its entirety.
  • 6P- C 6P gamut beyond SMPTE RP431-2
  • the first problem is the requirement to narrow the spectrum of the extended primaries.
  • the second problem is the complexity of designing a backwards compatible system using color primaries that are not related to current standards. But in some cases, there may be a need to extend the gamut beyond 6P-C and avoid these problems.
  • the cyan color primary position is located so that the gamut edge encompasses all of Pointer’s data set. In another embodiment, the cyan color primary position is a location that limits maximum saturation.
  • S6Pa Super 6Pa
  • Table 4 is a table of values for Super 6Pa.
  • the definition of x,y are described in ISO 11664-3:2012/CIE S 014 Part 3, which is incorporated herein by reference in its entirety.
  • the definition of u ’,v ’ are described in ISO 11664-5 : 2016/CIE S 014 Part 5, which is incorporated herein by reference in its entirety.
  • X defines each color primary as dominant color wavelength for RGB and complementary wavelengths CMY.
  • the saturation is expanded on the same hue angle as 6P-C as shown in FIG. 5.
  • this makes backward compatibility less complicated. However, this requires much more saturation (i.e., narrower spectra).
  • FIG. 5 illustrates Super 6Pb compared to Super 6Pa and 6P-C.
  • Table 5 is a table of values for Super 6Pb.
  • the definition of x,y are described in ISO 11664-3:2012/CIE S 014 Part 3, which is incorporated herein by reference in its entirety.
  • the definition of u ’,v ’ are described in ISO 11664-5 : 2016/CIE S 014 Part 5, which is incorporated herein by reference in its entirety.
  • X defines each color primary as dominant color wavelength for RGB and complementary wavelengths CMY.
  • a matrix is created from XYZ values of each of the primaries. As the XYZ values of the primaries change, the matrix changes. Additional details about the matrix are described below.
  • System 1 is comprised of an encode and decode system, which can be divided into base encoder and digitation, image data stacking, mapping into the standard data transport, readout, unstack, and finally image decoding.
  • the basic method of this system is to combine opposing color primaries within the three standard transport channels and identify them by their code value.
  • System 2 uses a sequential method where three color primaries are passed to the transport format as full bit level image data and inserted as normal. The three additional channels are delayed by one pixel and then placed into the transport instead of the first colors. This is useful in situations where quantizing artifacts may be critical to image performance.
  • this system is comprised of the six primaries (e.g., RGB plus a method to delay the CMY colors for injection), image resolution identification to allow for pixel count synchronization, start of video identification, and RGB Delay.
  • primaries e.g., RGB plus a method to delay the CMY colors for injection
  • image resolution identification to allow for pixel count synchronization
  • start of video identification e.g., start of video identification
  • RGB Delay e.g., a method to delay the CMY colors for injection
  • System 3 utilizes a dual link method where two wires are used.
  • a first set of three channels e.g., RGB
  • a second set of three channels e.g., CMY
  • System 1, System 2, or System 3 can be used as described. If four color components are used, two of the channels are set to “0”. If five color components are used, one of the channels is set to “0”.
  • this transportation method works for all primary systems described herein that include up to six color components.
  • System 1 fits within legacy SDI, CTA, and Ethernet transports.
  • System 1 has zero latency processing for conversion to an RGB display. However, System 1 is limited to 11 -bit words.
  • System 2 is advantageously operable to transport 6 channels using 16-bit words with no compression. Additionally, System 2 fits within newer SDI, CTA, and Ethernet transport formats. However, System 2 requires double bit rate speed. For example, a 4K image requires a data rate for an 8K RGB image.
  • System 3 is operable to transport up to 6 channels using 16-bit words with compression and at the same data required for a specific resolution.
  • a data rate for an RGB image is the same as for a 6P image using System 3.
  • System 3 is operable to transport up to 6 channels using 16-bit words with compression and at the same data required for a specific resolution.
  • a data rate for an RGB image is the same as for a 6P image using System 3.
  • System 3 is operable to transport up to 6 channels using 16-bit words with compression and at the same data required for a specific resolution.
  • a data rate for an RGB image is the same as for a 6P image using System 3.
  • System 3 is operable to transport up to 6 channels using 16-bit words with compression and at the same data required for a specific resolution.
  • R describes red data as linear light (e.g., without a non-linear function applied).
  • G describes green data as linear light.
  • B describes blue data as linear light.
  • C describes cyan data as linear light.
  • M describes magenta data as linear light.
  • Y c and/or Y describe yellow data as linear light.
  • R ’ describes red data as non-linear light (e.g., with a non-linear function applied).
  • G ’ describes green data as non-linear light.
  • B ’ describes blue data as non-linear light.
  • C ’ describes cyan data as non-linear light.
  • M describes magenta data as non-linear light.
  • Y c ’ and/or Y ’ describe yellow data as non-linear light.
  • Ye describes the luminance sum of RGBCMY data.
  • TRGB describes a System 2 encode that is the linear luminance sum of the RGB data.
  • TCMY describes a System 2 encode that is the linear luminance sum of the CMY data.
  • CR describes the data value of red after subtracting linear image luminance.
  • CB describes the data value of blue after subtracting linear image luminance.
  • Cc describes the data value of cyan after subtracting linear image luminance.
  • CY describes the data value of yellow after subtracting linear image luminance.
  • RGB describes a System 2 encode that is the nonlinear luminance sum of the RGB data.
  • K CMY describes a System 2 encode that is the nonlinear luminance sum of the CMY data.
  • -Y describes the sum of RGB data subtracted from Ye.
  • C ’ R describes the data value of red after subtracting nonlinear image luminance.
  • C ’B describes the data value of blue after subtracting nonlinear image luminance.
  • C ’c describes the data value of cyan after subtracting nonlinear image luminance.
  • C ’Y describes the data value of yellow after subtracting nonlinear image luminance.
  • B+Y describes a System 1 encode that includes either blue or yellow data.
  • G+M describes a System 1 encode that includes either green or magenta data.
  • R+C describes a System 1 encode that includes either green or magenta data.
  • CR+CC describes a System 1 encode that includes either color difference data.
  • CB+CY describes a System 1 encode that includes either color difference data.
  • 4:4:4 describes full bandwidth sampling of a color in an RGB system.
  • 4:4:4:4:4 describes full sampling of a color in an RGBCMY system.
  • 4:2:2 describes an encode where a full bandwidth luminance channel (7) is used to carry image detail and the remaining components are half sampled as a Cb Cr encode.
  • 4:2:2:2:2 describes an encode where a full bandwidth luminance channel (7) is used to carry image detail and the remaining components are half sampled as a Cb Cr Cy Cc encode.
  • 4:2:0 describes a component system similar to 4:2:2, but where Cr and Cb samples alternate per line.
  • 4:2:0:2:0 describes a component system similar to 4:2:2, but where Cr, Cb, Cy, and Cc samples alternate per line.
  • Constant luminance is the signal process where luminance (7) values are calculated in linear light.
  • Non-constant luminance is the signal process where luminance (7) values are calculated in nonlinear light.
  • the multi-primary color system is compatible with legacy systems.
  • a backwards compatible multi-primary color system is defined by a sampling method.
  • the sampling method is 4:4:4.
  • the sampling method is 4:2:2.
  • the sampling method is 4:2:0.
  • new encode and decode systems are divided into the steps of performing base encoding and digitization, image data stacking, mapping into the standard data transport, readout, unstacking, and image decoding (“System 1”).
  • System 1 combines opposing color primaries within three standard transport channels and identifies them by their code value.
  • the processes are analog processes.
  • the processes are digital processes.
  • the sampling method for a multi-primary color system is a 4:4:4 sampling method. Black and white bits are redefined. In one embodiment, putting black at midlevel within each data word allows the addition of CMY color data.
  • FIG. 6 illustrates an embodiment of an encode and decode system for a multiprimary color system.
  • the multi-primary color encode and decode system is divided into a base encoder and digitation, image data stacking, mapping into the standard data transport, readout, unstack, and finally image decoding (“System 1”).
  • System 1 image decoding
  • the method of this system combines opposing color primaries within the three standard transport channels and identifies them by their code value.
  • the encode and decode for a multi-primary color system are analog-based.
  • the encode and decode for a multi-primary color system are digital-based.
  • System 1 is designed to be compatible with lower bandwidth systems and allows a maximum of 11 bits per channel and is limited to sending only three channels of up to six primaries at a time. In one embodiment, it does this by using a stacking system where either the color channel or the complementary channel is decoded depending on the bit level of that one channel.
  • FIG. 7 illustrates a sequential method where three color primaries are passed to the transport format as full bit level image data and inserted as normal (“System 2”).
  • System 2 The three additional channels are delayed by one pixel and then placed into the transport instead of the first colors.
  • This method is useful in situations where quantizing artifacts is critical to image performance.
  • this system is comprised of six primaries (RGBCMY), a method to delay the CMY colors for injection, image resolution identification to all for pixel count synchronization, start of video identification, RGB delay, and for YCCCCC systems, logic to select the dominant color primary.
  • RGBCMY six primaries
  • the advantage of System 2 is that full bit level video can be transported, but at double the normal data rate.
  • System 2 sequences on a pixel -to-pixel basis.
  • System 2A a quadrature method is also possible (“System 2A”) that is operable to transport six primaries in stereo or twelve primary image information.
  • Each quadrant of the frame contains three color primary data sets. These are combined in the display.
  • a first set of three primaries is displayed in the upper left quadrant
  • a second set of three primaries is displayed in the upper right quadrant
  • a third set of primaries is displayed in the lower left quadrant
  • a fourth set of primaries is displayed in lower right quadrant.
  • the first set of three primaries, the second set of three primaries, the third set of three primaries, and the fourth set of three primaries do not contain any overlapping primaries (i.e., twelve different primaries).
  • the first set of three primaries, the second set of three primaries, the third set of three primaries, and the fourth set of three primaries contain overlapping primaries (i.e., at least one primary is contained in more than one set of three primaries).
  • the first set of three primaries and the third set of three primaries contain the same primaries and the second set of three primaries and the fourth set of three primaries contain the same primaries.
  • FIG. 113A illustrates one embodiment of a quadrature method (“System 2A”).
  • a first set of three primaries e.g., RGB
  • a second set of three primaries e.g., CMY
  • a third set of three primaries e.g., GC, BM, and RY
  • a fourth set of three primaries e.g., MR, YG, and CB
  • FIG. 113A illustrates a backwards compatible 12P system, this is merely for illustrative purposes.
  • the present invention is not limited to the twelve primaries shown in FIG. 113 A. Additionally, alternative pixel arrangements are compatible with the present invention.
  • FIG. 113B illustrates another embodiment of a quadrature method (“System 2A”).
  • a first set of three primaries e.g., RGB
  • a second set of three primaries e.g., CMY
  • a third set of three primaries e.g., GC, BM, and RY
  • a fourth set of three primaries e.g., MR, YG, and CB
  • FIG. 8B illustrates a backwards compatible 12P system, this is merely for illustrative purposes.
  • the present invention is not limited to the twelve primaries shown in FIG. 113B. Additionally, alternative pixel arrangements are compatible with the present invention.
  • FIG. 113C illustrates yet another embodiment of a quadrature method (“System 2A”).
  • a first set of three primaries e.g., RGB
  • a second set of three primaries e.g., CMY
  • a third set of three primaries e.g., GC, BM, and RY
  • a fourth set of three primaries e.g., MR, YG, and CB
  • FIG. 113C illustrates a backwards compatible 12P system, this is merely for illustrative purposes.
  • the present invention is not limited to the twelve primaries shown in FIG. 113C. Additionally, alternative pixel arrangements are compatible with the present invention.
  • FIG. 114A illustrates an embodiment of a quadrature method (“System 2A”) in stereo.
  • a first set of three primaries e.g., RGB
  • a second set of three primaries e.g., CMY
  • a third set of three primaries e.g., RGB
  • a fourth set of three primaries e.g., CMY
  • This embodiment allows for separation of the left eye with the first set of three primaries and the second set of three primaries and the right eye with the third set of three primaries and the fourth set of three primaries.
  • FIG. 114B illustrates another embodiment of a quadrature method (“System 2 A”) in stereo.
  • Alternative pixel arrangements and primaries are compatible with the present invention.
  • FIG. 114C illustrates yet another embodiment of a quadrature method (“System 2A”) in stereo.
  • System 2A quadrature method
  • Alternative pixel arrangements and primaries are compatible with the present invention.
  • System 2A allows for the ability to display multiple primaries (e.g., 12P and 6P) on a conventional monitor. Additionally, System 2A allows for a simplistic viewing of false color, which is useful in the production process and allows for visualizing relationships between colors. It also allows for display of multiple projectors (e.g., a first projector, a second projector, a third projector, and a fourth projector).
  • multiple projectors e.g., a first projector, a second projector, a third projector, and a fourth projector.
  • FIG. 8 illustrates one embodiment of a system encode and decode process using a dual link method (“System 3”).
  • System 3 utilizes a dual link method where two wires are used.
  • RGB is sent to link A and non-RGB primaries (e.g., CMY) are sent to link B. After arriving at the image destination, the two links are recombined.
  • Alternative primaries are compatible with the present invention.
  • System 3 is simpler and more straight forward than Systems 1 and 2.
  • the advantage with this system is that adoption is simply to format non-RGB primaries (e.g., CMY) on a second link. So, in one example, for an SDI design, RGB is sent on a standard SDI stream just as it is currently done. There is no modification to the transport and this link is operable to be sent to any RGB display requiring only the compensation for the luminance difference because the non-RGB primaries (e.g., CMY components) are not included. Data for the non-RGB primaries (e.g., CMY data) is transported in the same manner as RGB data. This data is then combined in the display to make up a 6P image.
  • non-RGB primaries e.g., CMY
  • the system requires two wires to move one image.
  • This system is operable to work with most any format including SMPTE ST292, 424, 2082, and 2110. It also is operable to work with dual HDMI/CTA connections.
  • the system includes at least one transfer function (e.g., OETF, EOTF).
  • FIG. 9 illustrates one embodiment of an encoding process using a dual link method. Alternative numbers of primaries and alternative primaries are compatible with the present invention.
  • FIG. 10 illustrates one embodiment of a decoding process using a dual link method.
  • Alternative numbers of primaries and alternative primaries are compatible with the present invention.
  • Color is generally defined by three component data levels (e.g., RGB, YCbCr).
  • a serial data stream must accommodate a word for each color contributor (e.g., R, G, B).
  • Use of more than three primaries requires accommodations to fit this data based on an RGB concept. This is why System 1, System 2, and System 3 use stacking, sequencing, and/or dual links. Multiple words are required to define a single pixel, which is inefficient because not all values are needed.
  • color is defined as a colorimetric coordinate.
  • every color is defined by three words.
  • Serial systems are already based on three color contributors (e.g., RGB).
  • System 4 preferably uses XYZ or Yxy as the three color contributors.
  • System 4 preferably uses two colorimetric coordinates and a luminance or a luma.
  • System 4 includes, but is not limited to, Yxy, L*a*b*, ICTCP, YCbCr, YUV, Yu'v', YPbPr, YIQ, and/or XYZ.
  • System 4 uses color contributors that are independent of a white point and/or a reference white value.
  • System 4 uses color contributors that are not independent of a white point and/or a reference white value (e.g., YCbCr, L*a*b*).
  • System 4 uses color contributors that require at least one known primaries (e.g., ICTCP).
  • L*C*h or other non-rectangular coordinate systems e.g., cylindrical, polar
  • 0 when converting Yxy to a polar system, 0 is restricted from 0 to 90 degrees because x and y are always non-negative.
  • the 0 angle is expanded by applying a transform (e.g., an affine transform) to x, y data wherein the x, y values of the white point of the system (e.g., D65) are subtracted from the x, y data such that the x, y data includes negative values.
  • 0 ranges from 0 to 360 degrees and the polar plot of the Yxy data is operable to occupy more than one quadrant.
  • XYZ has been used in cinema for over 10 years.
  • XYZ needs 16-bit float and 32- bit float encode or a minimum of 12 bits for gamma or log encoded images for better quality.
  • Transport of XYZ must be accomplished using a 4:4:4 sample system. Less than a 4:4:4 sample system causes loss of image detail because Y is used as a coordinate along with X and Z and carries color information, not a value.
  • X and Z are not orthogonal to Y and, therefore, also include luminance information.
  • converting to Yxy or Yu'v' concentrates the luminance in Y only, leaving two independent and pure chromaticity values.
  • X, Y, and Z are used to calculate x and y.
  • X, Y, and Z are used to calculate u' and v'.
  • I or L* components are used instead of Y, wherein I and/or L* data are created using gamma functions.
  • I is created using a 0.5 gamma function
  • L* is created using a 1/3 gamma function.
  • additional gamma encoding is not applied to the data as part of transport.
  • the system is operable to use any two independent colorimetric coordinates with similar properties to x and y, u’ and v’ , and/or u and v.
  • the two independent colorimetric coordinates are x and y and the system is a Yxy system.
  • the two colorimetric coordinates are u' and v' and the system is a Yu'v' system.
  • the two independent colorimetric coordinates e.g., x and y
  • this also provides an advantage for subsampling (e.g., 4:2:2, 4:2:0 and 4:1: 1).
  • other systems e.g., ICTCP and L*a*b*
  • a conversion matrix using the white point of [1,1,1] is operable to be used for ICTCP and L*a*b*, which would remove the white point reference.
  • the white point reference is operable to then be recaptured because it is the white point of [1,1,1] in XYZ space.
  • the image data includes a reference to at least one white point.
  • System 1, System 2, and System 3 use a YCbCr expansion to transport six color primary data sets, and the same transport (e.g., a YCbCr expansion) is operable to accommodate the image information as Yxy where Y is the luminance information and x,y describe CIE 1931 color coordinates in the half sample segments of the data stream (e.g., 4:2:2).
  • Y is the luminance information
  • x,y describe CIE 1931 color coordinates in the half sample segments of the data stream (e.g., 4:2:2).
  • x,y are fully sampled (e.g., 4:4:4).
  • the sampling rate is 4:2:0 or 4: 1 : 1.
  • the same transport is operable to accommodate the information as luminance and colorimetric coordinates other than x,y.
  • the same transport is operable to accommodate data set using one channel of luminance data and two channels of colorimetric data.
  • the same transport is operable to accommodate the image information as Yu'v' with full sampling (e.g., 4:4:4) or partial sampling (e.g., 4:2:2, 4:2:0, 4: 1:1).
  • the same transport is used with full sampling (e.g., XYZ).
  • x,y have no reference to any primaries because x,y are explicit colorimetric positions.
  • x and y are chromaticity coordinates such that x and y can be used to define a gamut of visible color.
  • u' and v' are explicit colorimetric positions. It is possible to define a gamut of visible color in other formats (e.g., L*a*b*, ICTCP, YCbCr), but it is not always trivial.
  • the display is operable to reproduce an x,y color within a certain range of Y values, wherein the range is a function of the primaries.
  • an image can be sent as linear data (e.g., without anon-linear function applied) with anon-linear function (e.g., opto- optical transfer function (OOTF)) added after the image is received, rather than requiring a non-linear function (e.g., OOTF) applied to the signal.
  • OOTF opto- optical transfer function
  • FIG. 115 illustrates one embodiment of a Yxy encode with an opto-electronic transfer function (OETF).
  • Image data is acquired in any format operable to be converted to XYZ data (e.g., RGB, RGBCMY, CMYK).
  • XYZ data is then converted to Yxy data, and the Yxy data is processed through an OETF.
  • the processed Yxy data is then converted to a standardized transportation format for mapping and readout.
  • x and y remain as independent colorimetric coordinates and the non-linear function (e.g., OETF, log, gamma, PQ) is only applied to Y, thus avoiding compression or loss of colorimetric data.
  • the OETF is described in ITU-R BT.2100 or ITU-R BT.1886.
  • Y is orthogonal to x and y, and remains orthogonal to x and y even when a non-linear function is applied.
  • System 4 is compatible with a plurality of data formats including data formats using one luminance coordinate and two colorimetric coordinates.
  • FIG. 116 illustrates one embodiment of a Yxy encode without an OETF.
  • Image data is acquired in any format operable to be converted to XYZ data (e.g., RGB, RGBCMY,
  • CMYK complementary metal-oxide-semiconductor
  • the XYZ data is then converted to Yxy data, and then converted to a standardized transportation format for mapping and readout.
  • FIG. 14 shows a Yxy encode
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 117 illustrates one embodiment of a Yxy decode with an electro-optical transfer function (EOTF).
  • EOTF electro-optical transfer function
  • System 4 is operable to be used with a plurality of data formats.
  • the matrices are as follows: [00312] In an embodiment where the color gamut used is a SMPTE RP431-2 color gamut, the matrices are as follows:
  • the matrices are as follows:
  • FIG. 118 illustrates one embodiment of a Yxy decode without an EOTF.
  • the Yxy data is then converted to the XYZ data.
  • the XYZ data is operable to be converted to multiple data formats including, but not limited to, RGB, CMYK, 6P (e.g., 6P-B, 6P-C), and gamuts including at least four primaries through at least twelve primaries.
  • FIG. 118 shows a Yxy encode
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 119 illustrates one embodiment of a 4:2:2 Yxy encode with an OETF.
  • a full bandwidth luminance channel (T) is used to carry image detail and the remaining color coordinate components (e.g., x,y) are half sampled.
  • the Yxy data undergoes a 4:2:2 encode.
  • Other encoding methods e.g., 4:4:4, 4:2:0, 4:1:1 are compatible with the present invention.
  • Other quantization methods and bit depths are also compatible with the present invention.
  • the bit depth is 8 bits, 10 bits, 12 bits, 14 bits, and/or 16 bits.
  • the Yxy values are sampled as floats.
  • FIG. 120 illustrates one embodiment of a 4:2:2 Yxy encode without an OETF.
  • the Yxy data undergoes a 4:2:2 encode.
  • Other encoding methods e.g., 4:4:4, 4:2:0, 4:1:1 are compatible with the present invention.
  • FIG. 120 shows a Yxy encode
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 121 illustrates one embodiment of a 4:4:4 Yxy encode with an OETF.
  • a full bandwidth luminance channel (T) is used to carry image detail and the remaining color coordinate components (e.g., x,y) are also fully sampled.
  • the Yxy data undergoes a 4:4:4 encode.
  • Other encoding methods e.g., 4:2:2, 4:2:0, 4:1:1 are compatible with the present invention.
  • FIG. 121 shows a Yxy encode
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 122 illustrates one embodiment of a 4:4:4 Yxy encode without an OETF.
  • the Yxy data undergoes a 4:4:4 encode.
  • Other encoding methods e.g., 4:2:2, 4:2:0, 4:1:1 are compatible with the present invention.
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 123 illustrates sample placements of Yxy system components for a 4:2:2 pixel mapping.
  • a plurality of pixels e.g., P00-P35
  • the first subscript number refers to a row number and the second subscript number refers to a column number.
  • E/ NT00 is the luma and the color components are x INT00 and y /JVT00 .
  • F/ NT01 is the luma.
  • F/ NT10 is the luma and the color components are x ]NT10 and yiNTio-
  • F/ NT11 is the luma.
  • the luma and the color components e.g., the set of image data
  • a particular pixel e.g., Poo
  • the data is sent linearly as luminance (e.g., YINTOO).
  • FIG. 123 includes Yxy system components
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 124 illustrates sample placements of Yxy system components for a 4:2:0 pixel mapping.
  • a plurality of pixels e.g., P00-P35
  • the first subscript number refers to a row number and the second subscript number refers to a column number.
  • Y/ NT00 is the luma and the color components are x INT00 and y WTO o-
  • F/ NT01 is the luma.
  • F/ NT10 is the luma.
  • F/ NT11 is the luma.
  • the luma and the color components corresponding to a particular pixel is used to calculate color and brightness of subpixels.
  • a particular pixel e.g., Poo
  • the data is sent linearly as luminance (e.g., YINTOO).
  • FIG. 124 includes Yxy system components, System 4 is operable to be used with a plurality of data formats.
  • the set of image data includes pixel mapping data.
  • the pixel mapping data includes a subsample of the set of values in a color space.
  • the color space is a Yxy color space (e.g., 4:2:2).
  • the pixel mapping data includes an alignment of the set of values in the color space (e.g., Yxy color space, Yu'v').
  • Table 6 illustrates mapping to SMPTE ST2110 for 4:2:2 sampling of Yxy data.
  • Table 7 illustrates mapping to SMPTE ST2110 for 4:4:4 linear and non-linear sampling of Yxy data.
  • the present invention is compatible with a plurality of data formats (e.g., Yu'v') and not restricted to Yxy data.
  • FIG. 125 illustrates one embodiment of a SMPTE ST292 Yxy system mapping.
  • Y/ NT is placed in the Y data segments
  • x INT is placed in the Cr data segments
  • y INT is placed in the Cb data segments.
  • luminance or luma is placed in the Y data segments
  • a first colorimetric coordinate is placed in the Cr data segments
  • a second colorimetric coordinate is placed in the Cb data segments.
  • FIG. 126 illustrates one embodiment of a SMPTE ST2082 Yxy system mapping.
  • Y/ NT is placed in the G data segments
  • x ]NT is placed in the R data segments
  • y INT is placed in the B data segments.
  • luminance or luma is placed in the G data segments
  • a first colorimetric coordinate is placed in the R data segments
  • a second colorimetric coordinate is placed in the B data segments.
  • FIG. 127 illustrates one embodiment of Yxy inserted into a CTA 861 stream.
  • FIG. 127 shows a Yxy system mapping
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 128 illustrates one embodiment of a Yxy decode with an EOTF.
  • a non-linear function is applied to the luminance to create a luma.
  • the nonlinear function is not applied to the two colorimetric coordinates.
  • System 4 is operable to be used with a plurality of data formats.
  • FIG. 129 illustrates one embodiment of a Yxy decode without an EOTF.
  • data is sent linearly as luminance.
  • a non-linear function e.g., EOTF
  • System 4 is operable to be used with a plurality of data formats.
  • XYZ is used as the basis of ACES for cinematographers and allows for the use of colors outside of the ITU-R BT.709 and/or the P3 color spaces, encompassing all of the CIE color space.
  • XYZ Colorists often work in XYZ, so there is widespread familiarity with XYZ. Further, XYZ is used for other standards (e.g., JPEG 2000, Digital Cinema Initiatives (DCI)), which could be easily adapted for System 4. Additionally, most color spaces use XYZ as the basis for conversion, so the conversions between XYZ and most color spaces are well understood and documented. Many professional displays also have XYZ option as a color reference function.
  • DCI Digital Cinema Initiatives
  • the image data converter includes at least one look-up table (LUT).
  • the at least one look-up table maps out of gamut colors to zero.
  • the at least one look-up table maps out of gamut colors to a periphery of visible colors.
  • TRANSFER FUNCTIONS [00334] The system design minimizes limitations to use standard transfer functions for both encode and/or decode processes. Current practices used in standards include, but are not limited to, ITU-R BT.1886, ITU-R BT.2020, SMPTE ST274, SMPTE ST296, SMPTE ST2084, and ITU-R BT.2100. These standards are compatible with this system and require no modification.
  • Encoding and decoding multi-primary (e.g., 6P, RGBC) images is formatted into several different configurations to adapt to image transport frequency limitations.
  • the highest quality transport is obtained by keeping all components as multi-primary (e.g., RGBCMY) components. This uses the highest sampling frequencies and requires the most signal bandwidth.
  • An alternate method is to sum the image details in a luminance channel at full bandwidth and then send the color difference signals at half or quarter sampling (e.g., Y Cr Cb Cc Cy). This allows a similar image to pass through lower bandwidth transports.
  • An IPT system is a similar idea to the Yxy system with several exceptions.
  • An IPT system or an ICTCP system is still an extension of XYZ and is operable to be derived from RGB and multiprimary (e.g., RGBCMY, RGBC) color coordinates.
  • RGBCMY, RGBC multiprimary
  • An IPT color description can be substituted within a 4:4:4 sampling structure, but XYZ has already been established and does not require the same level of calculations.
  • For an ICTCP transport system similar substitutions can be made. However, both substitution systems are limited in that anon-linear function (e.g., OOTF) is contained in all three components.
  • the non-linear function can be removed for IPT or ICTCP, the derivation would still be based on a set of RGB primaries with a white point reference. Removing the non-linear function may also alter the bit depth noise and compressibility.
  • FIG. 130A illustrates one embodiment of an IPT 4:4:4 encode.
  • FIG. 130B illustrates one embodiment of an IPT 4:4:4 decode.
  • FIG. 131 A illustrates one embodiment of an ICTCP 4:2:2 encode.
  • FIG. 13 IB illustrates one embodiment of an ICTCP 4:2:2 decode.
  • Transfer functions used in systems 1, 2, and 3 are generally framed around two basic implementations.
  • the transfer functions are defined within two standards.
  • the OETF is defined in ITU-R BT.709-6, table 1, row 1.2.
  • the inverse function, the EOTF is defined in ITU-R BT.1886.
  • PQ perceptual quantizer
  • HLG hybrid log-gamma
  • System 4 is operable to use any of the transfer functions, which can be applied to the Y component.
  • a new method has been developed: a 'A gamma function.
  • the A gamma function allows for a single calculation from the luminance (e.g., Y) component of the signal (e.g., Yxy signal) to the display.
  • the A gamma function is designed for data efficiency, not as an optical transform function.
  • the A gamma function is used instead of a nonlinear function (e.g., OETF or EOTF).
  • signal input to the A gamma function is assumed to be linear and constrained between values of 0 and 1.
  • the A gamma function is optimized for 10-bit transport and/or 12-bit transport.
  • the A gamma function is optimized for 14-bit transport and/or 16-bit transport.
  • the A gamma function is optimized for 8-bit transport.
  • a typical implementation applies an inverse of the A gamma function, which linearizes the signal. A conversion to a display gamut is then applied.
  • FIG. 134 illustrates one embodiment of a A gamma function.
  • using the 'A gamma function with the display gamma combines the functions into a single step rather than utilizing a two-step conversion process.
  • at least one tone curve is applied after the A gamma function.
  • the A gamma function advantageously provides ease to convert to and from linear values. Given that all color and tone mapping has to be done in the linear domain, having a simple to implement conversion is desirable and makes the conversion to and from linear values easier and simpler.
  • FIG. 135 illustrates a graph of maximum quantizing error using the A gamma function.
  • the maximum quantizing error from an original 16-bit image (black trace) to a 10- bit (blue trace) signal is shown in the graph.
  • the maximum quantizing error is less than 0.1% (e.g., 0.0916%) for 16-bit to 10-bit conversion using the A gamma function. This does not include any camera log functions designed into a camera.
  • the graph also shows the maximum quantizing error from the original 16-bit image to a 12-bit (red trace) signal and a 14-bit (green trace) signal.
  • a A gamma is ideal for converting images with 16-bit (e.g., 16-bit float) values to 12-bit (e.g., 12-bit integer) values
  • a 1/3 gamma provides equivalent performance in terms of peak signal-to-noise ratio (PSNR).
  • PSNR peak signal-to-noise ratio
  • the 1/3 gamma conversion from 16-bit float maintains the same performance as A gamma.
  • an equation for finding an optimum value of gamma is: [00348]
  • the Minimum Float Value is based on the IEEE Standard for
  • the range of image values is normalized to between 0 and 1.
  • the range of image values is preferably normalized to between 0 and 1 and then the gamma function is applied.
  • FIG. 108 illustrates one embodiment of a 1/3 gamma function.
  • FIG. 11 illustrates one embodiment of a six-primary color system encode using a 4:4:4 sampling method.
  • This process is accomplished by processing multiprimary (e.g., RGBCMY) video information through a standard Optical Electronic Transfer Function (OETF) (e.g., ITU-R BT.709-6), digitizing the video information as four samples per pixel, and quantizing the video information as 11 -bit or 9-bit.
  • OETF Optical Electronic Transfer Function
  • the multi-primary (e.g., RGBCMY) video information is processed through a standard Optical Optical Transfer Function (OOTF).
  • the multi-primary (e.g., RGBCMY) video information is processed through a Transfer Function (TF) other than OETF or OOTF.
  • TFs consist of two components, a Modulation Transfer Function (MTF) and a Phase Transfer Function (PTF).
  • MTF Modulation Transfer Function
  • PTF Phase Transfer Function
  • the MTF is a measure of the ability of an optical system to transfer various levels of detail from object to image. In one embodiment, performance is measured in terms of contrast (degrees of gray), or of modulation, produced for a perfect source of that detail level.
  • the PTF is a measure of the relative phase in the image(s) as a function of frequency. A relative phase change of 180°, for example, indicates that black and white in the image are reversed. This phenomenon occurs when the TF becomes negative.
  • MTF is measured using discrete frequency generation. In one embodiment, MTF is measured using continuous frequency generation. In another embodiment, MTF is measured using image scanning. In another embodiment, MTF is measured using waveform analysis.
  • the six-primary color system is for a 12-bit serial system.
  • Current practices normally set black at bit value 0 and white at bit value 4095 for 12-bit video.
  • the bit defining black is moved to bit value 2048.
  • the new encode has RGB values starting at bit value 2048 for black and bit value 4095 for white and non-RGB primary (e.g., CMY) values starting at bit value 2047 for black and bit value 0 as white.
  • the six- primary color system is for a 10-bit serial system.
  • FIG. 12 illustrates one embodiment for a method to package six channels of primary information into the three standard primary channels used in current serial video standards by modifying bit numbers for a 12-bit SDI and a 10-bit SDI.
  • FIG. 13 illustrates a simplified diagram estimating perceived viewer sensation as code values define each hue angle.
  • TABLE 8 and TABLE 9 list bit assignments for computer, production, and broadcast for a 12-bit system and a 10-bit system, respectively.
  • “Computer” refers to bit assignments compatible with CTA 861-G, November 2016, which is incorporated herein by reference in its entirety.
  • “Production” and/or “Broadcast” refer to bit assignments compatible with SMPTE ST 2082-0 (2016), SMPTE ST 2082-1 (2015), SMPTE ST 2082-10 (2015), SMPTE ST 2082-11 (2016), SMPTE ST 2082-12 (2016), SMPTE ST 2110-10 (2017), SMPTE ST 2110-20 (2017), SMPTE ST 2110-21 (2017), SMPTE ST 2110-30 (2017), SMPTE ST 2110-31 (2016), and/or SMPTE ST 2110-40 (2018), each of which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-6, which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-5, which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-4, which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-3, which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-2, which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-6, which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-5, which is incorporated herein by reference in its entirety.
  • the OETF process is defined in ITU-R BT.709-4, which is incorporated
  • the encoder is a non-constant luminance encoder. In another embodiment, the encoder is a constant luminance encoder.
  • FIG. 14 illustrates one embodiment for a method of stacking/encoding six-primary color information using a 4:4:4 video system.
  • Image data must be assembled according the serial system used. This is not a conversion process, but instead is a packing/ stacking process.
  • the packing/ stacking process is for a six-primary color system using a 4:4:4 sampling method.
  • FIG. 15 illustrates one embodiment for a method of unstacking/decoding six- primary color information using a 4:4:4 video system.
  • the RGB channels and the non-RGB primary (e.g., CMY) channels are combined into one 12-bit word and sent to a standardized transport format.
  • the standardized transport format is SMPTE ST424 SDI.
  • the decode is for a non-constant luminance, six- primary color system.
  • the decode is for a constant luminance, six- primary color system.
  • an electronic optical transfer function (EOTF) (e.g., ITU-R BT.1886) coverts image data back to linear for display.
  • the EOTF is defined in ITU-R BT.1886 (2011), which is incorporated herein by reference in its entirety.
  • FIG. 16 illustrates one embodiment of a 4:4:4 decoder.
  • System 2 uses sequential mapping to the standard transport format, so it includes a delay for the non-RGB (e.g., CMY) data.
  • the non-RGB (e.g., CMY) data is recovered in the decoder by delaying the RGB data. Since there is no stacking process, the full bit level video can be transported. For displays that are using optical filtering, this RGB delay could be removed and the process of mapping image data to the correct filter could be eliminated by assuming this delay with placement of the optical filter and the use of sequential filter colors.
  • Two methods can be used based on the type of optical filter used. Since this system is operating on a horizontal pixel sequence, some vertical compensation is required and pixels are rectangular.
  • RGBCMY multiprimary
  • CMY non-RGB
  • FIG. 18 allows for square pixels, but the non-RGB (e.g., CMY) components require a line delay for synchronization.
  • Other patterns eliminating the white subpixel are also compatible with the present invention.
  • FIG. 19 illustrates an embodiment of the present invention for sending six primary colors to a standardized transport format using a 4:4:4 encoder according to System 2.
  • Encoding is straight forward with a path for RGB sent directly to the transport format.
  • RGB data is mapped to each even numbered data segment in the transport.
  • Non-RGB (e.g., CMY) data is mapped to each odd numbered segment.
  • CMY complementary metal-oxide-semiconductor
  • “Computer” refers to bit assignments compatible with CTA 861 -G, November 2016, which is incorporated herein by reference in its entirety.
  • “Production” and/or “Broadcast” refer to bit assignments compatible with SMPTE ST 2082-0 (2016), SMPTE ST 2082-1 (2015), SMPTE ST 2082-10 (2015), SMPTE ST 2082-11 (2016), SMPTE ST 2082-12 (2016), SMPTE ST 2110-10 (2017), SMPTE ST 2110-20 (2017), SMPTE ST 2110-21 (2017), SMPTE ST 2110-30 (2017), SMPTE ST 2110-
  • the decode adds a pixel delay to the RGB data to realign the channels to a common pixel timing.
  • EOTF is applied and the output is sent to the next device in the system.
  • Metadata based on the standardized transport format is used to identify the format and image resolution so that the unpacking from the transport can be synchronized.
  • FIG. 20 shows one embodiment of a decoding with a pixel delay.
  • the decoding is 4:4:4 decoding.
  • the six- primary color decoder is in the signal path, where 11 -bit values for RGB are arranged above bit value 2048, while non-RGB (e.g., CMY) levels are arranged below bit value 2047 as libit. If the same data set is sent to a display and/or process that is not operable for six-primary color processing, the image data is assumed as black at bit value 0 as a full 12-bit word. Decoding begins by tapping image data prior to the unstacking process.
  • the packing/stacking process is for a six-primary color system using a 4:2:2 sampling method.
  • the standard method of converting from six primaries (e.g., RGBCMY) to a luminance and a set of color difference signals requires the addition of at least one new image designator.
  • the encoding and/or decoding process is compatible with transport through SMPTE ST 292-0 (2011), SMPTE ST 292-1 (2011, 2012, and/or 2018), SMPTE ST 292-2 (2011), SMPTE ST 2022-1 (2007), SMPTE ST 2022-2 (2007), SMPTE ST 2022-3 (2010), SMPTE ST 2022-4 (2011), SMPTE ST 2022-5 (2012 and/or 2013), SMPTE ST 2022-6 (2012), SMPTE ST 2022-7 (2013), and/or and CTA 861-G (2106), each of which is incorporated herein by reference in its entirety.
  • an electronic luminance component Y
  • the first component is: Ey 6 .
  • Ey 6 For an RGBCMY system, itcan be described as:
  • At least two new color components are disclosed. These are designated as Cc and Cy components.
  • the at least two new color components include a method to compensate for luminance and enable the system to function with older Y Cb Cr infrastructures. In one embodiment, adjustments are made to Cb and Cr in a Y Cb Cr infrastructure since the related level of luminance is operable for division over more components.
  • magenta is a sum of blue and red.
  • magenta is resolved as a calculation, not as optical data.
  • both the camera side and the monitor side of the system use magenta filters.
  • magenta would appear as a very deep blue which would include a narrow bandwidth primary, resulting in metameric issues from using narrow spectral components.
  • magenta as an integer value is resolved using the following equation:
  • the six-primary color system using a non-constant luminance encode for use with a 4:2:2 sampling method is compatible with transport through SMPTE ST 292-0 (2011), SMPTE ST 292-1 (2011, 2012, and/or 2018), SMPTE ST 292-2 (2011), SMPTE ST 2022-1 (2007), SMPTE ST 2022-2 (2007), SMPTE ST 2022-3 (2010), SMPTE ST 2022-4 (2011), SMPTE ST 2022-5 (2012 and/or 2013), SMPTE ST 2022-6 (2012), SMPTE ST 2022-7 (2013), and/or and CTA 861-G (2106), each of which is incorporated herein by reference in its entirety.
  • FIG. 21 illustrates one embodiment of an encode process for 4:2:2 video for packaging five channels of information into the standard three- channel designs. For 4:2:2, a similar method to the 4:4:4 system is used to package five channels of information into the standard three-channel designs used in current serial video standards.
  • FIG. 21 illustrates 12-bit SDI and 10-bit SDI encoding for a 4:2:2 system. TABLE 14 and TABLE 15 list bit assignments for a 12-bit and 10-bit system, respectively.
  • “Computer” refers to bit assignments compatible with CTA 861-G, November 2016, which is incorporated herein by reference in its entirety.
  • “Production” and/or “Broadcast” refer to bit assignments compatible with SMPTE ST 2082-0 (2016), SMPTE ST 2082-1 (2015), SMPTE ST 2082-10 (2015), SMPTE ST 2082-11 (2016), SMPTE ST 2082-12 (2016), SMPTE ST 2110-10 (2017), SMPTE ST 2110-20 (2017),
  • FIG. 22 illustrates one embodiment for a non-constant luminance encoding process for a six-primary color system.
  • the design of this process is similar to the designs used in current RGB systems.
  • Input video is sent to the Optical Electronic Transfer Function (OETF) process and then to the E Yf encoder.
  • OETF Optical Electronic Transfer Function
  • the output of this encoder includes all of the image detail information. In one embodiment, all of the image detail information is output as a monochrome image.
  • FIG. 23 illustrates one embodiment of a packaging process for a six-primary color system. These components are then sent to the packing/ stacking process. Components EC'Y-INT ar
  • FIG. 24 illustrates a 4:2:2 unstack process for a six-primary color system.
  • the image data is extracted from the serial format through the normal processes as defined by the serial data format standard.
  • the serial data format standard uses a 4:2:2 sampling structure.
  • the serial data format standard is SMPTE ST292.
  • the color difference components are separated and formatted back to valid 11 -bit data. Components E C ' Y-INT and E ⁇ C-INT are inverted so that bit value 2047 defines peak color luminance.
  • FIG. 25 illustrates one embodiment of a process to inversely quantize each individual color and pass the data through an electronic optical function transfer (EOTF) in a non-constant luminance system.
  • EOTF electronic optical function transfer
  • the individual color components, as well as E Y ' _ INT are inversely quantized and summed to breakout each individual color.
  • Magenta is then calculated and E Y _ INT is combined with these colors to resolve green.
  • EOTF Electronic Optical Transfer Function
  • the decoding is 4:2:2 decoding. This decode follows the same principles as the 4:4:4 decoder.
  • a luminance channel is used instead of discrete color channels.
  • image data is still taken prior to unstack from the EC'B-INT + EC'Y-INT ar
  • a 4:2:2 decoder a new component, called EL Y , is used to subtract the luminance levels that are present from the CMY channels from the E C ' B-INT + E C ' Y-INT and E BR-INT + E BC -INT components.
  • the resulting output is now the R and B image components of the EOTF process.
  • EL Y is also sent to the G matrix to convert the luminance and color difference components to a green output.
  • R’G’B’ is input to the EOTF process and output as GRGB, RRGB, and BRGB.
  • the decoder is a legacy RGB decoder for non-constant luminance systems.
  • the standard is SMPTE ST292. In one embodiment, the standard is SMPTE RP431-2. In one embodiment, the standard is ITU-R BT.2020. In another embodiment, the standard is SMPTE RP431-1. In another embodiment, the standard is ITU-R BT.1886. In another embodiment, the standard is SMPTE ST274. In another embodiment, the standard is SMPTE ST296. In another embodiment, the standard is SMPTE ST2084. In yet another embodiment, the standard is ITU-R BT.2100. In yet another embodiment, the standard is SMPTE ST424. In yet another embodiment, the standard is SMPTE ST425. In yet another embodiment, the standard is SMPTE ST2110.
  • FIG. 26 illustrates one embodiment of a constant luminance encode for a six- primary color system.
  • FIG. 27 illustrates one embodiment of a constant luminance decode for a six-primary color system.
  • the process for constant luminance encode and decode are very similar. The main difference being that the management of E ⁇ is linear.
  • the encode and decode processes stack into the standard serial data streams in the same way as is present in a non-constant luminance, six-primary color system. In one embodiment, the stacker design is the same as with the non-constant luminance system.
  • System 2 operation is using a sequential method of mapping to the standard transport instead of the method in System 1 where pixel data is combined to two color primaries in one data set as an 11 -bit word.
  • the advantage of System 1 is that there is no change to the standard transport.
  • the advantage of System 2 is that full bit level video can be transported, but at double the normal data rate.
  • YRGB and YCMY are used to define the luminance value for RGB as one group and CMY for the other.
  • Alternative primaries are compatible with the present invention.
  • FIG. 28 illustrates one example of 4:2:2 non-constant luminance encoding.
  • RGB and CMY components are mapped at different time intervals, there is no requirement for a stacking process and data is fed directly to the transport format.
  • the development of the separate color difference components is identical to System 1.
  • Alternative primaries are compatible with the present invention.
  • the encoder for System 2 takes the formatted color components in the same way as System 1. Two matrices are used to build two luminance channels. YRGB contains the luminance value for the RGB color primaries. YCMY contains the luminance value for the CMY color primaries. A set of delays are used to sequence the proper channel for YRGB, YCMY, and the RBCY channels. Because the RGB and non-RGB (e.g., CMY) components are mapped at different time intervals, there is no requirement for a stacking process, and data is fed directly to the transport format. The development of the separate color difference components is identical to System 1. The Encoder for System 2 takes the formatted color components in the same way as System 1.
  • YRGB contains the luminance value for the RGB color primaries
  • YCMY contains the luminance value for the CMY color primaries. This sequences YRGB, CR, and CC channels into the even segments of the standardized transport and YCMY, CB, and CY into the odd numbered segments. Since there is no combining color primary channels, full bit levels can be used limited only by the design of the standardized transport method. In addition, for use in matrix driven displays, there is no change to the input processing and only the method of outputting the correct color is required if the filtering or emissive subpixel is also placed sequentially.
  • Timing for the sequence is calculated by the source format descriptor which then flags the start of video and sets the pixel timing.
  • FIG. 29 illustrates one embodiment of a non-constant luminance decoding system.
  • Decoding uses timing synchronization from the format descriptor and start of video flags that are included in the payload ID, SDP, or EDID tables. This starts the pixel clock for each horizontal line to identify which set of components are routed to the proper part of the decoder. A pixel delay is used to realign the color primarily data of each subpixel.
  • YRGB and YCMY are combined to assemble a new Ye component which is used to decode the CR, CB, CC, CY, and CM components into RGBCMY.
  • the constant luminance system is not different from the non-constant luminance system in regard to operation. The difference is that the luminance calculation is done as a linear function instead of including the OOTF.
  • FIG. 30 illustrates one embodiment of a 4:2:2 constant luminance encoding system.
  • FIG. 31 illustrates one embodiment of a 4:2:2 constant luminance decoding system.
  • the six-primary color system uses a 4:2:0 sampling system.
  • the 4:2:0 format is widely used in H.262/MPEG-2, H.264/MPEG-4 Part 10 and VC-1 compression.
  • the process defined in SMPTE RP2050-1 provides a direct method to convert from a 4:2:2 sample structure to a 4:2:0 structure.
  • a 4:2:0 video decoder and encoder are connected via a 4:2:2 serial interface
  • the 4:2:0 data is decoded and converted to 4:2:2 by up-sampling the color difference component.
  • the 4:2:0 video encoder the 4:2:2 video data is converted to 4:2:0 video data by down-sampling the color difference component.
  • FIG. 32 illustrates one embodiment of a raster encoding diagram of sample placements for a six-primary color 4:2:0 progressive scan system.
  • horizontal lines show the raster on a display matrix.
  • Vertical lines depict drive columns. The intersection of these is a pixel calculation. Data around a particular pixel is used to calculate color and brightness of the subpixels.
  • Each “X” shows placement timing of the E Y INT sample. Red dots depict placement of the E C ' R-INT + E BC -INT sample. Blue triangles show placement of the E C ' B-INT + E C ' Y-INT sample.
  • the raster is an RGB raster. In another embodiment, the raster is a RGBCMY raster.
  • image data is split across three color channels in a transport system.
  • the image data is read as six- primary data.
  • the image data is read as RGB data.
  • the axis of modulation for each channel is considered as values describing two colors (e.g., blue and yellow) for a six-primary system or as a single color (e.g., blue) for an RGB system. This is based on where black is referenced.
  • black is decoded at a mid-level value. In an RGB system, the same data stream is used, but black is referenced at bit zero, not a mid-level.
  • the RGB values encoded in the 6P stream are based on ITU-R BT.709.
  • the RGB values encoded are based on SMPTE RP431.
  • these two embodiments require almost no processing to recover values for legacy display.
  • the decoding is for a 4:4:4 system.
  • the assumption of black places the correct data with each channel. If the 6P decoder is in the signal path, 11 -bit values for RGB are arranged above bit value 2048, while CMY level are arranged below bit value 2047 as 11 -bit. However, if this same data set is sent to a display or process that is does not understand 6P processing, then that image data is assumed as black at bit value 0 as a full 12-bit word.
  • FIG. 33 illustrates one embodiment of the six-primary color unstack process in a 4:2:2 video system.
  • Decoding starts by tapping image data prior to the unstacking process.
  • the input to the 6P unstack will map as shown in FIG. 34.
  • the output of the 6P decoder will map as shown in FIG. 35. This same data is sent uncorrected as the legacy RGB image data.
  • the interpretation of the RGB decode will map as shown in FIG. 36.
  • the decoding is for a 4:2:2 system.
  • This decode uses the same principles as the 4:4:4 decoder, but because a luminance channel is used instead of discrete color channels, the processing is modified. Legacy image data is still taken prior to unstack from the E C ' B-INT + E C ' Y-INT and E C ' R-INT + E C ' C-INT channels as shown in FIG. 37.
  • FIG. 38 illustrates one embodiment of a non-constant luminance decoder with a legacy process.
  • the dotted box marked (1) shows the process where a new component called EL y is used to subtract the luminance levels that are present from the CMY channels from the EC'B-INT + E C ' Y-INT and E C ' R-INT + E BC -INT components as shown in box (2).
  • the resulting output is now the R and B image components of the EOTF process.
  • EL y is also sent to the G matrix to convert the luminance and color difference components to a green output as shown in box (3).
  • R’G’B ’ is input to the EOTF process and output as GRGB, RRGB, and BRGB.
  • the decoder is a legacy RGB decoder for non-constant luminance systems.
  • the process is very similar with the exception that green is calculated as linear as shown in FIG. 39.
  • the six-primary color system outputs a legacy RGB image. This requires a matrix output to be built at the very end of the signal path.
  • FIG. 40 illustrates one embodiment of a legacy RGB image output at the end of the signal path.
  • the design logic of the C, M, and Y primaries is in that they are substantially equal in saturation and placed at substantially inverted hue angles compared to R, G, and B primaries, respectively.
  • substantially equal in saturation refers to a ⁇ 10% difference in saturation values for the C, M, and Y primaries in comparison to saturation values for the R, G, and B primaries, respectively.
  • substantially equal in saturation covers additional percentage differences in saturation values falling within the ⁇ 10% difference range.
  • substantially equal in saturation further covers a ⁇ 7.5% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; a ⁇ 5% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; a ⁇ 2% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; a ⁇ 1% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; and/or a ⁇ 0.5% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively.
  • the C, M, and Y primaries are equal in saturation to the R, G, and B primaries, respectively.
  • the cyan primary is equal in saturation to the red primary
  • the magenta primary is equal in saturation to the green primary
  • the yellow primary is equal in saturation to the blue primary.
  • the saturation values of the C, M, and Y primaries are not required to be substantially equal to their corollary primary saturation value among the R, G, and B primaries, but are substantially equal in saturation to a primary other than their corollary R, G, or B primary value.
  • the C primary saturation value is not required to be substantially equal in saturation to the R primary saturation value, but rather is substantially equal in saturation to the G primary saturation value and/or the B primary saturation value.
  • two different color saturations are used, wherein the two different color saturations are based on standardized gamuts already in use.
  • substantially inverted hue angles refers to a ⁇ 10% angle range from an inverted hue angle (e.g., 180 degrees).
  • substantially inverted hue angles cover additional percentage differences within the ⁇ 10% angle range from an inverted hue angle.
  • substantially inverted hue angles further covers a ⁇ 7.5% angle range from an inverted hue angle, a ⁇ 5% angle range from an inverted hue angle, a ⁇ 2% angle range from an inverted hue angle, a ⁇ 1% angle range from an inverted hue angle, and/or a ⁇ 0.5% angle range from an inverted hue angle.
  • the C, M, and Y primaries are placed at inverted hue angles (e.g., 180 degrees) compared to the R, G, and B primaries, respectively.
  • the gamut is the ITU-R BT.709-6 gamut. In another embodiment, the gamut is the SMPTE RP431-2 gamut.
  • the unstack process includes output as six, 11 -bit color channels that are separated and delivered to a decoder.
  • To convert an image from a six-primary color system to an RGB image at least two matrices are used.
  • One matrix is a 3x3 matrix converting a six- primary color system image to XYZ values.
  • a second matrix is a 3x3 matrix for converting from XYZ to the proper RGB color space.
  • XYZ values represent additive color space values, where XYZ matrices represent additive color space matrices.
  • Additive color space refers to the concept of describing a color by stating the amounts of primaries that, when combined, create light of that color.
  • each channel will drive each color.
  • the non-RGB (e.g., CMY) channels are ignored and only the RGB channels are displayed.
  • An element of operation is that both systems drive from the black area.
  • all are coded as bit value 0 being black and bit value 2047 being peak color luminance.
  • This process can also be reversed in a situation where an RGB source can feed a six-primary display.
  • the six-primary display would then have no information for the non-RGB (e.g., CMY) channels and would display the input in a standard RGB gamut.
  • FIG. 41 illustrates one embodiment of six-primary color output using a non-constant luminance decoder.
  • FIG. 42 illustrates one embodiment of a legacy RGB process within a six-primary color system.
  • the design of this matrix is a modification of the CIE process to convert RGB to XYZ.
  • u ’v’ values are converted back to CIE 1931 xyz values using the following formulas:
  • the gamut is SMPTE RP431-2.
  • the mapping for RGBCMY values for a SMPTE RP431-2 (6P-C) gamut are:
  • RGB saturation values SR, SG, and SB.
  • the results from the second operation are inverted and multiplied with the white point XYZ values.
  • the color gamut used is an ITU-R BT.709-6 color gamut. The values calculate as:
  • the color gamut is a SMPTE RP431-2 color gamut.
  • the values calculate as:
  • the XYZ matrix must converted to the correct standard color space.
  • the color gamut used is an ITU-R BT709.6 color gamut
  • the matrices are as follows:
  • the matrices are as follows:
  • ICiCp is a color representation format specified in the Rec. ITU-R BT.2100 standard that is used as a part of the color image pipeline in video and digital photography systems for high dynamic range (HDR) and wide color gamut (WCG) imagery.
  • the I (intensity) component is a luma component that represents the brightness of the video.
  • CT and Cp are blue-yellow (“tritanopia”) and red-green (“protanopia”) chroma components.
  • the format is derived from an associated RGB color space by a coordination transformation that includes two matrix transformations and an intermediate non-linear transfer function, known as a gamma pre-correction. The transformation produces three signals: I, CT, and Cp.
  • the ITP transformation can be used with RGB signals derived from either the perceptual quantizer (PQ) or hybrid log-gamma (HLG) nonlinearity functions.
  • PQ curve is described in ITU-
  • FIG. 43 illustrates one embodiment of packing six-primary color system image data into an ICjCp (ITP) format.
  • RGB image data is converted to an XYZ matrix.
  • the XYZ matrix is then converted to an LMS matrix.
  • the LMS matrix is then sent to an optical electronic transfer function (OETF).
  • OETF optical electronic transfer function
  • FIG. 44 illustrates one embodiment of a six-primary color system converting
  • RGBCMY image data into XYZ image data for an ITP format (e.g., 6P-B, 6P-C).
  • ITP format e.g., 6P-B, 6P-C
  • RGBCMY image data into XYZ image data for an ITP format e.g., 6P-B, 6P-C
  • this is modified by replacing the RGB to XYZ matrix with a process to convert RGBCMY to XYZ. This is the same method as described in the legacy RGB process.
  • the new matrix is as follows for an ITU-R BT.709-6 (6P-B) color gamut:
  • RGBCMY data based on an ITU-R BT.709-6 color gamut, is converted to an
  • the resulting XYZ matrix is converted to an LMS matrix, which is sent to an
  • the LMS matrix is converted to an ITP matrix.
  • the resulting ITP matrix is as follows: [00444]
  • the LMS matrix is sent to an Optical Optical Transfer Function (OOTF).
  • OOTF Optical Optical Transfer Function
  • the LMS matrix is sent to a Transfer Function other than OOTF or OETF.
  • the RGBCMY data is based on the SMPTE ST431-2 (6P- C) color gamut.
  • the matrices for an embodiment using the SMPTE ST431-2 color gamut are as follows:
  • the resulting ITP matrix is:
  • the decode process uses the standard ITP decode process, as the SRSGSB cannot be easily inverted. This makes it difficult to recover the six RGBCMY components from the ITP encode. Therefore, the display is operable to use the standard ICtCp decode process as described in the standards and is limited to just RGB output.
  • the system is operable to convert image data incorporating five primary colors.
  • the five primary colors include Red (R), Green (G), Blue (G), Cyan (C), and Yellow (Y), collectively referred to as RGBCY.
  • the five primary colors include Red (R), Green (G), Blue (B), Cyan (C), and Magenta (M), collectively referred to as RGBCM.
  • the five primary colors do not include Magenta (M).
  • the five primary colors include Red (R), Green (G), Blue (B), Cyan (C), and Orange (O), collectively referred to as RGBCO.
  • RGBCO primaries provide optimal spectral characteristics, transmittance characteristics, and makes use of a D65 white point. See, e.g., Moon-Cheol Kim et al., Wide Color Gamut Five Channel Multi-Primary for HDTV Application, Journal of Imaging Sci. & Tech. Vol. 49, No. 6, Nov./Dec. 2005, at 594- 604, which is hereby incorporated by reference in its entirety.
  • F M. C
  • F a tristimulus color vector
  • F (X, Y, Z) T
  • C a linear display control vector
  • C (Cl, C2, C3, C4, C5) T .
  • a gamut volume is calculated for a set of given control vectors on the gamut boundary.
  • the control vectors are converted into CIELAB uniform color space.
  • matrix M is non-square
  • the matrix inversion requires splitting the color gamut into a specified number of pyramids, with the base of each pyramid representing an outer surface and where the control vectors are calculated using linear equation for each given XYZ triplet present within each pyramid.
  • the conversion process is normalized.
  • a decision tree is created in order to determine which set of primaries are best to define a specified color.
  • a specified color is defined by multiple sets of primaries.
  • the system of the present invention uses a combination of parallel processing for adjacent pyramids and at least one algorithm for verifying solutions by checking constraint conditions.
  • the system uses a parallel computing algorithm.
  • the system uses a sequential algorithm.
  • the system uses a brightening image transformation algorithm.
  • the system uses a darkening image transformation algorithm.
  • the system uses an inverse sinusoidal contrast transformation algorithm.
  • the system uses a hyperbolic tangent contrast transformation algorithm. In yet another embodiment, the system uses a sine contrast transformation execution times algorithm. In yet another embodiment, the system uses a linear feature extraction algorithm. In yet another embodiment, the system uses a JPEG2000 encoding algorithm. In yet another embodiment, the system uses a parallelized arithmetic algorithm. In yet another embodiment, the system uses an algorithm other than those previously mentioned. In yet another embodiment, the system uses any combination of the aforementioned algorithms.
  • Each encode and/or decode system fits into existing video serial data streams that have already been established and standardized. This is key to industry acceptance. Encoder and/or decoder designs require little or no modification for a six-primary color system to map to these standard serial formats.
  • FIG. 45 illustrates one embodiment of a six-primary color system mapping to a SMPTE ST424 standard serial format.
  • the SMPTE ST424/ST425 set of standards allow very high sampling systems to be passed through a single cable. This is done by using alternating data streams, each containing different components of the image.
  • image formats are limited to RGB due to the absence of a method to send a full bandwidth Y signal.
  • the process for mapping a six-primary color system to a SMPTE ST425 format is the same as mapping to a SMPTE ST424 format.
  • To fit a six-primary color system into a SMPTE ST425/424 stream involves the following substitutions: G ] ' NT + M ] ' NT is placed in the Green data segments, R I ' NT + C I ' NT is placed in the Red data segments, and B I ' NT + T/ WT is placed into the Blue data segments.
  • FIG. 46 illustrates one embodiment of an SMPTE 424 6P readout.
  • System 2 requires twice the data rate as System 1, so it is not compatible with SMPTE 424. However, it maps easily into SMPTE ST2082 using a similar mapping sequence. In one example, System 2 is used to have the same data speed defined for 8K imaging to show a 4K image.
  • sub-image and data stream mapping occur as shown in SMPTE ST2082.
  • An image is broken into four sub-images, and each sub-image is broken up into two data streams (e.g., sub-image 1 is broken up into data stream 1 and data stream 2).
  • the data streams are put through a multiplexer and then sent to the interface as shown in FIG. 47.
  • FIG. 48 and FIG. 49 illustrate serial digital interfaces for a six-primary color system using the SMPTE ST2082 standard.
  • the six-primary color system data is RGBCMY data, which is mapped to the SMPTE ST2082 standard (FIG. 48). Data streams 1, 3, 5, and 7 follow the pattern shown for data stream 1. Data streams 2, 4, 6, and 8 follow the pattern shown for data stream 2.
  • the six-primary color system data is YRGB YCMY CR CB CC CY data, which is mapped to the SMPTE ST2082 standard (FIG. 49). Data streams 1, 3, 5, and 7 follow the pattern shown for data stream 1. Data streams 2, 4, 6, and 8 follow the pattern shown for data stream 2.
  • the standard serial format is SMPTE ST292.
  • SMPTE ST292 is an older standard than ST424 and is a single wire format for 1.5GB video, whereas ST424 is designed for up to 3GB video.
  • ST292 can identify the payload ID of SMPTE ST352, it is constrained to only accepting an image identified by a hex value, Oh. All other values are ignored. Due to the bandwidth and identifications limitations in ST292, a component video six-primary color system incorporates a full bit level luminance component.
  • Ey 6-INT is placed in the Y data segments
  • E C ' b-INT + E C ' y-INT is placed in the Cb data segments
  • Ec r-INT + E C ' c-INT is placed in the Cr data segments.
  • the standard serial format is SMPTE ST352.
  • SMPTE ST292 and ST424 Serial Digital Interface (SDI) formats include payload identification (ID) metadata to help the receiving device identify the proper image parameters.
  • ID payload identification
  • the standard is the SMPTE ST352 standard.
  • FIG. 50 illustrates one embodiment of an SMPTE ST292 6P mapping.
  • FIG. 51 illustrates one embodiment of an SMPTE ST292 6P readout.
  • FIG. 52 illustrates modifications to the SMPTE ST352 standards for a six-primary color system.
  • Hex code “Bh” identifies a constant luminance source and flag “Fh” indicates the presence of a six-primary color system.
  • Fh is used in combination with at least one other identifier located in byte 3.
  • the Fh flag is set to 0 if the image data is formatted as System 1 and the Fh flag is set to 1 if the image data is formatted as System 2.
  • the standard serial format is SMPTE ST2082. Where a six-primary color system requires more data, it may not always be compatible with SMPTE ST424. However, it maps easily into SMPTE ST2082 using the same mapping sequence. This usage would have the same data speed defined for 8K imaging in order to display a 4K image.
  • the standard serial format is SMPTE ST2022. Mapping to ST2022 is similar to mapping to ST292 and ST242, but as an ETHERNET format. The output of the stacker is mapped to the media pay load based on Real-time Transport Protocol (RTP) 3550, established by the Internet Engineering Task Force (IETF).
  • RTP Real-time Transport Protocol
  • RTP provides end- to-end network transport functions suitable for applications transmitting real-time data, including, but not limited to, audio, video, and/or simulation data, over multicast or unicast network services.
  • the data transport is augmented by a control protocol (RTCP) to allow monitoring of the data delivery in a manner scalable to large multicast networks, and to provide control and identification functionality.
  • RTCP control protocol
  • FIG. 53 illustrates one embodiment of a modification for a six-primary color system using the SMPTE ST2202 standard.
  • SMPTE ST2202-6:2012 HBRMT
  • ST2022 relies on header information to correctly configure the media payload. Parameters for this are established within the payload header using the video source format fields including, but not limited to, MAP, FRAME, FRATE, and/or SAMPLE.
  • MAP, FRAME, and FRATE remain as described in the standard.
  • MAP is used to identify if the input is ST292 or ST425 (RGB or Y Cb Cr).
  • SAMPLE is operable for modification to identify that the image is formatted as a six-primary color system image.
  • the image data is sent using flag “Oh” (unknown/unspecified).
  • the standard is SMPTE ST2110.
  • SMPTE ST2110 is a relatively new standard and defines moving video through an Internet system. The standard is based on development from the IETF and is described under RFC3550. Image data is described through “pgroup” construction. Each pgroup consists of an integer number of octets.
  • a sample definition is RGB or YCbCr and is described in metadata.
  • the metadata format uses a Session Description Protocol (SDP) format.
  • SDP Session Description Protocol
  • pgroup construction is defined for 4:4:4, 4:2:2, and 4:2:0 sampling as 8- bit, 10-bit, 12-bit, and in some cases 16-bit and 16-bit floating point wording.
  • six-primary color image data is limited to a 10-bit depth. In another embodiment, six-primary color image data is limited to a 12-bit depth. Where more than one sample is used, it is described as a set. For example, 4:4:4 sampling for blue, as anon-linear RGB set, is described as CO’B, Cl’B, C2’B, C3’B, and C4’B. The lowest number index being left most within the image. In another embodiment, the method of substitution is the same method used to map six-primary color content into the ST2110 standard.
  • the standard is SMPTE ST2110.
  • SMPTE ST2110-20 describes the construction for each pgroup.
  • six-primary color system content arrives for mapping as non-linear data for the SMPTE ST2110 standard.
  • six-primary color system content arrives for mapping as linear data for the SMPTE ST2110 standard.
  • FIG. 54 illustrates a table of 4:4:4 sampling for a six-primary color system for a 10-bit video system. For 4:4:4 10-bit video, 15 octets are used and cover 4 pixels.
  • FIG. 55 illustrates a table of 4:4:4 sampling for a six-primary color system for a 12-bit video system.
  • 9 octets are used and cover 2 pixels before restarting the sequence.
  • Non-linear RGBCMY image data would arrive as: G I ' NT + M I ' NT , R I ' NT + C[ NT , and B ] ' NT + Y/NT- Component substitution would follow what has been described for SMPTE ST424, where G I ' NT + M I ' NT is placed in the Green data segments, R I ' NT + is placed in the Red data segments, and B I ' NT + is placed in the Blue data segments.
  • the sequence described in the standard is shown as R0’, GO’, B0’, Rl’, GE, Bl’, etc.
  • FIG. 57 illustrates sample placements of six-primary system components for a 4:2:2 sampling system image. This follows the substitutions illustrated in FIG. 56, using a 4:2:2 sampling system.
  • FIG. 58 illustrates sequence substitutions for 10-bit and 12-bit video in 4:2:0 sampling systems using a Y Cb Cr Cc Cy color space.
  • Components are delivered to a pgroup including, but not limited to, EY 6-INT , Ec b -i N T + Ec'y- INT , and Ec r -i NT + E C ' c-INT .
  • EY 6-INT EY 6-INT
  • Ec b -i N T + Ec'y- INT Ec r -i NT + E C ' c-INT
  • E C ' c-INT E C ' c-INT
  • FIG. 59 illustrates sample placements of six-primary system components for a 4:2:0 sampling system image. This follows the substitutions illustrated in FIG. 58, using a 4:2:0 sampling system.
  • FIG. 60 illustrates modifications to SMPTE ST2110-20 for a 10-bit six-primary color system in 4:4:4 video.
  • SMPTE ST2110-20 describes the construction of each “pgroup”. Normally, six-primary color system data and/or content would arrive for mapping as nonlinear. However, with the present system there is no restriction on mapping data and/or content. For 4:4:4, 10-bit video, 15 octets are used and cover 4 pixels before restarting the sequence.
  • Non-linear, six-primary color system image data would arrive as G ] ' NT , B I ’ NT , R I ' NT , M ] ' NT , Y/ NT , and C/ NT .
  • the sequence described in the standard is shown as R0’, GO’, B0’, RE, GE, Bl’, etc.
  • FIG. 61 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:4:4 video.
  • 4:4:4 12-bit video, 9 octets are used and cover 2 pixels before restarting the sequence.
  • Non-linear, six-primary color system image data would arrive as G ] ' NT , R I ' NT , M I ' NT , Y/ NT , and C/ NT .
  • the sequence described in the standard is shown as R0’, GO’, B0’, RE, GE, BE, etc.
  • FIG. 62 illustrates modifications to SMPTE ST2110-20 for a 10-bit six primary color system in 4:2:2 video.
  • Components that are delivered to a SMPTE ST2110 pgroup include, but are not limited to, E Yr g b-INT , E Ycym-INT , Ecb-iNT? Ecr-iNT? E Cy-INT , and E C ' c-]NT .
  • E Yr g b-INT E Ycym-INT
  • Ecb-iNT? Ecr-iNT? E Cy-INT E C ' c-]NT .
  • FIG. 63 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:2:0 video.
  • Components that are delivered to a SMPTE ST2110 pgroup are the same as with the 4:2:2 method.
  • For 4:2:0 10-bit video, 15 octets are used and cover 8 pixels before restarting the sequence.
  • Table 16 summarizes mapping to SMPTE ST2110 for 4:2:2:2 and 4:2:0:2:0 sampling for System 1 and Table 17 summaries mapping to SMPTE ST2110 for 4:4:4:4:4 sampling (linear and non-linear) for System 1.
  • Table 18 summarizes mapping to SMPTE ST2110 for 4:2:2:2 sampling for System 2 and Table 19 summaries mapping to SMPTE ST2110 for 4:4:4:4:4 sampling (linear and non-linear) for System 2.
  • SDP is derived from IETF RFC 4566 which sets parameters including, but not limited to, bit depth and sampling parameters.
  • SDP parameters are contained within the RTP payload.
  • SDP parameters are contained within the media format and transport protocol. This payload information is transmitted as text. Therefore, modifications for the additional sampling identifiers requires the addition of new parameters for the sampling statement.
  • SDP parameters include, but are not limited to, color channel data, image data, framerate data, a sampling standard, a flag indicator, an active picture size code, a timestamp, a clock frequency, a frame count, a scrambling indicator, and/or a video format indicator.
  • the additional parameters include, but are not limited to, RGBCMY-4:4:4, YBRCY-4:2:2, and YBRCY- 4:2:0.
  • the additional parameters include, but are not limited to, CLYBRCY-4:2:2 and CLYBRCY-4:2:0.
  • 6PB1 defines 6P with a color gamut limited to ITU-R BT.709 formatted as System 1
  • 6PB2 defines 6P with a color gamut limited to ITU-R BT.709 formatted as System 2
  • 6PB3 defines 6P with a color gamut limited to ITU-R BT.709 formatted as System 3
  • 6PC1 defines 6P with a color gamut limited to SMPTE RP 431-2 formatted as System 1
  • 6PC2 defines 6P with a color gamut limited to SMPTE RP 431-2 formatted as System 2
  • 6PC3 defines 6P with a color gamut limited to SMPTE RP 431-2 formatted as System 3
  • 6PS1 defines 6P with a color gamut as Super 6P formatted as System
  • 6PS2 defines 6P with a color gamut as Super 6P formatted as System 2
  • 6PS3 defines 6P with a color gamut as Super 6P formatted as System 3.
  • Colorimetry can also be defined between a six-primary color system using the ITU-R BT.709-6 standard and the SMPTE ST431-2 standard, or colorimetry can be left defined as is standard for the desired standard.
  • the six-primary color system is integrated with a Consumer Technology Association (CTA) 861-based system.
  • CTA-861 establishes protocols, requirements, and recommendations for the utilization of uncompressed digital interfaces by consumer electronics devices including, but not limited to, digital televisions (DTVs), digital cable, satellite or terrestrial set-top boxes (STBs), and related peripheral devices including, but not limited to, DVD players and/or recorders, and other related Sources or Sinks.
  • DTVs digital televisions
  • STBs satellite or terrestrial set-top boxes
  • peripheral devices including, but not limited to, DVD players and/or recorders, and other related Sources or Sinks.
  • TMDS transition- minimized differential signaling
  • DVI Digital Visual Interface
  • HDMI High-Definition Multimedia Interface
  • TMDS is similar to low-voltage differential signaling (LVDS) in that it uses differential signaling to reduce electromagnetic interference (EMI), enabling faster signal transfers with increased accuracy.
  • EMI electromagnetic interference
  • TMDS uses a twisted pair for noise reduction, rather than a coaxial cable that is conventional for carrying video signals. Similar to LVDS, data is transmitted serially over the data link. When transmitting video data, and using HDMI, three TMDS twisted pairs are used to transfer video data.
  • each pixel packet is limited to 8 bits only. For bit depths higher than 8 bits, fragmented packs are used. This arrangement is no different than is already described in the current CTA-861 standard.
  • the system alters the AVI Infoframe Data to identify content.
  • AVI Infoframe Data is shown in Table 10 of CTA 861-G.
  • FIG. 64 illustrates the current RGB sampling structure for 4:4:4 sampling video data transmission.
  • video data is sent through three TMDS line pairs.
  • FIG. 65 illustrates a six-primary color sampling structure, RGBCMY, using System 1 for 4:4:4 sampling video data transmission.
  • the six-primary color sampling structure complies with CTA 861-G, November 2016, Consumer Technology Association, which is incorporated herein by reference in its entirety.
  • FIG. 66 illustrates an example of System 2 to RGBCMY 4:4:4 transmission.
  • FIG. 67 illustrates current Y Cb Cr 4:2:2 sampling transmission as non-constant luminance.
  • FIG. 68 illustrates a six-primary color system (System 1) using Y Cr Cb Cc Cy 4:2:2 sampling transmission as non-constant luminance.
  • FIG. 69 illustrates an example of a System 2 to Y Cr Cb Cc Cy 4:2:2 Transmission as non-constant luminance.
  • the Y Cr Cb Cc Cy 4:2:2 sampling transmission complies with CTA 861-G, November 2016, Consumer Technology Association.
  • FIG. 70 illustrates current Y Cb Cr 4:2:0 sampling transmission.
  • FIG. 71 illustrates a six-primary color system (System 1) using Y Cr Cb Cc Cy 4:2:0 sampling transmission.
  • HDMI sampling systems include Extended Display Identification Data (EDID) metadata.
  • EDID metadata describes the capabilities of a display device to a video source.
  • the data format is defined by a standard published by the Video Electronics Standards Association (VESA).
  • VESA Video Electronics Standards Association
  • the EDID data structure includes, but is not limited to, manufacturer name and serial number, product type, phosphor or filter type, timings supported by the display, display size, luminance data, and/or pixel mapping data.
  • the EDID data structure is modifiable and modification requires no additional hardware and/or tools.
  • EDID information is transmitted between the source device and the display through a display data channel (DDC), which is a collection of digital communication protocols created by VESA.
  • DDC display data channel
  • EDID providing the display information
  • DDC providing the link between the display and the source
  • the two accompanying standards enable an information exchange between the display and source.
  • VESA has assigned extensions for EDID.
  • Such extensions include, but are not limited to, timing extensions (00), additional time data black (CEA EDID Timing Extension (02)), video timing block extensions (VTB-EXT (10)), EDID 2.0 extension (20), display information extension (DI-EXT (40)), localized string extension (LS-EXT (50)), microdisplay interface extension (MI-EXT (60)), display ID extension (70), display transfer characteristics data block (DTCDB (A7, AF, BF)), block map (F0), display device data block (DDDB (FF)), and/or extension defined by monitor manufacturer (FF).
  • SDP parameters include data corresponding to a payload identification (ID) and/or EDID information.
  • FIG. 72 illustrates a dual stack LCD projection system for a six-primary color system.
  • the display is comprised of a dual stack of projectors. This display uses two projectors stacked on top of one another or placed side by side.
  • the optical paths of the projectors are aligned manually.
  • the two projectors are automatically aligned with internal software.
  • Each projector is similar, with the only difference being the color filters in each unit.
  • a first projector creates an RGB image while a second projector creates a CMY image.
  • the two projectors create a four-primary color display system.
  • the four-primary color system is an RGBC color system.
  • the four-primary color system is an RG1G2B system wherein the two Green primaries are within the 520-550nm wavelength range.
  • the four-primary color system is a RGBW system.
  • the two projectors create a five- primary color display system.
  • the five-primary display system includes a D65 white point.
  • the five-primary color display system includes a Yellow primary and/or a Cyan primary.
  • the five-primary color display system includes two Green primaries within the 520-550nm wavelength range. Refresh and pixel timings are synchronized, enabling a mechanical alignment between the two units so that each pixel overlays the same position between projector units.
  • the input signals to the projectors include a timing reference to synchronize the output images.
  • the outputs of the two projectors are passed through a half-silvered mirror to create one image.
  • the two projectors are Liquid-Crystal Display (LCD) projectors.
  • the two projectors are Digital Light Processing (DLP) projectors.
  • the two projectors are Liquid-Crystal on Silicon (LCOS) projectors.
  • the two projectors are Light-Emitting Diode (LED) projectors.
  • the display system includes colored LEDs for each of the primary colors in the system. In another embodiment, at least one of the primary colors is displayed using a combination of LEDs of other primary colors.
  • a 3D look-up table (LUT) is designed to map the signal data to the specific capabilities of the projector system.
  • the display is comprised of a single projector.
  • a single projector six-primary color system requires the addition of a second cross block assembly for the additional colors.
  • a single projector e.g., single LCD projector
  • the single projector six-primary color system includes a cyan dichroic mirror, an orange dichroic mirror, a blue dichroic mirror, a red dichroic mirror, and two additional standard mirrors.
  • the single projector six- primary color system includes at least four mirrors (e.g., at least six mirrors).
  • the single projector creates a four-primary color display.
  • the single projector creates a five-primary color display.
  • FIG. 74 illustrates a six-primary color system using a single projector and reciprocal mirrors.
  • the display is comprised of a single projector unit working in combination with at first set of at least six reciprocal mirrors, a second set of at least six reciprocal mirrors, and at least six LCD units.
  • Light from at least one light source emits towards the first set of at least six reciprocal mirrors.
  • one or more of the at least one light source is a Xenon lamp.
  • one or more of the at least one light source is a Hi -Pressure Mercury lamp (UHPHg).
  • FIG. 132 shows the emissive spectra of Xenon lamps and UHPHg lamps.
  • the first set of at least six reciprocal mirrors reflects light towards at least one of the at least six LCD units.
  • the at least six LCD units include, but are not limited to, a Green LCD, a Yellow LCD, a Cyan, LCD, a Red LCD, a Magenta LCD, and/or a Blue LCD.
  • Output from each of the at least six LCDs is received by the second set of at least six reciprocal mirrors. Output from the second set of at least six reciprocal mirrors is sent to the single projector unit.
  • Image data output by the single projector unit is output as a six-primary color system.
  • more than one projector is used.
  • prisms reflect light towards the LCD units and the single projector unit.
  • a combination of prisms and reciprocal mirrors reflect light towards the LCD units and the single projector unit.
  • the single projector has fewer than six LCD units.
  • the display is comprised of a dual stack Digital Micromirror Device (DMD) projector system.
  • FIG. 75 illustrates one embodiment of a dual stack DMD projector system. In this system, two projectors are stacked on top of one another.
  • the dual stack DMD projector system uses a spinning wheel filter.
  • the filter systems are illuminated by a xenon lamp.
  • each projector has two lamps and two identical color wheels.
  • the first projector uses an RGB, while the second projector uses a CMY filter set.
  • the first projector uses an RGB filter set, while the second projector uses a CMY filter set.
  • the first projector uses a rich color filter wheel that includes RGB filters and the second projector uses a cyan filter.
  • the first projector uses a high-brightness filter wheel and the second projector uses a cyan filter.
  • the wheels for each projector unit are preferably synchronized using an input video sync and/or a projector-to-projector sync, and timed so that the inverted colors are output of each projector at the same time.
  • the sync signal is part of the input signal data that is delivered to each projector.
  • the projectors are phosphor wheel systems.
  • a yellow phosphor wheel spins in time with a DMD imager to output sequential RG from a blue laser illuminator.
  • the second projector is designed the same, but uses a cyan phosphor wheel.
  • the output from the second projector becomes sequential BG.
  • the color wheel includes a cyan phosphor segment that is excited by blue light as described in U.S. Patent Application No. 14/163,985, filed January 24, 2014, now U.S. Patent No. 9,470,886, which is incorporated herein by reference in its entirety.
  • the output of both projectors is YRGGCB.
  • Magenta is developed by synchronizing the yellow and cyan wheels to overlap the flashing DMD.
  • the display is a single DMD projector solution.
  • a single DMD device is coupled with an RGB diode light source system.
  • the DMD projector uses LEDs.
  • the DMD projector includes CMY diodes.
  • the DMD projector creates CMY primaries using a double flashing technique.
  • the DMD projector is a single-chip DMD projector. The chip is synchronized with the LED lamps.
  • the DMD projector is a multichip DMD projector with one chip for each primary color LED in the system. An optical chain is used to split the light to the respective chips.
  • the single DMD projector has an RGBCMY color wheel.
  • the color wheel is a rich color wheel with color wheel segments (e.g., six segments).
  • the color wheel segments include Red, Green, Blue, Cyan, Magenta, and Yellow.
  • the color wheel segments include Magenta, Yellow, Orange, Cyan, Blue, and Green.
  • the color wheel is a high brightness color wheel with color wheel segments (e.g., six segments).
  • the color wheel segments include Red, Green, Blue, Cyan, Yellow, and White.
  • FIG. 76 illustrates one embodiment of a single DMD projector solution.
  • FIG. 77 illustrates one embodiment of a six-primary color system using a white OLED display.
  • the display is a white OLED monitor.
  • Current emissive monitor and/or television designs use a white emissive OLED array covered by a color filter. Changes to this type of display only require a change to pixel indexing and new six color primary filters. Different color filter arrays are used, placing each subpixel in a position that provides the least light restrictions, most color accuracy, and off axis display.
  • the optical filter for the OLED display uses a horizontal pixel sequence with rectangular pixels and vertical compensation.
  • the pixels are square.
  • the optical filter pattern does not include a white subpixel.
  • FIG. 78 illustrates one embodiment of an optical filter array for a white OLED display.
  • FIG. 79 illustrates one embodiment of a matrix of an LCD drive for a six-primary color system with a backlight illuminated LCD monitor.
  • the display is a backlight illuminated LCD display.
  • the design of an LCD display involves adding the CMY subpixels. Matrix drives for the CMY subpixels are similar to the RGB matrix drives. With the advent of 8K LCD televisions, it is technically feasible to change the matrix drive and optical filter and have a 4K six-primary color TV.
  • FIG. 80 illustrates one embodiment of an optical filter array for a six-primary color system with a backlight illuminated LCD monitor.
  • the optical filter array includes the additional CMY subpixels.
  • each pixel in the six-primary color system is a hexagonal shape.
  • Each hexagonal pixel is divided into six equilateral triangles and each of the primaries in the six-primary color system is displayed by one of the six equilateral triangles as described in U.S. Patent Application No. 12/005,931, filed July 3, 2008, which is incorporated herein by reference in its entirety.
  • each pixel is divided into six subpixels of the same size and area arranged in two rows of three columns. In another embodiment, each pixel is divided into six subpixels of the same size and area arranged in three rows of two columns. In yet another embodiment, each pixel is divided into six subpixels of the same size and area arranged in one row. In yet another embodiment, each pixel is divided into six subpixels of the same size and area arranged in one column. The luminance and intensity of each subpixel is dependent on the luminance and intensity of the adjacent subpixels in order to minimize the distinct visibility of individual subpixel and pixel structures. In one embodiment, complementary primary color subpixels are adjacent to each other to eliminate visual artifacts.
  • each pixel is divided into subpixels of different sizes and areas.
  • the size and number of subpixels for each primary color minimize blue and cyan spatial resolution without affecting the overall resolution of the display as described in U.S. Patent Application No. 12/909,742, filed October 21, 2010, now U.S. Patent No. 8,451,405, which is incorporated herein by reference in its entirety.
  • each pixel unit is divided into two subpixel units wherein one of the two subpixels is an RGB color and the other subpixel is the complementary CMY color of the first subpixel as described in U.S. Patent Application No. 12/229,845, filed March 5, 2009, which is incorporated herein by reference in its entirety.
  • each pixel includes at least one white subpixel to eliminate visual artifacts.
  • the at least one white subpixel includes a D65 white subpixel, a D60 white subpixel, a D45 white subpixel, a D27 white subpixel, and/or a D25 white subpixel.
  • using a D65 white subpixel eliminates most of the problems with metamerism.
  • the at least one white subpixel is a single white subpixel that matches the white point (e.g., a D65 white subpixel for a D65 white point).
  • the at least one white subpixel is at least two white subpixels.
  • the at least two white subpixels are preferably separated such that a linear combination of the at least two white subpixels covers a desired white Kelvin range.
  • the at least two white subpixels include a D65 white subpixel and a D27 white subpixel.
  • the at least two white subpixels include a D65 white subpixel and a D25 white subpixel.
  • the at least two white subpixels includes three white subpixels.
  • the three white subpixels include a D65 white subpixel, a D45 white subpixel, and a D27 white subpixel.
  • the three white subpixels include a
  • the mid-Kelvin white subpixel includes a green bias.
  • the green bias compensates for the slight magenta shift (e.g., when going from D25 to D65 with the straight line between the two points below the blackbody locus). Colors near the white locus and beyond are then a combination of the at least two white subpixels (e.g., two white subpixels, three white subpixels). A majority of colors will have a white component that is broad band.
  • the resultant spectra of a mixture of color primaries and white primaries will also be broad band with an extent dependent on an amount of the at least one white primary.
  • a higher broad band character of light results in fewer metameric problems. This is due to a white point being comprised of a combination of color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.) in anon-white subpixel system.
  • Total luminance is then related to intensities of the color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.).
  • colors such as vibrantly colored pastels are attained by using the color primaries to “color shift” a bright white to the pastel.
  • a fine balance of the color primaries is required, and small changes in a ratio of the color primaries will produce an unwanted color shift.
  • a system with at least one white subpixel is more tolerant to minor variations of intensity of the color primaries.
  • the white point of the six-primary color system changes depending on the display or the display mode.
  • the addition of white subpixels widens the bandwidth of the filter for each non-white primary.
  • each pixel is composed of fewer than six primary colors from the 6P gamut.
  • the display is composed of alternating and repeating subpixel patterns.
  • the display is composed of nonrepeating subpixel patterns.
  • the subpixel colors in a pixel and in adjacent pixels are arranged to minimize the spatial distance between colors that have maximal color distance from each other as described in U.S. Patent Application No. 10/543,511, filed January 13, 2003, now U.S. Patent No. 8,228,275, which is incorporated herein by reference in its entirety.
  • each pixel is one single primary color from the 6P gamut.
  • patterns of pixels are repeated across the display to minimize visibility of individual pixel structures as described in U.S. Patent Application No. 13/512,914, filed November 25, 2010, which is incorporated herein by reference in its entirety.
  • the display includes at least one perovskite.
  • the at least one perovskite is a lead halide perovskite.
  • the at least one perovskite is used as a quantum dot nanocrystal.
  • the at least one perovskite is a perovskite polymer bead. When light shines through the perovskite polymer bead, the color changes depending on the composition of the perovskite polymer bead (e.g., green, red, etc.).
  • the at least one perovskite is incorporated into a perovskite LED. Examples of perovskite LEDs are described in Lin, K., et al. (2016).
  • the at least one perovskite is 3D printed. See, e.g., Zhou, Nanjia, Yehonadav Bekenstein, CarissaN. Eisler, Dandan Zhang, Adam M. Schwartzberg, Peidong Yang, A. Paul Alivisatos, and Jennifer A. Lewis. 2019. “Perovskite Nanowire-Block Copolymer Composites With Digitally Programmable Polarization Anisotropy.” Science Advances, which is incorporated herein by reference in its entirety.
  • the display is a direct emissive assembled display.
  • the design for a direct emissive assembled display includes a matrix of color emitters grouped as a six-color system. Individual channel inputs drive each Quantum Dot (QD) element illuminator and/or micro LED element.
  • QD Quantum Dot
  • the quantum dots modulate light according to image data as described in U.S. Patent Application No. 15/905,085, filed February 26, 2018, now U.S. Patent No. 10,373,574, which is incorporated herein by reference in its entirety.
  • FIG. 81 illustrates an array for a Quantum Dot (QD) display device.
  • FIG. 82 illustrates one embodiment of an array for a six-primary color system for use with a direct emissive assembled display.
  • the display system is a dual-panel display system with two wide-gamut RGB displays.
  • One display has a Cyan filter and the other display has a clear Neutral -density filter.
  • the two displays are aligned and the outputs are passed through a half- silvered mirror to create an RGB-Cyan display on a view screen.
  • FIG. 133 illustrates one embodiment of the dual-panel display system using a Cyan filter.
  • FIG. 83 illustrates one embodiment of a six-primary color system in an emissive display that does not incorporate color filtered subpixels.
  • this can be modified for a six-primary color system by expanding the RGB or WRGB filter arrangement to an RGBCMY matrix.
  • the white subpixel could be removed as the luminance of the three additional primaries will replace it.
  • the CMY primaries are defined relative to the RGB primaries, and the intensities of the CMY primaries are dependent on the white point of the RGB system.
  • SDI video is input through an SDI decoder.
  • the SDI decoder outputs to a Y CrCbCcCy -RGBCMY converter.
  • the converter outputs RGBCMY data, with the luminance component (Y) subtracted. RGBCMY data is then converted to RGB data.
  • RGB data is sent to a scale sync generation component, receives adjustments to image controls, contrast, brightness, chroma, and saturation, is sent to a color correction component, and output to the display panel as LVDS data.
  • the SDI decoder outputs to an SDI Y-R switch component.
  • the SDI Y-R switch component outputs RGBCMY data.
  • the RGBCMY data is sent to a scale sync generation component, receives adjustments to image controls, contrast, brightness, chroma, and saturation, is sent to a color correction component, and output to a display panel as LVDS data.
  • the display uses narrow band illumination technologies.
  • the display is a laser display.
  • the display includes light emitting diodes (LEDs).
  • the LEDs include, but are not limited to, pumped phosphor LEDs, perovskite LEDs, organic LEDs (OLEDs), micro LEDs, and/or nanorods.
  • the display uses other narrow band systems (e.g., narrow filtered broad band light).
  • the multi-primary systems of the present invention provide an extended gamut along a right side of the CIE 1976 curve, which is important for flesh tones. Flesh tones are important for entertainment, medical, and/or scientific purposes.
  • the multi-primary systems of the present invention provide an extended gamut in the cyan region of the CIE 1976 curve.
  • the extension into the cyan area as well as into the shorter green area expands the reproduction of foliage, ice, and other natural items.
  • the multi-primary system of the present invention includes at least four primaries.
  • a first wavelength corresponding to a first primary is 460nm
  • a second wavelength corresponding to a second primary is 493nm
  • a third wavelength corresponding to a third primary is 540nm
  • a fourth wavelength corresponding to a fourth primary is 640nm as shown in Table 20.
  • the first wavelength, the second wavelength, the third wavelength, and/or the fourth wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, and/or the fourth wavelength is within ⁇ 5% of the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, and/or the fourth wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, and/or the fourth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 84 illustrates a graph of the four primaries listed in Table 20 with respect to CIE 1931.
  • the at least four primaries encompass 75.57% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 84.
  • the at least four primaries encompass at least 75% of the total area covered between 400nm and 700 nm for CIE 1931.
  • the at least four primaries encompass at least 70% of the total area covered between 400nm and 700nm for CIE 1931.
  • the at least four primaries encompass at least 65% of the total area covered between 400nm and 700nm for CIE 1931.
  • the at least four primaries include at least one white emitter.
  • the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
  • the at least one white emitter is a single white emitter that matches the white point (e.g., a D65 white emitter for a D65 white point).
  • the at least one white emitter is at least two white emitters. The at least two white emitters are preferably separated such that a linear combination of the at least two white emitters covers a desired white Kelvin range.
  • the at least two white emitters include a D65 white emitter and a D27 white emitter.
  • the at least two white emitters include a D65 white emitter and a D25 white emitter.
  • the at least two white emitters include three white emitters.
  • the three white emitters include a D65 white emitter, a D45 white emitter, and a D27 white emitter.
  • the three white emitters include a D65 white emitter, a mid-Kelvin white emitter (e.g., D45), and a D27 white emitter.
  • the mid-Kelvin white emitter includes a green bias.
  • the green bias compensates for the slight magenta shift (e.g., when going from D25 to D65 with the straight line between the two points below the blackbody locus).
  • Colors near the white locus and beyond are then a combination of the at least two white emitters (e.g., two white emitters, three white emitters).
  • a majority of colors will have a white component that is broad band. Therefore, the resultant spectra of a mixture of color primaries and white primaries will also be broad band with an extent dependent on an amount of the at least one white primary.
  • a higher broad band character of light results in fewer metameric problems. This is due to a white point being comprised of a combination of color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.) in a non-white emitter system. Total luminance is then related to intensities of the color primaries (e g., RGB, CMY, RGBC, RGBCMY, etc ).
  • a white emitter is included, increased luminance can be achieved separate from the color primaries. Additionally, colors such as vibrantly colored pastels are attained by using the color primaries to “color shift” a bright white to the pastel. Alternatively, a fine balance of the color primaries is required, and small changes in a ratio of the color primaries will produce an unwanted color shift. Thus, a system with at least one white emitter is more tolerant to minor variations of intensity of the color primaries.
  • the at least four primaries include RGBC, RGBW, or RG1G2B (i.e., a first green primary and a second green primary).
  • the at least four primaries include RGBY.
  • the multi-primary system of the present invention includes at least five primaries.
  • a first wavelength corresponding to a first primary is 460nm
  • a second wavelength corresponding to a second primary is 485nm
  • a third wavelength corresponding to a third primary is 510nm
  • a fourth wavelength corresponding to a fourth primary is 535nm
  • a fifth wavelength corresponding to a fifth primary is 640nm as shown in Table 21.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, and/or the fifth wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, and/or the fifth wavelength is within ⁇ 5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, and/or the fifth wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, the fourth primary, and/or the fifth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-1 OOnm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 85 illustrates a graph of the five primaries listed in Table 21 with respect to CIE 1931.
  • the at least four primaries encompass 87.55% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 85.
  • the at least five primaries encompass at least 87% of the total area covered between 400nm and 700 nm for CIE 1931.
  • the at least five primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least five primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least five primaries encompass at least 75% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least five primaries include at least one white emitter.
  • the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
  • using a D65 white emitter eliminates most of the problems with metamerism.
  • the at least five primaries include RGBCY, RGBCW, RG1G2BW (i.e., a first green primary and a second green primary), RGBW1W2 (i.e., a first white emitter and a second white emitter), or RG1G2BY (i.e., a first green primary and a second green primary).
  • the multi-primary system of the present invention includes at least six primaries.
  • a first wavelength corresponding to a first primary is 460nm
  • a second wavelength corresponding to a second primary is 490nm
  • a third wavelength corresponding to a third primary is 506nm
  • a fourth wavelength corresponding to a fourth primary is 520nm
  • a fifth wavelength corresponding to a fifth primary is 545nm
  • a sixth wavelength corresponding to a sixth primary is 640nm as shown in Table 22.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, and/or the sixth wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, and/or the sixth wavelength is within ⁇ 5% of the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, and/or the sixth wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, the fourth primary, the fifth primary, and/or the sixth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 86 illustrates a graph of the six primaries listed in Table 22 with respect to CIE 1931.
  • the at least six primaries encompass 91.11% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 86.
  • the at least six primaries encompass at least 90% of a total area covered between 400nm and 700 nm for CIE 1931.
  • the at least six primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least six primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least six primaries encompass at least 75% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least six primaries include at least one white emitter.
  • the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
  • the at least six primaries include RGBCMY, RGBCW1W2, RG1G2BW1W2 (i.e., a first green primary, a second green primary, a first white emitter, and a second white emitter), RGBW1W2W3 (i.e., a first white emitter, a second white emitter, and a third white emitter), or RGlG2BCY(i.e., a first green primary and a second green primary).
  • the multi-primary system of the present invention includes at least seven primaries.
  • a first wavelength corresponding to a first primary is 460nm
  • a second wavelength corresponding to a second primary is 480nm
  • a third wavelength corresponding to a third primary is 495nm
  • a fourth wavelength corresponding to a fourth primary is 508nm
  • a fifth wavelength corresponding to a fifth primary is 520nm
  • a sixth wavelength corresponding to a sixth primary is 540nm
  • a seventh wavelength corresponding to a seventh primary is 640nm as shown in Table 23.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, and/or the seventh wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, and/or the seventh wavelength is within ⁇ 5% of the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, and/or the seventh wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, and/or the seventh primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 87 illustrates a graph of the seven primaries listed in Table 23 with respect to CIE 1931.
  • the at least four primaries encompass 91.93% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 87.
  • the at least seven primaries encompass at least 90% of a total area covered between 400nm and 700 nm for CIE 1931.
  • the at least seven primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least seven primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least seven primaries encompass at least 75% of atotal area covered between 400nm and 700nm for CIE 1931.
  • the at least seven primaries include at least one white emitter.
  • the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
  • the multi-primary system of the present invention includes at least eight primaries.
  • a first wavelength corresponding to a first primary is 460nm
  • a second wavelength corresponding to a second primary is 480nm
  • a third wavelength corresponding to a third primary is 495nm
  • a fourth wavelength corresponding to a fourth primary is 500nm
  • a fifth wavelength corresponding to a fifth primary is 51 Inm
  • a sixth wavelength corresponding to a sixth primary is 521nm
  • a seventh wavelength corresponding to a seventh primary is 545nm
  • an eighth wavelength corresponding to an eighth primary is 640nm as shown in Table 24.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, and/or the eighth wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, and/or the eighth wavelength is within ⁇ 5% of the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, and/or the eighth wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, and/or the eighth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 88 illustrates a graph of the eight primaries listed in Table 24 with respect to CIE 1931.
  • the at least eight primaries encompass 92.55% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 88.
  • the at least eight primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least eight primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least eight primaries encompass at least 80% of a total area covered between 400nm and 700nm.
  • the at least eight primaries encompass at least 75% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least eight primaries include at least one white emitter.
  • the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
  • the multi-primary system of the present invention includes at least nine primaries.
  • the multi-primary system of the present invention includes at least ten primaries.
  • a first wavelength corresponding to a first primary is 440nm
  • a second wavelength corresponding to a second primary is 470nm
  • a third wavelength corresponding to a third primary is 485nm
  • a fourth wavelength corresponding to a fourth primary is 493nm
  • a fifth wavelength corresponding to a fifth primary is 502nm
  • a sixth wavelength corresponding to a sixth primary is 512nm
  • a seventh wavelength corresponding to a seventh primary is 520nm
  • an eighth wavelength corresponding to an eighth primary is 535nm
  • a ninth wavelength corresponding to a ninth primary is 550nm
  • a tenth wavelength corresponding to a tenth primary is 660nm as shown in Table 25.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, and/or the tenth wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, and/or the tenth wavelength is within ⁇ 5% of the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, and/or the tenth wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, the eighth primary, the ninth primary, and/or the tenth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 89 illustrates a graph of the ten primaries listed in Table 25 with respect to CIE 1931.
  • the at least ten primaries encompass 97.16% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 89.
  • the at least ten primaries encompass at least 95% of a total area covered between 400nm and 700 nm for CIE 1931.
  • the at least ten primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least ten primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least ten primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least ten primaries include at least one white emitter.
  • the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
  • the multi-primary system of the present invention includes at least eleven primaries. [00568] AT LEAST TWELVE PRIMARIES
  • the multi-primary system of the present invention includes at least twelve primaries.
  • a first wavelength corresponding to a first primary is 440nm
  • a second wavelength corresponding to a second primary is 470nm
  • a third wavelength corresponding to a third primary is 485nm
  • a fourth wavelength corresponding to a fourth primary is 493nm
  • a fifth wavelength corresponding to a fifth primary is 500nm
  • a sixth wavelength corresponding to a sixth primary is 505nm
  • a seventh wavelength corresponding to a seventh primary is 51 Inm
  • an eighth wavelength corresponding to an eighth primary is 517nm
  • a ninth wavelength corresponding to a ninth primary is 523nm
  • a tenth wavelength corresponding to a tenth primary is 535nm
  • an eleventh wavelength corresponding to an eleventh primary is 550nm
  • a twelfth wavelength corresponding to a twelfth primary is 670nm as shown in Table 26.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ⁇ 5% of the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, the eighth primary, the ninth primary, the tenth primary, and/or the twelfth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 90 illustrates a graph of the twelve primaries listed in Table 26 with respect to CIE 1931.
  • the at least twelve primaries encompass 97.91% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 90.
  • the at least twelve primaries encompass at least 95% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least twelve primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least twelve primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931.
  • a first wavelength corresponding to a first primary is 400nm
  • a second wavelength corresponding to a second primary is 468nm
  • a third wavelength corresponding to a third primary is 484nm
  • a fourth wavelength corresponding to a fourth primary is 493nm
  • a fifth wavelength corresponding to a fifth primary is 500nm
  • a sixth wavelength corresponding to a sixth primary is 506nm
  • a seventh wavelength corresponding to a seventh primary is 512nm
  • an eighth wavelength corresponding to an eighth primary is 518nm
  • a ninth wavelength corresponding to a ninth primary is 524nm
  • a tenth wavelength corresponding to a tenth primary is 535nm
  • an eleventh wavelength corresponding to an eleventh primary is 556nm
  • a twelfth wavelength corresponding to a twelfth primary is 700nm as shown in Table 27.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ⁇ 5% of the value listed in the table below.
  • the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ⁇ 2% of the value listed in the table below.
  • the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, the eighth primary, the ninth primary, the tenth primary, and/or the twelfth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
  • FIG. 91 illustrates a graph of the twelve primaries listed in Table 27 with respect to CIE 1931.
  • the at least twelve primaries encompass 99.14% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 91.
  • the at least twelve primaries encompass at least 97.5% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least twelve primaries encompass at least 95% of a total area covered between 400nm and 700nm for CIE 1931.
  • the at least twelve primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931.
  • a twelve primary system is backwards compatible with 6P-C.
  • the twelve primary system includes a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary as shown in Table 28.
  • the magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the green-cyan primary, the green primary, the yellow-green primary, the yellow primary, the red-yellow primary, the red primary, and/or the magenta-red primary is approximately (e.g., within ⁇ 10%) the value listed in the table below.
  • the magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the green-cyan primary, the green primary, the yellow-green primary, the yellow primary, the red-yellow primary, the red primary, and/or the magenta-red primary is within ⁇ 5% of the value listed in the table below.
  • magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the green-cyan primary, the green primary, the yellow-green primary, the yellow primary, the red-yellow primary, the red primary, and/or the magenta-red primary is within ⁇ 2% of the value listed in the table below.
  • the magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the greencyan primary, the green primary, the yellow-green primary, the yellow primary, the red- yellow primary, the red primary, and/or the magenta-red primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width).
  • the bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences. [00576] TABLE 28
  • FIG. 92 illustrates a graph of the twelve primaries listed in Table 28 with respect to CIE 1931.
  • the at least twelve primaries include at least one white emitter.
  • the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
  • the multi-primary system has a larger volume than that described in ITU-R BT.2020, which is detailed in ITU-R BT.2020 (2015) and ITU-R BT.2100 (2016).
  • ITU-R BT.2020 covers 75.8% of the CIE 1931 color space, which is described in CIE (1932). Commission intemationale de 1'Eclairage proceedings, 1931. Cambridge: Cambridge University Press and Smith, Thomas; Guild, John (1931-32). "The C.I.E. colorimetric standards and their use". Transactions of the Optical Society. 33 (3): 73-
  • ITU-R BT.2020 has a red primary at (0.708, 0.292), a green primary at (0.17, 0.797), and a blue primary at (0.131, 0.046).
  • FIG. 102A illustrates a front view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
  • FIG. 102B illustrates a normal orthogonal view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
  • FIG. 102C illustrates a top view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
  • the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid.
  • the multi-primary system has a larger volume than that described by DCI-P3 (“P3”), which is detailed in SMPTE EG 432-1 (2010) and SMPTE RP 431-2 (2011), each of which is incorporated herein by reference in its entirety.
  • DCI-P3 covers 45.5% of the CIE 1931 color space.
  • DCI-P3 with a D65 white point has a red primary at (0.680, 0.320), a green primary at (0.265, 0.690), and a blue primary at (0.150, 0.060).
  • FIG. 103A illustrates a front view of a three-dimensional plot of DCI-P3 in XYZ space.
  • FIG. 103B illustrates a normal orthogonal view of a three-dimensional plot of DCI-P3 in XYZ space.
  • FIG. 103C illustrates atop view of a three-dimensional plot of DCI-P3 in XYZ space.
  • the multi-primary system has the primary values listed in Table 3 (“6P-C”).
  • FIG. 104A illustrates a front view of 6P-C in XYZ space.
  • FIG. 104B illustrates a normal orthogonal view of 6P-C in XYZ space.
  • FIG. 104C illustrates atop view of 6P-C in XYZ space.
  • FIG. 105A illustrates a front view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
  • FIG. 105B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
  • FIG. 105C illustrates atop view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
  • the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid.
  • the volume of 6P-C is a rhombic cuboid with extensions beyond ITU-R
  • the extension on the far Y side of the rhombic cuboid is a triangular prism.
  • the extension toward the near X side is hexagonal prism.
  • FIG. 106A illustrates a front view of DCI-P3 (red) and 6P-C (green) in XYZ space.
  • FIG. 106B illustrates a normal orthogonal view of DCI-P3 (red) and 6P-C (green) in XYZ space.
  • FIG. 106C illustrates a top view of DCI-P3 (red) and 6P-C (green) in XYZ space.
  • the multi-primary system has four primaries with a red primary at about (0.6433, 0.3192), a green primary at about (0.3244, 0.6300), a blue primary at about (0.1513, 0.0748), and a cyan primary at about (0.0729, 0.3953) (“4P”).
  • FIG. 107A illustrates a front view of 4P in XYZ space.
  • FIG. 107B illustrates a normal orthogonal view of 4P in XYZ space.
  • FIG. 107C illustrates atop view of 4P in XYZ space.
  • FIG. 108A illustrates a front view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space.
  • FIG. 108B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space.
  • FIG. 108C illustrates atop view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space.
  • the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid.
  • the volume of 4P is a rhombic cuboid with extensions beyond ITU-R BT.2020.
  • the extension on the far Y side of the rhombic cuboid is a triangular prism.
  • the extension toward the near X side is hexagonal prism.
  • FIG. 109A illustrates a front view of DCI-P3 (red) and 4P (blue) in XYZ space.
  • FIG. 109B illustrates a normal orthogonal view of DCI-P3 (red) and 4P (blue) in XYZ space.
  • FIG. 109C illustrates atop view of DCI-P3 (red) and 4P (blue) in XYZ space.
  • the multi-primary system has four primaries with a red primary at about (0.6822, 0.3137), a green primary at about (0.2680, 0.7070), a blue primary at about (0.1367, 0.0543), and a cyan primary at about (0.0731, 0.3244) (“4P-N”).
  • FIG. 110A illustrates a front view of 4P-N in XYZ space.
  • FIG. HOB illustrates a normal orthogonal view of 4P-N in XYZ space.
  • FIG. 110C illustrates a top view of 4P-N in XYZ space.
  • FIG. 111 A illustrates a front view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
  • FIG. 11 IB illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
  • FIG. 111 C illustrates a top view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
  • the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid.
  • the volume of 4P-N is a rhombic cuboid with extensions beyond ITU-R BT.2020.
  • the extension on the far Y side of the rhombic cuboid is a triangular prism.
  • the extension toward the near X side is hexagonal prism.
  • FIG. 112A illustrates a front view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
  • FIG. 112B illustrates a normal orthogonal view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
  • FIG. 112C illustrates a top view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
  • the system is operable to display an image on a viewing device (e.g., display).
  • the image includes colors outside of an ITU-R BT.2020 color gamut, a P3 color gamut, and/or an ITU-R BT.709 color gamut.
  • the ITU-R BT.2020 color gamut is described in ITU-R BT.2020-2 (2015), which is incorporated herein by reference in its entirety.
  • the P3 color gamut is described in SMPTE-EG-0432-1 (2010), which is incorporated herein by reference in its entirety.
  • the ITU-R BT.709 color gamut is described in ITU-R BT.709-6 (2015), which is incorporated herein by reference in its entirety.
  • the image preferably includes colors outside of the ITU-R BT.2020 color gamut.
  • the ITU-R BT.2020 color gamut covers 75.8% of the CIE 1931 color space.
  • the ITU-R BT.2020 color gamut is defined as a triangle having a first vertex at (0.170, 0.797), a second vertex at (0.708, 0.292), and a third vertex at (0.131, 0.046).
  • the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.131, 0.046), and a third vertex at about (0.0454, 0.295) within a CIE 1931 color space.
  • the third vertex corresponds to a wavelength of about 490 nm.
  • this provides an expanded color gamut in the cyan region.
  • the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.708, 0.292), and a third vertex at about (0.266, 0.724) within a CIE 1931 color space.
  • this provides an expanded color gamut in the yellow region.
  • the third vertex corresponds to a wavelength of about 545 nm.
  • the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.708, 0.292), a second vertex at (0.131, 0.046), and a third vertex at about (0.718, 0.281) within a CIE 1931 color space.
  • the third vertex corresponds to a wavelength of about 640 nm.
  • this provides an expanded color gamut in the magenta region.
  • the colors outside of the ITU-R BT.2020 color gamut, the P3 color gamut, and/or the ITU-R BT.709 color gamut are obtained from a camera (e.g., video, still image) operable to obtain the colors. Additionally or alternatively, the image is modified from an original image to include the colors outside of the ITU-R BT.2020 color gamut, the P3 color gamut, and/or the ITU-R BT.709 color gamut. Colorists routinely push colors to places they were not in an original image. If colorists are given an even larger gamut, they can push color to a greater extent, even far beyond what the “real” colors actually were.
  • Including the colors outside of the ITU-R BT.2020 color gamut, the P3 color gamut, and/or the ITU-R BT.709 color gamut provides more color fidelity that the eye can see while still within the CIE diagram boundaries. For example, astronauts often see colors in space that are unable to be reproduced using current display technology. Additionally, when processing images for virtual production, the wider color gamut provides not only an opportunity to provide additional colors (e.g., cyan), but also produces more accurate flesh tones on the complementary side as well due to the wider color gamut.
  • additional colors e.g., cyan
  • the image is recognizable when separated into an RGB image and a CMY image in an RGBCMY system.
  • the RGB image and the CMY image preferably have no artifacts.
  • the image is produced from a conversion to XYZ coordinates from an original image using at least four triads. Each of the at least four triads includes three of the at least four primaries.
  • the XYZ coordinates are multiplied by at least four XYZ-to-triad matrices to determine one or more of the at least four triads in which the XYZ coordinates are located.
  • a sum of primary components of the one or more of the at least four triads is determined on a per-component basis and the sum is divided by a number of the one or more of the at least four triads.
  • the at least four primaries include at least four color primaries and a virtual primary (e.g., white point).
  • each of the at least four triads includes two adjacent primaries of the at least four color primaries and the virtual primary.
  • the image is preferably produced using an algorithm that minimizes and/or avoids non-matches, non-smoothness, and/or spurious matches.
  • non-matches result when a first combination of primaries and a second combination of primaries have equal XYZ coordinates, and may appear slightly different to viewers although they appear the same to the standard observer.
  • non-smoothness result when a color scale is perceived as abruptly changing by a viewer despite being a continuous curve through the gamut due to combinations of primaries to create the color scale.
  • spurious matches result from conditions (e.g., ambient lighting conditions, filters) that cause a first color and a second color having different XYZ coordinates to appear the same.
  • the image is modified from an original image to include a digital watermark.
  • the digital watermark is outside of the ITU-R BT.2020 color gamut.
  • the digital watermark is compressed, collapsed, and/or mapped to an edge of the smaller color gamut such that it is not visible and/or not detectable when displayed on a viewing device with a smaller color gamut than ITU-R BT.2020.
  • the digital watermark is not visible and/or not detectable when displayed on a viewing device with an ITU-R BT.2020 color gamut.
  • the digital watermark is a watermark image (e.g., logo), alphanumeric text (e.g., unique identification code), and/or a modification of pixels.
  • the digital watermark is invisible to the naked eye.
  • the digital watermark is perceptible when decoded by an algorithm.
  • the algorithm uses an encryption key to decode the digital watermark.
  • the digital watermark is visible in a non-obtrusive manner (e.g., at the bottom right of the screen).
  • the digital watermark is preferably detectable after size compression, scaling, cropping, and/or screenshots.
  • the digital watermark is an imperceptible change in sound and/or video.
  • the image is operable to be displayed on a viewing device.
  • the viewing device is a smartphone, a tablet, a laptop screen, a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a miniLED display, a microLED display, a liquid crystal display (LCD), a quantum dot display, a quantum nano emitting diode (QNED) device, a personal gaming device, a virtual reality (VR) device and/or an augmented reality (AR) device, an LED wall, a wearable display (e.g., VR/AR headset), and/or at least one projector.
  • the at least one projector includes more than one aligned and/or synchronized projector (e.g., manually, automatically via software).
  • the viewing device is foldable and/or flexible.
  • the viewing device is operable to display colors outside of an ITU-R BT.2020 color gamut, a P3 color gamut, and/or an ITU-R BT.709 color gamut.
  • the viewing device preferably is operable to display colors outside of the ITU-R BT.2020 color gamut.
  • the ITU-R BT.2020 gamut covers 75.8% of the CIE 1931 color space.
  • the viewing device is preferably operable to display at least 76% of the CIE 1931 color space. In a more preferred embodiment, the viewing device is operable to display at least 80% of the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 85% of the CIE 1931 color space.
  • the viewing device is operable to display at least 90% of the CIE 1931 color space. In yet another embodiment, the viewing device is operable to display at least 95% of the CIE 1931 color space. In still another embodiment, the viewing device is operable to display at least 97% of the CIE 1931 color space.
  • the viewing device is constructed and configured to display at least four primaries. Increasing the number of primaries in the viewing device to at least four primaries increases color accuracy of the viewing device relative to conventional RGB displays. Additionally, this allows for accurate display of colors that are traditionally difficult to reproduce on conventional RGB displays.
  • the viewing devices includes at least one component to provide the at least four primaries (e.g., at least one color wheel, a plurality of LEDs, etc.).
  • the at least four primaries include red, green, blue, and cyan.
  • the at least four primaries include red, green, blue, cyan, and yellow.
  • the at least four primaries include red, green, blue, cyan, yellow, and magenta.
  • the at least four primaries include red, a first green, a second green, and blue.
  • the at least four primaries includes at least one white primary.
  • teal is a color that is difficult to reproduce using conventional RGB displays. Adding a cyan primary increases the color accuracy of teal and the sensitivity of the display to colors in the region between green and blue on an RGBC display when compared to a conventional RGB display.
  • the viewing device is preferably operable to display flesh tones with increased color accuracy.
  • flesh tones are important for entertainment, medical, and/or scientific purposes.
  • the ability to identify and detect flesh tones is important for diagnostic imaging related to the skin and other organs (e.g., brain, lungs, etc.).
  • a person’s skin tone can vary slightly due to a number of factors, but the two main influences are health and emotion.
  • the human visual system has been optimized to detect small changes in skin reflectivity due to blood flow and oxygenation.
  • the M (green) and L (red) cones are operable to detect these changes. There is a long-standing, unmet need for an extended gamut providing more accurate flesh tones.
  • the viewing device is an RGBCMY viewing device.
  • the viewing device preferably includes a yellow primary.
  • the viewing device has a red primary with a longer wavelength than 615 nm. Flesh tones often appear yellowish or reddish after color correction. Additionally, skin often appears shiny after color correction. Increasing a cyan component and/or a magenta component improves the color accuracy of the flesh tones and reduces the shiny appearance of skin.
  • the viewing device is preferably operable to display natural surfaces (e.g., natural reflective surfaces) with increased color accuracy.
  • natural surfaces e.g., natural reflective surfaces
  • the multi-primary systems of the present invention provide an extended gamut in the cyan region.
  • the extension into the cyan area as well as into the shorter wavelength green area expands the reproduction of foliage, water, ice, and other natural items.
  • the viewing device includes pixels in a hexagonal shape.
  • the viewing device includes six primaries and each pixel in the six-primary color system is a hexagonal shape.
  • Each hexagonal pixel is divided into six equilateral triangles and each of the primaries in the six-primary color system is displayed by one of the six equilateral triangles as described in U.S. Patent Application No. 12/005,931, filed July 3, 2008, which is incorporated herein by reference in its entirety.
  • each pixel in the viewing device is comprised of subpixels of the same size and area arranged in at least one row and/or at least one column.
  • each pixel is divided into six subpixels of the same size and area arranged in two rows of three columns for a six-primary color system.
  • each pixel is divided into six subpixels of the same size and area arranged in three rows of two columns.
  • each pixel is divided into six subpixels of the same size and area arranged in one row.
  • each pixel is divided into six subpixels of the same size and area arranged in one column.
  • the luminance and intensity of each subpixel is dependent on the luminance and intensity of the adjacent subpixels in order to minimize the distinct visibility of individual subpixel and pixel structures.
  • complementary primary color subpixels are adjacent to each other to eliminate visual artifacts.
  • each pixel is divided into subpixels of different shapes, sizes, and/or areas.
  • the size and number of subpixels for each primary color minimize blue and cyan spatial resolution without affecting the overall resolution of the viewing device as described in U.S. Patent Application No. 12/909,742, filed October 21, 2010, now U.S. Patent No. 8,451,405, which is incorporated herein by reference in its entirety.
  • each pixel unit is divided into two subpixel units wherein one of the two subpixels is a first set of primaries and the other subpixel is a second set of primaries.
  • the second set of primaries is complementary to the first set of primaries.
  • one of the two subpixels is an RGB color and the other subpixel is the complementary CMY color of the first subpixel as described in U.S. Patent Application No. 12/229,845, filed March 5, 2009, which is incorporated herein by reference in its entirety.
  • each pixel includes at least one white subpixel to eliminate visual artifacts.
  • the at least one white subpixel includes a D65 white subpixel, a D60 white subpixel, a D45 white subpixel, a D27 white subpixel, and/or a D25 white subpixel.
  • using a D65 white subpixel eliminates most of the problems with metamerism.
  • the at least one white subpixel is a single white subpixel that matches the white point (e.g., a D65 white subpixel for a D65 white point).
  • the at least one white subpixel is at least two white subpixels.
  • the at least two white subpixels are preferably separated such that a linear combination of the at least two white subpixels covers a desired white Kelvin range.
  • the at least two white subpixels include a D65 white subpixel and a D27 white subpixel.
  • the at least two white subpixels include a D65 white subpixel and a D25 white subpixel.
  • the at least two white subpixels includes three white subpixels.
  • the three white subpixels include a D65 white subpixel, a D45 white subpixel, and a D27 white subpixel.
  • the three white subpixels include a D65 white subpixel, a mid-Kelvin white subpixel (e.g., D45), and a D27 white subpixel.
  • the mid-Kelvin white subpixel includes a green bias.
  • the green bias compensates for the slight magenta shift (e.g., when going from D25 to D65 with the straight line between the two points below the blackbody locus).
  • Colors near the white locus and beyond are then a combination of the at least two white subpixels (e.g., two white subpixels, three white subpixels).
  • a majority of colors will have a white component that is broad band. Therefore, the resultant spectra of a mixture of color primaries and white primaries will also be broad band with an extent dependent on an amount of the at least one white primary.
  • a higher broad band character of light results in fewer metameric problems. This is due to a white point being comprised of a combination of color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.) in anon-white subpixel system. Total luminance is then related to intensities of the color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.).
  • the white point of the multi-primary color system changes depending on the viewing device or the display mode.
  • the addition of white subpixels widens the bandwidth of the filter for each non-white primary.
  • each pixel is formed of fewer than the at least four primaries (e.g., three of four primaries, four of five primaries, five of six primaries, etc.). In one embodiment, each pixel is composed of fewer than six primary colors from the 6P gamut.
  • the viewing device is composed of alternating and repeating subpixel patterns. In another embodiment, the viewing device is composed of nonrepeating subpixel patterns.
  • the subpixel colors in a pixel and in adjacent pixels are arranged to minimize the spatial distance between colors that have maximal color distance from each other as described in U.S. Patent Application No. 10/543,511, filed January 13, 2003, now U.S. Patent No. 8,228,275, which is incorporated herein by reference in its entirety.
  • each pixel is one single primary color from the multi-primary system (e.g., 6P gamut).
  • patterns of pixels are repeated across the viewing device to minimize visibility of individual pixel structures as described in U.S.
  • the viewing device includes at least one perovskite.
  • the at least one perovskite is a lead halide perovskite.
  • the at least one perovskite is used as a quantum dot nanocrystal.
  • the at least one perovskite is a perovskite polymer bead. When light shines through the perovskite polymer bead, the color changes depending on the composition of the perovskite polymer bead (e.g., green, red, etc.).
  • the at least one perovskite is incorporated into a perovskite LED.
  • perovskite LEDs are described in Lin, K., et al. (2016). Perovskite light-emitting diodes with external quantum efficiency exceeding 20 per cent. Nature, 562(7726), 245-248, which is incorporated herein by reference in its entirety.
  • the at least one perovskite is 3D printed. See, e.g., Zhou, Nanjia, Yehonadav Bekenstein, CarissaN. Eisler, Dandan Zhang, Adam M. Schwartzberg, Peidong Yang, A. Paul Alivisatos, and Jennifer A. Lewis. 2019. “Perovskite Nanowire-Block Copolymer Composites With Digitally Programmable Polarization Anisotropy.” Science Advances, which is incorporated herein by reference in its entirety.
  • the viewing device is a direct emissive assembled display.
  • the design for a direct emissive assembled display includes a matrix of color emitters grouped as a multi -primary color system (e.g., 6P system). Individual channel inputs drive each Quantum Dot (QD) element illuminator and/or micro LED element.
  • QD Quantum Dot
  • the quantum dots modulate light according to image data as described in U.S. Patent Application No. 15/905,085, filed February 26, 2018, now U.S. Patent No.
  • System 1, System 2, or System 3 can be used as previously described. If four color components are used, two of the channels are set to “0”. If five color components are used, one of the channels is set to “0”.
  • 93 shows one embodiment of transportation of twelve individual color channels using the example in Table 28 with a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary on a first link (Link A) and a second link (Link B).
  • FIG. 94A shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a first link (Link A).
  • FIG. 94B shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a second link (Link B).
  • FIG. 95 A shows one embodiment of a 4:2:2 Constant Luminance Encode for a first link (Link A).
  • FIG. 95B shows one embodiment of a 4:2:2 Constant Luminance Encode for a second link (Link B).
  • FIG. 96A shows one embodiment of a 4:4:4 Encode for a first link (Link A).
  • FIG. 96B shows one embodiment of a 4:4:4 Encode for a second link (Link B).
  • FIG. 97A shows one embodiment of component mapping into SMPTE 2081-1 for a first link (Link A).
  • FIG. 97B shows one embodiment of component mapping into SMPTE 2081-1 for a second link (Link B).
  • FIG. 98A shows one embodiment of R,G,B,C,M,Y,GC,MR,BM,YG,RY,CB mapping into SMPTE 2081-1 for a first link (Link A).
  • FIG. 98B shows one embodiment of R,G,B,C,M,Y,GC,MR,BM,YG,RY,CB mapping into SMPTE 2081-1 for a second link (Link B).
  • FIG. 99A shows one embodiment of a 4:2:2 Non-Constant Luminance Decode for a first link (Link A).
  • FIG. 99B shows one embodiment of a 4:2:2 Non-Constant Luminance
  • FIG. 100A shows one embodiment of a 4:2:2 Constant Luminance Decode for a first link (Link A).
  • FIG. 100B shows one embodiment of a 4:2:2 Constant Luminance Decode for a second link (Link B).
  • FIG. 101A shows one embodiment of a 4:4:4 Decode for a first link (Link A).
  • FIG. 101B shows one embodiment of a 4:4:4 Decode for a second link (Link B).
  • the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter, and a display system, wherein the display system and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the display system.
  • the image data converter includes a digital interface.
  • the digital interface is operable to encode and decode the set of image data.
  • the system further includes at least one transfer function (TF) for processing the set of image data.
  • the system further includes a set of Session Description Protocol (SDP) parameters.
  • SDP Session Description Protocol
  • the set of SDP parameters is modifiable.
  • the display system includes a Liquid Crystal Display (LCD) projector, wherein the LCD projector is operable to transmit light through a plurality of LCD units using at least one prism and/or at least one reciprocal mirror.
  • the display system includes a Digital Micromirror Device (DMD) projector.
  • the DMD projector includes at least one DMD chip, wherein the at least one DMD chip is synchronized with at least one light source.
  • the display system is operable to use a combination of primary color display elements and/or a combination of primary color light sources to display a different primary color.
  • the image data converter includes an alignment signal to synchronize and align (e.g., mechanically align) at least two projectors.
  • the display system includes an apparatus to combine the output display of the at least two projectors, thereby creating a combined output display.
  • the at least four primary color values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary.
  • the set of image data further includes a bit level, a first set of color channel data, and a second set of color channel data.
  • the image data converter is operable to create a combined set of color channel data from the first set of color channel data and the second set of color channel data for display on the display system.
  • the combined set of color channel data has a combined bit level equal to the bit level of the set of image data.
  • the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter, and a display system, wherein the display system includes at least one light source, wherein the display system and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the display system.
  • the image data converter includes a digital interface.
  • the digital interface is operable to encode and decode the set of image data.
  • the system further includes at least one transfer function (TF) for processing the set of image data.
  • the system further includes a set of TF
  • Session Description Protocol SDP
  • the set of SDP parameters is modifiable.
  • the at least one light source includes at least one Light-Emitting Diode (LED).
  • the at least one light source includes a Xenon lamp.
  • the at least one light source includes a blue laser system.
  • the at least four primary color values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary.
  • the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter, and a display system, wherein the display system includes at least one display screen, wherein the at least one display screen comprises a plurality of pixels, wherein each of the plurality of pixels is divided into a plurality of subpixels, wherein the display system and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the display system.
  • the image data converter includes a digital interface.
  • the digital interface is operable to encode and decode the set of image data.
  • the system further includes at least one transfer function (TF) for processing the set of image data.
  • the system further includes a set of Session Description Protocol (SDP) parameters.
  • the set of SDP parameters is modifiable.
  • the at least one display screen includes a Liquid Crystal Display (LCD) display screen, a Light-Emitting Diode (LED) display screen, and/or a Quantum Dot (QD) display screen.
  • the at least one display screen includes at least one white subpixel, at least two green subpixels, at least one cyan subpixel, at least one magenta subpixel, and/or at least one yellow subpixels.
  • the at least four primary color values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary.
  • the at least one display screen includes at least one perovskite.
  • the at least one display screen includes at least two display screens, wherein the display system includes a mirror apparatus (e.g., half-silvered mirror apparatus), and wherein the mirror apparatus is operable to combine the at least two display screens on a view screen.
  • the display system includes an expanded filter arrangement, and wherein the set of image data includes Low-Voltage Differential Signaling (LVDS) data.
  • LVDS Low-Voltage Differential Signaling
  • the present invention provides a system for displaying a primary color system including a set of image data, an image data converter, and at least one display device, wherein the set of image data includes primary color data for at least four primary color values, wherein the at least one display device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device.
  • the system further includes at least one transfer function (TF) for processing the set of image data.
  • the system further includes a set of Session Description Protocol (SDP) parameters.
  • the set of image data includes a first set of color channel data and a second set of color channel data.
  • the image data converter further includes a first link component and a second link component.
  • the first link component is operable to transport the first set of color channel data to the at least one display device and the second link component is operable to transport the second set of color channel data to the at least one display device in parallel with the first link component.
  • the at least one display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the at least one display device is based on the set of image data.
  • the at least four primary color values include at least one white emitter.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 493nm, a third primary at approximately 540nm, and a fourth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 485nm, a third primary at approximately 510nm, a fourth primary at approximately 535nm, and a fifth primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 490nm, a third primary at approximately 506nm, a fourth primary at approximately 520nm, a fifth primary at approximately 545nm, and a sixth primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 508nm, a fifth primary at approximately 520nm, a sixth primary at approximately 540nm, and a seventh primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 500nm, a fifth primary at approximately 511nm, a sixth primary at approximately 521nm, a seventh primary at approximately 545nm, and an eighth primary at approximately 640nm.
  • the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 502nm, a sixth primary at approximately 512nm, a seventh primary at approximately 520nm, an eighth primary at approximately 535nm, a ninth primary at approximately 550nm, and a tenth primary at approximately 660nm.
  • the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 505nm, a seventh primary at approximately 511nm, an eighth primary at approximately 517nm, a ninth primary at approximately 523nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 550nm, and a twelfth primary at approximately 670nm.
  • the at least four primaries include a first primary at approximately 400nm, a second primary at approximately 468nm, a third primary at approximately 484nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 506nm, a seventh primary at approximately 512nm, an eighth primary at approximately 518nm, a ninth primary at approximately 524nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 556nm, and a twelfth primary at approximately 700nm.
  • the at least four primaries include a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary.
  • the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data.
  • the set of SDP parameters is modifiable.
  • the first set of color channel data is converted by the first link component and the second set of color channel data is converted by the second link component, and wherein the first set of color channel data and the second set of color channel data are combined to form the set of image data for display on the single display device.
  • the system further includes a standardized transport format, wherein the first link component includes a first standardized transport format link and wherein the second link component includes a second standardized transport format link, wherein the standardized transport format is operable to receive the first set of image data and the second set of image data using the first standardized transport format link and the second standardized transport format link, and wherein the first standardized transport format link and the second standardized transport format link are operable to combine the first set of image data and the second set of image data into a combined set of image data.
  • the present invention provides a system for displaying a primary color system including a set of image data, an image data converter, and at least one display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least one display device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device.
  • the image data converter includes a digital interface.
  • the digital interface is operable to encode and decode the set of image data.
  • the system further includes at least one transfer function (TF) for processing the set of image data.
  • the system further includes a set of Session Description Protocol (SDP) parameters.
  • SDP Session Description Protocol
  • the set of SDP parameters is modifiable.
  • the set of image data includes a first set of color channel data and a second set of color channel data.
  • the image data converter further includes a first link component and a second link component.
  • the first link component is operable to transport the first set of color channel data to the at least one display device and the second link component is operable to transport the second set of color channel data to the at least one display device in parallel with the first link component.
  • the set of SDP parameters is modified based on the conversion.
  • the present invention provides a system for displaying a primary color system including a set of image data, an image data converter, and at least one display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one display device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device.
  • the image data converter includes a digital interface.
  • the digital interface is operable to encode and decode the set of image data.
  • the system further includes at least one transfer function (TF) for processing the set of image data.
  • TF transfer function
  • the system further includes a set of Session Description Protocol (SDP) parameters.
  • the set of image data includes a first set of color channel data and a second set of color channel data.
  • the image data converter further includes a first link component and a second link component.
  • the first link component is operable to transport the first set of color channel data to the at least one display device and the second link component is operable to transport the second set of color channel data to the at least one display device in parallel with the first link component.
  • the at least one white emitter includes a white emitter matching a white point of the primary color system. In one embodiment, the at least one white emitter includes at least three white emitters.
  • the at least three white emitters each have a different color temperature.
  • the at least one white emitter includes a midKelvin white emitter.
  • the mid-Kelvin white emitter is modified to include a green bias.
  • the present invention provides a system for displaying a primary color system including a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, an image data converter, wherein the image data converter includes a digital interface, and wherein the digital interface is operable to encode and decode the set of values in Yxy color space, and at least one display device, wherein the at least one display device and the image data converter are in network communication, wherein the encode and the decode include transportation of the set of image data as Yxy data, and wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x and y, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device.
  • the at least one display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the at least one display device is based on the set of image data.
  • the image data converter is operable to convert the set of primary color signals to the set of values in Yxy color space. In one embodiment, the image data converter is operable to convert the set of values in Yxy color space to a plurality of color gamuts. In one embodiment, the image data converter is operable to fully sample the Yxy data related to the luminance Y and subsample the Yxy data related to the two colorimetric coordinates x and y.
  • the Yxy data related to the luminance Y and the two colorimetric coordinates x and y are fully sampled.
  • the set of image data is integrated into a standardized transportation format.
  • the set of values in Yxy color space includes a reference to at least one white point.
  • the Yxy data includes floating points.
  • the encode includes converting the set of primary color signals to XYZ data and then converting the XYZ data to create the set of values in Yxy color space.
  • the decode includes converting the Yxy data to XYZ data and then converting the XYZ data to a format operable to display on the at least one display device.
  • the set of image data is transported linearly without a non-linear function applied to the luminance Y.
  • the present invention provides a system for displaying a primary color system including a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, an image data converter, wherein the image data converter includes a digital interface, and wherein the digital interface is operable to encode and decode the set of values in Yxy color space, at least one non-linear function for processing the set of values in Yxy color space, and at least one display device, wherein the at least one display device and the image data converter are in network communication, wherein the at least one non-linear function is not applied to the colorimetric coordinates x and y, and wherein the at least one non-linear function is applied to the luminance Y, thereby creating a luma Y', wherein the encode and the decode include
  • the at least one non-linear function includes at least one of a gamma function, a log function, a perceptual quantizer (PQ) function, an opto-electronic transfer function (OETF), an opto-optical transfer function (OOTF), and/or an electro-optical transfer function (EOTF).
  • the image data converter applies one or more of the at least one non-linear function to encode the set of values in Yxy color space.
  • the image data converter applies one or more of the at least one non-linear function to decode the set of values in Yxy color space.
  • the image data converter includes a look-up table.
  • the present invention provides a system for displaying a primary color system including a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, an image data converter, wherein the image data converter includes a digital interface, and wherein the digital interface is operable to encode and decode the set of values in Yxy color space, a set of Session Description Protocol (SDP) parameters, wherein the set of SDP parameters is modifiable, and at least one display device, wherein the set of image data further includes pixel mapping data, wherein the at least one display device and the image data converter are in network communication, wherein the encode and the decode include transportation of the set of image data as Yxy data, and wherein the Yxy data is related to the two colorimetric coordinates and the luminance
  • SDP Session Description Protocol
  • the present invention provides a method for displaying a primary color system including providing a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, encoding the set of image data in Yxy color space using a digital interface of an image data converter, wherein the image data converter is in network communication with at least one display device, decoding the set of image data in Yxy color space using the digital interface of the image data converter, and the image data converter converting the set of image data for display on the at least one display device, wherein the encoding and the decoding include transportation of the set of image data as Yxy data, wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x andy.
  • the present invention provides a system for displaying a digital representation of an image including the image and a viewing device, wherein the image includes colors outside of an ITU-R BT.2020 color gamut, wherein the viewing device is operable to display the digital representation of the image and at least 80% of a total area covered between about 400nm and about 700nm by an International Commission on Illumination (CIE) 1931 color space, and wherein the viewing device is operable to display the colors outside of the ITU-R BT.2020 color gamut.
  • CIE International Commission on Illumination
  • the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.131, 0.046), and a third vertex at about (0.0454, 0.295) within the CIE 1931 color space.
  • the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.708, 0.292), and a third vertex at about (0.266, 0.724) within the CIE 1931 color space.
  • the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at about (0.708, 0.292), a second vertex at (0.131, 0.046), and a third vertex at about (0.719, 0.281) within the CIE 1931 color space.
  • the viewing device is selected from the group consisting of a smartphone, a tablet, a laptop screen, a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a miniLED display, a microLED display, a liquid crystal display (LCD), a quantum dot display, a quantum nano emitting diode (QNED) device, a personal gaming device, a virtual reality (VR) device and/or an augmented reality (AR) device, an LED wall, a wearable display (e.g., VR/AR headset), and at least one projector.
  • a smartphone a tablet, a laptop screen
  • a light emitting diode (LED) display an organic light emitting diode (OLED) display
  • miniLED organic light emitting diode
  • microLED miniLED display
  • LCD liquid crystal display
  • QNED quantum dot display
  • QNED quantum nano emitting diode
  • a personal gaming device e.g., a virtual reality (VR) device
  • the viewing device is operable to display at least 85% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 90% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 95% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 97% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space.
  • the viewing device is operable to display at least four primaries, and wherein the at least four primaries include red, green, blue, and cyan. In one embodiment, the viewing device is operable to display at least five primaries, and wherein the at least five primaries include red, green, blue, cyan, and yellow. In one embodiment, the viewing device is operable to display at least six primaries, and wherein the at least six primaries include red, green, blue, cyan, yellow, and magenta. In one embodiment, the viewing device is operable to display at least one white primary.
  • the system further includes a set of Session Description Protocol (SDP) parameters, wherein the SDP parameters include color channel data, image data, framerate data, a sampling standard, a flag indicator, an active picture size code, a timestamp, a clock frequency, a frame count, a scrambling indicator, and/or a video format indicator.
  • SDP Session Description Protocol
  • the image is modified from an original image to include the colors outside of the ITU-R BT.2020 color gamut.
  • the present invention provides a system for displaying a digital representation of an image including the image, a set of image data corresponding to the image, and a viewing device, wherein the viewing device is constructed and configured to provide a cyan primary, wherein the image includes colors inside of a first color gamut, wherein the colors inside of the first color gamut are outside of an ITU-R BT.2020 color gamut, wherein the viewing device is operable to display the digital representation of the image and at least 80% of a total area covered between about 400nm and about 700nm by an
  • the viewing device is operable to display the colors outside of the ITU-R BT.2020 color gamut.
  • the set of image data occupies a larger volume in the CIE 1931 color space than the ITU-R BT.2020 color gamut.
  • the set of image data is compressed and/or truncated when the set of image data is mapped to a second color gamut, wherein the second color gamut is not equivalent to the first color gamut.
  • the present invention provides a method for displaying a digital representation of an image including providing a viewing device, wherein the viewing device includes at least one component to provide a cyan primary, providing the image and a set of image data corresponding to the image to the viewing device, wherein the image includes colors outside of an ITU-R BT.2020 color gamut, and displaying the digital representation of the image on the viewing device, wherein the displaying the digital representation of the image on the viewing device includes displaying the colors outside of the ITU-R BT.2020 color gamut, wherein the viewing device is operable to display at least 80% of a total area covered between about 400nm and about 700nm by an International Commission on Illumination (CIE) 1931 color space.
  • the method further includes modifying the image from an original image to include the colors outside of the ITU-R BT.2020 color gamut.
  • the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and at least one viewing device, wherein the at least one viewing device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one viewing device.
  • SDP Session Description Protocol
  • the at least four primary values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary.
  • the at least four primary values include a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary.
  • the at least one viewing device is selected from the group consisting of a smartphone, a tablet, a laptop screen, a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a miniLED display, a microLED display, a liquid crystal display (LCD), a quantum dot display, a quantum nano emitting diode (QNED) device, a personal gaming device, a virtual reality (VR) device and/or an augmented reality (AR) device, an LED wall, a wearable display, and at least one projector.
  • the at least one viewing device includes at least one perovskite.
  • the set of SDP parameters is modifiable, and wherein once the set of image data has been converted by the image data converter for the at least one viewing device, the set of SDP parameters is modified based on the conversion.
  • the at least one viewing device includes at least one white emitter, wherein the at least one white emitter includes a mid-Kelvin white emitter, and wherein the mid-Kelvin white emitter is modified to include a green bias.
  • the image data converter includes an alignment signal to synchronize and align at least two projectors, and wherein the at least one viewing device includes an apparatus to combine the output display of the at least two projectors, thereby creating a combined output display.
  • the set of image data includes a first set of color channel data and a second set of color channel data
  • the image data converter further includes a first link component and a second link component, wherein the first link component is operable to transport the first set of color channel data to the at least one viewing device, wherein the second link component is operable to transport the second set of color channel data to the at least one viewing device in parallel with the first link component.
  • the primary color data corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, wherein the two colorimetric coordinates x and y are orthogonal to the luminance Y, wherein the encode and the decode include transportation of Yxy data, wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x and y, and wherein the Yxy data includes pixel mapping data.
  • the system further includes at least one non-linear function for processing the set of values in Yxy color space, wherein the at least one non-linear function is not applied to the colorimetric coordinates x and y, and wherein the at least one non-linear function is applied to the luminance Y, thereby creating a luma Y'.
  • the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes primary color data for at least four primary color values, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with at least one viewing device, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the at least one viewing device.
  • the primary color data corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, wherein the two colorimetric coordinates x and y are orthogonal to the luminance Y, wherein the encode and the decode includes transportation of Yxy data, wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x and y, and wherein the Yxy data includes pixel mapping data.
  • the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and at least one viewing device, wherein the at least one viewing device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one viewing device.
  • SDP Session Description Protocol
  • the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes primary color data for at least four primary color values, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with at least one viewing device, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the at least one viewing device.
  • the present invention provides a system for displaying a digital representation of an image including the image and a viewing device, wherein the image includes colors outside of an ITU-R BT.2020 color gamut, wherein the viewing device is operable to display the digital representation of the image and at least 80% of a total area covered between about 400nm and about 700nm by an International Commission on Illumination (CIE) 1931 color space, and wherein the viewing device is operable to display the colors outside of the ITU-R BT.2020 color gamut.
  • FIG. 137 is a schematic diagram of an embodiment of the invention illustrating a computer system, generally described as 800, having a network 810, a plurality of computing devices 820, 830, 840, a server 850, and a database 870.
  • the server 850 is constructed, configured, and coupled to enable communication over a network 810 with a plurality of computing devices 820, 830, 840.
  • the server 850 includes a processing unit 851 with an operating system 852.
  • the operating system 852 enables the server 850 to communicate through network 810 with the remote, distributed user devices.
  • Database 870 may house an operating system 872, memory 874, and programs 876.
  • the system 800 includes a network 810 for distributed communication via a wireless communication antenna 812 and processing by at least one mobile communication computing device 830.
  • wireless and wired communication and connectivity between devices and components described herein include wireless network communication such as WI-FI, WORLDWIDE INTEROPERABILITY FOR MICROWAVE ACCESS (WIMAX), Radio Frequency (RF) communication including RF identification (RFID), NEAR FIELD COMMUNICATION (NFC), BLUETOOTH including BLUETOOTH LOW ENERGY (BLE), ZIGBEE, Infrared (IR) communication, cellular communication, satellite communication, Universal Serial Bus (USB), Ethernet communications, communication via fiber-optic cables, coaxial cables, twisted pair cables, and/or any other type of wireless or wired communication.
  • WI-FI WI-FI
  • RF Radio Frequency
  • RFID RF identification
  • NFC NEAR FIELD COMMUNICATION
  • BLUETOOTH including BLUETOOTH LOW ENERGY (BLE)
  • ZIGBEE Infrared
  • IR Infrared
  • the system 800 is a virtualized computing system capable of executing any or all aspects of software and/or application components presented herein on the computing devices 820, 830, 840.
  • the computer system 800 may be implemented using hardware or a combination of software and hardware, either in a dedicated computing device, or integrated into another entity, or distributed across multiple entities or computing devices.
  • the computing devices 820, 830, 840 are intended to represent various forms of electronic devices including at least a processor and a memory, such as a server, blade server, mainframe, mobile phone, personal digital assistant (PDA), smartphone, desktop computer, notebook computer, tablet computer, workstation, laptop, and other similar computing devices.
  • PDA personal digital assistant
  • the computing device 820 includes components such as a processor 860, a system memory 862 having a random access memory (RAM) 864 and a read-only memory (ROM) 866, and a system bus 868 that couples the memory 862 to the processor 860.
  • the computing device 830 may additionally include components such as a storage device 890 for storing the operating system 892 and one or more application programs 894, a network interface unit 896, and/or an input/output controller 898. Each of the components may be coupled to each other through at least one bus 868.
  • the input/output controller 898 may receive and process input from, or provide output to, a number of other devices 899, including, but not limited to, alphanumeric input devices, mice, electronic styluses, display units, touch screens, signal generation devices (e.g., speakers), or printers.
  • other devices 899 including, but not limited to, alphanumeric input devices, mice, electronic styluses, display units, touch screens, signal generation devices (e.g., speakers), or printers.
  • the processor 860 may be a general- purpose microprocessor (e.g., a central processing unit (CPU)), a graphics processing unit (GPU), a microcontroller, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA), a Programmable Logic Device (PLD), a controller, a state machine, gated or transistor logic, discrete hardware components, or any other suitable entity or combinations thereof that can perform calculations, process instructions for execution, and/or other manipulations of information.
  • a general- purpose microprocessor e.g., a central processing unit (CPU)
  • GPU graphics processing unit
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • FPGA Field Programmable Gate Array
  • PLD Programmable Logic Device
  • multiple processors 860 and/or multiple buses 868 may be used, as appropriate, along with multiple memories 862 of multiple types (e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core).
  • multiple computing devices may be connected, with each device providing portions of the necessary operations (e.g., a server bank, a group of blade servers, or a multiprocessor system).
  • a server bank e.g., a server bank, a group of blade servers, or a multiprocessor system.
  • some steps or methods may be performed by circuitry that is specific to a given function.
  • the computer system 800 may operate in a networked environment using logical connections to local and/or remote computing devices 820, 830, 840 through a network 810.
  • a computing device 830 may connect to a network 810 through a network interface unit 896 connected to a bus 868.
  • Computing devices may communicate communication media through wired networks, direct-wired connections or wirelessly, such as acoustic, RF, or infrared, through an antenna 897 in communication with the network antenna 812 and the network interface unit 896, which may include digital signal processing circuitry when necessary.
  • the network interface unit 896 may provide for communications under various modes or protocols.
  • the instructions may be implemented in hardware, software, firmware, or any combinations thereof.
  • a computer readable medium may provide volatile or non-volatile storage for one or more sets of instructions, such as operating systems, data structures, program modules, applications, or other data embodying any one or more of the methodologies or functions described herein.
  • the computer readable medium may include the memory 862, the processor 860, and/or the storage media 890 and may be a single medium or multiple media (e.g., a centralized or distributed computer system) that store the one or more sets of instructions 900.
  • Non-transitory computer readable media includes all computer readable media, with the sole exception being a transitory, propagating signal per se.
  • the instructions 900 may further be transmitted or received over the network 810 via the network interface unit 896 as communication media, which may include a modulated data signal such as a carrier wave or other transport mechanism and includes any deliver media.
  • modulated data signal means a signal that has one or more of its characteristics changed or set in a manner as to encode information in the signal.
  • Storage devices 890 and memory 862 include, but are not limited to, volatile and non-volatile media such as cache, RAM, ROM, EPROM, EEPROM, FLASH memory, or other solid state memory technology, discs (e.g., digital versatile discs (DVD), HD-DVD, BLU-RAY, compact disc (CD), or CD-ROM) or other optical storage; magnetic cassettes, magnetic tape, magnetic disk storage, floppy disks, or other magnetic storage devices; or any other medium that can be used to store the computer readable instructions and which can be accessed by the computer system 800.
  • volatile and non-volatile media such as cache, RAM, ROM, EPROM, EEPROM, FLASH memory, or other solid state memory technology
  • discs e.g., digital versatile discs (DVD), HD-DVD, BLU-RAY, compact disc (CD), or CD-ROM
  • CD-ROM compact disc
  • magnetic cassettes magnetic tape
  • magnetic disk storage floppy disks
  • magnetic storage devices or any other medium that can be used
  • the computer system 800 is within a cloud-based network.
  • the server 850 is a designated physical server for distributed computing devices 820, 830, and 840.
  • the server 850 is a cloud-based server platform.
  • the cloud-based server platform hosts serverless functions for distributed computing devices 820, 830, and 840.
  • the computer system 800 is within an edge computing network.
  • the server 850 is an edge server
  • the database 870 is an edge database.
  • the edge server 850 and the edge database 870 are part of an edge computing platform.
  • the edge server 850 and the edge database 870 are designated to distributed computing devices 820, 830, and 840.
  • the edge server 850 and the edge database 870 are not designated for computing devices 820, 830, and 840.
  • the distributed computing devices 820, 830, and 840 are connected to an edge server in the edge computing network based on proximity, availability, latency, bandwidth, and/or other factors.
  • the computer system 800 may not include all of the components shown in FIG. 137 may include other components that are not explicitly shown in FIG. 137 or may utilize an architecture completely different than that shown in FIG. 137.
  • the various illustrative logical blocks, modules, elements, circuits, and algorithms described in connection with the embodiments discussed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate the interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application (e.g., arranged in a different order or positioned in a different way), but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.

Abstract

Systems and methods for a multi-primary color system for display. A multi-primary color system increases the number of primary colors available in a color system and color system equipment. Increasing the number of primary colors reduces metameric errors from viewer to viewer. One embodiment of the multi-primary color system includes Red, Green, Blue, Cyan, Yellow, and Magenta primaries. The systems of the present invention maintain compatibility with existing color systems and equipment and provide systems for backwards compatibility with older color systems.

Description

SYSTEM AND METHOD FOR A MULTI-PRIMARY WIDE GAMUT COLOR SYSTEM
CROSS REFERENCES TO RELATED APPLICATIONS
[0001] This application relates to and claims priority from the following applications. This application claims priority from U.S. Application No. 17/060,917, filed October 1, 2020, U.S. Application No. 17/082,741, filed October 28, 2020, U.S. Application No. 17/076,383, filed October 21, 2020, and U.S. Application No. 17/209,959, filed March 23, 2021, each of which is incorporated herein by reference.
[0002] U.S. Application No. 17/060,917 is a continuation-in-part of U.S. Application No. 17/009,408, filed September 1, 2020, which is a continuation-in-part of U.S. Application No. 16/887,807, filed May 29, 2020, which is a continuation-in-part of U.S. Application No.
16/860,769, filed April 28, 2020, which is a continuation-in-part of U.S. Application No.
16/853,203, filed April 20, 2020, which is a continuation-in-part of U.S. Patent Application No. 16/831,157, filed March 26, 2020, which is a continuation of U.S. Patent Application No. 16/659,307, filed October 21, 2019, now U.S. Patent No. 10,607,527, which is related to and claims priority from U.S. Provisional Patent Application No. 62/876,878, filed July 22, 2019, U.S. Provisional Patent Application No. 62/847,630, filed May 14, 2019, U.S. Provisional Patent Application No. 62/805,705, filed February 14, 2019, and U.S. Provisional Patent Application No. 62/750,673, filed October 25, 2018, each of which is incorporated herein by reference in its entirety.
[0003] U.S. Application No. 17/082,741 is a continuation-in-part of U.S. Application No. 17/009,408, filed September 1, 2020, which is a continuation-in-part of U.S. Application No. 16/887,807, filed May 29, 2020, which is a continuation-in-part of U.S. Application No.
16/860,769, filed April 28, 2020, which is a continuation-in-part of U.S. Application No.
16/853,203, filed April 20, 2020, which is a continuation-in-part of U.S. Patent Application
No. 16/831,157, filed March 26, 2020, which is a continuation of U.S. Patent Application No. 16/659,307, filed October 21, 2019, now U.S. Patent No. 10,607,527, which is related to and claims priority from U.S. Provisional Patent Application No. 62/876,878, filed July 22, 2019, U.S. Provisional Patent Application No. 62/847,630, filed May 14, 2019, U.S. Provisional Patent Application No. 62/805,705, filed February 14, 2019, and U.S. Provisional Patent Application No. 62/750,673, filed October 25, 2018, each of which is incorporated herein by reference in its entirety.
[0004] U.S. Application No. 17/076,383 is a continuation-in-part of U.S. Application No. 17/009,408, filed September 1, 2020, which is a continuation-in-part of U.S. Application No. 16/887,807, filed May 29, 2020, which is a continuation-in-part of U.S. Application No.
16/860,769, filed April 28, 2020, which is a continuation-in-part of U.S. Application No.
16/853,203, filed April 20, 2020, which is a continuation-in-part of U.S. Patent Application No. 16/831,157, filed March 26, 2020, which is a continuation of U.S. Patent Application No. 16/659,307, filed October 21, 2019, now U.S. Patent No. 10,607,527, which is related to and claims priority from U.S. Provisional Patent Application No. 62/876,878, filed July 22, 2019, U.S. Provisional Patent Application No. 62/847,630, filed May 14, 2019, U.S. Provisional Patent Application No. 62/805,705, filed February 14, 2019, and U.S. Provisional Patent Application No. 62/750,673, filed October 25, 2018, each of which is incorporated herein by reference in its entirety.
[0005] U.S. Application No. 17/209,959 is a continuation-in-part of U.S. Application No. 17/082,741, filed October 28, 2020, which is a continuation-in-part of U.S. Application No. 17/009,408, filed September 1, 2020, which is a continuation-in-part of U.S. Application No. 16/887,807, filed May 29, 2020, which is a continuation-in-part of U.S. Application No.
16/860,769, filed April 28, 2020, which is a continuation-in-part of U.S. Application No. 16/853,203, filed April 20, 2020, which is a continuation-in-part of U.S. Patent Application No. 16/831,157, filed March 26, 2020, which is a continuation of U.S. Patent Application No. 16/659,307, filed October 21, 2019, now U.S. Patent No. 10,607,527, which is related to and claims priority from U.S. Provisional Patent Application No. 62/876,878, filed July 22, 2019, U.S. Provisional Patent Application No. 62/847,630, filed May 14, 2019, U.S. Provisional Patent Application No. 62/805,705, filed February 14, 2019, and U.S. Provisional Patent Application No. 62/750,673, filed October 25, 2018, each of which is incorporated herein by reference in its entirety.
BACKGROUND OF THE INVENTION
[0006] 1. Field of the Invention
[0007] The present invention relates to color systems, and more specifically to a wide gamut color system with an increased number of primary colors.
[0008] 2. Description of the Prior Art
[0009] It is generally known in the prior art to provide for an increased color gamut system within a display.
[0010] Prior art patent documents include the following:
[0011] U.S. Patent Publication No. 20200144327 for Light emitting diode module and display device by inventors Lee, et al., filed June 27, 2019 and published May 7, 2020, is directed to a light emitting diode module that includes a cell array including first to fourth light emitting diode cells, each cell having a first conductive semiconductor layer, an active layer, and a second conductive semiconductor layer, the cell array having a first surface and a second surface opposite to the first surface; first to fourth light adjusting portions on the second surface of the cell array to respectively correspond to the first to fourth light emitting diode cells, to provide red light, first green light, second green light, and blue light, respectively; light blocking walls between the first to fourth light adjusting portions to isolate the first to fourth light adjusting portions from one another; and an electrode portion on the first surface of the cell array, and electrically connected to the first to fourth light emitting diode cells to selectively drive the first to fourth light emitting diode cells.
[0012] U.S. Patent No. 10,847,498 for Display device and electronic device by inventors Nakamura, et al., filed April 10, 2020 and issued November 24, 2020, is directed to a display panel that includes a plurality of light-emitting elements. Light emitted from a first lightemitting element has a CIE 1931 chromaticity coordinate x of greater than 0.680 and less than or equal to 0.720 and a CIE 1931 chromaticity coordinate y of greater than or equal to 0.260 and less than or equal to 0.320. Light emitted from a second light-emitting element has a CIE 1931 chromaticity coordinate x of greater than or equal to 0.130 and less than or equal to 0.250 and a CIE 1931 chromaticity coordinate y of greater than 0.710 and less than or equal to 0.810. Light emitted from a third light-emitting element has a CIE 1931 chromaticity coordinate x of greater than or equal to 0.120 and less than or equal to 0.170 and a CIE 1931 chromaticity coordinate y of greater than or equal to 0.020 and less than 0.060.
[0013] U.S. Patent No. 10,504,437 for Display panel, control method thereof, display device and display system for anti -peeping display by inventors Zhang, et al., filed May 26, 2016 and issued December 10, 2019, is directed to a display panel, a control method thereof, a display device and a display system comprising such a display panel. The display panel includes a plurality of pixel units. Each pixel unit has a plurality of subpixels Each subpixel has a display subpixel and an interference subpixel. Additionally, the interference subpixel and the display subpixel are different in at least one of color and gray scale. The display panel also includes a first control unit configured to control the display subpixel to be switched on during a first period of time in each display period, and to control the interference subpixel to be switched off during the first period of time in each display period and switched on during a second period of time in each display period. [0014] U.S. Patent Publication No. 20200128220 for Image processing method and apparatus, electronic device, and computer storage medium by inventors Bao, et al., filed December 19, 2019 and published April 23, 2020, is directed to an image processing method including: obtaining a facial skin tone area in an image to be processed; filtering the image to be processed to obtain a filtered smooth image; obtaining a high-frequency image based on the smooth image and the image to be processed; obtaining a facial skin tone high-frequency image based on the high-frequency image and a facial skin tone mask; and superimposing the high-frequency image and the image to be processed based on the facial skin tone mask and preset first superimposition strength in a luma channel, and superimposing a luma channel signal of the facial skin tone high-frequency image onto a luma channel signal of the image to be processed, to obtain a first image.
[0015] U.S. Patent Publication No. 20200209678 for Reflective pixel unit, reflective display panel and display apparatus by inventors Hsu, et al., filed April 17, 2019 and published July 2, 2020, is directed to a reflective pixel unit, a reflective display panel and a display apparatus. The reflective pixel unit includes a substrate, a reflective plate on the substrate, and a reflective filter layer on a side of the reflective plate facing away from the substrate. The reflective filter layer is configured such that a surface of the reflective filter layer facing away from the reflective plate receives visible light and reflects a part of light having wavelengths within a specific range in the visible light, and allows another part of the light having wavelengths within the specific range to pass through the reflective filter layer to the reflective plate. The reflective plate is configured to reflect the another part of the light having wavelengths within the specific range passed through the reflective filter layer.
[0016] U.S. Patent No. 10,222,263 for RGB value calculation device by inventor Yasuyuki Shigezane, filed February 6, 2017 and issued March 5, 2019, is directed to a microcomputer that equally divides the circumference of an RGB circle into 6xn (n is an integer of 1 or more) parts, and calculates an RGB value of each divided color. (255, 0, 0) is stored as a reference RGB value of a reference color in a ROM in the microcomputer. The microcomputer converts the reference RGB value depending on an angular difference of the RGB circle between a designated color whose RGB value is to be found and the reference color, and assumes the converted RGB value as an RGB value of the designated color.
[0017] U.S. Patent No. 9,373,305 for Semiconductor device, image processing system and program by inventor Hiorfumi Kawaguchi, filed May 29, 2015 and issued June 21, 2016, is directed to an image process device including a display panel operable to provide an input interface for receiving an input of an adjustment value of at least a part of color attributes of each vertex of n axes (n is an integer equal to or greater than 3) serving as adjustment axes in an RGB color space, and an adjustment data generation unit operable to calculate the degree of influence indicative of a following index of each of the n-axis vertices, for each of the n axes, on a basis of distance between each of the n-axis vertices and a target point which is an arbitrary lattice point in the RGB color space, and operable to calculate adjusted coordinates of the target point in the RGB color space.
[0018] U.S. Publication No. 20130278993 for Color-mixing bi-primary color systems for displays by inventor Heikenfeld, et.al, filed September 1, 2011 and published October 24, 2013, is directed to a display pixel. The pixel includes first and second substrates arranged to define a channel. A fluid is located within the channel and includes a first colorant and a second colorant. The first colorant has a first charge and a color. The second colorant has a second charge that is opposite in polarity to the first charge and a color that is complimentary to the color of the first colorant. A first electrode, with a voltage source, is operably coupled to the fluid and configured to moving one or both of the first and second colorants within the fluid and alter at least one spectral property of the pixel. [0019] U.S. Patent No. 8,599,226 for Device and method of data conversion for wide gamut displays by inventor Ben-Chorin, et. al, filed February 13, 2012 and issued December 3, 2013, is directed to a method and system for converting color image data from a, for example, three-dimensional color space format to a format usable by an n-primary display, wherein n is greater than or equal to 3. The system may define a two-dimensional sub-space having a plurality of two-dimensional positions, each position representing a set of n primary color values and a third, scaleable coordinate value for generating an n-primary display input signal. Furthermore, the system may receive a three-dimensional color space input signal including out-of range pixel data not reproducible by a three-primary additive display, and may convert the data to side gamut color image pixel data suitable for driving the wide gamut color display.
[0020] U.S. Patent No. 8,081,835 for Multiprimary color sub-pixel rendering with metameric filtering by inventor Elliot, et. al, filed July 13, 2010 and issued December 20, 2011, is directed to systems and methods of rendering image data to multi primary displays that adjusts image data across metamers as herein disclosed. The metamer filtering may be based upon input image content and may optimize sub-pixel values to improve image rendering accuracy or perception. The optimizations may be made according to many possible desired effects. One embodiment comprises a display system comprising: a display, said display capable of selecting from a set of image data values, said set comprising at least one metamer; an input image data unit; a spatial frequency detection unit, said spatial frequency detection unit extracting a spatial frequency characteristic from said input image data; and a selection unit, said unit selecting image data from said metamer according to said spatial frequency characteristic.
[0021] U.S. Patent No. 7,916,939 for High brightness wide gamut display by inventor Roth, et. al, filed November 30, 2009 and issued March 29, 2011, is directed to a device to produce a color image, the device including a color filtering arrangement to produce at least four colors, each color produced by a filter on a color filtering mechanism having a relative segment size, wherein the relative segment sizes of at least two of the primary colors differ. [0022] U.S. Patent No. 6,769,772 for Six color display apparatus having increased color gamut by inventor Roddy, et. al, filed October 11, 2002 and issued August 3, 2004, is directed to a display system for digital color images using six color light sources or two or more multicolor LED arrays or OLEDs to provide an expanded color gamut. Apparatus uses two or more spatial light modulators, which may be cycled between two or more color light sources or LED arrays to provide a six-color display output. Pairing of modulated colors using relative luminance helps to minimize flicker effects.
SUMMARY OF THE INVENTION
[0023] It is an object of this invention to provide an enhancement to the current RGB systems or a replacement for them.
[0024] In one embodiment, the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter, wherein the single display device and the image data converter are in network communication, wherein the image data converter further includes a first link component and a second link component, wherein the image data converter is operable to convert the set of image data for display on the single display device, and wherein the first link component is operable to transport the first set of color channel data to the single display device and wherein the second link component is operable to transport the second set of color channel data to the single display device in parallel with the first link component. In one embodiment, the single display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the single display device is based on the set of image data. [0025] In another embodiment, the present invention provides system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, wherein the set of SDP parameters is modifiable, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a midKelvin white emitter, wherein the single display device and the image data converter are in network communication, wherein the image data converter further includes a first link component and a second link component, wherein the image data converter is operable to convert the set of image data for display on the single display device, wherein the first link component is operable to transport the first set of color channel data to the single display device and wherein the second link component is operable to transport the second set of color channel data to the single display device in parallel with the first link component, and wherein once the set of image data has been converted by the image data converter for the single display device, the set of SDP parameters is modified based on the conversion.
[0026] In yet another embodiment, the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with a single display device, and wherein the image data converter further includes a first link component and a second link component, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the single display device using the image data converter, transporting the first set of color channel data to the single display device using the first link component, and transporting the second set of color channel data to the single display device using the first link component in parallel with the first link component, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter.
[0027] These and other aspects of the present invention will become apparent to those skilled in the art after a reading of the following description of the preferred embodiment when considered with the drawings, as they support the claimed invention.
BRIEF DESCRIPTION OF THE DRAWINGS [0028] The patent or application file contains at least one drawing executed in color.
Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
[0029] FIG. 1 illustrates one embodiment of a six primary system including a red primary, a green primary, a blue primary, a cyan primary, a magenta primary, and a yellow primary (“6P-B”) compared to ITU-R BT.709-6.
[0030] FIG. 2 illustrates another embodiment of a six primary system including a red primary, a green primary, a blue primary, a cyan primary, a magenta primary, and a yellow primary (“6P-C”) compared to SMPTE RP431 -2 for a D60 white point.
[0031] FIG. 3 illustrates yet another embodiment of a six primary system including a red primary, a green primary, a blue primary, a cyan primary, a magenta primary, and a yellow primary (“6P-C”) compared to SMPTE RP431 -2 for a D65 white point.
[0032] FIG. 4 illustrates Super 6Pa compared to 6P-C.
[0033] FIG. 5 illustrates Super 6Pb compared to Super 6Pa and 6P-C.
[0034] FIG. 6 illustrates an embodiment of an encode and decode system for a multiprimary color system.
[0035] FIG. 7 illustrates a sequential method where three color primaries are passed to the transport format as full bit level image data and inserted as normal (“System 2”).
[0036] FIG. 8 illustrates one embodiment of a system encode and decode process using a dual link method (“System 3”).
[0037] FIG. 9 illustrates one embodiment of an encoding process using a dual link method.
[0038] FIG. 10 illustrates one embodiment of a decoding process using a dual link method. [0039] FIG. 11 illustrates one embodiment of a six-primary color system encode using a 4:4:4 sampling method.
[0040] FIG. 12 illustrates one embodiment for a method to package six channels of primary information into the three standard primary channels used in current serial video standards by modifying bit numbers for a 12-bit SDI and a 10-bit SDI.
[0041] FIG. 13 illustrates a simplified diagram estimating perceived viewer sensation as code values define each hue angle.
[0042] FIG. 14 illustrates one embodiment for a method of stacking/encoding six-primary color information using a 4:4:4 video system.
[0043] FIG. 15 illustrates one embodiment for a method of unstacking/decoding six- primary color information using a 4:4:4 video system.
[0044] FIG. 16 illustrates one embodiment of a 4:4:4 decoder for a six-primary color system.
[0045] FIG. 17 illustrates one embodiment of an optical filter.
[0046] FIG. 18 illustrates another embodiment of an optical filter.
[0047] FIG. 19 illustrates an embodiment of the present invention for sending six primary colors to a standardized transport format.
[0048] FIG. 20 illustrates one embodiment of a decode process adding a pixel delay to the RGB data for realigning the channels to a common pixel timing.
[0049] FIG. 21 illustrates one embodiment of an encode process for 4:2:2 video for packaging five channels of information into the standard three-channel designs.
[0050] FIG. 22 illustrates one embodiment for a non-constant luminance encode for a six- primary color system.
[0051] FIG. 23 illustrates one embodiment of a packaging process for a six-primary color system. [0052] FIG. 24 illustrates a 4:2:2 unstack process for a six-primary color system.
[0053] FIG. 25 illustrates one embodiment of a process to inversely quantize each individual color and pass the data through an electronic optical function transfer (EOTF) in a non-constant luminance system.
[0054] FIG. 26 illustrates one embodiment of a constant luminance encode for a six- primary color system.
[0055] FIG. 27 illustrates one embodiment of a constant luminance decode for a six- primary color system.
[0056] FIG. 28 illustrates one example of 4:2:2 non-constant luminance encoding.
[0057] FIG. 29 illustrates one embodiment of a non-constant luminance decoding system.
[0058] FIG. 30 illustrates one embodiment of a 4:2:2 constant luminance encoding system.
[0059] FIG. 31 illustrates one embodiment of a 4:2:2 constant luminance decoding system.
[0060] FIG. 32 illustrates a raster encoding diagram of sample placements for a six- primary color system.
[0061] FIG. 33 illustrates one embodiment of the six-primary color unstack process in a 4:2:2 video system.
[0062] FIG. 34 illustrates one embodiment of mapping input to the six-primary color system unstack process.
[0063] FIG. 35 illustrates one embodiment of mapping the output of a six-primary color system decoder.
[0064] FIG. 36 illustrates one embodiment of mapping the RGB decode for a six-primary color system. [0065] FIG. 37 illustrates one embodiment of an unstack system for a six-primary color system.
[0066] FIG. 38 illustrates one embodiment of a legacy RGB decoder for a six-primary, non-constant luminance system.
[0067] FIG. 39 illustrates one embodiment of a legacy RGB decoder for a six-primary, constant luminance system.
[0068] FIG. 40 illustrates one embodiment of a six-primary color system with output to a legacy RGB system.
[0069] FIG. 41 illustrates one embodiment of six-primary color output using a nonconstant luminance decoder.
[0070] FIG. 42 illustrates one embodiment of a legacy RGB process within a six-primary color system.
[0071] FIG. 43 illustrates one embodiment of packing six-primary color system image data into an ICjCp (ITP) format.
[0072] FIG. 44 illustrates one embodiment of a six-primary color system converting RGBCMY image data into XYZ image data for an ITP format.
[0073] FIG. 45 illustrates one embodiment of six-primary color mapping with SMPTE ST424.
[0074] FIG. 46 illustrates one embodiment of a six-primary color system readout for a SMPTE ST424 standard.
[0075] FIG. 47 illustrates a process of 2160p transport over 12G-SDI.
[0076] FIG. 48 illustrates one embodiment for mapping RGBCMY data to the SMPTE ST2082 standard for a six-primary color system.
[0077] FIG. 49 illustrates one embodiment for mapping YRGB YCMY CR CB CC CY data to the SMPTE ST2082 standard for a six-primary color system. [0078] FIG. 50 illustrates one embodiment for mapping six-primary color system data using the SMPTE ST292 standard.
[0079] FIG. 51 illustrates one embodiment of the readout for a six-primary color system using the SMPTE ST292 standard.
[0080] FIG. 52 illustrates modifications to the SMPTE ST352 standards for a six-primary color system.
[0081] FIG. 53 illustrates modifications to the SMPTE ST2022 standard for a six-primary color system.
[0082] FIG. 54 illustrates a table of 4:4:4 sampling for a six-primary color system for a 10-bit video system.
[0083] FIG. 55 illustrates a table of 4:4:4 sampling for a six-primary color system for a 12-bit video system.
[0084] FIG. 56 illustrates sequence substitutions for 10-bit and 12-bit video in 4:2:2 sampling systems in a Y Cb Cr Cc Cy color space.
[0085] FIG. 57 illustrates sample placements of six-primary system components for a 4:2:2 sampling system image.
[0086] FIG. 58 illustrates sequence substitutions for 10-bit and 12-bit video in 4:2:0 sampling systems using a Y Cb Cr Cc Cy color space.
[0087] FIG. 59 illustrates sample placements of six-primary system components for a 4:2:0 sampling system image.
[0088] FIG. 60 illustrates modifications to SMPTE ST2110-20 for a 10-bit six-primary color system in 4:4:4 video.
[0089] FIG. 61 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:4:4 video. [0090] FIG. 62 illustrates modifications to SMPTE ST2110-20 for a 10-bit six primary color system in 4:2:2 video.
[0091] FIG. 63 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:2:0 video.
[0092] FIG. 64 illustrates an RGB sampling transmission for a 4:4:4 sampling system.
[0093] FIG. 65 illustrates a RGBCMY sampling transmission for a 4:4:4 sampling system.
[0094] FIG. 66 illustrates an example of System 2 to RGBCMY 4:4:4 transmission.
[0095] FIG. 67 illustrates a Y Cb Cr sampling transmission using a 4:2:2 sampling system.
[0096] FIG. 68 illustrates a Y Cr Cb Cc Cy sampling transmission using a 4:2:2 sampling system.
[0097] FIG. 69 illustrates an example of a System 2 to Y Cr Cb Cc Cy 4:2:2 Transmission as non-constant luminance.
[0098] FIG. 70 illustrates a Y Cb Cr sampling transmission using a 4:2:0 sampling system.
[0099] FIG. 71 illustrates a Y Cr Cb Cc Cy sampling transmission using a 4:2:0 sampling system.
[00100] FIG. 72 illustrates a dual stack LCD projection system for a six-primary color system.
[00101] FIG. 73 illustrates one embodiment of a single projector.
[00102] FIG. 74 illustrates a six-primary color system using a single projector and reciprocal mirrors.
[00103] FIG. 75 illustrates a dual stack DMD projection system for a six-primary color system. [00104] FIG. 76 illustrates one embodiment of a single DMD projector solution.
[00105] FIG. 77 illustrates one embodiment of a color filter array for a six-primary color system with a white OLED monitor.
[00106] FIG. 78 illustrates one embodiment of an optical filter array for a six-primary color system with a white OLED monitor.
[00107] FIG. 79 illustrates one embodiment of a matrix of an LCD drive for a six-primary color system with a backlight illuminated LCD monitor.
[00108] FIG. 80 illustrates one embodiment of an optical filter array for a six-primary color system with a backlight illuminated LCD monitor.
[00109] FIG. 81 illustrates an array for a Quantum Dot (QD) display device.
[00110] FIG. 82 illustrates one embodiment of an array for a six-primary color system for use with a direct emissive assembled display.
[00111] FIG. 83 illustrates one embodiment of a six-primary color system in an emissive display that does not incorporate color filtered subpixels.
[00112] FIG. 84 illustrates a graph of one embodiment of a four primary system with respect to CIE 1931.
[00113] FIG. 85 illustrates a graph of one embodiment of a five primary system with respect to CIE 1931.
[00114] FIG. 86 illustrates a graph of one embodiment of a six primary system with respect to CIE 1931.
[00115] FIG. 87 illustrates a graph of one embodiment of a seven primary system with respect to CIE 1931.
[00116] FIG. 88 illustrates a graph of one embodiment of an eight primary system with respect to CIE 1931. [00117] FIG. 89 illustrates a graph of one embodiment of a ten primary system with respect to CIE 1931.
[00118] FIG. 90 illustrates a graph of one embodiment of a twelve primary system with respect to CIE 1931.
[00119] FIG. 91 illustrates a graph of another embodiment of a twelve primary system with respect to CIE 1931.
[00120] FIG. 92 illustrates a graph of a twelve primary system that is backwards compatible with 6P-C with respect to CIE 1931.
[00121] FIG. 93 shows one embodiment of transportation of twelve individual color channels on a first link (Link A) and a second link (Link B).
[00122] FIG. 94A shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a first link (Link A).
[00123] FIG. 94B shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a second link (Link B).
[00124] FIG. 95 A shows one embodiment of a 4:2:2 Constant Luminance Encode for a first link (Link A).
[00125] FIG. 95B shows one embodiment of a 4:2:2 Constant Luminance Encode for a second link (Link B).
[00126] FIG. 96A shows one embodiment of a 4:4:4 Encode for a first link (Link A).
[00127] FIG. 96B shows one embodiment of a 4:4:4 Encode for a second link (Link B).
[00128] FIG. 97A shows one embodiment of component mapping into SMPTE 2081-1 for a first link (Link A).
[00129] FIG. 97B shows one embodiment of component mapping into SMPTE 2081-1 for a second link (Link B). [00130] FIG. 98A shows one embodiment of a twelve primary system mapping into SMPTE 2081-1 for a first link (Link A).
[00131] FIG. 98B shows one embodiment of the twelve primary system mapping into
SMPTE 2081-1 for a second link (Link B).
[00132] FIG. 99A shows one embodiment of a 4:2:2 Non-Constant Luminance Decode for a first link (Link A).
[00133] FIG. 99B shows one embodiment of a 4:2:2 Non-Constant Luminance Decode for a second link (Link B).
[00134] FIG. 100A shows one embodiment of a 4:2:2 Constant Luminance Decode for a first link (Link A).
[00135] FIG. 100B shows one embodiment of a 4:2:2 Constant Luminance Decode for a second link (Link B).
[00136] FIG. 101A shows one embodiment of a 4:4:4 Decode for a first link (Link A).
[00137] FIG. 101B shows one embodiment of a 4:4:4 Decode for a second link (Link B).
[00138] FIG. 102A illustrates a front view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
[00139] FIG. 102B illustrates a normal orthogonal view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
[00140] FIG. 102C illustrates a top view of a three-dimensional plot of ITU-R BT.2020 in XYZ space.
[00141] FIG. 103A illustrates a front view of a three-dimensional plot of DCI-P3 in XYZ space.
[00142] FIG. 103B illustrates a normal orthogonal view of a three-dimensional plot of DCI-P3 in XYZ space. [00143] FIG. 103C illustrates a top view of a three-dimensional plot of DCI-P3 in XYZ space.
[00144] FIG. 104A illustrates a front view of 6P-C in XYZ space.
[00145] FIG. 104B illustrates a normal orthogonal view of 6P-C in XYZ space.
[00146] FIG. 104C illustrates a top view of 6P-C in XYZ space.
[00147] FIG. 105A illustrates a front view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
[00148] FIG. 105B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
[00149] FIG. 105C illustrates a top view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space.
[00150] FIG. 106A illustrates a front view of DCI-P3 (red) and 6P-C (green) in XYZ space.
[00151] FIG. 106B illustrates a normal orthogonal view of DCI-P3 (red) and 6P-C (green) in XYZ space.
[00152] FIG. 106C illustrates a top view of DCI-P3 (red) and 6P-C (green) in XYZ space.
[00153] FIG. 107A illustrates a front view of 4P in XYZ space.
[00154] FIG. 107B illustrates a normal orthogonal view of 4P in XYZ space.
[00155] FIG. 107C illustrates a top view of 4P in XYZ space.
[00156] FIG. 108A illustrates a front view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space.
[00157] FIG. 108B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space.
[00158] FIG. 108C illustrates a top view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space [00159] FIG. 109A illustrates a front view of DCI-P3 (red) and 4P (blue) in XYZ space.
[00160] FIG. 109B illustrates a normal orthogonal view of DCI-P3 (red) and 4P (blue) in
XYZ space.
[00161] FIG. 109C illustrates a top view of DCI-P3 (red) and 4P (blue) in XYZ space.
[00162] FIG. 110A illustrates a front view of 4P-N in XYZ space.
[00163] FIG. HOB illustrates a normal orthogonal view of 4P-N in XYZ space.
[00164] FIG. 110C illustrates a top view of 4P-N in XYZ space.
[00165] FIG. 111 A illustrates a front view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
[00166] FIG. 11 IB illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
[00167] FIG. 111C illustrates a top view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space.
[00168] FIG. 112A illustrates a front view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
[00169] FIG. 112B illustrates a normal orthogonal view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
[00170] FIG. 112C illustrates a top view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
[00171] FIG. 113A illustrates one embodiment of a quadrature method (“System 2A”).
[00172] FIG. 113B illustrates another embodiment of a quadrature method (“System 2A”).
[00173] FIG. 113C illustrates yet another embodiment of a quadrature method (“System
2A”).
[00174] FIG. 114A illustrates an embodiment of a stereo quadrature method (“System
2A”).
[00175] FIG. 114B illustrates another embodiment of a stereo quadrature method (“System 2A”). [00176] FIG. 114C illustrates yet another embodiment of a stereo quadrature method
(“System 2A”).
[00177] FIG. 115 illustrates one embodiment of a Yxy encode with an OETF.
[00178] FIG. 116 illustrates one embodiment of a Yxy encode without an OETF.
[00179] FIG. 117 illustrates one embodiment of a Yxy decode with an electro-optical transfer function (EOTF).
[00180] FIG. 118 illustrates one embodiment of a Yxy decode without an EOTF.
[00181] FIG. 119 illustrates one embodiment of a 4:2:2 Yxy encode with an OETF.
[00182] FIG. 120 illustrates one embodiment of a 4:2:2 Yxy encode without an OETF.
[00183] FIG. 121 illustrates one embodiment of a 4:4:4 Yxy encode with an OETF.
[00184] FIG. 122 illustrates one embodiment of a 4:4:4 Yxy encode without an OETF.
[00185] FIG. 123 illustrates sample placements of Yxy system components for a 4:2:2 pixel mapping.
[00186] FIG. 124 illustrates sample placements of Yxy system components for a 4:2:0 pixel mapping.
[00187] FIG. 125 illustrates one embodiment of a SMPTE ST292 Yxy system mapping.
[00188] FIG. 126 illustrates one embodiment of a SMPTE ST2082 Yxy system mapping.
[00189] FIG. 127 illustrates one embodiment of Yxy inserted into a CTA 861 stream.
[00190] FIG. 128 illustrates one embodiment of a Yxy decode with an EOTF.
[00191] FIG. 129 illustrates one embodiment of a Yxy decode without an EOTF.
[00192] FIG. 130A illustrates one embodiment of an IPT 4:4:4 encode.
[00193] FIG. 130B illustrates one embodiment of an IPT 4:4:4 decode.
[00194] FIG. 131 A illustrates one embodiment of an ICTCP 4:2:2 encode.
[00195] FIG. 13 IB illustrates one embodiment of an ICTCP 4:2:2 decode.
[00196] FIG. 132 illustrates the emissive spectra of Xenon lamps and UHPHg lamps. [00197] FIG. 133 illustrates one embodiment of the dual-panel display system using a Cyan filter.
[00198] FIG. 134 illustrates one embodiment of a Vi gamma function.
[00199] FIG. 135 illustrates a graph of maximum quantizing error using the Vi gamma function.
[00200] FIG. 136 illustrates one embodiment of a 1/3 gamma function.
[00201] FIG. 137 is a schematic diagram of an embodiment of the invention illustrating a computer system.
DETAILED DESCRIPTION
[00202] The present invention is generally directed to a multi-primary color system.
[00203] In one embodiment, the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter, wherein the single display device and the image data converter are in network communication, wherein the image data converter further includes a first link component and a second link component, wherein the image data converter is operable to convert the set of image data for display on the single display device, and wherein the first link component is operable to transport the first set of color channel data to the single display device and wherein the second link component is operable to transport the second set of color channel data to the single display device in parallel with the first link component. In one embodiment, the single display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the single display device is based on the set of image data. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 493nm, a third primary at approximately 540nm, and a fourth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 485nm, a third primary at approximately 510nm, a fourth primary at approximately 535nm, and a fifth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 490nm, a third primary at approximately 506nm, a fourth primary at approximately 520nm, a fifth primary at approximately 545nm, and a sixth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 508nm, a fifth primary at approximately 520nm, a sixth primary at approximately 540nm, and a seventh primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 500nm, a fifth primary at approximately 51 Inm, a sixth primary at approximately 521nm, a seventh primary at approximately 545nm, and an eighth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 502nm, a sixth primary at approximately 512nm, a seventh primary at approximately 520nm, an eighth primary at approximately 535nm, a ninth primary at approximately 550nm, and a tenth primary at approximately 660nm. In one embodiment, the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 505nm, a seventh primary at approximately 511nm, an eighth primary at approximately 517nm, a ninth primary at approximately 523nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 550nm, and a twelfth primary at approximately 670nm. In one embodiment, the at least four primaries include a first primary at approximately 400nm, a second primary at approximately 468nm, a third primary at approximately 484nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 506nm, a seventh primary at approximately 512nm, an eighth primary at approximately 518nm, a ninth primary at approximately 524nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 556nm, and a twelfth primary at approximately 700nm. In one embodiment, the at least four primaries include a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary. In one embodiment, the set of SDP parameters is modifiable. In one embodiment, the mid-Kelvin white emitter is modified to include a green bias. In one embodiment, the first set of color channel data is converted by the first link component and the second set of color channel data is converted by the second link component, and wherein the first set of color channel data and the second set of color channel data are combined to form the set of image data for display on the single display device. In one embodiment, the system further includes a standardized transport format, wherein the first link component includes a first standardized transport format link and wherein the second link component includes a second standardized transport format link, wherein the standardized transport format is operable to receive the first set of image data and the second set of image data using the first standardized transport format link and the second standardized transport format link, and wherein the first standardized transport format link and the second standardized transport format link are operable to combine the first set of image data and the second set of image data into a combined set of image data.
[00204] In another embodiment, the present invention provides system for displaying a primary color system including a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, wherein the set of SDP parameters is modifiable, and a display device, wherein the display device is a single display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a midKelvin white emitter, wherein the single display device and the image data converter are in network communication, wherein the image data converter further includes a first link component and a second link component, wherein the image data converter is operable to convert the set of image data for display on the single display device, wherein the first link component is operable to transport the first set of color channel data to the single display device and wherein the second link component is operable to transport the second set of color channel data to the single display device in parallel with the first link component, and wherein once the set of image data has been converted by the image data converter for the single display device, the set of SDP parameters is modified based on the conversion. In one embodiment, the mid-Kelvin white emitter is modified to include a green bias. In one embodiment, the at least one white emitter includes a white emitter matching a white point of the primary color system.
[00205] In yet another embodiment, the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with a single display device, and wherein the image data converter further includes a first link component and a second link component, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the single display device using the image data converter, transporting the first set of color channel data to the single display device using the first link component, and transporting the second set of color channel data to the single display device using the first link component in parallel with the first link component, wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter. In one embodiment, the at least one white emitter includes a white emitter matching a white point of the primary color system. [00206] The present invention relates to color systems. A multitude of color systems are known, but they continue to suffer numerous issues. As imaging technology is moving forward, there has been a significant interest in expanding the range of colors that are replicated on electronic displays. Enhancements to the television system have expanded from the early CCIR 601 standard to ITU-R BT.709-6, to SMPTE RP431-2, and ITU-R BT.2020. Each one has increased the gamut of visible colors by expanding the distance from the reference white point to the position of the Red (R), Green (G), and Blue (B) color primaries (collectively known as “RGB”) in chromaticity space. While this approach works, it has several disadvantages. When implemented in content presentation, issues arise due to the technical methods used to expand the gamut of colors seen (typically using a more-narrow emissive spectrum) can result in increased viewer metameric errors and require increased power due to lower illumination source. These issues increase both capital and operational costs.
[00207] With the current available technologies, displays are limited in respect to their range of color and light output. There are many misconceptions regarding how viewers interpret the display output technically versus real-world sensations viewed with the human eye. The reason we see more than just the three emitting primary colors is because the eye combines the spectral wavelengths incident on it into the three bands. Humans interpret the radiant energy (spectrum and amplitude) from a display and process it so that an individual color is perceived. The display does not emit a color or a specific wavelength that directly relates to the sensation of color. It simply radiates energy at the same spectrum which humans sense as light and color. It is the observer who interprets this energy as color.
[00208] When the CIE 2° standard observer was established in 1931, common understanding of color sensation was that the eye used red, blue, and green cone receptors (James Maxwell & James Forbes 1855). Later with the Munsell vision model (Munsell 1915), Munsell described the vision system to include three separate components: luminance, hue, and saturation. Using RGB emitters or filters, these three primary colors are the components used to produce images on today’s modem electronic displays.
[00209] There are three primary physical variables that affect sensation of color. These are the spectral distribution of radiant energy as it is absorbed into the retina, the sensitivity of the eye in relation to the intensity of light landing on the retinal pigment epithelium, and the distribution of cones within the retina. The distribution of cones (e.g., L cones, M cones, and S cones) varies considerably from person to person.
[00210] Enhancements in brightness have been accomplished through larger backlights or higher efficiency phosphors. Encoding of higher dynamic ranges is addressed using higher range, more perceptually uniform electro-optical transfer functions to support these enhancements to brightness technology, while wider color gamuts are produced by using narrow bandwidth emissions. Narrower bandwidth emitters result in the viewer experiencing higher color saturation. But there can be a disconnect between how saturation is produced and how it is controlled. What is believed to occur when changing saturation is that increasing color values of a color primary represents an increase to saturation. This is not true, as changing saturation requires the variance of a color primary spectral output as parametric. There are no variable spectrum displays available to date as the technology to do so has not been commercially developed, nor has the new infrastructure required to support this been discussed.
[00211] Instead, the method that a display changes for viewer color sensation is by changing color luminance. As data values increase, the color primary gets brighter. Changes to color saturation are accomplished by varying the brightness of all three primaries and taking advantage of the dominant color theory. [00212] Expanding color primaries beyond RGB has been discussed before. There have been numerous designs of multi-primary displays. For example, SHARP has attempted this with their four-color QUATTRON TV systems by adding a yellow color primary and developing an algorithm to drive it. Another four primary color display was proposed by Matthew Brennesholtz which included an additional cyan primary, and a six primary display was described by Yan Xiong, Fei Deng, Shan Xu, and Sufang Gao of the School of Physics and Optoelectric Engineering at the Yangtze University Jingzhou China. In addition, AU OPTRONICS has developed a five primary display technology. SONY has also recently disclosed a camera design featuring RGBCMY (red, green, blue, cyan, magenta, and yellow) and RGBCMYW (red, green, blue cyan, magenta, yellow, and white) sensors.
[00213] Actual working displays have been shown publicly as far back as the late 1990’s, including samples from Tokyo Polytechnic University, Nagoya City University, and Genoa Technologies. However, all of these systems are exclusive to their displays, and any additional color primary information is limited to the display’s internal processing.
[00214] Additionally, the Visual Arts System for Archiving and Retrieval of Images (VASARI) project developed a colorimetric scanner system for direct digital imaging of paintings. The system provides more accurate coloring than conventional film, allowing it to replace film photography. Despite the project beginning in 1989, technical developments have continued. Additional information is available at (last accessed March 30, 2020), which is incorporated herein by reference in its entirety.
[00215] None of the prior art discloses developing additional color primary information outside of the display. Moreover, the system driving the display is often proprietary to the demonstration. In each of these executions, nothing in the workflow is included to acquire or generate additional color primary information. The development of a multi-primary color system is not complete if the only part of the system that supports the added primaries is within the display itself.
[00216] Referring now to the drawings in general, the illustrations are for the purpose of describing one or more preferred embodiments of the invention and are not intended to limit the invention thereto.
[00217] Additional details about multi-primary systems are available in U.S. Patent No. 10,607,527; 10,950,160; 10,950,161; 10,950,162; 10,997,896; 11,011,098; 11,017,708; 11,030,934; 11,037,480; 11,037,481; 11,037,482; 11,043,157; 11,049,431; 11,062,638; 11,062,639; 11,069,279; 11,069,280; and 11,100,838 and U.S. Publication Nos.
20200251039, 20210233454, and 20210209990, each of which is incorporated herein by reference in its entirety.
[00218] Traditional displays include three primaries: red, green, and blue. The multiprimary systems of the present invention include at least four primaries. The at least four primaries preferably include at least one red primary, at least one green primary, and/or at least one blue primary. In one embodiment, the at least four primaries include a cyan primary, a magenta primary, and/or a yellow primary. In one embodiment, the at least four primaries include at least one white primary.
[00219] In one embodiment, the multi-primary system includes six primaries. In one preferred embodiment, the six primaries include a red (R) primary, a green (G) primary, a blue (B) primary, a cyan (C) primary, a magenta (M) primary, and a yellow (Y) primary, often referred to as “RGBCMY”. However, the systems and methods of the present invention are not restricted to RGBCMY, and alternative primaries are compatible with the present invention.
[00220] 6P-B [00221] 6P-B is a color set that uses the same RGB values that are defined in the ITU-R BT.709-6 television standard. The gamut includes these RGB primary colors and then adds three more color primaries orthogonal to these based on the white point. The white point used in 6P-B is D65 (ISO 11664-2).
[00222] In one embodiment, the red primary has a dominant wavelength of 609nm, the yellow primary has a dominant wavelength of 571nm, the green primary has a dominant wavelength of 552nm, the cyan primary has a dominant wavelength of 491nm, and the blue primary has a dominant wavelength of 465nm as shown in Table 1. In one embodiment, the dominant wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the dominant wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the dominant wavelength is within ±2% of the value listed in the table below.
[00223] TABLE 1
X y u’ v’ A,
W (D65) 0.3127 0.3290 0.1978 0.4683
R 0.6400 0.3300 0.4507 0.5228 609nm
G 0.3000 0.6000 0.1250 0.5625 552nm
B 0.1500 0.0600 0.1754 0.1578 464nm
C 0.1655 0.3270 0.1041 0.4463 491nm
M 0.3221 0.1266 0.3325 0.2940
Y 0.4400 0.5395 0.2047 0.5649 571nm
[00224] FIG. 1 illustrates 6P-B compared to ITU-R BT.709-6.
[00225] 6P-C
[00226] 6P-C is based on the same RGB primaries defined in SMPTE RP431-2 projection recommendation. Each gamut includes these RGB primary colors and then adds three more color primaries orthogonal to these based on the white point. The white point used in 6P-B is D65 (ISO 11664-2). Two versions of 6P-C are used. One is optimized for a D60 white point (SMPTE ST2065-1), and the other is optimized for a D65 white point. Additional information about white points is available in ISO 11664-2:2007 “Colorimetry — Part 2: CIE standard illuminants” and "ST 2065-1:2012 - SMPTE Standard - Academy Color Encoding Specification (ACES)," in ST 2065-1:2012, pp.1-23, 17 April 2012, doi: 10.5594/SMPTE.ST2065-1.2012, each of which is incorporated herein by reference in its entirety.
[00227] In one embodiment, the red primary has a dominant wavelength of 615nm, the yellow primary has a dominant wavelength of 570nm, the green primary has a dominant wavelength of 545nm, the cyan primary has a dominant wavelength of 493nm, and the blue primary has a dominant wavelength of 465nm as shown in Table 2. In one embodiment, the dominant wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the dominant wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the dominant wavelength is within ±2% of the value listed in the table below.
[00228] TABLE 2
X y u’ v’ I
W (D60) 0.3217 0.3377 0.2008 0.4742
R 0.6800 0.3200 0.4964 0.5256 615nm
G 0.2650 0.6900 0.0980 0.5777 545nm
B 0.1500 0.0600 0.1754 0.1579 465nm
C 0.1627 0.3419 0.0960 0.4540 493nm
M 0.3523 0.1423 0.3520 0.3200
Y 0.4502 0.5472 0.2078 0.5683 570nm
[00229] FIG. 2 illustrates 6P-C compared to SMPTE RP431-2 for a D60 white point.
[00230] In one embodiment, the red primary has a dominant wavelength of 615nm, the yellow primary has a dominant wavelength of 570nm, the green primary has a dominant wavelength of 545nm, the cyan primary has a dominant wavelength of 423nm, and the blue primary has a dominant wavelength of 465nm as shown in Table 3. In one embodiment, the dominant wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the dominant wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the dominant wavelength is within ±2% of the value listed in the table below.
[00231] TABLE 3
X y u’ v’ A,
W (D65) 0.3127 0.3290 0.1978 0.4683
R 0.6800 0.3200 0.4964 0.5256 615nm
G 0.2650 0.6900 0.0980 0.5777 545nm
B 0.1500 0.0600 0.1754 0.1579 465nm
C 0.1617 0.3327 0.0970 0.4490 492nm
M 0.3383 0.1372 0.3410 0.3110
Y 0.4470 0.5513 0.2050 0.5689 570nm
[00232] FIG. 3 illustrates 6P-C compared to SMPTE RP431-2 for a D65 white point.
[00233] SUPER 6P
[00234] One of the advantages of ITU-R BT.2020 is that it can include all of the Pointer colors and that increasing primary saturation in a six-color primary design could also do this. Pointer is described in “The Gamut of Real Surface Colors, M.R. Pointer”, Published in Colour Research and Application Volume #5, Issue #3 (1980), which is incorporated herein by reference in its entirety. However, extending the 6P gamut beyond SMPTE RP431-2 (“6P- C”) adds two problems. The first problem is the requirement to narrow the spectrum of the extended primaries. The second problem is the complexity of designing a backwards compatible system using color primaries that are not related to current standards. But in some cases, there may be a need to extend the gamut beyond 6P-C and avoid these problems. If the goal is to encompass Pointer’s data set, then it is possible to keep most of the 6P-C system and only change the cyan color primary position. In one embodiment, the cyan color primary position is located so that the gamut edge encompasses all of Pointer’s data set. In another embodiment, the cyan color primary position is a location that limits maximum saturation. With 6P-C, cyan is positioned as u’=0.096, v’=0.454. In one embodiment of Super 6P, cyan is moved to u’=0.075, v’=0.430 (“Super 6Pa” (S6Pa)). Advantageously, this creates anew gamut that covers Pointer’s data set almost in its entirety. FIG. 4 illustrates Super 6Pa compared to 6P-C.
[00235] Table 4 is a table of values for Super 6Pa. The definition of x,y are described in ISO 11664-3:2012/CIE S 014 Part 3, which is incorporated herein by reference in its entirety. The definition of u ’,v ’ are described in ISO 11664-5 : 2016/CIE S 014 Part 5, which is incorporated herein by reference in its entirety. X defines each color primary as dominant color wavelength for RGB and complementary wavelengths CMY.
[00236] TABLE 4
X y u’ v’ 1
W (D60) 0.3217 0.3377 0.2008 0.4742
W (D65) 0.3127 0.3290 0.1978 0.4683
R 0.6800 0.3200 0.4964 0.5256 615nm
G 0.2650 0.6900 0.0980 0.5777 545nm
B 0.1500 0.0600 0.1754 0.1579 465nm
C 0.1211 0.3088 0.0750 0.4300 490nm
M 0.3523 0.1423 0.3520 0.3200
Y 0.4502 0.5472 0.2078 0.5683 570nm
[00237] In an alternative embodiment, the saturation is expanded on the same hue angle as 6P-C as shown in FIG. 5. Advantageously, this makes backward compatibility less complicated. However, this requires much more saturation (i.e., narrower spectra). In another embodiment of Super 6P, cyan is moved to u’=0.067, v’=0.449 (“Super 6Pb” (S6Pb)).
Additionally, FIG. 5 illustrates Super 6Pb compared to Super 6Pa and 6P-C.
[00238] Table 5 is a table of values for Super 6Pb. The definition of x,y are described in ISO 11664-3:2012/CIE S 014 Part 3, which is incorporated herein by reference in its entirety. The definition of u ’,v ’ are described in ISO 11664-5 : 2016/CIE S 014 Part 5, which is incorporated herein by reference in its entirety. X defines each color primary as dominant color wavelength for RGB and complementary wavelengths CMY.
[00239] TABLE 5
X y u’ v’ 1
W (ACES D60) 0.32168 0.33767 0.2008 0.4742
W (D65) 0.3127 0.3290 0.1978 0.4683
R 0.6800 0.3200 0.4964 0.5256 615nm
G 0.2650 0.6900 0.0980 0.5777 545nm
B 0.1500 0.0600 0.1754 0.1579 465nm
C 0.1156 0.3442 0.0670 0.4490 493nm
M 0.3523 0.1423 0.3520 0.3200
Y 0.4502 0.5472 0.2078 0.5683 570nm
[00240] In a preferred embodiment, a matrix is created from XYZ values of each of the primaries. As the XYZ values of the primaries change, the matrix changes. Additional details about the matrix are described below.
[00241] FORMATTING AND TRANSPORTATION OF MULTI-PRIMARY SIGNALS [00242] The present invention includes three different methods to format video for transport: System 1, System 2, and System 3. System 1 is comprised of an encode and decode system, which can be divided into base encoder and digitation, image data stacking, mapping into the standard data transport, readout, unstack, and finally image decoding. In one embodiment, the basic method of this system is to combine opposing color primaries within the three standard transport channels and identify them by their code value.
[00243] System 2 uses a sequential method where three color primaries are passed to the transport format as full bit level image data and inserted as normal. The three additional channels are delayed by one pixel and then placed into the transport instead of the first colors. This is useful in situations where quantizing artifacts may be critical to image performance.
In one embodiment, this system is comprised of the six primaries (e.g., RGB plus a method to delay the CMY colors for injection), image resolution identification to allow for pixel count synchronization, start of video identification, and RGB Delay.
[00244] System 3 utilizes a dual link method where two wires are used. In one embodiment, a first set of three channels (e.g., RGB) are sent to link A and a second set of three channels (e.g., CMY) is sent to link B. Once they arrive at the image destination, they are recombined.
[00245] To transport up to six color components (e.g., four, five, or six), System 1, System 2, or System 3 can be used as described. If four color components are used, two of the channels are set to “0”. If five color components are used, one of the channels is set to “0”. Advantageously, this transportation method works for all primary systems described herein that include up to six color components.
[00246] COMPARISON OF THREE SYSTEMS
[00247] Advantageously, System 1 fits within legacy SDI, CTA, and Ethernet transports.
Additionally, System 1 has zero latency processing for conversion to an RGB display. However, System 1 is limited to 11 -bit words.
[00248] System 2 is advantageously operable to transport 6 channels using 16-bit words with no compression. Additionally, System 2 fits within newer SDI, CTA, and Ethernet transport formats. However, System 2 requires double bit rate speed. For example, a 4K image requires a data rate for an 8K RGB image.
[00249] In comparison, System 3 is operable to transport up to 6 channels using 16-bit words with compression and at the same data required for a specific resolution. For example, a data rate for an RGB image is the same as for a 6P image using System 3. However, System
3 requires a twin cable connection within the video system. [00250] NOMENCLATURE
[00251] In one embodiment, a standard video nomenclature is used to better describe each system.
[00252] R describes red data as linear light (e.g., without a non-linear function applied). G describes green data as linear light. B describes blue data as linear light. C describes cyan data as linear light. M describes magenta data as linear light. Yc and/or Y describe yellow data as linear light.
[00253] R ’ describes red data as non-linear light (e.g., with a non-linear function applied). G ’ describes green data as non-linear light. B ’ describes blue data as non-linear light. C ’ describes cyan data as non-linear light. M’ describes magenta data as non-linear light. Yc ’ and/or Y ’ describe yellow data as non-linear light.
[00254] Ye describes the luminance sum of RGBCMY data. TRGB describes a System 2 encode that is the linear luminance sum of the RGB data. TCMY describes a System 2 encode that is the linear luminance sum of the CMY data.
[00255] CR describes the data value of red after subtracting linear image luminance. CB describes the data value of blue after subtracting linear image luminance. Cc describes the data value of cyan after subtracting linear image luminance. CY describes the data value of yellow after subtracting linear image luminance.
[00256] E’RGB describes a System 2 encode that is the nonlinear luminance sum of the RGB data. K ’CMY describes a System 2 encode that is the nonlinear luminance sum of the CMY data. -Y describes the sum of RGB data subtracted from Ye.
[00257] C ’R describes the data value of red after subtracting nonlinear image luminance.
C ’B describes the data value of blue after subtracting nonlinear image luminance. C ’c describes the data value of cyan after subtracting nonlinear image luminance. C ’Y describes the data value of yellow after subtracting nonlinear image luminance. [00258] B+Y describes a System 1 encode that includes either blue or yellow data. G+M describes a System 1 encode that includes either green or magenta data. R+C describes a System 1 encode that includes either green or magenta data.
[00259] CR+CC describes a System 1 encode that includes either color difference data. CB+CY describes a System 1 encode that includes either color difference data.
[00260] 4:4:4 describes full bandwidth sampling of a color in an RGB system. 4:4:4:4:4:4 describes full sampling of a color in an RGBCMY system. 4:2:2 describes an encode where a full bandwidth luminance channel (7) is used to carry image detail and the remaining components are half sampled as a Cb Cr encode. 4:2:2:2:2 describes an encode where a full bandwidth luminance channel (7) is used to carry image detail and the remaining components are half sampled as a Cb Cr Cy Cc encode. 4:2:0 describes a component system similar to 4:2:2, but where Cr and Cb samples alternate per line. 4:2:0:2:0 describes a component system similar to 4:2:2, but where Cr, Cb, Cy, and Cc samples alternate per line.
[00261] Constant luminance is the signal process where luminance (7) values are calculated in linear light. Non-constant luminance is the signal process where luminance (7) values are calculated in nonlinear light.
[00262] DERIVING COLOR COMPONENTS
[00263] When using a color difference method (4:2:2), several components need specific processing so that they can be used in lower frequency transports. These are derived as:
[00272] SYSTEM 1
[00273] In one embodiment, the multi-primary color system is compatible with legacy systems. A backwards compatible multi-primary color system is defined by a sampling method. In one embodiment, the sampling method is 4:4:4. In one embodiment, the sampling method is 4:2:2. In another embodiment, the sampling method is 4:2:0. In one embodiment of a backwards compatible multi-primary color system, new encode and decode systems are divided into the steps of performing base encoding and digitization, image data stacking, mapping into the standard data transport, readout, unstacking, and image decoding (“System 1”). In one embodiment, System 1 combines opposing color primaries within three standard transport channels and identifies them by their code value. In one embodiment of a backwards compatible multi-primary color system, the processes are analog processes. In another embodiment of a backwards compatible multi-primary color system, the processes are digital processes.
[00274] In one embodiment, the sampling method for a multi-primary color system is a 4:4:4 sampling method. Black and white bits are redefined. In one embodiment, putting black at midlevel within each data word allows the addition of CMY color data.
[00275] FIG. 6 illustrates an embodiment of an encode and decode system for a multiprimary color system. In one embodiment, the multi-primary color encode and decode system is divided into a base encoder and digitation, image data stacking, mapping into the standard data transport, readout, unstack, and finally image decoding (“System 1”). In one embodiment, the method of this system combines opposing color primaries within the three standard transport channels and identifies them by their code value. In one embodiment, the encode and decode for a multi-primary color system are analog-based. In another embodiment, the encode and decode for a multi-primary color system are digital-based.
System 1 is designed to be compatible with lower bandwidth systems and allows a maximum of 11 bits per channel and is limited to sending only three channels of up to six primaries at a time. In one embodiment, it does this by using a stacking system where either the color channel or the complementary channel is decoded depending on the bit level of that one channel.
[00276] SYSTEM 2
[00277] FIG. 7 illustrates a sequential method where three color primaries are passed to the transport format as full bit level image data and inserted as normal (“System 2”). The three additional channels are delayed by one pixel and then placed into the transport instead of the first colors. This method is useful in situations where quantizing artifacts is critical to image performance. In one embodiment, this system is comprised of six primaries (RGBCMY), a method to delay the CMY colors for injection, image resolution identification to all for pixel count synchronization, start of video identification, RGB delay, and for YCCCCC systems, logic to select the dominant color primary. The advantage of System 2 is that full bit level video can be transported, but at double the normal data rate.
[00278] SYSTEM 2A
[00279] System 2 sequences on a pixel -to-pixel basis. However, a quadrature method is also possible (“System 2A”) that is operable to transport six primaries in stereo or twelve primary image information. Each quadrant of the frame contains three color primary data sets. These are combined in the display. A first set of three primaries is displayed in the upper left quadrant, a second set of three primaries is displayed in the upper right quadrant, a third set of primaries is displayed in the lower left quadrant, and a fourth set of primaries is displayed in lower right quadrant. In one embodiment, the first set of three primaries, the second set of three primaries, the third set of three primaries, and the fourth set of three primaries do not contain any overlapping primaries (i.e., twelve different primaries). Alternatively, the first set of three primaries, the second set of three primaries, the third set of three primaries, and the fourth set of three primaries contain overlapping primaries (i.e., at least one primary is contained in more than one set of three primaries). In one embodiment, the first set of three primaries and the third set of three primaries contain the same primaries and the second set of three primaries and the fourth set of three primaries contain the same primaries.
[00280] FIG. 113A illustrates one embodiment of a quadrature method (“System 2A”). In the example shown in FIG. 113A, a first set of three primaries (e.g., RGB) is displayed in the upper left quadrant, a second set of three primaries (e.g., CMY) is displayed in the upper right quadrant, a third set of three primaries (e.g., GC, BM, and RY) is displayed in the lower left quadrant, and a fourth set of three primaries (e.g., MR, YG, and CB) is displayed in the lower right quadrant. Although the example shown in FIG. 113A illustrates a backwards compatible 12P system, this is merely for illustrative purposes. The present invention is not limited to the twelve primaries shown in FIG. 113 A. Additionally, alternative pixel arrangements are compatible with the present invention.
[00281] FIG. 113B illustrates another embodiment of a quadrature method (“System 2A”). In the example shown in FIG. 113B, a first set of three primaries (e.g., RGB) is displayed in the upper left quadrant, a second set of three primaries (e.g., CMY) is displayed in the upper right quadrant, a third set of three primaries (e.g., GC, BM, and RY) is displayed in the lower left quadrant, and a fourth set of three primaries (e.g., MR, YG, and CB) is displayed in the lower right quadrant. Although the example shown in FIG. 8B illustrates a backwards compatible 12P system, this is merely for illustrative purposes. The present invention is not limited to the twelve primaries shown in FIG. 113B. Additionally, alternative pixel arrangements are compatible with the present invention.
[00282] FIG. 113C illustrates yet another embodiment of a quadrature method (“System 2A”). In the example shown in FIG. 113C, a first set of three primaries (e.g., RGB) is displayed in the upper left quadrant, a second set of three primaries (e.g., CMY) is displayed in the upper right quadrant, a third set of three primaries (e.g., GC, BM, and RY) is displayed in the lower left quadrant, and a fourth set of three primaries (e.g., MR, YG, and CB) is displayed in the lower right quadrant. Although the example shown in FIG. 113C illustrates a backwards compatible 12P system, this is merely for illustrative purposes. The present invention is not limited to the twelve primaries shown in FIG. 113C. Additionally, alternative pixel arrangements are compatible with the present invention.
[00283] FIG. 114A illustrates an embodiment of a quadrature method (“System 2A”) in stereo. In the example shown in FIG. 114A, a first set of three primaries (e.g., RGB) is displayed in the upper left quadrant, a second set of three primaries (e.g., CMY) is displayed in the upper right quadrant, a third set of three primaries (e.g., RGB) is displayed in the lower left quadrant, and a fourth set of three primaries (e.g., CMY) is displayed in the lower right quadrant. This embodiment allows for separation of the left eye with the first set of three primaries and the second set of three primaries and the right eye with the third set of three primaries and the fourth set of three primaries. Alternatively, a first set of three primaries (e.g., RGB) is displayed in the upper left quadrant, a second set of three primaries (e.g., RGB) is displayed in the upper right quadrant, a third set of three primaries (e.g., CMY) is displayed in the lower left quadrant, and a fourth set of three primaries (e.g., CMY) is displayed in the lower right quadrant. Alternative pixel arrangements and primaries are compatible with the present invention. [00284] FIG. 114B illustrates another embodiment of a quadrature method (“System 2 A”) in stereo. Alternative pixel arrangements and primaries are compatible with the present invention.
[00285] FIG. 114C illustrates yet another embodiment of a quadrature method (“System 2A”) in stereo. Alternative pixel arrangements and primaries are compatible with the present invention.
[00286] Advantageously, System 2A allows for the ability to display multiple primaries (e.g., 12P and 6P) on a conventional monitor. Additionally, System 2A allows for a simplistic viewing of false color, which is useful in the production process and allows for visualizing relationships between colors. It also allows for display of multiple projectors (e.g., a first projector, a second projector, a third projector, and a fourth projector).
[00287] SYSTEM 3
[00288] FIG. 8 illustrates one embodiment of a system encode and decode process using a dual link method (“System 3”). System 3 utilizes a dual link method where two wires are used. In one embodiment, RGB is sent to link A and non-RGB primaries (e.g., CMY) are sent to link B. After arriving at the image destination, the two links are recombined. Alternative primaries are compatible with the present invention.
[00289] System 3 is simpler and more straight forward than Systems 1 and 2. The advantage with this system is that adoption is simply to format non-RGB primaries (e.g., CMY) on a second link. So, in one example, for an SDI design, RGB is sent on a standard SDI stream just as it is currently done. There is no modification to the transport and this link is operable to be sent to any RGB display requiring only the compensation for the luminance difference because the non-RGB primaries (e.g., CMY components) are not included. Data for the non-RGB primaries (e.g., CMY data) is transported in the same manner as RGB data. This data is then combined in the display to make up a 6P image. The downside is that the system requires two wires to move one image. This system is operable to work with most any format including SMPTE ST292, 424, 2082, and 2110. It also is operable to work with dual HDMI/CTA connections. In one embodiment, the system includes at least one transfer function (e.g., OETF, EOTF).
[00290] FIG. 9 illustrates one embodiment of an encoding process using a dual link method. Alternative numbers of primaries and alternative primaries are compatible with the present invention.
[00291] FIG. 10 illustrates one embodiment of a decoding process using a dual link method. Alternative numbers of primaries and alternative primaries are compatible with the present invention.
[00292] SYSTEM 4
[00293] Color is generally defined by three component data levels (e.g., RGB, YCbCr). A serial data stream must accommodate a word for each color contributor (e.g., R, G, B). Use of more than three primaries requires accommodations to fit this data based on an RGB concept. This is why System 1, System 2, and System 3 use stacking, sequencing, and/or dual links. Multiple words are required to define a single pixel, which is inefficient because not all values are needed.
[00294] In a preferred embodiment, color is defined as a colorimetric coordinate. Thus, every color is defined by three words. Serial systems are already based on three color contributors (e.g., RGB). System 4 preferably uses XYZ or Yxy as the three color contributors. System 4 preferably uses two colorimetric coordinates and a luminance or a luma. In one embodiment, System 4 includes, but is not limited to, Yxy, L*a*b*, ICTCP, YCbCr, YUV, Yu'v', YPbPr, YIQ, and/or XYZ. In a preferred embodiment, System 4 uses color contributors that are independent of a white point and/or a reference white value. Alternatively, System 4 uses color contributors that are not independent of a white point and/or a reference white value (e.g., YCbCr, L*a*b*). In another embodiment, System 4 uses color contributors that require at least one known primaries (e.g., ICTCP). In yet another embodiment, L*C*h or other non-rectangular coordinate systems (e.g., cylindrical, polar) are compatible with the present invention. In one embodiment, a polar system is defined from Yxy by converting x,y to a hue angle (e.g., 0 = arctan(y/x)) and a magnitude vector (e.g., r) that is similar to C* in an L*C*h polar system. However, when converting Yxy to a polar system, 0 is restricted from 0 to 90 degrees because x and y are always non-negative. In one embodiment, the 0 angle is expanded by applying a transform (e.g., an affine transform) to x, y data wherein the x, y values of the white point of the system (e.g., D65) are subtracted from the x, y data such that the x, y data includes negative values. Thus, 0 ranges from 0 to 360 degrees and the polar plot of the Yxy data is operable to occupy more than one quadrant. [00295] XYZ has been used in cinema for over 10 years. XYZ needs 16-bit float and 32- bit float encode or a minimum of 12 bits for gamma or log encoded images for better quality. Transport of XYZ must be accomplished using a 4:4:4 sample system. Less than a 4:4:4 sample system causes loss of image detail because Y is used as a coordinate along with X and Z and carries color information, not a value. Further, X and Z are not orthogonal to Y and, therefore, also include luminance information. Advantageously, converting to Yxy or Yu'v' concentrates the luminance in Y only, leaving two independent and pure chromaticity values. In one embodiment, X, Y, and Z are used to calculate x and y. Alternatively, X, Y, and Z are used to calculate u' and v'.
[00296] However, if Y or an equivalent component is used as a luminance value with two independent colorimetric coordinates (e.g., x and y, u' and v', u and v, etc.) used to describe color, then a system using subsampling is possible because of differing visual sensitivity to color and luminance. In one embodiment, I or L* components are used instead of Y, wherein I and/or L* data are created using gamma functions. As anon-limiting example, I is created using a 0.5 gamma function, while L* is created using a 1/3 gamma function. In these embodiments, additional gamma encoding is not applied to the data as part of transport. The system is operable to use any two independent colorimetric coordinates with similar properties to x and y, u’ and v’ , and/or u and v. In a preferred embodiment, the two independent colorimetric coordinates are x and y and the system is a Yxy system. In another preferred embodiment, the two colorimetric coordinates are u' and v' and the system is a Yu'v' system. Advantageously, the two independent colorimetric coordinates (e.g., x and y) are independent of a white point. Further, this reduces the complexity of the system when compared to XYZ, which includes a luminance value for all three channels (i.e., X, Y, and Z). Further, this also provides an advantage for subsampling (e.g., 4:2:2, 4:2:0 and 4:1: 1). In one embodiment, other systems (e.g., ICTCP and L*a*b*) require a white point in calculations. However, a conversion matrix using the white point of [1,1,1] is operable to be used for ICTCP and L*a*b*, which would remove the white point reference. The white point reference is operable to then be recaptured because it is the white point of [1,1,1] in XYZ space. In a preferred embodiment, the image data includes a reference to at least one white point.
[00297] Current technology uses components derived from the legacy NTSC television system. Encoding described in SMPTE, ITU, and CTA standards includes methods using subsampling as 4:2:2, 4:2:0, and 4:1:1. Advantageously, this allows for color transportation of more than three primaries, including, but not limited to, at least four primaries, at least five primaries, at least six primaries, at least seven primaries, at least eight primaries, at least nine primaries, at least ten primaries, at least eleven primaries, and/or at least twelve primaries (e.g., through a SMPTE ST292 or an HDMI 1.2 transport).
[00298] System 1, System 2, and System 3 use a YCbCr expansion to transport six color primary data sets, and the same transport (e.g., a YCbCr expansion) is operable to accommodate the image information as Yxy where Y is the luminance information and x,y describe CIE 1931 color coordinates in the half sample segments of the data stream (e.g., 4:2:2). Alternatively, x,y are fully sampled (e.g., 4:4:4). In yet another embodiment, the sampling rate is 4:2:0 or 4: 1 : 1. In still another embodiment, the same transport is operable to accommodate the information as luminance and colorimetric coordinates other than x,y. In one embodiment, the same transport is operable to accommodate data set using one channel of luminance data and two channels of colorimetric data. Alternatively, the same transport is operable to accommodate the image information as Yu'v' with full sampling (e.g., 4:4:4) or partial sampling (e.g., 4:2:2, 4:2:0, 4: 1:1). In one embodiment, the same transport is used with full sampling (e.g., XYZ).
[00299] Advantageously, there is no need to add more channels, nor is there any need to separate the luminance information from the color components. Further, for example, x,y have no reference to any primaries because x,y are explicit colorimetric positions. In the Yxy space, x and y are chromaticity coordinates such that x and y can be used to define a gamut of visible color. Similarly, in the Yu'v' space, u' and v' are explicit colorimetric positions. It is possible to define a gamut of visible color in other formats (e.g., L*a*b*, ICTCP, YCbCr), but it is not always trivial. To determine if a color is visible in Yxy space, it must be determined if the sum of x and y is greater than or equal to zero. If not, the color is not visible. If the x,y point is within the CIE x,y locus (CIE horseshoe), the color is visible. If not, the color is not visible. The Y value plays a role especially in a display. In one embodiment, the display is operable to reproduce an x,y color within a certain range of Y values, wherein the range is a function of the primaries. Another advantage is that an image can be sent as linear data (e.g., without anon-linear function applied) with anon-linear function (e.g., opto- optical transfer function (OOTF)) added after the image is received, rather than requiring a non-linear function (e.g., OOTF) applied to the signal. This allows for a much simpler encode and decode system. In one embodiment, only Y, L*, or I are altered by a non-linear function.
Alternatively, Y, L*, or I are sent linearly (e.g., without anon-linear function applied). [00300] FIG. 115 illustrates one embodiment of a Yxy encode with an opto-electronic transfer function (OETF). Image data is acquired in any format operable to be converted to XYZ data (e.g., RGB, RGBCMY, CMYK). The XYZ data is then converted to Yxy data, and the Yxy data is processed through an OETF. The processed Yxy data is then converted to a standardized transportation format for mapping and readout. Advantageously, x and y remain as independent colorimetric coordinates and the non-linear function (e.g., OETF, log, gamma, PQ) is only applied to Y, thus avoiding compression or loss of colorimetric data. In one embodiment, the OETF is described in ITU-R BT.2100 or ITU-R BT.1886. Advantageously, Y is orthogonal to x and y, and remains orthogonal to x and y even when a non-linear function is applied. Although the example shown includes Yxy data, System 4 is compatible with a plurality of data formats including data formats using one luminance coordinate and two colorimetric coordinates.
[00301] There are many different RGB sets so the matrix used to convert the image data from a set of RGB primaries to XYZ will involve a specific solution given the RGB values:
[00302] In an embodiment where the image data is 6P-B data, the following equation is used to convert to XYZ data: [00303] In an embodiment where the image data is 6P-C data with a D60 white point, the following equation is used to convert to XYZ data:
[00306] FIG. 116 illustrates one embodiment of a Yxy encode without an OETF. Image data is acquired in any format operable to be converted to XYZ data (e.g., RGB, RGBCMY,
CMYK). The XYZ data is then converted to Yxy data, and then converted to a standardized transportation format for mapping and readout. Although the example in FIG. 14 shows a Yxy encode, System 4 is operable to be used with a plurality of data formats.
[00307] FIG. 117 illustrates one embodiment of a Yxy decode with an electro-optical transfer function (EOTF). After mapping and readout, the data is processed through an EOTF to yield the Yxy data. The Yxy data is then converted back to the XYZ data. The XYZ data is operable to be converted to multiple data formats including, but not limited to, RGB, CMYK,
6P (e.g., 6P-B, 6P-C), and gamuts including at least four primaries through at least twelve primaries. Although the example in FIG. 117 shows a Yxy decode, System 4 is operable to be used with a plurality of data formats.
[00308] Finally, the XYZ data must converted to the correct standard color space. In an embodiment where the color gamut used is a 6P-B color gamut, the following equations are used:
[00309] In an embodiment where the color gamut used is a 6P-C color gamut with a D60 white point, the following equations are used:
[00310] In another embodiment where the color used is a 6P-C color gamut with a D65 white point, the following equations are used:
[00311] In an embodiment where the color gamut used is an ITU-R BT709.6 color gamut, the matrices are as follows: [00312] In an embodiment where the color gamut used is a SMPTE RP431-2 color gamut, the matrices are as follows:
[00313] In an embodiment where the color gamut used is an ITU-R BT.2020/2100 color gamut, the matrices are as follows:
[00314] To convert the Yxy data to the XYZ data, the following equations are used:
[00315] FIG. 118 illustrates one embodiment of a Yxy decode without an EOTF. After mapping and readout, the Yxy data is then converted to the XYZ data. The XYZ data is operable to be converted to multiple data formats including, but not limited to, RGB, CMYK, 6P (e.g., 6P-B, 6P-C), and gamuts including at least four primaries through at least twelve primaries. Although the example in FIG. 118 shows a Yxy encode, System 4 is operable to be used with a plurality of data formats.
[00316] FIG. 119 illustrates one embodiment of a 4:2:2 Yxy encode with an OETF. A full bandwidth luminance channel (T) is used to carry image detail and the remaining color coordinate components (e.g., x,y) are half sampled. In the example shown in FIG. 119, the Yxy data undergoes a 4:2:2 encode. Other encoding methods (e.g., 4:4:4, 4:2:0, 4:1:1) are compatible with the present invention. Other quantization methods and bit depths are also compatible with the present invention. In one embodiment, the bit depth is 8 bits, 10 bits, 12 bits, 14 bits, and/or 16 bits. In one embodiment, the Yxy values are sampled as floats.
Although the example in FIG. 119 shows a Yxy decode, System 4 is operable to be used with a plurality of data formats. [00317] FIG. 120 illustrates one embodiment of a 4:2:2 Yxy encode without an OETF. In the example shown in FIG. 120, the Yxy data undergoes a 4:2:2 encode. Other encoding methods (e.g., 4:4:4, 4:2:0, 4:1:1) are compatible with the present invention. Although the example in FIG. 120 shows a Yxy encode, System 4 is operable to be used with a plurality of data formats.
[00318] FIG. 121 illustrates one embodiment of a 4:4:4 Yxy encode with an OETF. A full bandwidth luminance channel (T) is used to carry image detail and the remaining color coordinate components (e.g., x,y) are also fully sampled. In the example shown in FIG. 121, the Yxy data undergoes a 4:4:4 encode. Other encoding methods (e.g., 4:2:2, 4:2:0, 4:1:1) are compatible with the present invention. Although the example in FIG. 121 shows a Yxy encode, System 4 is operable to be used with a plurality of data formats.
[00319] FIG. 122 illustrates one embodiment of a 4:4:4 Yxy encode without an OETF. In the example shown in FIG. 122, the Yxy data undergoes a 4:4:4 encode. Other encoding methods (e.g., 4:2:2, 4:2:0, 4:1:1) are compatible with the present invention. Although the example in FIG. 122 shows a Yxy encode, System 4 is operable to be used with a plurality of data formats.
[00320] FIG. 123 illustrates sample placements of Yxy system components for a 4:2:2 pixel mapping. A plurality of pixels (e.g., P00-P35) is shown in FIG. 123. The first subscript number refers to a row number and the second subscript number refers to a column number. For pixel Poo, E/NT00 is the luma and the color components are xINT00 and y/JVT00. For pixel P01, F/NT01 is the luma. For pixel Pio, F/NT10 is the luma and the color components are x]NT10 and yiNTio- For pixel Pn, F/NT11 is the luma. In one embodiment, the luma and the color components (e.g., the set of image data) corresponding to a particular pixel (e.g., Poo) is used to calculate color and brightness of subpixels. Although the example shown in FIG. 123 includes luma, it is equally possible that the data is sent linearly as luminance (e.g., YINTOO). Further, although the example in FIG. 123 includes Yxy system components, System 4 is operable to be used with a plurality of data formats.
[00321] FIG. 124 illustrates sample placements of Yxy system components for a 4:2:0 pixel mapping. A plurality of pixels (e.g., P00-P35) is shown in FIG. 124. The first subscript number refers to a row number and the second subscript number refers to a column number. For pixel Poo, Y/NT00 is the luma and the color components are xINT00 and yWTOo- For pixel P01, F/NT01 is the luma. For pixel Pio, F/NT10 is the luma. For pixel Pn, F/NT11 is the luma. In one embodiment, the luma and the color components corresponding to a particular pixel (e.g., Poo) is used to calculate color and brightness of subpixels. Although the example shown in FIG. 124 includes luma, it is equally possible that the data is sent linearly as luminance (e.g., YINTOO). Further, Although the example in FIG. 124 includes Yxy system components, System 4 is operable to be used with a plurality of data formats.
[00322] In one embodiment, the set of image data includes pixel mapping data. In one embodiment, the pixel mapping data includes a subsample of the set of values in a color space. In a preferred embodiment, the color space is a Yxy color space (e.g., 4:2:2). In one embodiment, the pixel mapping data includes an alignment of the set of values in the color space (e.g., Yxy color space, Yu'v').
[00323] Table 6 illustrates mapping to SMPTE ST2110 for 4:2:2 sampling of Yxy data. Table 7 illustrates mapping to SMPTE ST2110 for 4:4:4 linear and non-linear sampling of Yxy data. The present invention is compatible with a plurality of data formats (e.g., Yu'v') and not restricted to Yxy data.
[00324] TABLE 6 [00325] TABLE 7
[00326] FIG. 125 illustrates one embodiment of a SMPTE ST292 Yxy system mapping.
To fit a Yxy system into a SMPTE ST292 stream involves the following substitutions: Y/NT is placed in the Y data segments, xINT is placed in the Cr data segments, and yINT is placed in the Cb data segments. In a preferred embodiment, luminance or luma is placed in the Y data segments, a first colorimetric coordinate is placed in the Cr data segments, and a second colorimetric coordinate is placed in the Cb data segments. Although the example in FIG. 125 shows a Yxy system mapping, System 4 is operable to be used with a plurality of data formats (e.g., Yu'v').
[00327] FIG. 126 illustrates one embodiment of a SMPTE ST2082 Yxy system mapping. To fit a Yxy system into a SMPTE ST292 stream involves the following substitutions: Y/NT is placed in the G data segments, x]NT is placed in the R data segments, and yINT is placed in the B data segments. In a preferred embodiment, luminance or luma is placed in the G data segments, a first colorimetric coordinate is placed in the R data segments, and a second colorimetric coordinate is placed in the B data segments. Although the example in FIG. 126 shows a Yxy system mapping, System 4 is operable to be used with a plurality of data formats (e.g., Yu'v'). [00328] FIG. 127 illustrates one embodiment of Yxy inserted into a CTA 861 stream. Although the example in FIG. 127 shows a Yxy system mapping, System 4 is operable to be used with a plurality of data formats.
[00329] FIG. 128 illustrates one embodiment of a Yxy decode with an EOTF. In one embodiment, a non-linear function is applied to the luminance to create a luma. The nonlinear function is not applied to the two colorimetric coordinates. Although the example in FIG. 128 shows a Yxy decode, System 4 is operable to be used with a plurality of data formats.
[00330] FIG. 129 illustrates one embodiment of a Yxy decode without an EOTF. In one embodiment, data is sent linearly as luminance. A non-linear function (e.g., EOTF) is not applied to the luminance or the two colorimetric coordinates. Although the example in FIG. 129 shows a Yxy decode, System 4 is operable to be used with a plurality of data formats. [00331] Advantageously, XYZ is used as the basis of ACES for cinematographers and allows for the use of colors outside of the ITU-R BT.709 and/or the P3 color spaces, encompassing all of the CIE color space. Colorists often work in XYZ, so there is widespread familiarity with XYZ. Further, XYZ is used for other standards (e.g., JPEG 2000, Digital Cinema Initiatives (DCI)), which could be easily adapted for System 4. Additionally, most color spaces use XYZ as the basis for conversion, so the conversions between XYZ and most color spaces are well understood and documented. Many professional displays also have XYZ option as a color reference function.
[00332] In one embodiment, the image data converter includes at least one look-up table (LUT). In one embodiment, the at least one look-up table maps out of gamut colors to zero. In one embodiment, the at least one look-up table maps out of gamut colors to a periphery of visible colors.
[00333] TRANSFER FUNCTIONS [00334] The system design minimizes limitations to use standard transfer functions for both encode and/or decode processes. Current practices used in standards include, but are not limited to, ITU-R BT.1886, ITU-R BT.2020, SMPTE ST274, SMPTE ST296, SMPTE ST2084, and ITU-R BT.2100. These standards are compatible with this system and require no modification.
[00335] Encoding and decoding multi-primary (e.g., 6P, RGBC) images is formatted into several different configurations to adapt to image transport frequency limitations. The highest quality transport is obtained by keeping all components as multi-primary (e.g., RGBCMY) components. This uses the highest sampling frequencies and requires the most signal bandwidth. An alternate method is to sum the image details in a luminance channel at full bandwidth and then send the color difference signals at half or quarter sampling (e.g., Y Cr Cb Cc Cy). This allows a similar image to pass through lower bandwidth transports.
[00336] An IPT system is a similar idea to the Yxy system with several exceptions. An IPT system or an ICTCP system is still an extension of XYZ and is operable to be derived from RGB and multiprimary (e.g., RGBCMY, RGBC) color coordinates. An IPT color description can be substituted within a 4:4:4 sampling structure, but XYZ has already been established and does not require the same level of calculations. For an ICTCP transport system, similar substitutions can be made. However, both substitution systems are limited in that anon-linear function (e.g., OOTF) is contained in all three components. Although the non-linear function can be removed for IPT or ICTCP, the derivation would still be based on a set of RGB primaries with a white point reference. Removing the non-linear function may also alter the bit depth noise and compressibility.
[00337] For transport, simple substitutions can be made using the foundation of what is described with transport of XYZ for the use of IPT in current systems as well as the current standards used for ICTCP. [00338] FIG. 130A illustrates one embodiment of an IPT 4:4:4 encode.
[00339] FIG. 130B illustrates one embodiment of an IPT 4:4:4 decode.
[00340] FIG. 131 A illustrates one embodiment of an ICTCP 4:2:2 encode.
[00341] FIG. 13 IB illustrates one embodiment of an ICTCP 4:2:2 decode.
[00342] Transfer functions used in systems 1, 2, and 3 are generally framed around two basic implementations. For images displaying using a standard dynamic range, the transfer functions are defined within two standards. The OETF is defined in ITU-R BT.709-6, table 1, row 1.2. The inverse function, the EOTF, is defined in ITU-R BT.1886. For high dynamic range imaging, the perceptual quantizer (PQ) and hybrid log-gamma (HLG) curves are described in ITU-R BT.2100-2: 2018, table 4.
[00343] System 4 is operable to use any of the transfer functions, which can be applied to the Y component. However, to improve compatibility and to simplify conversion between standard transfer functions, a new method has been developed: a 'A gamma function. Advantageously, the A gamma function allows for a single calculation from the luminance (e.g., Y) component of the signal (e.g., Yxy signal) to the display. Advantageously, the A gamma function is designed for data efficiency, not as an optical transform function. In one embodiment, the A gamma function is used instead of a nonlinear function (e.g., OETF or EOTF). In one embodiment, signal input to the A gamma function is assumed to be linear and constrained between values of 0 and 1. In one embodiment, the A gamma function is optimized for 10-bit transport and/or 12-bit transport. Alternatively, the A gamma function is optimized for 14-bit transport and/or 16-bit transport. In an alternative embodiment, the A gamma function is optimized for 8-bit transport. A typical implementation applies an inverse of the A gamma function, which linearizes the signal. A conversion to a display gamut is then applied.
[00344] FIG. 134 illustrates one embodiment of a A gamma function. [00345] In one embodiment, for a source n = y/L and for a display L = n2. In another embodiment, a display gamma is calculated as L = n ^, where A is a desired final EOTF. Advantageously, using the 'A gamma function with the display gamma combines the functions into a single step rather than utilizing a two-step conversion process. In one embodiment, at least one tone curve is applied after the A gamma function. The A gamma function advantageously provides ease to convert to and from linear values. Given that all color and tone mapping has to be done in the linear domain, having a simple to implement conversion is desirable and makes the conversion to and from linear values easier and simpler.
[00346] FIG. 135 illustrates a graph of maximum quantizing error using the A gamma function. The maximum quantizing error from an original 16-bit image (black trace) to a 10- bit (blue trace) signal is shown in the graph. In the embodiment shown in the graph, the maximum quantizing error is less than 0.1% (e.g., 0.0916%) for 16-bit to 10-bit conversion using the A gamma function. This does not include any camera log functions designed into a camera. The graph also shows the maximum quantizing error from the original 16-bit image to a 12-bit (red trace) signal and a 14-bit (green trace) signal.
[00347] While a A gamma is ideal for converting images with 16-bit (e.g., 16-bit float) values to 12-bit (e.g., 12-bit integer) values, for other data sets a 1/3 gamma provides equivalent performance in terms of peak signal-to-noise ratio (PSNR). For HDR content, which has a wider luminance dynamic range (e.g., up to 1000 cd/m2), the 1/3 gamma conversion from 16-bit float maintains the same performance as A gamma. In one embodiment, an equation for finding an optimum value of gamma is: [00348] In one embodiment, the Minimum Float Value is based on the IEEE Standard for
Floating-Point Arithmetic (IEEE 754) (July 2019), which is incorporated herein by reference in its entirety. In one embodiment, the range of image values is normalized to between 0 and 1. The range of image values is preferably normalized to between 0 and 1 and then the gamma function is applied.
[00349] For example, for an HDR system (e.g., with a luminance dynamic range of 1000- 4000 cd/m2), the above equation becomes:
[00350] FIG. 108 illustrates one embodiment of a 1/3 gamma function.
[00351] SIX-PRIMARY COLOR ENCODE USING A 4:4:4 SAMPLING METHOD [00352] FIG. 11 illustrates one embodiment of a six-primary color system encode using a 4:4:4 sampling method.
[00353] Subjective testing during the development and implementation of the current digital cinema system (DCI Version 1.2) showed that perceptible quantizing artifacts were not noticeable with system bit resolutions higher than 11 bits. Current serial digital transport systems support 12 bits. Remapping six color components to a 12-bit stream is accomplished by lowering the bit limit to 11 bits (values 0 to 2047) for 12-bit serial systems or 9 bits (values 0 to 512) for 10-bit serial systems. This process is accomplished by processing multiprimary (e.g., RGBCMY) video information through a standard Optical Electronic Transfer Function (OETF) (e.g., ITU-R BT.709-6), digitizing the video information as four samples per pixel, and quantizing the video information as 11 -bit or 9-bit.
[00354] In another embodiment, the multi-primary (e.g., RGBCMY) video information is processed through a standard Optical Optical Transfer Function (OOTF). In yet another embodiment, the multi-primary (e.g., RGBCMY) video information is processed through a Transfer Function (TF) other than OETF or OOTF. TFs consist of two components, a Modulation Transfer Function (MTF) and a Phase Transfer Function (PTF). The MTF is a measure of the ability of an optical system to transfer various levels of detail from object to image. In one embodiment, performance is measured in terms of contrast (degrees of gray), or of modulation, produced for a perfect source of that detail level. The PTF is a measure of the relative phase in the image(s) as a function of frequency. A relative phase change of 180°, for example, indicates that black and white in the image are reversed. This phenomenon occurs when the TF becomes negative.
[00355] There are several methods for measuring MTF. In one embodiment, MTF is measured using discrete frequency generation. In one embodiment, MTF is measured using continuous frequency generation. In another embodiment, MTF is measured using image scanning. In another embodiment, MTF is measured using waveform analysis.
[00356] In one embodiment, the six-primary color system is for a 12-bit serial system. Current practices normally set black at bit value 0 and white at bit value 4095 for 12-bit video. In order to package six colors into the existing three-serial streams, the bit defining black is moved to bit value 2048. Thus, the new encode has RGB values starting at bit value 2048 for black and bit value 4095 for white and non-RGB primary (e.g., CMY) values starting at bit value 2047 for black and bit value 0 as white. In another embodiment, the six- primary color system is for a 10-bit serial system.
[00357] FIG. 12 illustrates one embodiment for a method to package six channels of primary information into the three standard primary channels used in current serial video standards by modifying bit numbers for a 12-bit SDI and a 10-bit SDI. FIG. 13 illustrates a simplified diagram estimating perceived viewer sensation as code values define each hue angle. TABLE 8 and TABLE 9 list bit assignments for computer, production, and broadcast for a 12-bit system and a 10-bit system, respectively. In one embodiment, “Computer” refers to bit assignments compatible with CTA 861-G, November 2016, which is incorporated herein by reference in its entirety. In one embodiment, “Production” and/or “Broadcast” refer to bit assignments compatible with SMPTE ST 2082-0 (2016), SMPTE ST 2082-1 (2015), SMPTE ST 2082-10 (2015), SMPTE ST 2082-11 (2016), SMPTE ST 2082-12 (2016), SMPTE ST 2110-10 (2017), SMPTE ST 2110-20 (2017), SMPTE ST 2110-21 (2017), SMPTE ST 2110-30 (2017), SMPTE ST 2110-31 (2018), and/or SMPTE ST 2110-40 (2018), each of which is incorporated herein by reference in its entirety.
[00358] TABLE 8: 12-B it Assignments
[00360] In one embodiment, the OETF process is defined in ITU-R BT.709-6, which is incorporated herein by reference in its entirety. In one embodiment, the OETF process is defined in ITU-R BT.709-5, which is incorporated herein by reference in its entirety. In another embodiment, the OETF process is defined in ITU-R BT.709-4, which is incorporated herein by reference in its entirety. In yet another embodiment, the OETF process is defined in ITU-R BT.709-3, which is incorporated herein by reference in its entirety. In yet another embodiment, the OETF process is defined in ITU-R BT.709-2, which is incorporated herein by reference in its entirety. In yet another embodiment, the OETF process is defined in ITU-
R BT.709-1, which is incorporated herein by reference in its entirety.
[00361] In one embodiment, the encoder is a non-constant luminance encoder. In another embodiment, the encoder is a constant luminance encoder.
[00362] SIX-PRIMARY COLOR PACKING/STACKING USING A 4:4:4 SAMPLING METHOD
[00363] FIG. 14 illustrates one embodiment for a method of stacking/encoding six-primary color information using a 4:4:4 video system. Image data must be assembled according the serial system used. This is not a conversion process, but instead is a packing/ stacking process. In one embodiment, the packing/ stacking process is for a six-primary color system using a 4:4:4 sampling method.
[00364] FIG. 15 illustrates one embodiment for a method of unstacking/decoding six- primary color information using a 4:4:4 video system. In one embodiment, the RGB channels and the non-RGB primary (e.g., CMY) channels are combined into one 12-bit word and sent to a standardized transport format. In one embodiment, the standardized transport format is SMPTE ST424 SDI. In one embodiment, the decode is for a non-constant luminance, six- primary color system. In another embodiment, the decode is for a constant luminance, six- primary color system. In yet another embodiment, an electronic optical transfer function (EOTF) (e.g., ITU-R BT.1886) coverts image data back to linear for display. In one embodiment, the EOTF is defined in ITU-R BT.1886 (2011), which is incorporated herein by reference in its entirety. FIG. 16 illustrates one embodiment of a 4:4:4 decoder.
[00365] System 2 uses sequential mapping to the standard transport format, so it includes a delay for the non-RGB (e.g., CMY) data. The non-RGB (e.g., CMY) data is recovered in the decoder by delaying the RGB data. Since there is no stacking process, the full bit level video can be transported. For displays that are using optical filtering, this RGB delay could be removed and the process of mapping image data to the correct filter could be eliminated by assuming this delay with placement of the optical filter and the use of sequential filter colors. [00366] Two methods can be used based on the type of optical filter used. Since this system is operating on a horizontal pixel sequence, some vertical compensation is required and pixels are rectangular. This can be either as a line double repeat using the same multiprimary (e.g., RGBCMY) data to fill the following line as shown in FIG. 17, or could be separated as RGB on line one and non-RGB (e.g., CMY) on line two as shown in FIG. 18. The format shown in FIG. 18 allows for square pixels, but the non-RGB (e.g., CMY) components require a line delay for synchronization. Other patterns eliminating the white subpixel are also compatible with the present invention.
[00367] FIG. 19 illustrates an embodiment of the present invention for sending six primary colors to a standardized transport format using a 4:4:4 encoder according to System 2.
Encoding is straight forward with a path for RGB sent directly to the transport format. RGB data is mapped to each even numbered data segment in the transport. Non-RGB (e.g., CMY) data is mapped to each odd numbered segment. Because different resolutions are used in all of the standardized transport formats, there must be identification for what they are so that the start of each horizontal line and horizontal pixel count can be identified to time the RGB/non- RGB (e.g., CMY) mapping to the transport. The identification is the same as currently used in each standardized transport function. TABLE 10, TABLE 11, TABLE 12, and TABLE 13 list 16-bit assignments, 12-bit assignments, 10-bit assignments, and 8-bit assignments, respectively. In one embodiment, “Computer” refers to bit assignments compatible with CTA 861 -G, November 2016, which is incorporated herein by reference in its entirety. In one embodiment, “Production” and/or “Broadcast” refer to bit assignments compatible with SMPTE ST 2082-0 (2016), SMPTE ST 2082-1 (2015), SMPTE ST 2082-10 (2015), SMPTE ST 2082-11 (2016), SMPTE ST 2082-12 (2016), SMPTE ST 2110-10 (2017), SMPTE ST 2110-20 (2017), SMPTE ST 2110-21 (2017), SMPTE ST 2110-30 (2017), SMPTE ST 2110-
31 (2018), and/or SMPTE ST 2110-40 (2018), each of which is incorporated herein by reference in its entirety.
[00368] TABLE 10: 16-Bit Assignments [00372] The decode adds a pixel delay to the RGB data to realign the channels to a common pixel timing. EOTF is applied and the output is sent to the next device in the system. Metadata based on the standardized transport format is used to identify the format and image resolution so that the unpacking from the transport can be synchronized. FIG. 20 shows one embodiment of a decoding with a pixel delay.
[00373] In one embodiment, the decoding is 4:4:4 decoding. With this method, the six- primary color decoder is in the signal path, where 11 -bit values for RGB are arranged above bit value 2048, while non-RGB (e.g., CMY) levels are arranged below bit value 2047 as libit. If the same data set is sent to a display and/or process that is not operable for six-primary color processing, the image data is assumed as black at bit value 0 as a full 12-bit word. Decoding begins by tapping image data prior to the unstacking process.
[00374] SIX-PRIMARY COLOR ENCODE USING A 4:2:2 SAMPLING METHOD [00375] In one embodiment, the packing/stacking process is for a six-primary color system using a 4:2:2 sampling method. In order to fit the new six-primary color system into a lower bandwidth serial system, while maintaining backwards compatibility, the standard method of converting from six primaries (e.g., RGBCMY) to a luminance and a set of color difference signals requires the addition of at least one new image designator. In one embodiment, the encoding and/or decoding process is compatible with transport through SMPTE ST 292-0 (2011), SMPTE ST 292-1 (2011, 2012, and/or 2018), SMPTE ST 292-2 (2011), SMPTE ST 2022-1 (2007), SMPTE ST 2022-2 (2007), SMPTE ST 2022-3 (2010), SMPTE ST 2022-4 (2011), SMPTE ST 2022-5 (2012 and/or 2013), SMPTE ST 2022-6 (2012), SMPTE ST 2022-7 (2013), and/or and CTA 861-G (2106), each of which is incorporated herein by reference in its entirety. [00376] In order for the system to package all of the image while supporting both six- primary and legacy displays, an electronic luminance component (Y) must be derived. The first component is: Ey6. For an RGBCMY system, itcan be described as:
[00377] Critical to getting back to legacy display compatibility, value E_' Y is described as:
[00378] In addition, at least two new color components are disclosed. These are designated as Cc and Cy components. The at least two new color components include a method to compensate for luminance and enable the system to function with older Y Cb Cr infrastructures. In one embodiment, adjustments are made to Cb and Cr in a Y Cb Cr infrastructure since the related level of luminance is operable for division over more components. These new components are as follows:
[00379] Within such a system, it is not possible to define magenta as a wavelength. This is because the green vector in CIE 1976 passes into, and beyond, the CIE designated purple line. Magenta is a sum of blue and red. Thus, in one embodiment, magenta is resolved as a calculation, not as optical data. In one embodiment, both the camera side and the monitor side of the system use magenta filters. In this case, if magenta were defined as a wavelength, it would not land at the point described. Instead, magenta would appear as a very deep blue which would include a narrow bandwidth primary, resulting in metameric issues from using narrow spectral components. In one embodiment, magenta as an integer value is resolved using the following equation:
[00380] The above equation assists in maintaining the fidelity of a magenta value while minimizing any metameric errors. This is advantageous over prior art, where magenta appears instead as a deep blue instead of the intended primary color value.
[00381] SIX-PRIMARY NON-CONSTANT LUMINANCE ENCODE USING A 4:2:2 SAMPLING METHOD
[00382] In one embodiment, the six-primary color system using a non-constant luminance encode for use with a 4:2:2 sampling method. In one embodiment, the encoding process and/or decoding process is compatible with transport through SMPTE ST 292-0 (2011), SMPTE ST 292-1 (2011, 2012, and/or 2018), SMPTE ST 292-2 (2011), SMPTE ST 2022-1 (2007), SMPTE ST 2022-2 (2007), SMPTE ST 2022-3 (2010), SMPTE ST 2022-4 (2011), SMPTE ST 2022-5 (2012 and/or 2013), SMPTE ST 2022-6 (2012), SMPTE ST 2022-7 (2013), and/or and CTA 861-G (2106), each of which is incorporated herein by reference in its entirety.
[00383] Current practices use a non-constant luminance path design, which is found in all the video systems currently deployed. FIG. 21 illustrates one embodiment of an encode process for 4:2:2 video for packaging five channels of information into the standard three- channel designs. For 4:2:2, a similar method to the 4:4:4 system is used to package five channels of information into the standard three-channel designs used in current serial video standards. FIG. 21 illustrates 12-bit SDI and 10-bit SDI encoding for a 4:2:2 system. TABLE 14 and TABLE 15 list bit assignments for a 12-bit and 10-bit system, respectively. In one embodiment, “Computer” refers to bit assignments compatible with CTA 861-G, November 2016, which is incorporated herein by reference in its entirety. In one embodiment, “Production” and/or “Broadcast” refer to bit assignments compatible with SMPTE ST 2082-0 (2016), SMPTE ST 2082-1 (2015), SMPTE ST 2082-10 (2015), SMPTE ST 2082-11 (2016), SMPTE ST 2082-12 (2016), SMPTE ST 2110-10 (2017), SMPTE ST 2110-20 (2017),
SMPTE ST 2110-21 (2017), SMPTE ST 2110-30 (2017), SMPTE ST 2110-31 (2018), and/or
SMPTE ST 2110-40 (2018), each of which is incorporated herein by reference in its entirety.
[00384] TABLE 14: 12-Bit Assignments
[00386] FIG. 22 illustrates one embodiment for a non-constant luminance encoding process for a six-primary color system. The design of this process is similar to the designs used in current RGB systems. Input video is sent to the Optical Electronic Transfer Function (OETF) process and then to the EYf encoder. The output of this encoder includes all of the image detail information. In one embodiment, all of the image detail information is output as a monochrome image.
[00387] The output is then subtracted from ER' , EB' , Ec' , and EY to make the following color difference components: ECR> EC'B> ECC> EC'Y
These components are then half sampled (x2) while E^ is fully sampled (x4).
[00388] FIG. 23 illustrates one embodiment of a packaging process for a six-primary color system. These components are then sent to the packing/ stacking process. Components EC'Y-INT ar|d ECC-INT are inverted so that bit 0 now defines peak luminance for the corresponding component. In one embodiment, this is the same packaging process performed with the 4:4:4 sampling method design, resulting in two 11 -bit components combining into one 12-bit component.
[00389] SIX-PRIMARY NON-CONSTANT LUMINANCE DECODE USING A 4:2:2 SAMPLING METHOD
[00390] FIG. 24 illustrates a 4:2:2 unstack process for a six-primary color system. In one embodiment, the image data is extracted from the serial format through the normal processes as defined by the serial data format standard. In another embodiment, the serial data format standard uses a 4:2:2 sampling structure. In yet another embodiment, the serial data format standard is SMPTE ST292. The color difference components are separated and formatted back to valid 11 -bit data. Components EC'Y-INT and E^C-INT are inverted so that bit value 2047 defines peak color luminance.
[00391] FIG. 25 illustrates one embodiment of a process to inversely quantize each individual color and pass the data through an electronic optical function transfer (EOTF) in a non-constant luminance system. The individual color components, as well as EY' _INT, are inversely quantized and summed to breakout each individual color. Magenta is then calculated and EY _INTis combined with these colors to resolve green. These calculations then go back through an Electronic Optical Transfer Function (EOTF) process to output the six- primary color system. [00392] In one embodiment, the decoding is 4:2:2 decoding. This decode follows the same principles as the 4:4:4 decoder. However, in 4:2:2 decoding, a luminance channel is used instead of discrete color channels. Here, image data is still taken prior to unstack from the EC'B-INT + EC'Y-INT ar|d ^CR-INT + ^CC-INT channels. With a 4:2:2 decoder, a new component, called ELY, is used to subtract the luminance levels that are present from the CMY channels from the EC'B-INT + EC'Y-INT and EBR-INT + EBC-INT components. The resulting output is now the R and B image components of the EOTF process. ELY is also sent to the G matrix to convert the luminance and color difference components to a green output. Thus, R’G’B’ is input to the EOTF process and output as GRGB, RRGB, and BRGB. In another embodiment, the decoder is a legacy RGB decoder for non-constant luminance systems.
[00393] In one embodiment, the standard is SMPTE ST292. In one embodiment, the standard is SMPTE RP431-2. In one embodiment, the standard is ITU-R BT.2020. In another embodiment, the standard is SMPTE RP431-1. In another embodiment, the standard is ITU-R BT.1886. In another embodiment, the standard is SMPTE ST274. In another embodiment, the standard is SMPTE ST296. In another embodiment, the standard is SMPTE ST2084. In yet another embodiment, the standard is ITU-R BT.2100. In yet another embodiment, the standard is SMPTE ST424. In yet another embodiment, the standard is SMPTE ST425. In yet another embodiment, the standard is SMPTE ST2110.
[00394] SIX-PRIMARY CONSTANT LUMINANCE DECODE USING A 4:2:2 SAMPLING METHOD
[00395] FIG. 26 illustrates one embodiment of a constant luminance encode for a six- primary color system. FIG. 27 illustrates one embodiment of a constant luminance decode for a six-primary color system. The process for constant luminance encode and decode are very similar. The main difference being that the management of E^ is linear. The encode and decode processes stack into the standard serial data streams in the same way as is present in a non-constant luminance, six-primary color system. In one embodiment, the stacker design is the same as with the non-constant luminance system.
[00396] System 2 operation is using a sequential method of mapping to the standard transport instead of the method in System 1 where pixel data is combined to two color primaries in one data set as an 11 -bit word. The advantage of System 1 is that there is no change to the standard transport. The advantage of System 2 is that full bit level video can be transported, but at double the normal data rate.
[00397] The difference between the systems is the use of two Y channels in System 2. In one embodiment, YRGB and YCMY are used to define the luminance value for RGB as one group and CMY for the other. Alternative primaries are compatible with the present invention.
[00398] FIG. 28 illustrates one example of 4:2:2 non-constant luminance encoding.
Because the RGB and CMY components are mapped at different time intervals, there is no requirement for a stacking process and data is fed directly to the transport format. The development of the separate color difference components is identical to System 1. Alternative primaries are compatible with the present invention.
[00399] The encoder for System 2 takes the formatted color components in the same way as System 1. Two matrices are used to build two luminance channels. YRGB contains the luminance value for the RGB color primaries. YCMY contains the luminance value for the CMY color primaries. A set of delays are used to sequence the proper channel for YRGB, YCMY, and the RBCY channels. Because the RGB and non-RGB (e.g., CMY) components are mapped at different time intervals, there is no requirement for a stacking process, and data is fed directly to the transport format. The development of the separate color difference components is identical to System 1. The Encoder for System 2 takes the formatted color components in the same way as System 1. Two matrices are used to build two luminance channels: YRGB contains the luminance value for the RGB color primaries and YCMY contains the luminance value for the CMY color primaries. This sequences YRGB, CR, and CC channels into the even segments of the standardized transport and YCMY, CB, and CY into the odd numbered segments. Since there is no combining color primary channels, full bit levels can be used limited only by the design of the standardized transport method. In addition, for use in matrix driven displays, there is no change to the input processing and only the method of outputting the correct color is required if the filtering or emissive subpixel is also placed sequentially.
[00400] Timing for the sequence is calculated by the source format descriptor which then flags the start of video and sets the pixel timing.
[00401] FIG. 29 illustrates one embodiment of a non-constant luminance decoding system. Decoding uses timing synchronization from the format descriptor and start of video flags that are included in the payload ID, SDP, or EDID tables. This starts the pixel clock for each horizontal line to identify which set of components are routed to the proper part of the decoder. A pixel delay is used to realign the color primarily data of each subpixel. YRGB and YCMY are combined to assemble a new Ye component which is used to decode the CR, CB, CC, CY, and CM components into RGBCMY.
[00402] The constant luminance system is not different from the non-constant luminance system in regard to operation. The difference is that the luminance calculation is done as a linear function instead of including the OOTF. FIG. 30 illustrates one embodiment of a 4:2:2 constant luminance encoding system. FIG. 31 illustrates one embodiment of a 4:2:2 constant luminance decoding system.
[00403] SIX-PRIMARY COLOR SYSTEM USING A 4:2:0 SAMPLING SYSTEM [00404] In one embodiment, the six-primary color system uses a 4:2:0 sampling system. The 4:2:0 format is widely used in H.262/MPEG-2, H.264/MPEG-4 Part 10 and VC-1 compression. The process defined in SMPTE RP2050-1 provides a direct method to convert from a 4:2:2 sample structure to a 4:2:0 structure. When a 4:2:0 video decoder and encoder are connected via a 4:2:2 serial interface, the 4:2:0 data is decoded and converted to 4:2:2 by up-sampling the color difference component. In the 4:2:0 video encoder, the 4:2:2 video data is converted to 4:2:0 video data by down-sampling the color difference component.
[00405] There typically exists a color difference mismatch between the 4:2:0 video data from the 4:2:0 video data to be encoded. Several stages of codec concatenation are common through the processing chain. As a result, color difference signal mismatch between 4:2:0 video data input to 4:2:0 video encoder and 4:2:0 video output from 4:2:0 video decoder is accumulated and the degradation becomes visible.
[00406] FILTERING WITHIN A SIX-PRIMARY COLOR SYSTEM USING A 4:2:0 SAMPLING METHOD
[00407] When a 4:2:0 video decoder and encoder are connected via a serial interface, 4:2:0 data is decoded and the data is converted to 4:2:2 by up-sampling the color difference component, and then the 4:2:2 video data is mapped onto a serial interface. In the 4:2:0 video encoder, the 4:2:2 video data from the serial interface is converted to 4:2:0 video data by down-sampling the color difference component. At least one set of filter coefficients exists for 4:2:074:2:2 up-sampling and 4:2:274:2:0 down-sampling. The at least one set of filter coefficients provide minimally degraded 4:2:0 color difference signals in concatenated operations.
[00408] FILTER COEFFICIENTS IN A SIX-PRIMARY COLOR SYSTEM USING A 4:2:0 SAMPLING METHOD
[00409] FIG. 32 illustrates one embodiment of a raster encoding diagram of sample placements for a six-primary color 4:2:0 progressive scan system. Within this compression process, horizontal lines show the raster on a display matrix. Vertical lines depict drive columns. The intersection of these is a pixel calculation. Data around a particular pixel is used to calculate color and brightness of the subpixels. Each “X” shows placement timing of the EY INT sample. Red dots depict placement of the EC'R-INT + EBC-INT sample. Blue triangles show placement of the EC'B-INT + EC'Y-INT sample.
[00410] In one embodiment, the raster is an RGB raster. In another embodiment, the raster is a RGBCMY raster.
[00411] SIX-PRIMARY COLOR SYSTEM BACKWARDS COMPATIBILITY
[00412] By designing the color gamut within the saturation levels of standard formats and using inverse color primary positions, it is easy to resolve an RGB image with minimal processing. In one embodiment for six-primary encoding, image data is split across three color channels in a transport system. In one embodiment, the image data is read as six- primary data. In another embodiment, the image data is read as RGB data. By maintaining a standard white point, the axis of modulation for each channel is considered as values describing two colors (e.g., blue and yellow) for a six-primary system or as a single color (e.g., blue) for an RGB system. This is based on where black is referenced. In one embodiment of a six-primary color system, black is decoded at a mid-level value. In an RGB system, the same data stream is used, but black is referenced at bit zero, not a mid-level.
[00413] In one embodiment, the RGB values encoded in the 6P stream are based on ITU-R BT.709. In another embodiment, the RGB values encoded are based on SMPTE RP431. Advantageously, these two embodiments require almost no processing to recover values for legacy display.
[00414] Two decoding methods are proposed. The first is a preferred method that uses very limited processing, negating any issues with latency. The second is a more straightforward method using a set of matrices at the end of the signal path to conform the 6P image to RGB. [00415] In one embodiment, the decoding is for a 4:4:4 system. In one embodiment, the assumption of black places the correct data with each channel. If the 6P decoder is in the signal path, 11 -bit values for RGB are arranged above bit value 2048, while CMY level are arranged below bit value 2047 as 11 -bit. However, if this same data set is sent to a display or process that is does not understand 6P processing, then that image data is assumed as black at bit value 0 as a full 12-bit word.
[00416] FIG. 33 illustrates one embodiment of the six-primary color unstack process in a 4:2:2 video system. Decoding starts by tapping image data prior to the unstacking process. The input to the 6P unstack will map as shown in FIG. 34. The output of the 6P decoder will map as shown in FIG. 35. This same data is sent uncorrected as the legacy RGB image data. The interpretation of the RGB decode will map as shown in FIG. 36.
[00417] Alternatively, the decoding is for a 4:2:2 system. This decode uses the same principles as the 4:4:4 decoder, but because a luminance channel is used instead of discrete color channels, the processing is modified. Legacy image data is still taken prior to unstack from the EC'B-INT + EC'Y-INT and EC'R-INT + EC'C-INT channels as shown in FIG. 37.
[00418] FIG. 38 illustrates one embodiment of a non-constant luminance decoder with a legacy process. The dotted box marked (1) shows the process where a new component called EL yis used to subtract the luminance levels that are present from the CMY channels from the EC'B-INT + EC' Y-INT and EC'R-INT + EBC-INT components as shown in box (2). The resulting output is now the R and B image components of the EOTF process. EL yis also sent to the G matrix to convert the luminance and color difference components to a green output as shown in box (3). Thus, R’G’B ’ is input to the EOTF process and output as GRGB, RRGB, and BRGB. In another embodiment, the decoder is a legacy RGB decoder for non-constant luminance systems. [00419] For a constant luminance system, the process is very similar with the exception that green is calculated as linear as shown in FIG. 39.
[00420] SIX-PRIMARY COLOR SYSTEM USING A MATRIX OUTPUT
[00421] In one embodiment, the six-primary color system outputs a legacy RGB image. This requires a matrix output to be built at the very end of the signal path. FIG. 40 illustrates one embodiment of a legacy RGB image output at the end of the signal path. The design logic of the C, M, and Y primaries is in that they are substantially equal in saturation and placed at substantially inverted hue angles compared to R, G, and B primaries, respectively. In one embodiment, substantially equal in saturation refers to a ±10% difference in saturation values for the C, M, and Y primaries in comparison to saturation values for the R, G, and B primaries, respectively. In addition, substantially equal in saturation covers additional percentage differences in saturation values falling within the ±10% difference range. For example, substantially equal in saturation further covers a ±7.5% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; a ±5% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; a ±2% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; a ±1% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively; and/or a ±0.5% difference in saturation values for the C, M, and Y primaries in comparison to the saturation values for the R, G, and B primaries, respectively. In a preferred embodiment, the C, M, and Y primaries are equal in saturation to the R, G, and B primaries, respectively. For example, the cyan primary is equal in saturation to the red primary, the magenta primary is equal in saturation to the green primary, and the yellow primary is equal in saturation to the blue primary. [00422] In an alternative embodiment, the saturation values of the C, M, and Y primaries are not required to be substantially equal to their corollary primary saturation value among the R, G, and B primaries, but are substantially equal in saturation to a primary other than their corollary R, G, or B primary value. For example, the C primary saturation value is not required to be substantially equal in saturation to the R primary saturation value, but rather is substantially equal in saturation to the G primary saturation value and/or the B primary saturation value. In one embodiment, two different color saturations are used, wherein the two different color saturations are based on standardized gamuts already in use.
[00423] In one embodiment, substantially inverted hue angles refers to a ±10% angle range from an inverted hue angle (e.g., 180 degrees). In addition, substantially inverted hue angles cover additional percentage differences within the ±10% angle range from an inverted hue angle. For example, substantially inverted hue angles further covers a ±7.5% angle range from an inverted hue angle, a ±5% angle range from an inverted hue angle, a ±2% angle range from an inverted hue angle, a ±1% angle range from an inverted hue angle, and/or a ±0.5% angle range from an inverted hue angle. In a preferred embodiment, the C, M, and Y primaries are placed at inverted hue angles (e.g., 180 degrees) compared to the R, G, and B primaries, respectively.
[00424] In one embodiment, the gamut is the ITU-R BT.709-6 gamut. In another embodiment, the gamut is the SMPTE RP431-2 gamut.
[00425] The unstack process includes output as six, 11 -bit color channels that are separated and delivered to a decoder. To convert an image from a six-primary color system to an RGB image, at least two matrices are used. One matrix is a 3x3 matrix converting a six- primary color system image to XYZ values. A second matrix is a 3x3 matrix for converting from XYZ to the proper RGB color space. In one embodiment, XYZ values represent additive color space values, where XYZ matrices represent additive color space matrices. Additive color space refers to the concept of describing a color by stating the amounts of primaries that, when combined, create light of that color.
[00426] When a six-primary display is connected to the six-primary output, each channel will drive each color. When this same output is sent to an RGB display, the non-RGB (e.g., CMY) channels are ignored and only the RGB channels are displayed. An element of operation is that both systems drive from the black area. At this point in the decoder, all are coded as bit value 0 being black and bit value 2047 being peak color luminance. This process can also be reversed in a situation where an RGB source can feed a six-primary display. The six-primary display would then have no information for the non-RGB (e.g., CMY) channels and would display the input in a standard RGB gamut. FIG. 41 illustrates one embodiment of six-primary color output using a non-constant luminance decoder. FIG. 42 illustrates one embodiment of a legacy RGB process within a six-primary color system.
[00427] The design of this matrix is a modification of the CIE process to convert RGB to XYZ. First, u ’v’ values are converted back to CIE 1931 xyz values using the following formulas: [00429] In one embodiment, the gamut is SMPTE RP431-2. The mapping for RGBCMY values for a SMPTE RP431-2 (6P-C) gamut are:
[00430] Following mapping the RGBCMY values to a matrix, a white point conversion occurs:
[00431] For a six-primary color system using an ITU-R BT.709-6 (6P-B) color gamut, the white point is D65:
[00432] For a six-primary color system using a SMPTE RP431-2 (6P-C) color gamut, the white point is D60:
[00433] Following the white point conversion, a calculation is required for RGB saturation values, SR, SG, and SB. The results from the second operation are inverted and multiplied with the white point XYZ values. In one embodiment, the color gamut used is an ITU-R BT.709-6 color gamut. The values calculate as:
Where
[00434] In one embodiment, the color gamut is a SMPTE RP431-2 color gamut. The values calculate as:
Where
[00435] Next, a six-primary color-to-XYZ matrix must be calculated. For an embodiment where the color gamut is an ITU-R BT.709-6 color gamut, the calculation is as follows:
Wherein the resulting matrix is multiplied by the SRSGSB matrix:
[00436] For an embodiment where the color gamut is a SMPTE RP431-2 color gamut, the calculation is as follows:
Wherein the resulting matrix is multiplied by the SRSGSB matrix:
[00437] Finally, the XYZ matrix must converted to the correct standard color space. In an embodiment where the color gamut used is an ITU-R BT709.6 color gamut, the matrices are as follows:
[00438] In an embodiment where the color gamut used is a SMPTE RP431-2 color gamut, the matrices are as follows:
[00439] PACKING A SIX-PRIMARY COLOR SYSTEM INTO ICTCP
[00440] ICiCp (ITP) is a color representation format specified in the Rec. ITU-R BT.2100 standard that is used as a part of the color image pipeline in video and digital photography systems for high dynamic range (HDR) and wide color gamut (WCG) imagery. The I (intensity) component is a luma component that represents the brightness of the video. CT and Cp are blue-yellow (“tritanopia”) and red-green (“protanopia”) chroma components. The format is derived from an associated RGB color space by a coordination transformation that includes two matrix transformations and an intermediate non-linear transfer function, known as a gamma pre-correction. The transformation produces three signals: I, CT, and Cp. The ITP transformation can be used with RGB signals derived from either the perceptual quantizer (PQ) or hybrid log-gamma (HLG) nonlinearity functions. The PQ curve is described in ITU-
R BT2100-2:2018, Table 4, which is incorporated herein by reference in its entirety. [00441] FIG. 43 illustrates one embodiment of packing six-primary color system image data into an ICjCp (ITP) format. In one embodiment, RGB image data is converted to an XYZ matrix. The XYZ matrix is then converted to an LMS matrix. The LMS matrix is then sent to an optical electronic transfer function (OETF). The conversion process is represented below:
Output from the OETF is converted to ITP format. The resulting matrix is:
[00442] FIG. 44 illustrates one embodiment of a six-primary color system converting
RGBCMY image data into XYZ image data for an ITP format (e.g., 6P-B, 6P-C). For a six- primary color system, this is modified by replacing the RGB to XYZ matrix with a process to convert RGBCMY to XYZ. This is the same method as described in the legacy RGB process.
The new matrix is as follows for an ITU-R BT.709-6 (6P-B) color gamut:
[00443] RGBCMY data, based on an ITU-R BT.709-6 color gamut, is converted to an
XYZ matrix. The resulting XYZ matrix is converted to an LMS matrix, which is sent to an
OETF. Once processed by the OETF, the LMS matrix is converted to an ITP matrix. The resulting ITP matrix is as follows: [00444] In another embodiment, the LMS matrix is sent to an Optical Optical Transfer Function (OOTF). In yet another embodiment, the LMS matrix is sent to a Transfer Function other than OOTF or OETF.
[00445] In another embodiment, the RGBCMY data is based on the SMPTE ST431-2 (6P- C) color gamut. The matrices for an embodiment using the SMPTE ST431-2 color gamut are as follows:
The resulting ITP matrix is:
[00446] The decode process uses the standard ITP decode process, as the SRSGSB cannot be easily inverted. This makes it difficult to recover the six RGBCMY components from the ITP encode. Therefore, the display is operable to use the standard ICtCp decode process as described in the standards and is limited to just RGB output.
[00447] CONVERTING TO A FIVE-COLOR MULTI-PRIMARY DISPLAY
[00448] In one embodiment, the system is operable to convert image data incorporating five primary colors. In one embodiment, the five primary colors include Red (R), Green (G), Blue (G), Cyan (C), and Yellow (Y), collectively referred to as RGBCY. In another embodiment, the five primary colors include Red (R), Green (G), Blue (B), Cyan (C), and Magenta (M), collectively referred to as RGBCM. In one embodiment, the five primary colors do not include Magenta (M).
[00449] In one embodiment, the five primary colors include Red (R), Green (G), Blue (B), Cyan (C), and Orange (O), collectively referred to as RGBCO. RGBCO primaries provide optimal spectral characteristics, transmittance characteristics, and makes use of a D65 white point. See, e.g., Moon-Cheol Kim et al., Wide Color Gamut Five Channel Multi-Primary for HDTV Application, Journal of Imaging Sci. & Tech. Vol. 49, No. 6, Nov./Dec. 2005, at 594- 604, which is hereby incorporated by reference in its entirety.
[00450] In one embodiment, a five-primary color model is expressed as F = M. C, where F is equal to a tristimulus color vector, F = (X, Y, Z)T, and C is equal to a linear display control vector, C = (Cl, C2, C3, C4, C5)T. Thus, a conversion matrix for the five-primary color model is represented as
[00451] Using the above equation and matrix, a gamut volume is calculated for a set of given control vectors on the gamut boundary. The control vectors are converted into CIELAB uniform color space. However, because matrix M is non-square, the matrix inversion requires splitting the color gamut into a specified number of pyramids, with the base of each pyramid representing an outer surface and where the control vectors are calculated using linear equation for each given XYZ triplet present within each pyramid. By separating regions into pyramids, the conversion process is normalized. In one embodiment, a decision tree is created in order to determine which set of primaries are best to define a specified color. In one embodiment, a specified color is defined by multiple sets of primaries. In order to locate each pyramid, 2D chromaticity look-up tables are used, with corresponding pyramid numbers for input chromaticity values in xy or u ’v ’. Typical methods using pyramids require 1000 x 1000 address ranges in order to properly search the boundaries of adjacent pyramids with look-up table memory. The system of the present invention uses a combination of parallel processing for adjacent pyramids and at least one algorithm for verifying solutions by checking constraint conditions. In one embodiment, the system uses a parallel computing algorithm. In one embodiment, the system uses a sequential algorithm. In another embodiment, the system uses a brightening image transformation algorithm. In another embodiment, the system uses a darkening image transformation algorithm. In another embodiment, the system uses an inverse sinusoidal contrast transformation algorithm. In another embodiment, the system uses a hyperbolic tangent contrast transformation algorithm. In yet another embodiment, the system uses a sine contrast transformation execution times algorithm. In yet another embodiment, the system uses a linear feature extraction algorithm. In yet another embodiment, the system uses a JPEG2000 encoding algorithm. In yet another embodiment, the system uses a parallelized arithmetic algorithm. In yet another embodiment, the system uses an algorithm other than those previously mentioned. In yet another embodiment, the system uses any combination of the aforementioned algorithms.
[00452] MAPPING A SIX-PRIMARY COLOR SYSTEM INTO STANDARDIZED TRANSPORT FORMATS
[00453] Each encode and/or decode system fits into existing video serial data streams that have already been established and standardized. This is key to industry acceptance. Encoder and/or decoder designs require little or no modification for a six-primary color system to map to these standard serial formats.
[00454] FIG. 45 illustrates one embodiment of a six-primary color system mapping to a SMPTE ST424 standard serial format. The SMPTE ST424/ST425 set of standards allow very high sampling systems to be passed through a single cable. This is done by using alternating data streams, each containing different components of the image. For use with a six-primary color system transport, image formats are limited to RGB due to the absence of a method to send a full bandwidth Y signal.
[00455] The process for mapping a six-primary color system to a SMPTE ST425 format is the same as mapping to a SMPTE ST424 format. To fit a six-primary color system into a SMPTE ST425/424 stream involves the following substitutions: G]'NT + M]'NT is placed in the Green data segments, RI'NT + CI'NT is placed in the Red data segments, and BI'NT + T/WT is placed into the Blue data segments. FIG. 46 illustrates one embodiment of an SMPTE 424 6P readout.
[00456] System 2 requires twice the data rate as System 1, so it is not compatible with SMPTE 424. However, it maps easily into SMPTE ST2082 using a similar mapping sequence. In one example, System 2 is used to have the same data speed defined for 8K imaging to show a 4K image.
[00457] In one embodiment, sub-image and data stream mapping occur as shown in SMPTE ST2082. An image is broken into four sub-images, and each sub-image is broken up into two data streams (e.g., sub-image 1 is broken up into data stream 1 and data stream 2). The data streams are put through a multiplexer and then sent to the interface as shown in FIG. 47.
[00458] FIG. 48 and FIG. 49 illustrate serial digital interfaces for a six-primary color system using the SMPTE ST2082 standard. In one embodiment, the six-primary color system data is RGBCMY data, which is mapped to the SMPTE ST2082 standard (FIG. 48). Data streams 1, 3, 5, and 7 follow the pattern shown for data stream 1. Data streams 2, 4, 6, and 8 follow the pattern shown for data stream 2. In one embodiment, the six-primary color system data is YRGB YCMY CR CB CC CY data, which is mapped to the SMPTE ST2082 standard (FIG. 49). Data streams 1, 3, 5, and 7 follow the pattern shown for data stream 1. Data streams 2, 4, 6, and 8 follow the pattern shown for data stream 2.
[00459] In one embodiment, the standard serial format is SMPTE ST292. SMPTE ST292 is an older standard than ST424 and is a single wire format for 1.5GB video, whereas ST424 is designed for up to 3GB video. However, while ST292 can identify the payload ID of SMPTE ST352, it is constrained to only accepting an image identified by a hex value, Oh. All other values are ignored. Due to the bandwidth and identifications limitations in ST292, a component video six-primary color system incorporates a full bit level luminance component. To fit a six-primary color system into a SMPTE ST292 stream involves the following substitutions: Ey6-INT is placed in the Y data segments, EC'b-INT + EC'y-INT is placed in the Cb data segments, and Ecr-INT + EC'c-INT is placed in the Cr data segments. In another embodiment, the standard serial format is SMPTE ST352.
[00460] SMPTE ST292 and ST424 Serial Digital Interface (SDI) formats include payload identification (ID) metadata to help the receiving device identify the proper image parameters. The tables for this need modification by adding at least one flag identifying that the image source is a six-primary color RGB image. Therefore, six-primary color system format additions need to be added. In one embodiment, the standard is the SMPTE ST352 standard.
[00461] FIG. 50 illustrates one embodiment of an SMPTE ST292 6P mapping. FIG. 51 illustrates one embodiment of an SMPTE ST292 6P readout.
[00462] FIG. 52 illustrates modifications to the SMPTE ST352 standards for a six-primary color system. Hex code “Bh” identifies a constant luminance source and flag “Fh” indicates the presence of a six-primary color system. In one embodiment, Fh is used in combination with at least one other identifier located in byte 3. In another embodiment, the Fh flag is set to 0 if the image data is formatted as System 1 and the Fh flag is set to 1 if the image data is formatted as System 2.
[00463] In another embodiment, the standard serial format is SMPTE ST2082. Where a six-primary color system requires more data, it may not always be compatible with SMPTE ST424. However, it maps easily into SMPTE ST2082 using the same mapping sequence. This usage would have the same data speed defined for 8K imaging in order to display a 4K image. [00464] In another embodiment, the standard serial format is SMPTE ST2022. Mapping to ST2022 is similar to mapping to ST292 and ST242, but as an ETHERNET format. The output of the stacker is mapped to the media pay load based on Real-time Transport Protocol (RTP) 3550, established by the Internet Engineering Task Force (IETF). RTP provides end- to-end network transport functions suitable for applications transmitting real-time data, including, but not limited to, audio, video, and/or simulation data, over multicast or unicast network services. The data transport is augmented by a control protocol (RTCP) to allow monitoring of the data delivery in a manner scalable to large multicast networks, and to provide control and identification functionality. There are no changes needed in the formatting or mapping of the bit packing described in SMPTE ST 2022-6: 2012 (HBRMT), which is incorporated herein by reference in its entirety.
[00465] FIG. 53 illustrates one embodiment of a modification for a six-primary color system using the SMPTE ST2202 standard. For SMPTE ST2202-6:2012 (HBRMT), there are no changes needed in formatting or mapping of the bit packing. ST2022 relies on header information to correctly configure the media payload. Parameters for this are established within the payload header using the video source format fields including, but not limited to, MAP, FRAME, FRATE, and/or SAMPLE. MAP, FRAME, and FRATE remain as described in the standard. MAP is used to identify if the input is ST292 or ST425 (RGB or Y Cb Cr). SAMPLE is operable for modification to identify that the image is formatted as a six-primary color system image. In one embodiment, the image data is sent using flag “Oh” (unknown/unspecified).
[00466] In another embodiment, the standard is SMPTE ST2110. SMPTE ST2110 is a relatively new standard and defines moving video through an Internet system. The standard is based on development from the IETF and is described under RFC3550. Image data is described through “pgroup” construction. Each pgroup consists of an integer number of octets. In one embodiment, a sample definition is RGB or YCbCr and is described in metadata. In one embodiment, the metadata format uses a Session Description Protocol (SDP) format. Thus, pgroup construction is defined for 4:4:4, 4:2:2, and 4:2:0 sampling as 8- bit, 10-bit, 12-bit, and in some cases 16-bit and 16-bit floating point wording. In one embodiment, six-primary color image data is limited to a 10-bit depth. In another embodiment, six-primary color image data is limited to a 12-bit depth. Where more than one sample is used, it is described as a set. For example, 4:4:4 sampling for blue, as anon-linear RGB set, is described as CO’B, Cl’B, C2’B, C3’B, and C4’B. The lowest number index being left most within the image. In another embodiment, the method of substitution is the same method used to map six-primary color content into the ST2110 standard.
[00467] In another embodiment, the standard is SMPTE ST2110. SMPTE ST2110-20 describes the construction for each pgroup. In one embodiment, six-primary color system content arrives for mapping as non-linear data for the SMPTE ST2110 standard. In another embodiment, six-primary color system content arrives for mapping as linear data for the SMPTE ST2110 standard.
[00468] FIG. 54 illustrates a table of 4:4:4 sampling for a six-primary color system for a 10-bit video system. For 4:4:4 10-bit video, 15 octets are used and cover 4 pixels.
[00469] FIG. 55 illustrates a table of 4:4:4 sampling for a six-primary color system for a 12-bit video system. For 4:4:4 12-bit video, 9 octets are used and cover 2 pixels before restarting the sequence.
[00470] Non-linear RGBCMY image data would arrive as: GI'NT + MI'NT, RI'NT + C[NT, and B]'NT + Y/NT- Component substitution would follow what has been described for SMPTE ST424, where GI'NT + MI'NT is placed in the Green data segments, RI'NT + is placed in the Red data segments, and BI'NT + is placed in the Blue data segments. The sequence described in the standard is shown as R0’, GO’, B0’, Rl’, GE, Bl’, etc. [00471] FIG. 56 illustrates sequence substitutions for 10-bit and 12-bit video in 4:2:2 sampling systems in a Y Cb Cr Cc Cy color space. Components are delivered to a 4:2:2 pgroup including, but not limited to, EY6-INT, Ecb-iNT + Ecy-iNT, and Ecr-INT + ECC-INT- For 4:2:2 10-bit video, 5 octets are used and cover 2 pixels before restarting the sequence. For 4:2:2 12-bit video, 6 octets are used and cover 2 pixels before restarting the sequence.
Component substitution follows what has been described for SMPTE ST292, where Ey6-INT is placed in the Y data segments, Ecb-INT + EC'y-INT is placed in the Cb data segments, and EC'T-INT + ^CC-INT is placed in the Cr data segments. The sequence described in the standard is shown as CbO’, Y0’, CrO’, YF, CrF, Y3’, Cb2’, Y4’, Cr2’, Y5’, etc. In another embodiment, the video data is represented at a bit level other than 10-bit or 12-bit. In another embodiment, the sampling system is a sampling system other than 4:2:2. In another embodiment, the standard is STMPE ST2110.
[00472] FIG. 57 illustrates sample placements of six-primary system components for a 4:2:2 sampling system image. This follows the substitutions illustrated in FIG. 56, using a 4:2:2 sampling system.
[00473] FIG. 58 illustrates sequence substitutions for 10-bit and 12-bit video in 4:2:0 sampling systems using a Y Cb Cr Cc Cy color space. Components are delivered to a pgroup including, but not limited to, EY6-INT, Ecb-iNT + Ec'y-INT, and Ecr-iNT + EC'c-INT. For 4:2:0 10-bit video data, 15 octets are used and cover 8 pixels before restarting the sequence. For 4:2:0 12-bit video data, 9 octets are used and cover 4 pixels before restarting the sequence. Component substitution follows what is described in SMPTE ST292 where Ey6-INT is placed in the Y data segments, EC'b-INT + Ecy-INT is placed in the Cb data segments, and Ecr-INT + EC' C-INT is placed in the Cr data segments. The sequence described in the standard is shown as Y’00, Y’01, Y’, etc. [00474] FIG. 59 illustrates sample placements of six-primary system components for a 4:2:0 sampling system image. This follows the substitutions illustrated in FIG. 58, using a 4:2:0 sampling system.
[00475] FIG. 60 illustrates modifications to SMPTE ST2110-20 for a 10-bit six-primary color system in 4:4:4 video. SMPTE ST2110-20 describes the construction of each “pgroup”. Normally, six-primary color system data and/or content would arrive for mapping as nonlinear. However, with the present system there is no restriction on mapping data and/or content. For 4:4:4, 10-bit video, 15 octets are used and cover 4 pixels before restarting the sequence. Non-linear, six-primary color system image data would arrive as G]'NT, BINT, RI'NT, M]'NT, Y/NT, and C/NT. The sequence described in the standard is shown as R0’, GO’, B0’, RE, GE, Bl’, etc.
[00476] FIG. 61 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:4:4 video. For 4:4:4, 12-bit video, 9 octets are used and cover 2 pixels before restarting the sequence. Non-linear, six-primary color system image data would arrive as G]'NT, RI'NT, MI'NT, Y/NT, and C/NT. The sequence described in the standard is shown as R0’, GO’, B0’, RE, GE, BE, etc.
[00477] FIG. 62 illustrates modifications to SMPTE ST2110-20 for a 10-bit six primary color system in 4:2:2 video. Components that are delivered to a SMPTE ST2110 pgroup include, but are not limited to, EYrgb-INT, EYcym-INT, Ecb-iNT? Ecr-iNT? ECy-INT, and EC'c-]NT. For 4:2:2, 10-bit video, 5 octets are used and cover 2 pixels before restarting the sequence. For 4:2:2:2, 12-bit video, 6 octets are used and cover 2 pixels before restarting the sequence. Component substitution follows what is described for SMPTE ST292, where Eyrgb-iNT or EyCym-iNT are placed in the Y data segments, Ecr-INT or ECC-INT are placed in the Cr data segments, and EC'b-INT or EC'y-INT are placed in the Cb data segments. The sequence described in the standard is shown as Cb’O, Y’O, Cr’O, Y’l, Cb’l, Y’2, Cr’l, Y’3, Cb’2, Y’4, Cr’2, etc.
[00478] FIG. 63 illustrates modifications to SMPTE ST2110-20 for a 12-bit six-primary color system in 4:2:0 video. Components that are delivered to a SMPTE ST2110 pgroup are the same as with the 4:2:2 method. For 4:2:0, 10-bit video, 15 octets are used and cover 8 pixels before restarting the sequence. For 4:2:0, 12-bit video, 9 octets are used and cover 4 pixels before restarting the sequence. Component substitution follows what is described for SMPTE ST292, where Eyrgb-INT or Eycym-]NT placed in the Y data segments, Ecr-INT or EC' C-INT are placed in the Cr data segments, and EC'b-INT or EC'y-INT are placed in the Cb data segments. The sequence described in the standard is shown as Y’00, Y’01, Y’, etc.
[00479] Table 16 summarizes mapping to SMPTE ST2110 for 4:2:2:2:2 and 4:2:0:2:0 sampling for System 1 and Table 17 summaries mapping to SMPTE ST2110 for 4:4:4:4:4:4 sampling (linear and non-linear) for System 1.
[00480] TABLE 16 [00481] TABLE 17
[00482] Table 18 summarizes mapping to SMPTE ST2110 for 4:2:2:2:2 sampling for System 2 and Table 19 summaries mapping to SMPTE ST2110 for 4:4:4:4:4:4 sampling (linear and non-linear) for System 2.
[00483] TABLE 18
[00484] TABLE 19
[00485] SESSION DESCRIPTION PROTOCOL (SDP) MODIFICATION FOR A SIX- PRIMARY COLOR SYSTEM
[00486] SDP is derived from IETF RFC 4566 which sets parameters including, but not limited to, bit depth and sampling parameters. In one embodiment, SDP parameters are contained within the RTP payload. In another embodiment, SDP parameters are contained within the media format and transport protocol. This payload information is transmitted as text. Therefore, modifications for the additional sampling identifiers requires the addition of new parameters for the sampling statement. SDP parameters include, but are not limited to, color channel data, image data, framerate data, a sampling standard, a flag indicator, an active picture size code, a timestamp, a clock frequency, a frame count, a scrambling indicator, and/or a video format indicator. For non-constant luminance imaging, the additional parameters include, but are not limited to, RGBCMY-4:4:4, YBRCY-4:2:2, and YBRCY- 4:2:0. For constant luminance signals, the additional parameters include, but are not limited to, CLYBRCY-4:2:2 and CLYBRCY-4:2:0.
[00487] Additionally, differentiation is included with the colorimetry identifier in one embodiment. For example, 6PB1 defines 6P with a color gamut limited to ITU-R BT.709 formatted as System 1, 6PB2 defines 6P with a color gamut limited to ITU-R BT.709 formatted as System 2, 6PB3 defines 6P with a color gamut limited to ITU-R BT.709 formatted as System 3, 6PC1 defines 6P with a color gamut limited to SMPTE RP 431-2 formatted as System 1, 6PC2 defines 6P with a color gamut limited to SMPTE RP 431-2 formatted as System 2, 6PC3 defines 6P with a color gamut limited to SMPTE RP 431-2 formatted as System 3, 6PS1 defines 6P with a color gamut as Super 6P formatted as System
1, 6PS2 defines 6P with a color gamut as Super 6P formatted as System 2, and 6PS3 defines 6P with a color gamut as Super 6P formatted as System 3.
[00488] Colorimetry can also be defined between a six-primary color system using the ITU-R BT.709-6 standard and the SMPTE ST431-2 standard, or colorimetry can be left defined as is standard for the desired standard. For example, the SDP parameters for a 1920x1080 six-primary color system using the ITU-R BT.709-6 standard with a 10-bit signal as System 1 are as follows: m = video 30000 RTP/AVP 112, a = rtpmap: 112 raw/90000, a = fmtp: 112, sampling = YBRCY-4:2:2, width = 1920, height = 1080, exactframerate = 30000/1001, depth = 10, TCS = SDR, colorimetry = 6PB1, PM = 2110GPM, SSN = ST2110- 20:2017.
[00489] In one embodiment, the six-primary color system is integrated with a Consumer Technology Association (CTA) 861-based system. CTA-861 establishes protocols, requirements, and recommendations for the utilization of uncompressed digital interfaces by consumer electronics devices including, but not limited to, digital televisions (DTVs), digital cable, satellite or terrestrial set-top boxes (STBs), and related peripheral devices including, but not limited to, DVD players and/or recorders, and other related Sources or Sinks.
[00490] These systems are provided as parallel systems so that video content is parsed across several line pairs. This enables each video component to have its own transition- minimized differential signaling (TMDS) path. TMDS is a technology for transmitting highspeed serial data and is used by the Digital Visual Interface (DVI) and High-Definition Multimedia Interface (HDMI) video interfaces, as well as other digital communication interfaces. TMDS is similar to low-voltage differential signaling (LVDS) in that it uses differential signaling to reduce electromagnetic interference (EMI), enabling faster signal transfers with increased accuracy. In addition, TMDS uses a twisted pair for noise reduction, rather than a coaxial cable that is conventional for carrying video signals. Similar to LVDS, data is transmitted serially over the data link. When transmitting video data, and using HDMI, three TMDS twisted pairs are used to transfer video data.
[00491] In such a system, each pixel packet is limited to 8 bits only. For bit depths higher than 8 bits, fragmented packs are used. This arrangement is no different than is already described in the current CTA-861 standard.
[00492] Based on CTA extension Version 3, identification of a six-primary color transmission would be performed by the sink device (e.g., the monitor). Adding recognition of the additional formats would be flagged in the CTA Data Block Extended Tag Codes (byte 3). Since codes 33 and above are reserved, any two bits could be used to identify that the format is RGB, RGBCMY, Y Cb Cr, or Y Cb Cr Cc Cy and/or identify System 1 or System 2. Should byte 3 define a six-primary sampling format, and where the block 5 extension identifies byte 1 as ITU-R BT.709, then logic assigns as 6P-B. However, should byte 4 bit 7 identify colorimetry as DCI-P3, the color gamut would be assigned as 6P-C.
[00493] In one embodiment, the system alters the AVI Infoframe Data to identify content. AVI Infoframe Data is shown in Table 10 of CTA 861-G. In one embodiment, Y2=l, Yl=0, and Y0=0 identifies content as 6P 4:2:0:2:0. In another embodiment, Y2=l, Yl=0, and Y0=l identifies content as Y Cr Cb Cc Cy. In yet another embodiment, Y2=l, Yl=l, and Y0=0 identifies content as RGBCMY.
[00494] Byte 2 Cl=l, C0=l identifies extended colorimetry in Table 11 of CTA 861-G. Byte 3 EC2, ECI, ECO identifies additional colorimetry extension valid in Table 13 of CTA 861-G. Table 14 of CTA 861-G reserves additional extensions. In one embodiment, ACE3=1,
ACE2=0, ACE1=O, and ACE0=X identifies 6P-B. In one embodiment, ACE3=0, ACE2=1, ACE1=O, and ACE0=X identifies 6P-C. In one embodiment, ACE3=0, ACE2=0, ACE1=1, and ACE0=X identifies System 1. In one embodiment, ACE3=1, ACE2=1, ACE1=O, and ACE0=X identifies System 2.
[00495] FIG. 64 illustrates the current RGB sampling structure for 4:4:4 sampling video data transmission. For HDMI 4:4:4 sampling, video data is sent through three TMDS line pairs. FIG. 65 illustrates a six-primary color sampling structure, RGBCMY, using System 1 for 4:4:4 sampling video data transmission. In one embodiment, the six-primary color sampling structure complies with CTA 861-G, November 2016, Consumer Technology Association, which is incorporated herein by reference in its entirety. FIG. 66 illustrates an example of System 2 to RGBCMY 4:4:4 transmission. FIG. 67 illustrates current Y Cb Cr 4:2:2 sampling transmission as non-constant luminance. FIG. 68 illustrates a six-primary color system (System 1) using Y Cr Cb Cc Cy 4:2:2 sampling transmission as non-constant luminance. FIG. 69 illustrates an example of a System 2 to Y Cr Cb Cc Cy 4:2:2 Transmission as non-constant luminance. In one embodiment, the Y Cr Cb Cc Cy 4:2:2 sampling transmission complies with CTA 861-G, November 2016, Consumer Technology Association. FIG. 70 illustrates current Y Cb Cr 4:2:0 sampling transmission. FIG. 71 illustrates a six-primary color system (System 1) using Y Cr Cb Cc Cy 4:2:0 sampling transmission.
[00496] HDMI sampling systems include Extended Display Identification Data (EDID) metadata. EDID metadata describes the capabilities of a display device to a video source. The data format is defined by a standard published by the Video Electronics Standards Association (VESA). The EDID data structure includes, but is not limited to, manufacturer name and serial number, product type, phosphor or filter type, timings supported by the display, display size, luminance data, and/or pixel mapping data. The EDID data structure is modifiable and modification requires no additional hardware and/or tools.
[00497] EDID information is transmitted between the source device and the display through a display data channel (DDC), which is a collection of digital communication protocols created by VESA. With EDID providing the display information and DDC providing the link between the display and the source, the two accompanying standards enable an information exchange between the display and source.
[00498] In addition, VESA has assigned extensions for EDID. Such extensions include, but are not limited to, timing extensions (00), additional time data black (CEA EDID Timing Extension (02)), video timing block extensions (VTB-EXT (10)), EDID 2.0 extension (20), display information extension (DI-EXT (40)), localized string extension (LS-EXT (50)), microdisplay interface extension (MI-EXT (60)), display ID extension (70), display transfer characteristics data block (DTCDB (A7, AF, BF)), block map (F0), display device data block (DDDB (FF)), and/or extension defined by monitor manufacturer (FF).
[00499] In one embodiment, SDP parameters include data corresponding to a payload identification (ID) and/or EDID information.
[00500] MULTI-PRIMARY COLOR SYSTEM DISPLAY
[00501] FIG. 72 illustrates a dual stack LCD projection system for a six-primary color system. In one embodiment, the display is comprised of a dual stack of projectors. This display uses two projectors stacked on top of one another or placed side by side. In one embodiment, the optical paths of the projectors are aligned manually. In another embodiment, the two projectors are automatically aligned with internal software. Each projector is similar, with the only difference being the color filters in each unit. In one embodiment, a first projector creates an RGB image while a second projector creates a CMY image. In another embodiment, the two projectors create a four-primary color display system. In one embodiment, the four-primary color system is an RGBC color system. In another embodiment, the four-primary color system is an RG1G2B system wherein the two Green primaries are within the 520-550nm wavelength range. In one embodiment, the four-primary color system is a RGBW system. In yet another embodiment, the two projectors create a five- primary color display system. In one embodiment, the five-primary display system includes a D65 white point. In another embodiment, the five-primary color display system includes a Yellow primary and/or a Cyan primary. In one embodiment, the five-primary color display system includes two Green primaries within the 520-550nm wavelength range. Refresh and pixel timings are synchronized, enabling a mechanical alignment between the two units so that each pixel overlays the same position between projector units. In one embodiment, the input signals to the projectors include a timing reference to synchronize the output images. In one embodiment, the outputs of the two projectors are passed through a half-silvered mirror to create one image. In one embodiment, the two projectors are Liquid-Crystal Display (LCD) projectors. In another embodiment, the two projectors are Digital Light Processing (DLP) projectors. In yet another embodiment, the two projectors are Liquid-Crystal on Silicon (LCOS) projectors. In yet another embodiment, the two projectors are Light-Emitting Diode (LED) projectors. In one embodiment, the display system includes colored LEDs for each of the primary colors in the system. In another embodiment, at least one of the primary colors is displayed using a combination of LEDs of other primary colors. In one embodiment, a 3D look-up table (LUT) is designed to map the signal data to the specific capabilities of the projector system.
[00502] In one embodiment, the display is comprised of a single projector. A single projector six-primary color system requires the addition of a second cross block assembly for the additional colors. One embodiment of a single projector (e.g., single LCD projector) is shown in FIG. 73. In this embodiment, the single projector six-primary color system includes a cyan dichroic mirror, an orange dichroic mirror, a blue dichroic mirror, a red dichroic mirror, and two additional standard mirrors. In one embodiment, the single projector six- primary color system includes at least four mirrors (e.g., at least six mirrors). In one embodiment, the single projector creates a four-primary color display. In another embodiment, the single projector creates a five-primary color display.
[00503] FIG. 74 illustrates a six-primary color system using a single projector and reciprocal mirrors. In one embodiment, the display is comprised of a single projector unit working in combination with at first set of at least six reciprocal mirrors, a second set of at least six reciprocal mirrors, and at least six LCD units. Light from at least one light source emits towards the first set of at least six reciprocal mirrors. In one embodiment, one or more of the at least one light source is a Xenon lamp. In another embodiment, one or more of the at least one light source is a Hi -Pressure Mercury lamp (UHPHg). FIG. 132 shows the emissive spectra of Xenon lamps and UHPHg lamps. UHPHg lamps have a decrease in intensity in the Cyan region between 450nm to 500nm and in the Yellow region between 570-580 nm. For this reason, Xenon lamps are a preferable embodiment. The first set of at least six reciprocal mirrors reflects light towards at least one of the at least six LCD units. The at least six LCD units include, but are not limited to, a Green LCD, a Yellow LCD, a Cyan, LCD, a Red LCD, a Magenta LCD, and/or a Blue LCD. Output from each of the at least six LCDs is received by the second set of at least six reciprocal mirrors. Output from the second set of at least six reciprocal mirrors is sent to the single projector unit. Image data output by the single projector unit is output as a six-primary color system. In another embodiment, there are more than two sets of reciprocal mirrors. In another embodiment, more than one projector is used. In one embodiment, prisms reflect light towards the LCD units and the single projector unit. In another embodiment, a combination of prisms and reciprocal mirrors reflect light towards the LCD units and the single projector unit. In one embodiment, the single projector has fewer than six LCD units.
[00504] In another embodiment, the display is comprised of a dual stack Digital Micromirror Device (DMD) projector system. FIG. 75 illustrates one embodiment of a dual stack DMD projector system. In this system, two projectors are stacked on top of one another. In one embodiment, the dual stack DMD projector system uses a spinning wheel filter. In one embodiment, the filter systems are illuminated by a xenon lamp. In one embodiment, each projector has two lamps and two identical color wheels. In one embodiment, the first projector uses an RGB, while the second projector uses a CMY filter set. In another embodiment, the first projector uses an RGB filter set, while the second projector uses a CMY filter set. In one embodiment, the first projector uses a rich color filter wheel that includes RGB filters and the second projector uses a cyan filter. In another embodiment, the first projector uses a high-brightness filter wheel and the second projector uses a cyan filter. The wheels for each projector unit are preferably synchronized using an input video sync and/or a projector-to-projector sync, and timed so that the inverted colors are output of each projector at the same time. The sync signal is part of the input signal data that is delivered to each projector.
[00505] In one embodiment, the projectors are phosphor wheel systems. A yellow phosphor wheel spins in time with a DMD imager to output sequential RG from a blue laser illuminator. The second projector is designed the same, but uses a cyan phosphor wheel. The output from the second projector becomes sequential BG. In one embodiment, the color wheel includes a cyan phosphor segment that is excited by blue light as described in U.S. Patent Application No. 14/163,985, filed January 24, 2014, now U.S. Patent No. 9,470,886, which is incorporated herein by reference in its entirety. Combined, the output of both projectors is YRGGCB. Magenta is developed by synchronizing the yellow and cyan wheels to overlap the flashing DMD.
[00506] In another embodiment, the display is a single DMD projector solution. A single DMD device is coupled with an RGB diode light source system. In one embodiment, the DMD projector uses LEDs. In one embodiment, the DMD projector includes CMY diodes. In another embodiment, the DMD projector creates CMY primaries using a double flashing technique. In one embodiment, the DMD projector is a single-chip DMD projector. The chip is synchronized with the LED lamps. In another embodiment, the DMD projector is a multichip DMD projector with one chip for each primary color LED in the system. An optical chain is used to split the light to the respective chips. In one embodiment, the single DMD projector has an RGBCMY color wheel. In one embodiment, the color wheel is a rich color wheel with color wheel segments (e.g., six segments). In one embodiment, the color wheel segments include Red, Green, Blue, Cyan, Magenta, and Yellow. In another embodiment the color wheel segments include Magenta, Yellow, Orange, Cyan, Blue, and Green. In another embodiment, the color wheel is a high brightness color wheel with color wheel segments (e.g., six segments). In one embodiment, the color wheel segments include Red, Green, Blue, Cyan, Yellow, and White. FIG. 76 illustrates one embodiment of a single DMD projector solution.
[00507] FIG. 77 illustrates one embodiment of a six-primary color system using a white OLED display. In yet another embodiment, the display is a white OLED monitor. Current emissive monitor and/or television designs use a white emissive OLED array covered by a color filter. Changes to this type of display only require a change to pixel indexing and new six color primary filters. Different color filter arrays are used, placing each subpixel in a position that provides the least light restrictions, most color accuracy, and off axis display. In one embodiment, the optical filter for the OLED display uses a horizontal pixel sequence with rectangular pixels and vertical compensation. In another embodiment, the pixels are square. In one embodiment, the optical filter pattern does not include a white subpixel. FIG. 78 illustrates one embodiment of an optical filter array for a white OLED display.
[00508] FIG. 79 illustrates one embodiment of a matrix of an LCD drive for a six-primary color system with a backlight illuminated LCD monitor. In yet another embodiment, the display is a backlight illuminated LCD display. The design of an LCD display involves adding the CMY subpixels. Matrix drives for the CMY subpixels are similar to the RGB matrix drives. With the advent of 8K LCD televisions, it is technically feasible to change the matrix drive and optical filter and have a 4K six-primary color TV.
[00509] FIG. 80 illustrates one embodiment of an optical filter array for a six-primary color system with a backlight illuminated LCD monitor. The optical filter array includes the additional CMY subpixels.
[00510] In one embodiment, each pixel in the six-primary color system is a hexagonal shape. Each hexagonal pixel is divided into six equilateral triangles and each of the primaries in the six-primary color system is displayed by one of the six equilateral triangles as described in U.S. Patent Application No. 12/005,931, filed July 3, 2008, which is incorporated herein by reference in its entirety.
[00511] In one embodiment, each pixel is divided into six subpixels of the same size and area arranged in two rows of three columns. In another embodiment, each pixel is divided into six subpixels of the same size and area arranged in three rows of two columns. In yet another embodiment, each pixel is divided into six subpixels of the same size and area arranged in one row. In yet another embodiment, each pixel is divided into six subpixels of the same size and area arranged in one column. The luminance and intensity of each subpixel is dependent on the luminance and intensity of the adjacent subpixels in order to minimize the distinct visibility of individual subpixel and pixel structures. In one embodiment, complementary primary color subpixels are adjacent to each other to eliminate visual artifacts.
[00512] In one embodiment, each pixel is divided into subpixels of different sizes and areas. The size and number of subpixels for each primary color minimize blue and cyan spatial resolution without affecting the overall resolution of the display as described in U.S. Patent Application No. 12/909,742, filed October 21, 2010, now U.S. Patent No. 8,451,405, which is incorporated herein by reference in its entirety.
[00513] In another embodiment, each pixel unit is divided into two subpixel units wherein one of the two subpixels is an RGB color and the other subpixel is the complementary CMY color of the first subpixel as described in U.S. Patent Application No. 12/229,845, filed March 5, 2009, which is incorporated herein by reference in its entirety.
[00514] In one embodiment, each pixel includes at least one white subpixel to eliminate visual artifacts. In one embodiment, the at least one white subpixel includes a D65 white subpixel, a D60 white subpixel, a D45 white subpixel, a D27 white subpixel, and/or a D25 white subpixel. Advantageously, using a D65 white subpixel eliminates most of the problems with metamerism. In a preferred embodiment, the at least one white subpixel is a single white subpixel that matches the white point (e.g., a D65 white subpixel for a D65 white point). In another embodiment, the at least one white subpixel is at least two white subpixels. The at least two white subpixels are preferably separated such that a linear combination of the at least two white subpixels covers a desired white Kelvin range. In one embodiment, the at least two white subpixels include a D65 white subpixel and a D27 white subpixel. In another embodiment, the at least two white subpixels include a D65 white subpixel and a D25 white subpixel.
[00515] In yet another embodiment, the at least two white subpixels includes three white subpixels. In one embodiment, the three white subpixels include a D65 white subpixel, a D45 white subpixel, and a D27 white subpixel. Alternatively, the three white subpixels include a
D65 white subpixel, a mid-Kelvin white subpixel (e.g., D45), and a D27 white subpixel. In a preferred embodiment, the mid-Kelvin white subpixel includes a green bias. Advantageously, the green bias compensates for the slight magenta shift (e.g., when going from D25 to D65 with the straight line between the two points below the blackbody locus). Colors near the white locus and beyond are then a combination of the at least two white subpixels (e.g., two white subpixels, three white subpixels). A majority of colors will have a white component that is broad band. Therefore, the resultant spectra of a mixture of color primaries and white primaries will also be broad band with an extent dependent on an amount of the at least one white primary. A higher broad band character of light results in fewer metameric problems. This is due to a white point being comprised of a combination of color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.) in anon-white subpixel system. Total luminance is then related to intensities of the color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.). [00516] Advantageously, if at least one white subpixel is included, increased luminance can be achieved separate from the color primaries. Additionally, colors such as vibrantly colored pastels are attained by using the color primaries to “color shift” a bright white to the pastel. Alternatively, a fine balance of the color primaries is required, and small changes in a ratio of the color primaries will produce an unwanted color shift. Thus, a system with at least one white subpixel is more tolerant to minor variations of intensity of the color primaries. [00517] In one embodiment, the white point of the six-primary color system changes depending on the display or the display mode. In one embodiment, the addition of white subpixels widens the bandwidth of the filter for each non-white primary.
[00518] In one embodiment, each pixel is composed of fewer than six primary colors from the 6P gamut. The display is composed of alternating and repeating subpixel patterns. In another embodiment, the display is composed of nonrepeating subpixel patterns. [00519] In one embodiment, the subpixel colors in a pixel and in adjacent pixels are arranged to minimize the spatial distance between colors that have maximal color distance from each other as described in U.S. Patent Application No. 10/543,511, filed January 13, 2003, now U.S. Patent No. 8,228,275, which is incorporated herein by reference in its entirety.
[00520] In one embodiment, each pixel is one single primary color from the 6P gamut. In one embodiment, patterns of pixels are repeated across the display to minimize visibility of individual pixel structures as described in U.S. Patent Application No. 13/512,914, filed November 25, 2010, which is incorporated herein by reference in its entirety.
[00521] In one embodiment, the display includes at least one perovskite. In one embodiment, the at least one perovskite is a lead halide perovskite. In one embodiment, the at least one perovskite is used as a quantum dot nanocrystal. In one embodiment, the at least one perovskite is a perovskite polymer bead. When light shines through the perovskite polymer bead, the color changes depending on the composition of the perovskite polymer bead (e.g., green, red, etc.). In one embodiment, the at least one perovskite is incorporated into a perovskite LED. Examples of perovskite LEDs are described in Lin, K., et al. (2018).
Perovskite light-emitting diodes with external quantum efficiency exceeding 20 per cent. Nature, 562(7726), 245-248, which is incorporated herein by reference in its entirety. In one embodiment, the at least one perovskite is 3D printed. See, e.g., Zhou, Nanjia, Yehonadav Bekenstein, CarissaN. Eisler, Dandan Zhang, Adam M. Schwartzberg, Peidong Yang, A. Paul Alivisatos, and Jennifer A. Lewis. 2019. “Perovskite Nanowire-Block Copolymer Composites With Digitally Programmable Polarization Anisotropy.” Science Advances, which is incorporated herein by reference in its entirety.
[00522] In yet another embodiment, the display is a direct emissive assembled display.
The design for a direct emissive assembled display includes a matrix of color emitters grouped as a six-color system. Individual channel inputs drive each Quantum Dot (QD) element illuminator and/or micro LED element. In one embodiment, the quantum dots modulate light according to image data as described in U.S. Patent Application No. 15/905,085, filed February 26, 2018, now U.S. Patent No. 10,373,574, which is incorporated herein by reference in its entirety.
[00523] FIG. 81 illustrates an array for a Quantum Dot (QD) display device.
[00524] FIG. 82 illustrates one embodiment of an array for a six-primary color system for use with a direct emissive assembled display.
[00525] In one embodiment, the display system is a dual-panel display system with two wide-gamut RGB displays. One display has a Cyan filter and the other display has a clear Neutral -density filter. The two displays are aligned and the outputs are passed through a half- silvered mirror to create an RGB-Cyan display on a view screen. FIG. 133 illustrates one embodiment of the dual-panel display system using a Cyan filter.
[00526] FIG. 83 illustrates one embodiment of a six-primary color system in an emissive display that does not incorporate color filtered subpixels. For LCD and WOLED displays, this can be modified for a six-primary color system by expanding the RGB or WRGB filter arrangement to an RGBCMY matrix. For WRGB systems, the white subpixel could be removed as the luminance of the three additional primaries will replace it. In one embodiment, the CMY primaries are defined relative to the RGB primaries, and the intensities of the CMY primaries are dependent on the white point of the RGB system. SDI video is input through an SDI decoder. In one embodiment, the SDI decoder outputs to a Y CrCbCcCy -RGBCMY converter. The converter outputs RGBCMY data, with the luminance component (Y) subtracted. RGBCMY data is then converted to RGB data. This RGB data is sent to a scale sync generation component, receives adjustments to image controls, contrast, brightness, chroma, and saturation, is sent to a color correction component, and output to the display panel as LVDS data. In another embodiment the SDI decoder outputs to an SDI Y-R switch component. The SDI Y-R switch component outputs RGBCMY data. The RGBCMY data is sent to a scale sync generation component, receives adjustments to image controls, contrast, brightness, chroma, and saturation, is sent to a color correction component, and output to a display panel as LVDS data.
[00527] NARROW BAND ILLUMINATION
[00528] In one embodiment, the display uses narrow band illumination technologies. In one embodiment, the display is a laser display. In another embodiment, the display includes light emitting diodes (LEDs). In one embodiment, the LEDs include, but are not limited to, pumped phosphor LEDs, perovskite LEDs, organic LEDs (OLEDs), micro LEDs, and/or nanorods. In one embodiment, the display uses other narrow band systems (e.g., narrow filtered broad band light). Advantageously, the multi-primary systems of the present invention provide an extended gamut along a right side of the CIE 1976 curve, which is important for flesh tones. Flesh tones are important for entertainment, medical, and/or scientific purposes. There is a long-standing, unmet need for an extended gamut providing more accurate flesh tones. Additionally, the multi-primary systems of the present invention provide an extended gamut in the cyan region of the CIE 1976 curve. The extension into the cyan area as well as into the shorter green area expands the reproduction of foliage, ice, and other natural items.
[00529] AT LEAST FOUR PRIMARIES
[00530] As previously described, in one embodiment, the multi-primary system of the present invention includes at least four primaries. In one embodiment, a first wavelength corresponding to a first primary is 460nm, a second wavelength corresponding to a second primary is 493nm, a third wavelength corresponding to a third primary is 540nm, and a fourth wavelength corresponding to a fourth primary is 640nm as shown in Table 20. In one embodiment, the first wavelength, the second wavelength, the third wavelength, and/or the fourth wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the first wavelength, the second wavelength, the third wavelength, and/or the fourth wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, and/or the fourth wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, and/or the fourth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
[00531] TABLE 20
[00532] FIG. 84 illustrates a graph of the four primaries listed in Table 20 with respect to CIE 1931. In one embodiment, the at least four primaries encompass 75.57% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 84. In a preferred embodiment, the at least four primaries encompass at least 75% of the total area covered between 400nm and 700 nm for CIE 1931. Alternatively, the at least four primaries encompass at least 70% of the total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least four primaries encompass at least 65% of the total area covered between 400nm and 700nm for CIE 1931. [00533] In one embodiment, the at least four primaries include at least one white emitter.
In one embodiment, the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
Advantageously, using a D65 white emitter eliminates most of the problems with metamerism. In a preferred embodiment, the at least one white emitter is a single white emitter that matches the white point (e.g., a D65 white emitter for a D65 white point). In another embodiment, the at least one white emitter is at least two white emitters. The at least two white emitters are preferably separated such that a linear combination of the at least two white emitters covers a desired white Kelvin range. In one embodiment, the at least two white emitters include a D65 white emitter and a D27 white emitter. In another embodiment, the at least two white emitters include a D65 white emitter and a D25 white emitter.
[00534] In yet another embodiment, the at least two white emitters include three white emitters. In one embodiment, the three white emitters include a D65 white emitter, a D45 white emitter, and a D27 white emitter. Alternatively, the three white emitters include a D65 white emitter, a mid-Kelvin white emitter (e.g., D45), and a D27 white emitter. In a preferred embodiment, the mid-Kelvin white emitter includes a green bias. Advantageously, the green bias compensates for the slight magenta shift (e.g., when going from D25 to D65 with the straight line between the two points below the blackbody locus). Colors near the white locus and beyond are then a combination of the at least two white emitters (e.g., two white emitters, three white emitters). A majority of colors will have a white component that is broad band. Therefore, the resultant spectra of a mixture of color primaries and white primaries will also be broad band with an extent dependent on an amount of the at least one white primary. A higher broad band character of light results in fewer metameric problems. This is due to a white point being comprised of a combination of color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.) in a non-white emitter system. Total luminance is then related to intensities of the color primaries (e g., RGB, CMY, RGBC, RGBCMY, etc ).
[00535] Advantageously, if at least one white emitter is included, increased luminance can be achieved separate from the color primaries. Additionally, colors such as vibrantly colored pastels are attained by using the color primaries to “color shift” a bright white to the pastel. Alternatively, a fine balance of the color primaries is required, and small changes in a ratio of the color primaries will produce an unwanted color shift. Thus, a system with at least one white emitter is more tolerant to minor variations of intensity of the color primaries.
[00536] In another embodiment, the at least four primaries include RGBC, RGBW, or RG1G2B (i.e., a first green primary and a second green primary). Alternatively, the at least four primaries include RGBY.
[00537] AT LEAST FIVE PRIMARIES
[00538] As previously described, in one embodiment, the multi-primary system of the present invention includes at least five primaries. In one embodiment, a first wavelength corresponding to a first primary is 460nm, a second wavelength corresponding to a second primary is 485nm, a third wavelength corresponding to a third primary is 510nm, a fourth wavelength corresponding to a fourth primary is 535nm, and a fifth wavelength corresponding to a fifth primary is 640nm as shown in Table 21. In one embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, and/or the fifth wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, and/or the fifth wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, and/or the fifth wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, the fourth primary, and/or the fifth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-1 OOnm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
[00539] TABLE 21
[00540] FIG. 85 illustrates a graph of the five primaries listed in Table 21 with respect to CIE 1931. In one embodiment, the at least four primaries encompass 87.55% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 85. In a preferred embodiment, the at least five primaries encompass at least 87% of the total area covered between 400nm and 700 nm for CIE 1931. Alternatively, the at least five primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least five primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931. In yet another embodiment, the at least five primaries encompass at least 75% of a total area covered between 400nm and 700nm for CIE 1931.
[00541] In one embodiment, the at least five primaries include at least one white emitter. In one embodiment, the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter. Advantageously, using a D65 white emitter eliminates most of the problems with metamerism.
[00542] In another embodiment, the at least five primaries include RGBCY, RGBCW, RG1G2BW (i.e., a first green primary and a second green primary), RGBW1W2 (i.e., a first white emitter and a second white emitter), or RG1G2BY (i.e., a first green primary and a second green primary).
[00543] AT LEAST SIX PRIMARIES
[00544] As previously described, in one embodiment, the multi-primary system of the present invention includes at least six primaries. In one embodiment, a first wavelength corresponding to a first primary is 460nm, a second wavelength corresponding to a second primary is 490nm, a third wavelength corresponding to a third primary is 506nm, a fourth wavelength corresponding to a fourth primary is 520nm, a fifth wavelength corresponding to a fifth primary is 545nm, and a sixth wavelength corresponding to a sixth primary is 640nm as shown in Table 22. In one embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, and/or the sixth wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, and/or the sixth wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, and/or the sixth wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, the fourth primary, the fifth primary, and/or the sixth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
[00545] TABLE 22
[00546] FIG. 86 illustrates a graph of the six primaries listed in Table 22 with respect to CIE 1931. In one embodiment, the at least six primaries encompass 91.11% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 86. In a preferred embodiment, the at least six primaries encompass at least 90% of a total area covered between 400nm and 700 nm for CIE 1931. Alternatively, the at least six primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least six primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931. In yet another embodiment, the at least six primaries encompass at least 75% of a total area covered between 400nm and 700nm for CIE 1931.
[00547] In one embodiment, the at least six primaries include at least one white emitter. In one embodiment, the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
[00548] In another embodiment, the at least six primaries include RGBCMY, RGBCW1W2, RG1G2BW1W2 (i.e., a first green primary, a second green primary, a first white emitter, and a second white emitter), RGBW1W2W3 (i.e., a first white emitter, a second white emitter, and a third white emitter), or RGlG2BCY(i.e., a first green primary and a second green primary).
[00549] AT LEAST SEVEN PRIMARIES
[00550] As previously described, in one embodiment, the multi-primary system of the present invention includes at least seven primaries. In one embodiment, a first wavelength corresponding to a first primary is 460nm, a second wavelength corresponding to a second primary is 480nm, a third wavelength corresponding to a third primary is 495nm, a fourth wavelength corresponding to a fourth primary is 508nm, a fifth wavelength corresponding to a fifth primary is 520nm, a sixth wavelength corresponding to a sixth primary is 540nm, and a seventh wavelength corresponding to a seventh primary is 640nm as shown in Table 23. In one embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, and/or the seventh wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, and/or the seventh wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, and/or the seventh wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, and/or the seventh primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences. [00551] TABLE 23
[00552] FIG. 87 illustrates a graph of the seven primaries listed in Table 23 with respect to CIE 1931. In one embodiment, the at least four primaries encompass 91.93% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 87. In a preferred embodiment, the at least seven primaries encompass at least 90% of a total area covered between 400nm and 700 nm for CIE 1931. Alternatively, the at least seven primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least seven primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931. In yet another embodiment, the at least seven primaries encompass at least 75% of atotal area covered between 400nm and 700nm for CIE 1931.
[00553] In one embodiment, the at least seven primaries include at least one white emitter. In one embodiment, the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
[00554] AT LEAST EIGHT PRIMARIES
[00555] As previously described, in one embodiment, the multi-primary system of the present invention includes at least eight primaries. In one embodiment, a first wavelength corresponding to a first primary is 460nm, a second wavelength corresponding to a second primary is 480nm, a third wavelength corresponding to a third primary is 495nm, a fourth wavelength corresponding to a fourth primary is 500nm, a fifth wavelength corresponding to a fifth primary is 51 Inm, a sixth wavelength corresponding to a sixth primary is 521nm, a seventh wavelength corresponding to a seventh primary is 545nm, and an eighth wavelength corresponding to an eighth primary is 640nm as shown in Table 24. In one embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, and/or the eighth wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, and/or the eighth wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, and/or the eighth wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, and/or the eighth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
[00556] TABLE 24
[00557] FIG. 88 illustrates a graph of the eight primaries listed in Table 24 with respect to CIE 1931. In one embodiment, the at least eight primaries encompass 92.55% of atotal area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 88. In a preferred embodiment, the at least eight primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931. Alternatively, the at least eight primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least eight primaries encompass at least 80% of a total area covered between 400nm and 700nm. In yet another embodiment, the at least eight primaries encompass at least 75% of a total area covered between 400nm and 700nm for CIE 1931.
[00558] In one embodiment, the at least eight primaries include at least one white emitter. In one embodiment, the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
[00559] AT LEAST NINE PRIMARIES
[00560] As previously described, in one embodiment, the multi-primary system of the present invention includes at least nine primaries.
[00561] AT LEAST TEN PRIMARIES
[00562] As previously described, in one embodiment, the multi-primary system of the present invention includes at least ten primaries. In one embodiment, a first wavelength corresponding to a first primary is 440nm, a second wavelength corresponding to a second primary is 470nm, a third wavelength corresponding to a third primary is 485nm, a fourth wavelength corresponding to a fourth primary is 493nm, a fifth wavelength corresponding to a fifth primary is 502nm, a sixth wavelength corresponding to a sixth primary is 512nm, a seventh wavelength corresponding to a seventh primary is 520nm, an eighth wavelength corresponding to an eighth primary is 535nm, a ninth wavelength corresponding to a ninth primary is 550nm, and a tenth wavelength corresponding to a tenth primary is 660nm as shown in Table 25. In one embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, and/or the tenth wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, and/or the tenth wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, and/or the tenth wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, the eighth primary, the ninth primary, and/or the tenth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
[00563] TABLE 25
[00564] FIG. 89 illustrates a graph of the ten primaries listed in Table 25 with respect to CIE 1931. In one embodiment, the at least ten primaries encompass 97.16% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 89. In a preferred embodiment, the at least ten primaries encompass at least 95% of a total area covered between 400nm and 700 nm for CIE 1931. Alternatively, the at least ten primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least ten primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931. In yet another embodiment, the at least ten primaries encompass at least 80% of a total area covered between 400nm and 700nm for CIE 1931.
[00565] In one embodiment, the at least ten primaries include at least one white emitter. In one embodiment, the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
[00566] AT LEAST ELEVEN PRIMARIES
[00567] As previously described, in one embodiment, the multi-primary system of the present invention includes at least eleven primaries. [00568] AT LEAST TWELVE PRIMARIES
[00569] As previously described, in one embodiment, the multi-primary system of the present invention includes at least twelve primaries. In one embodiment, a first wavelength corresponding to a first primary is 440nm, a second wavelength corresponding to a second primary is 470nm, a third wavelength corresponding to a third primary is 485nm, a fourth wavelength corresponding to a fourth primary is 493nm, a fifth wavelength corresponding to a fifth primary is 500nm, a sixth wavelength corresponding to a sixth primary is 505nm, a seventh wavelength corresponding to a seventh primary is 51 Inm, an eighth wavelength corresponding to an eighth primary is 517nm, a ninth wavelength corresponding to a ninth primary is 523nm, a tenth wavelength corresponding to a tenth primary is 535nm, an eleventh wavelength corresponding to an eleventh primary is 550nm, and a twelfth wavelength corresponding to a twelfth primary is 670nm as shown in Table 26. In one embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is approximately (e.g., within ±10%) the value listed in the table below.
Alternatively, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, the eighth primary, the ninth primary, the tenth primary, and/or the twelfth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
[00570] TABLE 26
[00571] FIG. 90 illustrates a graph of the twelve primaries listed in Table 26 with respect to CIE 1931. In one embodiment, the at least twelve primaries encompass 97.91% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 90. In a preferred embodiment, the at least twelve primaries encompass at least 95% of a total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least twelve primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931. In yet another embodiment, the at least twelve primaries encompass at least 85% of a total area covered between 400nm and 700nm for CIE 1931.
[00572] In another embodiment, a first wavelength corresponding to a first primary is 400nm, a second wavelength corresponding to a second primary is 468nm, a third wavelength corresponding to a third primary is 484nm, a fourth wavelength corresponding to a fourth primary is 493nm, a fifth wavelength corresponding to a fifth primary is 500nm, a sixth wavelength corresponding to a sixth primary is 506nm, a seventh wavelength corresponding to a seventh primary is 512nm, an eighth wavelength corresponding to an eighth primary is 518nm, a ninth wavelength corresponding to a ninth primary is 524nm, a tenth wavelength corresponding to a tenth primary is 535nm, an eleventh wavelength corresponding to an eleventh primary is 556nm, and a twelfth wavelength corresponding to a twelfth primary is 700nm as shown in Table 27. In one embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ±5% of the value listed in the table below. In yet another embodiment, the first wavelength, the second wavelength, the third wavelength, the fourth wavelength, the fifth wavelength, the sixth wavelength, the seventh wavelength, the eighth wavelength, the ninth wavelength, the tenth wavelength, the eleventh wavelength, and/or the twelfth wavelength is within ±2% of the value listed in the table below. In one embodiment, the first primary, the second primary, the third primary, the fourth primary, the fifth primary, the sixth primary, the seventh primary, the eighth primary, the ninth primary, the tenth primary, and/or the twelfth primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences.
[00573] TABLE 27
A, (nm) X y z
400 0.173333 0.004796 0.821870
471 0.121467 0.062582 0.815951
484 0.073437 0.185017 0.741546
493 0.031757 0.363575 0.604668
500 0.008168 0.538398 0.453434
506 0.004646 0.675876 0.319478
512 0.022245 0.779613 0.198142
518 0.059329 0.829413 0.111258
524 0.106028 0.829166 0.064806
538 0.215042 0.765581 0.019377
556 0.344530 0.652011 0.003459
700 0.734705 0.265295 0.000000
[00574] FIG. 91 illustrates a graph of the twelve primaries listed in Table 27 with respect to CIE 1931. In one embodiment, the at least twelve primaries encompass 99.14% of a total area covered between 400nm and 700nm for CIE 1931 as shown in FIG. 91. In a preferred embodiment, the at least twelve primaries encompass at least 97.5% of a total area covered between 400nm and 700nm for CIE 1931. In another embodiment, the at least twelve primaries encompass at least 95% of a total area covered between 400nm and 700nm for CIE 1931. In yet another embodiment, the at least twelve primaries encompass at least 90% of a total area covered between 400nm and 700nm for CIE 1931.
[00575] In yet another embodiment, a twelve primary system is backwards compatible with 6P-C. In one embodiment, the twelve primary system includes a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary as shown in Table 28. In one embodiment, the magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the green-cyan primary, the green primary, the yellow-green primary, the yellow primary, the red-yellow primary, the red primary, and/or the magenta-red primary is approximately (e.g., within ±10%) the value listed in the table below. Alternatively, the magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the green-cyan primary, the green primary, the yellow-green primary, the yellow primary, the red-yellow primary, the red primary, and/or the magenta-red primary is within ±5% of the value listed in the table below. In yet another embodiment, the magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the green-cyan primary, the green primary, the yellow-green primary, the yellow primary, the red-yellow primary, the red primary, and/or the magenta-red primary is within ±2% of the value listed in the table below. In one embodiment, , the magenta primary, the blue-magenta primary, the blue primary, the cyan-blue primary, the cyan primary, the greencyan primary, the green primary, the yellow-green primary, the yellow primary, the red- yellow primary, the red primary, and/or the magenta-red primary has a bandwidth that is very narrow (e.g., l-10nm half band width), medium (e.g., 10-50nm half band width), and/or wide (e.g., 50-100nm half band width). The bandwidth chosen depends on the application. For example, wider bandwidth primaries will have fewer metameric perceptual differences. [00576] TABLE 28
[00577] FIG. 92 illustrates a graph of the twelve primaries listed in Table 28 with respect to CIE 1931.
[00578] In one embodiment, the at least twelve primaries include at least one white emitter. In one embodiment, the at least one white emitter includes a D65 white emitter, a D60 white emitter, a D45 white emitter, a D27 white emitter, and/or a D25 white emitter.
[00579] MULTI-PRIMARY VOLUMES
[00580] In a preferred embodiment, the multi-primary system has a larger volume than that described in ITU-R BT.2020, which is detailed in ITU-R BT.2020 (2015) and ITU-R BT.2100 (2018). ITU-R BT.2020 covers 75.8% of the CIE 1931 color space, which is described in CIE (1932). Commission intemationale de 1'Eclairage proceedings, 1931. Cambridge: Cambridge University Press and Smith, Thomas; Guild, John (1931-32). "The C.I.E. colorimetric standards and their use". Transactions of the Optical Society. 33 (3): 73-
134. doi: 10.1088/1475-4878/33/3/301, each of which is incorporated herein by reference in its entirety. Further, ITU-R BT.2020 has a red primary at (0.708, 0.292), a green primary at (0.17, 0.797), and a blue primary at (0.131, 0.046). FIG. 102A illustrates a front view of a three-dimensional plot of ITU-R BT.2020 in XYZ space. FIG. 102B illustrates a normal orthogonal view of a three-dimensional plot of ITU-R BT.2020 in XYZ space. FIG. 102C illustrates a top view of a three-dimensional plot of ITU-R BT.2020 in XYZ space. As can be seen from these plots, the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid. [00581] Additionally or alternatively, the multi-primary system has a larger volume than that described by DCI-P3 (“P3”), which is detailed in SMPTE EG 432-1 (2010) and SMPTE RP 431-2 (2011), each of which is incorporated herein by reference in its entirety. DCI-P3 covers 45.5% of the CIE 1931 color space. Further, DCI-P3 with a D65 white point has a red primary at (0.680, 0.320), a green primary at (0.265, 0.690), and a blue primary at (0.150, 0.060). FIG. 103A illustrates a front view of a three-dimensional plot of DCI-P3 in XYZ space. FIG. 103B illustrates a normal orthogonal view of a three-dimensional plot of DCI-P3 in XYZ space. FIG. 103C illustrates atop view of a three-dimensional plot of DCI-P3 in XYZ space.
[00582] In one embodiment, the multi-primary system has the primary values listed in Table 3 (“6P-C”). FIG. 104A illustrates a front view of 6P-C in XYZ space. FIG. 104B illustrates a normal orthogonal view of 6P-C in XYZ space. FIG. 104C illustrates atop view of 6P-C in XYZ space.
[00583] FIG. 105A illustrates a front view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space. FIG. 105B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space. FIG. 105C illustrates atop view of ITU-R BT.2020 (yellow) and 6P-C (green) in XYZ space. Again, the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid. The volume of 6P-C is a rhombic cuboid with extensions beyond ITU-R
BT.2020 (shown in green). The extension on the far Y side of the rhombic cuboid is a triangular prism. The extension toward the near X side is hexagonal prism.
[00584] FIG. 106A illustrates a front view of DCI-P3 (red) and 6P-C (green) in XYZ space. FIG. 106B illustrates a normal orthogonal view of DCI-P3 (red) and 6P-C (green) in XYZ space. FIG. 106C illustrates a top view of DCI-P3 (red) and 6P-C (green) in XYZ space.
[00585] In one embodiment, the multi-primary system has four primaries with a red primary at about (0.6433, 0.3192), a green primary at about (0.3244, 0.6300), a blue primary at about (0.1513, 0.0748), and a cyan primary at about (0.0729, 0.3953) (“4P”). FIG. 107A illustrates a front view of 4P in XYZ space. FIG. 107B illustrates a normal orthogonal view of 4P in XYZ space. FIG. 107C illustrates atop view of 4P in XYZ space.
[00586] FIG. 108A illustrates a front view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space. FIG. 108B illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space. FIG. 108C illustrates atop view of ITU-R BT.2020 (yellow) and 4P (blue) in XYZ space. Again, the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid. The volume of 4P is a rhombic cuboid with extensions beyond ITU-R BT.2020. The extension on the far Y side of the rhombic cuboid is a triangular prism. The extension toward the near X side is hexagonal prism.
[00587] FIG. 109A illustrates a front view of DCI-P3 (red) and 4P (blue) in XYZ space. FIG. 109B illustrates a normal orthogonal view of DCI-P3 (red) and 4P (blue) in XYZ space. FIG. 109C illustrates atop view of DCI-P3 (red) and 4P (blue) in XYZ space.
[00588] In one embodiment, the multi-primary system has four primaries with a red primary at about (0.6822, 0.3137), a green primary at about (0.2680, 0.7070), a blue primary at about (0.1367, 0.0543), and a cyan primary at about (0.0731, 0.3244) (“4P-N”). FIG. 110A illustrates a front view of 4P-N in XYZ space. FIG. HOB illustrates a normal orthogonal view of 4P-N in XYZ space. FIG. 110C illustrates a top view of 4P-N in XYZ space.
[00589] FIG. 111 A illustrates a front view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space. FIG. 11 IB illustrates a normal orthogonal view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space. FIG. 111 C illustrates a top view of ITU-R BT.2020 (yellow) and 4P-N (blue) in XYZ space. Again, the volume of ITU-R BT.2020 in XYZ space is a rhombic cuboid. The volume of 4P-N is a rhombic cuboid with extensions beyond ITU-R BT.2020. The extension on the far Y side of the rhombic cuboid is a triangular prism. The extension toward the near X side is hexagonal prism.
[00590] FIG. 112A illustrates a front view of DCI-P3 (red) and 4P-N (blue) in XYZ space. FIG. 112B illustrates a normal orthogonal view of DCI-P3 (red) and 4P-N (blue) in XYZ space. FIG. 112C illustrates a top view of DCI-P3 (red) and 4P-N (blue) in XYZ space.
[00591] TABLE 29
[00592] IMAGE
[00593] The system is operable to display an image on a viewing device (e.g., display). In one embodiment, the image includes colors outside of an ITU-R BT.2020 color gamut, a P3 color gamut, and/or an ITU-R BT.709 color gamut. The ITU-R BT.2020 color gamut is described in ITU-R BT.2020-2 (2015), which is incorporated herein by reference in its entirety. The P3 color gamut is described in SMPTE-EG-0432-1 (2010), which is incorporated herein by reference in its entirety. The ITU-R BT.709 color gamut is described in ITU-R BT.709-6 (2015), which is incorporated herein by reference in its entirety. [00594] The image preferably includes colors outside of the ITU-R BT.2020 color gamut. The ITU-R BT.2020 color gamut covers 75.8% of the CIE 1931 color space. The ITU-R BT.2020 color gamut is defined as a triangle having a first vertex at (0.170, 0.797), a second vertex at (0.708, 0.292), and a third vertex at (0.131, 0.046). In one embodiment, the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.131, 0.046), and a third vertex at about (0.0454, 0.295) within a CIE 1931 color space. The third vertex corresponds to a wavelength of about 490 nm. Advantageously, this provides an expanded color gamut in the cyan region. In another embodiment, the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.708, 0.292), and a third vertex at about (0.266, 0.724) within a CIE 1931 color space. Advantageously, this provides an expanded color gamut in the yellow region. The third vertex corresponds to a wavelength of about 545 nm. In yet another embodiment, the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.708, 0.292), a second vertex at (0.131, 0.046), and a third vertex at about (0.718, 0.281) within a CIE 1931 color space. The third vertex corresponds to a wavelength of about 640 nm. Advantageously, this provides an expanded color gamut in the magenta region. [00595] In one embodiment, the colors outside of the ITU-R BT.2020 color gamut, the P3 color gamut, and/or the ITU-R BT.709 color gamut are obtained from a camera (e.g., video, still image) operable to obtain the colors. Additionally or alternatively, the image is modified from an original image to include the colors outside of the ITU-R BT.2020 color gamut, the P3 color gamut, and/or the ITU-R BT.709 color gamut. Colorists routinely push colors to places they were not in an original image. If colorists are given an even larger gamut, they can push color to a greater extent, even far beyond what the “real” colors actually were. [00596] Including the colors outside of the ITU-R BT.2020 color gamut, the P3 color gamut, and/or the ITU-R BT.709 color gamut provides more color fidelity that the eye can see while still within the CIE diagram boundaries. For example, astronauts often see colors in space that are unable to be reproduced using current display technology. Additionally, when processing images for virtual production, the wider color gamut provides not only an opportunity to provide additional colors (e.g., cyan), but also produces more accurate flesh tones on the complementary side as well due to the wider color gamut.
[00597] In one embodiment, the image is recognizable when separated into an RGB image and a CMY image in an RGBCMY system. The RGB image and the CMY image preferably have no artifacts. In one embodiment, the image is produced from a conversion to XYZ coordinates from an original image using at least four triads. Each of the at least four triads includes three of the at least four primaries. The XYZ coordinates are multiplied by at least four XYZ-to-triad matrices to determine one or more of the at least four triads in which the XYZ coordinates are located. A sum of primary components of the one or more of the at least four triads is determined on a per-component basis and the sum is divided by a number of the one or more of the at least four triads. In one embodiment, the at least four primaries include at least four color primaries and a virtual primary (e.g., white point). In one embodiment, each of the at least four triads includes two adjacent primaries of the at least four color primaries and the virtual primary. Systems with more than four primaries are overdetermined, and using at least four triads to convert the image to the multi-primary system increases accuracy and reduces artifacts. Additional details regarding the at least four triads are included in U.S. Patent Application No. 17/180,441, which is incorporated herein by reference in its entirety. [00598] The image is preferably produced using an algorithm that minimizes and/or avoids non-matches, non-smoothness, and/or spurious matches. For example, non-matches result when a first combination of primaries and a second combination of primaries have equal XYZ coordinates, and may appear slightly different to viewers although they appear the same to the standard observer. Further, non-smoothness result when a color scale is perceived as abruptly changing by a viewer despite being a continuous curve through the gamut due to combinations of primaries to create the color scale. Additionally, spurious matches result from conditions (e.g., ambient lighting conditions, filters) that cause a first color and a second color having different XYZ coordinates to appear the same.
[00599] In one embodiment, the image is modified from an original image to include a digital watermark. In one embodiment, the digital watermark is outside of the ITU-R BT.2020 color gamut. In one embodiment, the digital watermark is compressed, collapsed, and/or mapped to an edge of the smaller color gamut such that it is not visible and/or not detectable when displayed on a viewing device with a smaller color gamut than ITU-R BT.2020. In another embodiment, the digital watermark is not visible and/or not detectable when displayed on a viewing device with an ITU-R BT.2020 color gamut. In one embodiment, the digital watermark is a watermark image (e.g., logo), alphanumeric text (e.g., unique identification code), and/or a modification of pixels. In one embodiment, the digital watermark is invisible to the naked eye. In a preferred embodiment, the digital watermark is perceptible when decoded by an algorithm. In one embodiment, the algorithm uses an encryption key to decode the digital watermark. In another embodiment, the digital watermark is visible in a non-obtrusive manner (e.g., at the bottom right of the screen). The digital watermark is preferably detectable after size compression, scaling, cropping, and/or screenshots. In yet another embodiment, the digital watermark is an imperceptible change in sound and/or video. [00600] DISPLAY
[00601] As previously described, the image is operable to be displayed on a viewing device. In one embodiment, the viewing device is a smartphone, a tablet, a laptop screen, a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a miniLED display, a microLED display, a liquid crystal display (LCD), a quantum dot display, a quantum nano emitting diode (QNED) device, a personal gaming device, a virtual reality (VR) device and/or an augmented reality (AR) device, an LED wall, a wearable display (e.g., VR/AR headset), and/or at least one projector. In one embodiment, the at least one projector includes more than one aligned and/or synchronized projector (e.g., manually, automatically via software). In one embodiment, the viewing device is foldable and/or flexible.
[00602] In one embodiment, the viewing device is operable to display colors outside of an ITU-R BT.2020 color gamut, a P3 color gamut, and/or an ITU-R BT.709 color gamut. The viewing device preferably is operable to display colors outside of the ITU-R BT.2020 color gamut. The ITU-R BT.2020 gamut covers 75.8% of the CIE 1931 color space. The viewing device is preferably operable to display at least 76% of the CIE 1931 color space. In a more preferred embodiment, the viewing device is operable to display at least 80% of the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 85% of the CIE 1931 color space. In another embodiment, the viewing device is operable to display at least 90% of the CIE 1931 color space. In yet another embodiment, the viewing device is operable to display at least 95% of the CIE 1931 color space. In still another embodiment, the viewing device is operable to display at least 97% of the CIE 1931 color space.
[00603] Advantageously, the viewing device is constructed and configured to display at least four primaries. Increasing the number of primaries in the viewing device to at least four primaries increases color accuracy of the viewing device relative to conventional RGB displays. Additionally, this allows for accurate display of colors that are traditionally difficult to reproduce on conventional RGB displays. The viewing devices includes at least one component to provide the at least four primaries (e.g., at least one color wheel, a plurality of LEDs, etc.). In one embodiment, the at least four primaries include red, green, blue, and cyan. In another embodiment, the at least four primaries include red, green, blue, cyan, and yellow. In yet another embodiment, the at least four primaries include red, green, blue, cyan, yellow, and magenta. In still another embodiment, the at least four primaries include red, a first green, a second green, and blue. In yet another embodiment, the at least four primaries includes at least one white primary. For example, teal is a color that is difficult to reproduce using conventional RGB displays. Adding a cyan primary increases the color accuracy of teal and the sensitivity of the display to colors in the region between green and blue on an RGBC display when compared to a conventional RGB display.
[00604] The viewing device is preferably operable to display flesh tones with increased color accuracy. As previously described, flesh tones are important for entertainment, medical, and/or scientific purposes. In particular, the ability to identify and detect flesh tones is important for diagnostic imaging related to the skin and other organs (e.g., brain, lungs, etc.). A person’s skin tone can vary slightly due to a number of factors, but the two main influences are health and emotion. The human visual system has been optimized to detect small changes in skin reflectivity due to blood flow and oxygenation. The M (green) and L (red) cones are operable to detect these changes. There is a long-standing, unmet need for an extended gamut providing more accurate flesh tones. In a preferred embodiment, the viewing device is an RGBCMY viewing device. The viewing device preferably includes a yellow primary. In one embodiment, the viewing device has a red primary with a longer wavelength than 615 nm. Flesh tones often appear yellowish or reddish after color correction. Additionally, skin often appears shiny after color correction. Increasing a cyan component and/or a magenta component improves the color accuracy of the flesh tones and reduces the shiny appearance of skin.
[00605] The viewing device is preferably operable to display natural surfaces (e.g., natural reflective surfaces) with increased color accuracy. As previously described, the multi-primary systems of the present invention provide an extended gamut in the cyan region. The extension into the cyan area as well as into the shorter wavelength green area expands the reproduction of foliage, water, ice, and other natural items. There is a long-standing, unmet need for an extended gamut providing more accurate reproduction of natural items.
[00606] In one embodiment, the viewing device includes pixels in a hexagonal shape. In one embodiment, the viewing device includes six primaries and each pixel in the six-primary color system is a hexagonal shape. Each hexagonal pixel is divided into six equilateral triangles and each of the primaries in the six-primary color system is displayed by one of the six equilateral triangles as described in U.S. Patent Application No. 12/005,931, filed July 3, 2008, which is incorporated herein by reference in its entirety.
[00607] In one embodiment, each pixel in the viewing device is comprised of subpixels of the same size and area arranged in at least one row and/or at least one column. In one embodiment, each pixel is divided into six subpixels of the same size and area arranged in two rows of three columns for a six-primary color system. In another embodiment, each pixel is divided into six subpixels of the same size and area arranged in three rows of two columns. In yet another embodiment, each pixel is divided into six subpixels of the same size and area arranged in one row. In yet another embodiment, each pixel is divided into six subpixels of the same size and area arranged in one column. The luminance and intensity of each subpixel is dependent on the luminance and intensity of the adjacent subpixels in order to minimize the distinct visibility of individual subpixel and pixel structures. In one embodiment, complementary primary color subpixels are adjacent to each other to eliminate visual artifacts.
[00608] In one embodiment, each pixel is divided into subpixels of different shapes, sizes, and/or areas. The size and number of subpixels for each primary color minimize blue and cyan spatial resolution without affecting the overall resolution of the viewing device as described in U.S. Patent Application No. 12/909,742, filed October 21, 2010, now U.S. Patent No. 8,451,405, which is incorporated herein by reference in its entirety.
[00609] In another embodiment, each pixel unit is divided into two subpixel units wherein one of the two subpixels is a first set of primaries and the other subpixel is a second set of primaries. In one embodiment, the second set of primaries is complementary to the first set of primaries. In one embodiment, one of the two subpixels is an RGB color and the other subpixel is the complementary CMY color of the first subpixel as described in U.S. Patent Application No. 12/229,845, filed March 5, 2009, which is incorporated herein by reference in its entirety.
[00610] In one embodiment, each pixel includes at least one white subpixel to eliminate visual artifacts. In one embodiment, the at least one white subpixel includes a D65 white subpixel, a D60 white subpixel, a D45 white subpixel, a D27 white subpixel, and/or a D25 white subpixel. Advantageously, using a D65 white subpixel eliminates most of the problems with metamerism. In a preferred embodiment, the at least one white subpixel is a single white subpixel that matches the white point (e.g., a D65 white subpixel for a D65 white point). In another embodiment, the at least one white subpixel is at least two white subpixels. The at least two white subpixels are preferably separated such that a linear combination of the at least two white subpixels covers a desired white Kelvin range. In one embodiment, the at least two white subpixels include a D65 white subpixel and a D27 white subpixel. In another embodiment, the at least two white subpixels include a D65 white subpixel and a D25 white subpixel.
[00611] In yet another embodiment, the at least two white subpixels includes three white subpixels. In one embodiment, the three white subpixels include a D65 white subpixel, a D45 white subpixel, and a D27 white subpixel. Alternatively, the three white subpixels include a D65 white subpixel, a mid-Kelvin white subpixel (e.g., D45), and a D27 white subpixel. In a preferred embodiment, the mid-Kelvin white subpixel includes a green bias. Advantageously, the green bias compensates for the slight magenta shift (e.g., when going from D25 to D65 with the straight line between the two points below the blackbody locus). Colors near the white locus and beyond are then a combination of the at least two white subpixels (e.g., two white subpixels, three white subpixels). A majority of colors will have a white component that is broad band. Therefore, the resultant spectra of a mixture of color primaries and white primaries will also be broad band with an extent dependent on an amount of the at least one white primary. A higher broad band character of light results in fewer metameric problems. This is due to a white point being comprised of a combination of color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.) in anon-white subpixel system. Total luminance is then related to intensities of the color primaries (e.g., RGB, CMY, RGBC, RGBCMY, etc.).
[00612] Advantageously, if at least one white subpixel is included, increased luminance can be achieved separate from the color primaries. Additionally, colors such as vibrantly colored pastels are attained by using the color primaries to “color shift” a bright white to the pastel. Alternatively, a fine balance of the color primaries is required, and small changes in a ratio of the color primaries will produce an unwanted color shift. Thus, a system with at least one white subpixel is more tolerant to minor variations of intensity of the color primaries. [00613] In one embodiment, the white point of the multi-primary color system changes depending on the viewing device or the display mode. In one embodiment, the addition of white subpixels widens the bandwidth of the filter for each non-white primary.
[00614] In one embodiment, each pixel is formed of fewer than the at least four primaries (e.g., three of four primaries, four of five primaries, five of six primaries, etc.). In one embodiment, each pixel is composed of fewer than six primary colors from the 6P gamut. The viewing device is composed of alternating and repeating subpixel patterns. In another embodiment, the viewing device is composed of nonrepeating subpixel patterns.
[00615] In one embodiment, the subpixel colors in a pixel and in adjacent pixels are arranged to minimize the spatial distance between colors that have maximal color distance from each other as described in U.S. Patent Application No. 10/543,511, filed January 13, 2003, now U.S. Patent No. 8,228,275, which is incorporated herein by reference in its entirety.
[00616] In one embodiment, each pixel is one single primary color from the multi-primary system (e.g., 6P gamut). In one embodiment, patterns of pixels are repeated across the viewing device to minimize visibility of individual pixel structures as described in U.S.
Patent Application No. 13/512,914, filed November 25, 2010, which is incorporated herein by reference in its entirety.
[00617] In one embodiment, the viewing device includes at least one perovskite. In one embodiment, the at least one perovskite is a lead halide perovskite. In one embodiment, the at least one perovskite is used as a quantum dot nanocrystal. In one embodiment, the at least one perovskite is a perovskite polymer bead. When light shines through the perovskite polymer bead, the color changes depending on the composition of the perovskite polymer bead (e.g., green, red, etc.). In one embodiment, the at least one perovskite is incorporated into a perovskite LED. Examples of perovskite LEDs are described in Lin, K., et al. (2018). Perovskite light-emitting diodes with external quantum efficiency exceeding 20 per cent. Nature, 562(7726), 245-248, which is incorporated herein by reference in its entirety. In one embodiment, the at least one perovskite is 3D printed. See, e.g., Zhou, Nanjia, Yehonadav Bekenstein, CarissaN. Eisler, Dandan Zhang, Adam M. Schwartzberg, Peidong Yang, A. Paul Alivisatos, and Jennifer A. Lewis. 2019. “Perovskite Nanowire-Block Copolymer Composites With Digitally Programmable Polarization Anisotropy.” Science Advances, which is incorporated herein by reference in its entirety.
[00618] In yet another embodiment, the viewing device is a direct emissive assembled display. The design for a direct emissive assembled display includes a matrix of color emitters grouped as a multi -primary color system (e.g., 6P system). Individual channel inputs drive each Quantum Dot (QD) element illuminator and/or micro LED element. In one embodiment, the quantum dots modulate light according to image data as described in U.S. Patent Application No. 15/905,085, filed February 26, 2018, now U.S. Patent No.
10,373,574, which is incorporated herein by reference in its entirety.
[00619] Additional details about the multi-primary system and the display are included in U.S. Application No. 17/180,441 and U.S. Patent Publication Nos. 20210027693, 20210020094, 20210035487, and 20210043127, each of which is incorporated herein by reference in its entirety.
[00620] TRANSPORTATION OF UP TO 6 PRIMARIES
[00621] To transport up to six color components (e.g., four, five, or six), System 1, System 2, or System 3 can be used as previously described. If four color components are used, two of the channels are set to “0”. If five color components are used, one of the channels is set to “0”.
[00622] TRANSPORTATION OF MORE THAN 6 PRIMARIES [00623] To allow for transportation of more than six color components (e.g., seven, eight, nine, ten, eleven, twelve), a hybrid of System 2 and System 3 is used to transport up to and including twelve individual color channels. FIG. 93 shows one embodiment of transportation of twelve individual color channels using the example in Table 28 with a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary on a first link (Link A) and a second link (Link B).
[00624] FIG. 94A shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a first link (Link A). FIG. 94B shows one embodiment of a 4:2:2 Non-Constant Luminance Encode for a second link (Link B).
[00625] FIG. 95 A shows one embodiment of a 4:2:2 Constant Luminance Encode for a first link (Link A). FIG. 95B shows one embodiment of a 4:2:2 Constant Luminance Encode for a second link (Link B).
[00626] FIG. 96A shows one embodiment of a 4:4:4 Encode for a first link (Link A). FIG. 96B shows one embodiment of a 4:4:4 Encode for a second link (Link B).
[00627] FIG. 97A shows one embodiment of component mapping into SMPTE 2081-1 for a first link (Link A). FIG. 97B shows one embodiment of component mapping into SMPTE 2081-1 for a second link (Link B).
[00628] FIG. 98A shows one embodiment of R,G,B,C,M,Y,GC,MR,BM,YG,RY,CB mapping into SMPTE 2081-1 for a first link (Link A). FIG. 98B shows one embodiment of R,G,B,C,M,Y,GC,MR,BM,YG,RY,CB mapping into SMPTE 2081-1 for a second link (Link B).
[00629] FIG. 99A shows one embodiment of a 4:2:2 Non-Constant Luminance Decode for a first link (Link A). FIG. 99B shows one embodiment of a 4:2:2 Non-Constant Luminance
Decode for a second link (Link B). [00630] FIG. 100A shows one embodiment of a 4:2:2 Constant Luminance Decode for a first link (Link A). FIG. 100B shows one embodiment of a 4:2:2 Constant Luminance Decode for a second link (Link B).
[00631] FIG. 101A shows one embodiment of a 4:4:4 Decode for a first link (Link A). FIG. 101B shows one embodiment of a 4:4:4 Decode for a second link (Link B).
[00632] If seven color components are used, five of the channels in the hybrid of System 2 and System 3 are set to “0”. If eight color components are used, four of the channels in the hybrid of System 2 and System 3 are set to “0”. If nine color components are used, three of the channels in the hybrid of System 2 and System 3 are set to “0”. If ten color components are used, two of the channels in the hybrid of System 2 and System 3 are set to “0”. If eleven color components are used, one of the channels in the hybrid of System 2 and System 3 is set to “0”.
[00633] In one embodiment, the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter, and a display system, wherein the display system and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the display system. In one embodiment, the image data converter includes a digital interface. In one embodiment, the digital interface is operable to encode and decode the set of image data. In one embodiment, the system further includes at least one transfer function (TF) for processing the set of image data. In one embodiment, the system further includes a set of Session Description Protocol (SDP) parameters. In one embodiment, the set of SDP parameters is modifiable. In one embodiment, the display system includes a Liquid Crystal Display (LCD) projector, wherein the LCD projector is operable to transmit light through a plurality of LCD units using at least one prism and/or at least one reciprocal mirror. In another embodiment, the display system includes a Digital Micromirror Device (DMD) projector. In one embodiment, the DMD projector includes at least one DMD chip, wherein the at least one DMD chip is synchronized with at least one light source. In one embodiment, the display system is operable to use a combination of primary color display elements and/or a combination of primary color light sources to display a different primary color. In one embodiment, the image data converter includes an alignment signal to synchronize and align (e.g., mechanically align) at least two projectors. In one embodiment, the display system includes an apparatus to combine the output display of the at least two projectors, thereby creating a combined output display. In one embodiment, the at least four primary color values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary. In one embodiment, the set of image data further includes a bit level, a first set of color channel data, and a second set of color channel data. In one embodiment, the image data converter is operable to create a combined set of color channel data from the first set of color channel data and the second set of color channel data for display on the display system. In one embodiment, the combined set of color channel data has a combined bit level equal to the bit level of the set of image data. [00634] In another embodiment, the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter, and a display system, wherein the display system includes at least one light source, wherein the display system and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the display system. In one embodiment, the image data converter includes a digital interface. In one embodiment, the digital interface is operable to encode and decode the set of image data. In one embodiment, the system further includes at least one transfer function (TF) for processing the set of image data. In one embodiment, the system further includes a set of
Session Description Protocol (SDP) parameters. In one embodiment, the set of SDP parameters is modifiable. In one embodiment, the at least one light source includes at least one Light-Emitting Diode (LED). In another embodiment, the at least one light source includes a Xenon lamp. In yet another embodiment, the at least one light source includes a blue laser system. In one embodiment, the at least four primary color values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary.
[00635] In yet another embodiment, the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter, and a display system, wherein the display system includes at least one display screen, wherein the at least one display screen comprises a plurality of pixels, wherein each of the plurality of pixels is divided into a plurality of subpixels, wherein the display system and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the display system. In one embodiment, the image data converter includes a digital interface. In one embodiment, the digital interface is operable to encode and decode the set of image data. In one embodiment, the system further includes at least one transfer function (TF) for processing the set of image data. In one embodiment, the system further includes a set of Session Description Protocol (SDP) parameters. In one embodiment, the set of SDP parameters is modifiable. In one embodiment, the at least one display screen includes a Liquid Crystal Display (LCD) display screen, a Light-Emitting Diode (LED) display screen, and/or a Quantum Dot (QD) display screen. In one embodiment, the at least one display screen includes at least one white subpixel, at least two green subpixels, at least one cyan subpixel, at least one magenta subpixel, and/or at least one yellow subpixels. In one embodiment, the at least four primary color values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary. In one embodiment, the at least one display screen includes at least one perovskite. In one embodiment, the at least one display screen includes at least two display screens, wherein the display system includes a mirror apparatus (e.g., half-silvered mirror apparatus), and wherein the mirror apparatus is operable to combine the at least two display screens on a view screen. In one embodiment, the display system includes an expanded filter arrangement, and wherein the set of image data includes Low-Voltage Differential Signaling (LVDS) data.
[00636] In one embodiment, the present invention provides a system for displaying a primary color system including a set of image data, an image data converter, and at least one display device, wherein the set of image data includes primary color data for at least four primary color values, wherein the at least one display device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device. In one embodiment, the system further includes at least one transfer function (TF) for processing the set of image data. In one embodiment, the system further includes a set of Session Description Protocol (SDP) parameters. In one embodiment, the set of image data includes a first set of color channel data and a second set of color channel data. In one embodiment, the image data converter further includes a first link component and a second link component. In one embodiment, the first link component is operable to transport the first set of color channel data to the at least one display device and the second link component is operable to transport the second set of color channel data to the at least one display device in parallel with the first link component. In one embodiment, the at least one display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the at least one display device is based on the set of image data. In one embodiment, the at least four primary color values include at least one white emitter. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 493nm, a third primary at approximately 540nm, and a fourth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 485nm, a third primary at approximately 510nm, a fourth primary at approximately 535nm, and a fifth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 490nm, a third primary at approximately 506nm, a fourth primary at approximately 520nm, a fifth primary at approximately 545nm, and a sixth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 508nm, a fifth primary at approximately 520nm, a sixth primary at approximately 540nm, and a seventh primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 500nm, a fifth primary at approximately 511nm, a sixth primary at approximately 521nm, a seventh primary at approximately 545nm, and an eighth primary at approximately 640nm. In one embodiment, the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 502nm, a sixth primary at approximately 512nm, a seventh primary at approximately 520nm, an eighth primary at approximately 535nm, a ninth primary at approximately 550nm, and a tenth primary at approximately 660nm. In one embodiment, the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 505nm, a seventh primary at approximately 511nm, an eighth primary at approximately 517nm, a ninth primary at approximately 523nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 550nm, and a twelfth primary at approximately 670nm. In one embodiment, the at least four primaries include a first primary at approximately 400nm, a second primary at approximately 468nm, a third primary at approximately 484nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 506nm, a seventh primary at approximately 512nm, an eighth primary at approximately 518nm, a ninth primary at approximately 524nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 556nm, and a twelfth primary at approximately 700nm. In one embodiment, the at least four primaries include a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary. In one embodiment, the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data. In one embodiment, the set of SDP parameters is modifiable. In one embodiment, the first set of color channel data is converted by the first link component and the second set of color channel data is converted by the second link component, and wherein the first set of color channel data and the second set of color channel data are combined to form the set of image data for display on the single display device. In one embodiment, the system further includes a standardized transport format, wherein the first link component includes a first standardized transport format link and wherein the second link component includes a second standardized transport format link, wherein the standardized transport format is operable to receive the first set of image data and the second set of image data using the first standardized transport format link and the second standardized transport format link, and wherein the first standardized transport format link and the second standardized transport format link are operable to combine the first set of image data and the second set of image data into a combined set of image data.
[00637] In another embodiment, the present invention provides a system for displaying a primary color system including a set of image data, an image data converter, and at least one display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least one display device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device. In one embodiment, the image data converter includes a digital interface. In one embodiment, the digital interface is operable to encode and decode the set of image data. In one embodiment, the system further includes at least one transfer function (TF) for processing the set of image data. In one embodiment, the system further includes a set of Session Description Protocol (SDP) parameters. In one embodiment, the set of SDP parameters is modifiable. In one embodiment, the set of image data includes a first set of color channel data and a second set of color channel data. In one embodiment, the image data converter further includes a first link component and a second link component. In one embodiment, the first link component is operable to transport the first set of color channel data to the at least one display device and the second link component is operable to transport the second set of color channel data to the at least one display device in parallel with the first link component. In one embodiment, once the set of image data has been converted by the image data converter for the at least one display device, the set of SDP parameters is modified based on the conversion. [00638] In yet another embodiment, the present invention provides a system for displaying a primary color system including a set of image data, an image data converter, and at least one display device, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter, wherein the at least one display device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device. In one embodiment, the image data converter includes a digital interface. In one embodiment, the digital interface is operable to encode and decode the set of image data. In one embodiment, the system further includes at least one transfer function (TF) for processing the set of image data. In one embodiment, the system further includes a set of Session Description Protocol (SDP) parameters. In one embodiment, the set of image data includes a first set of color channel data and a second set of color channel data. In one embodiment, the image data converter further includes a first link component and a second link component. In one embodiment, the first link component is operable to transport the first set of color channel data to the at least one display device and the second link component is operable to transport the second set of color channel data to the at least one display device in parallel with the first link component. In one embodiment, the at least one white emitter includes a white emitter matching a white point of the primary color system. In one embodiment, the at least one white emitter includes at least three white emitters. In one embodiment, the at least three white emitters each have a different color temperature. In one embodiment, the at least one white emitter includes a midKelvin white emitter. In one embodiment, the mid-Kelvin white emitter is modified to include a green bias.
[00639] In one embodiment, the present invention provides a system for displaying a primary color system including a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, an image data converter, wherein the image data converter includes a digital interface, and wherein the digital interface is operable to encode and decode the set of values in Yxy color space, and at least one display device, wherein the at least one display device and the image data converter are in network communication, wherein the encode and the decode include transportation of the set of image data as Yxy data, and wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x and y, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device. In one embodiment, the at least one display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the at least one display device is based on the set of image data. In one embodiment, the image data converter is operable to convert the set of primary color signals to the set of values in Yxy color space. In one embodiment, the image data converter is operable to convert the set of values in Yxy color space to a plurality of color gamuts. In one embodiment, the image data converter is operable to fully sample the Yxy data related to the luminance Y and subsample the Yxy data related to the two colorimetric coordinates x and y. In one embodiment, the Yxy data related to the luminance Y and the two colorimetric coordinates x and y are fully sampled. In one embodiment, the set of image data is integrated into a standardized transportation format. In one embodiment, the set of values in Yxy color space includes a reference to at least one white point. In one embodiment, the Yxy data includes floating points. In one embodiment, the encode includes converting the set of primary color signals to XYZ data and then converting the XYZ data to create the set of values in Yxy color space. In one embodiment, the decode includes converting the Yxy data to XYZ data and then converting the XYZ data to a format operable to display on the at least one display device. In one embodiment, the set of image data is transported linearly without a non-linear function applied to the luminance Y.
[00640] In another embodiment, the present invention provides a system for displaying a primary color system including a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, an image data converter, wherein the image data converter includes a digital interface, and wherein the digital interface is operable to encode and decode the set of values in Yxy color space, at least one non-linear function for processing the set of values in Yxy color space, and at least one display device, wherein the at least one display device and the image data converter are in network communication, wherein the at least one non-linear function is not applied to the colorimetric coordinates x and y, and wherein the at least one non-linear function is applied to the luminance Y, thereby creating a luma Y', wherein the encode and the decode include transportation of Yxy data, and wherein the Yxy data is related to the two colorimetric coordinates x and y and the luma Y', and wherein the image data converter is operable to convert the set of image data for display on the at least one display device. In one embodiment, the at least one non-linear function includes at least one of a gamma function, a log function, a perceptual quantizer (PQ) function, an opto-electronic transfer function (OETF), an opto-optical transfer function (OOTF), and/or an electro-optical transfer function (EOTF). In one embodiment, the image data converter applies one or more of the at least one non-linear function to encode the set of values in Yxy color space. In one embodiment, the image data converter applies one or more of the at least one non-linear function to decode the set of values in Yxy color space. In one embodiment, the image data converter includes a look-up table. [00641] In yet another embodiment, the present invention provides a system for displaying a primary color system including a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, an image data converter, wherein the image data converter includes a digital interface, and wherein the digital interface is operable to encode and decode the set of values in Yxy color space, a set of Session Description Protocol (SDP) parameters, wherein the set of SDP parameters is modifiable, and at least one display device, wherein the set of image data further includes pixel mapping data, wherein the at least one display device and the image data converter are in network communication, wherein the encode and the decode include transportation of the set of image data as Yxy data, and wherein the Yxy data is related to the two colorimetric coordinates and the luminance Y, and wherein the image data converter is operable to convert the set of image data for display on the at least one display device. In one embodiment, the pixel mapping data includes a subsample of the set of values in Yxy color space. In one embodiment, the pixel mapping data includes an alignment of the set of values in Yxy color space.
[00642] In still another embodiment, the present invention provides a method for displaying a primary color system including providing a set of image data including a set of primary color signals, wherein the set of primary color signals corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, encoding the set of image data in Yxy color space using a digital interface of an image data converter, wherein the image data converter is in network communication with at least one display device, decoding the set of image data in Yxy color space using the digital interface of the image data converter, and the image data converter converting the set of image data for display on the at least one display device, wherein the encoding and the decoding include transportation of the set of image data as Yxy data, wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x andy.
[00643] In one embodiment, the present invention provides a system for displaying a digital representation of an image including the image and a viewing device, wherein the image includes colors outside of an ITU-R BT.2020 color gamut, wherein the viewing device is operable to display the digital representation of the image and at least 80% of a total area covered between about 400nm and about 700nm by an International Commission on Illumination (CIE) 1931 color space, and wherein the viewing device is operable to display the colors outside of the ITU-R BT.2020 color gamut. In one embodiment, the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.131, 0.046), and a third vertex at about (0.0454, 0.295) within the CIE 1931 color space. In another embodiment, the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at (0.170, 0.797), a second vertex at (0.708, 0.292), and a third vertex at about (0.266, 0.724) within the CIE 1931 color space. In one embodiment, the colors outside of the ITU-R BT.2020 color gamut have a chromaticity within a triangle with a first vertex at about (0.708, 0.292), a second vertex at (0.131, 0.046), and a third vertex at about (0.719, 0.281) within the CIE 1931 color space. In one embodiment, the viewing device is selected from the group consisting of a smartphone, a tablet, a laptop screen, a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a miniLED display, a microLED display, a liquid crystal display (LCD), a quantum dot display, a quantum nano emitting diode (QNED) device, a personal gaming device, a virtual reality (VR) device and/or an augmented reality (AR) device, an LED wall, a wearable display (e.g., VR/AR headset), and at least one projector. In one embodiment, the viewing device is operable to display at least 85% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 90% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 95% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least 97% of a total area covered between about 400nm and about 700nm for the CIE 1931 color space. In one embodiment, the viewing device is operable to display at least four primaries, and wherein the at least four primaries include red, green, blue, and cyan. In one embodiment, the viewing device is operable to display at least five primaries, and wherein the at least five primaries include red, green, blue, cyan, and yellow. In one embodiment, the viewing device is operable to display at least six primaries, and wherein the at least six primaries include red, green, blue, cyan, yellow, and magenta. In one embodiment, the viewing device is operable to display at least one white primary. In one embodiment, the system further includes a set of Session Description Protocol (SDP) parameters, wherein the SDP parameters include color channel data, image data, framerate data, a sampling standard, a flag indicator, an active picture size code, a timestamp, a clock frequency, a frame count, a scrambling indicator, and/or a video format indicator. In one embodiment, the image is modified from an original image to include the colors outside of the ITU-R BT.2020 color gamut.
[00644] In another embodiment, the present invention provides a system for displaying a digital representation of an image including the image, a set of image data corresponding to the image, and a viewing device, wherein the viewing device is constructed and configured to provide a cyan primary, wherein the image includes colors inside of a first color gamut, wherein the colors inside of the first color gamut are outside of an ITU-R BT.2020 color gamut, wherein the viewing device is operable to display the digital representation of the image and at least 80% of a total area covered between about 400nm and about 700nm by an
International Commission on Illumination (CIE) 1931 color space, and wherein the viewing device is operable to display the colors outside of the ITU-R BT.2020 color gamut. In one embodiment, the set of image data occupies a larger volume in the CIE 1931 color space than the ITU-R BT.2020 color gamut. In one embodiment, the set of image data is compressed and/or truncated when the set of image data is mapped to a second color gamut, wherein the second color gamut is not equivalent to the first color gamut.
[00645] In yet another embodiment, the present invention provides a method for displaying a digital representation of an image including providing a viewing device, wherein the viewing device includes at least one component to provide a cyan primary, providing the image and a set of image data corresponding to the image to the viewing device, wherein the image includes colors outside of an ITU-R BT.2020 color gamut, and displaying the digital representation of the image on the viewing device, wherein the displaying the digital representation of the image on the viewing device includes displaying the colors outside of the ITU-R BT.2020 color gamut, wherein the viewing device is operable to display at least 80% of a total area covered between about 400nm and about 700nm by an International Commission on Illumination (CIE) 1931 color space. In one embodiment, the method further includes modifying the image from an original image to include the colors outside of the ITU-R BT.2020 color gamut.
[00646] In one embodiment, the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and at least one viewing device, wherein the at least one viewing device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one viewing device. In one embodiment, the at least four primary values include at least one white primary, at least two green primaries, at least one cyan primary, at least one magenta primary, and/or at least one yellow primary. In one embodiment, the at least four primary values include a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary. In one embodiment, the at least one viewing device is selected from the group consisting of a smartphone, a tablet, a laptop screen, a light emitting diode (LED) display, an organic light emitting diode (OLED) display, a miniLED display, a microLED display, a liquid crystal display (LCD), a quantum dot display, a quantum nano emitting diode (QNED) device, a personal gaming device, a virtual reality (VR) device and/or an augmented reality (AR) device, an LED wall, a wearable display, and at least one projector. In one embodiment, the at least one viewing device includes at least one perovskite. In one embodiment, the set of SDP parameters is modifiable, and wherein once the set of image data has been converted by the image data converter for the at least one viewing device, the set of SDP parameters is modified based on the conversion. In one embodiment, the at least one viewing device includes at least one white emitter, wherein the at least one white emitter includes a mid-Kelvin white emitter, and wherein the mid-Kelvin white emitter is modified to include a green bias. In one embodiment, the image data converter includes an alignment signal to synchronize and align at least two projectors, and wherein the at least one viewing device includes an apparatus to combine the output display of the at least two projectors, thereby creating a combined output display. In one embodiment, the set of image data includes a first set of color channel data and a second set of color channel data, wherein the image data converter further includes a first link component and a second link component, wherein the first link component is operable to transport the first set of color channel data to the at least one viewing device, wherein the second link component is operable to transport the second set of color channel data to the at least one viewing device in parallel with the first link component. In one embodiment, the primary color data corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, wherein the two colorimetric coordinates x and y are orthogonal to the luminance Y, wherein the encode and the decode include transportation of Yxy data, wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x and y, and wherein the Yxy data includes pixel mapping data. In one embodiment, the system further includes at least one non-linear function for processing the set of values in Yxy color space, wherein the at least one non-linear function is not applied to the colorimetric coordinates x and y, and wherein the at least one non-linear function is applied to the luminance Y, thereby creating a luma Y'.
[00647] In another embodiment, the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes primary color data for at least four primary color values, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with at least one viewing device, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the at least one viewing device. In one embodiment, the primary color data corresponds to a set of values in Yxy color space, wherein the set of values in Yxy color space includes two colorimetric coordinates x and y and a luminance Y, wherein the two colorimetric coordinates x and y are orthogonal to the luminance Y, wherein the encode and the decode includes transportation of Yxy data, wherein the Yxy data is related to the luminance Y and the two colorimetric coordinates x and y, and wherein the Yxy data includes pixel mapping data.
[00648] In one embodiment, the present invention provides a system for displaying a primary color system including a set of image data, wherein the set of image data includes primary color data for at least four primary color values, an image data converter wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data, a set of Session Description Protocol (SDP) parameters, and at least one viewing device, wherein the at least one viewing device and the image data converter are in network communication, and wherein the image data converter is operable to convert the set of image data for display on the at least one viewing device.
[00649] In another embodiment, the present invention provides a method for displaying a multi-primary color system including providing a set of image data, wherein the set of image data includes primary color data for at least four primary color values, encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with at least one viewing device, decoding the set of image data using the digital interface of the image data converter, and converting the set of image data for display on the at least one viewing device.
[00650] In yet another embodiment, the present invention provides a system for displaying a digital representation of an image including the image and a viewing device, wherein the image includes colors outside of an ITU-R BT.2020 color gamut, wherein the viewing device is operable to display the digital representation of the image and at least 80% of a total area covered between about 400nm and about 700nm by an International Commission on Illumination (CIE) 1931 color space, and wherein the viewing device is operable to display the colors outside of the ITU-R BT.2020 color gamut. [00651] FIG. 137 is a schematic diagram of an embodiment of the invention illustrating a computer system, generally described as 800, having a network 810, a plurality of computing devices 820, 830, 840, a server 850, and a database 870.
[00652] The server 850 is constructed, configured, and coupled to enable communication over a network 810 with a plurality of computing devices 820, 830, 840. The server 850 includes a processing unit 851 with an operating system 852. The operating system 852 enables the server 850 to communicate through network 810 with the remote, distributed user devices. Database 870 may house an operating system 872, memory 874, and programs 876. [00653] In one embodiment of the invention, the system 800 includes a network 810 for distributed communication via a wireless communication antenna 812 and processing by at least one mobile communication computing device 830. Alternatively, wireless and wired communication and connectivity between devices and components described herein include wireless network communication such as WI-FI, WORLDWIDE INTEROPERABILITY FOR MICROWAVE ACCESS (WIMAX), Radio Frequency (RF) communication including RF identification (RFID), NEAR FIELD COMMUNICATION (NFC), BLUETOOTH including BLUETOOTH LOW ENERGY (BLE), ZIGBEE, Infrared (IR) communication, cellular communication, satellite communication, Universal Serial Bus (USB), Ethernet communications, communication via fiber-optic cables, coaxial cables, twisted pair cables, and/or any other type of wireless or wired communication. In another embodiment of the invention, the system 800 is a virtualized computing system capable of executing any or all aspects of software and/or application components presented herein on the computing devices 820, 830, 840. In certain aspects, the computer system 800 may be implemented using hardware or a combination of software and hardware, either in a dedicated computing device, or integrated into another entity, or distributed across multiple entities or computing devices. [00654] By way of example, and not limitation, the computing devices 820, 830, 840 are intended to represent various forms of electronic devices including at least a processor and a memory, such as a server, blade server, mainframe, mobile phone, personal digital assistant (PDA), smartphone, desktop computer, notebook computer, tablet computer, workstation, laptop, and other similar computing devices. The components shown here, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the invention described and/or claimed in the present application. [00655] In one embodiment, the computing device 820 includes components such as a processor 860, a system memory 862 having a random access memory (RAM) 864 and a read-only memory (ROM) 866, and a system bus 868 that couples the memory 862 to the processor 860. In another embodiment, the computing device 830 may additionally include components such as a storage device 890 for storing the operating system 892 and one or more application programs 894, a network interface unit 896, and/or an input/output controller 898. Each of the components may be coupled to each other through at least one bus 868. The input/output controller 898 may receive and process input from, or provide output to, a number of other devices 899, including, but not limited to, alphanumeric input devices, mice, electronic styluses, display units, touch screens, signal generation devices (e.g., speakers), or printers.
[00656] By way of example, and not limitation, the processor 860 may be a general- purpose microprocessor (e.g., a central processing unit (CPU)), a graphics processing unit (GPU), a microcontroller, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA), a Programmable Logic Device (PLD), a controller, a state machine, gated or transistor logic, discrete hardware components, or any other suitable entity or combinations thereof that can perform calculations, process instructions for execution, and/or other manipulations of information. [00657] In another implementation, shown as 840 in FIG. 137 multiple processors 860 and/or multiple buses 868 may be used, as appropriate, along with multiple memories 862 of multiple types (e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core).
[00658] Also, multiple computing devices may be connected, with each device providing portions of the necessary operations (e.g., a server bank, a group of blade servers, or a multiprocessor system). Alternatively, some steps or methods may be performed by circuitry that is specific to a given function.
[00659] According to various embodiments, the computer system 800 may operate in a networked environment using logical connections to local and/or remote computing devices 820, 830, 840 through a network 810. A computing device 830 may connect to a network 810 through a network interface unit 896 connected to a bus 868. Computing devices may communicate communication media through wired networks, direct-wired connections or wirelessly, such as acoustic, RF, or infrared, through an antenna 897 in communication with the network antenna 812 and the network interface unit 896, which may include digital signal processing circuitry when necessary. The network interface unit 896 may provide for communications under various modes or protocols.
[00660] In one or more exemplary aspects, the instructions may be implemented in hardware, software, firmware, or any combinations thereof. A computer readable medium may provide volatile or non-volatile storage for one or more sets of instructions, such as operating systems, data structures, program modules, applications, or other data embodying any one or more of the methodologies or functions described herein. The computer readable medium may include the memory 862, the processor 860, and/or the storage media 890 and may be a single medium or multiple media (e.g., a centralized or distributed computer system) that store the one or more sets of instructions 900. Non-transitory computer readable media includes all computer readable media, with the sole exception being a transitory, propagating signal per se. The instructions 900 may further be transmitted or received over the network 810 via the network interface unit 896 as communication media, which may include a modulated data signal such as a carrier wave or other transport mechanism and includes any deliver media. The term “modulated data signal” means a signal that has one or more of its characteristics changed or set in a manner as to encode information in the signal. [00661] Storage devices 890 and memory 862 include, but are not limited to, volatile and non-volatile media such as cache, RAM, ROM, EPROM, EEPROM, FLASH memory, or other solid state memory technology, discs (e.g., digital versatile discs (DVD), HD-DVD, BLU-RAY, compact disc (CD), or CD-ROM) or other optical storage; magnetic cassettes, magnetic tape, magnetic disk storage, floppy disks, or other magnetic storage devices; or any other medium that can be used to store the computer readable instructions and which can be accessed by the computer system 800.
[00662] In one embodiment, the computer system 800 is within a cloud-based network. In one embodiment, the server 850 is a designated physical server for distributed computing devices 820, 830, and 840. In one embodiment, the server 850 is a cloud-based server platform. In one embodiment, the cloud-based server platform hosts serverless functions for distributed computing devices 820, 830, and 840.
[00663] In another embodiment, the computer system 800 is within an edge computing network. The server 850 is an edge server, and the database 870 is an edge database. The edge server 850 and the edge database 870 are part of an edge computing platform. In one embodiment, the edge server 850 and the edge database 870 are designated to distributed computing devices 820, 830, and 840. In one embodiment, the edge server 850 and the edge database 870 are not designated for computing devices 820, 830, and 840. The distributed computing devices 820, 830, and 840 are connected to an edge server in the edge computing network based on proximity, availability, latency, bandwidth, and/or other factors.
[00664] It is also contemplated that the computer system 800 may not include all of the components shown in FIG. 137 may include other components that are not explicitly shown in FIG. 137 or may utilize an architecture completely different than that shown in FIG. 137. The various illustrative logical blocks, modules, elements, circuits, and algorithms described in connection with the embodiments discussed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate the interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application (e.g., arranged in a different order or positioned in a different way), but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
[00665] The above-mentioned examples are provided to serve the purpose of clarifying the aspects of the invention, and it will be apparent to one skilled in the art that they do not serve to limit the scope of the invention. By nature, this invention is highly adjustable, customizable and adaptable. The above-mentioned examples are just some of the many configurations that the mentioned components can take on. All modifications and improvements have been deleted herein for the sake of conciseness and readability but are properly within the scope of the present invention.

Claims

CLAIMS The invention claimed is:
1. A system for displaying a primary color system, comprising: a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data; an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data; a set of Session Description Protocol (SDP) parameters; and a display device, wherein the display device is a single display device; wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter; wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter; wherein the single display device and the image data converter are in network communication; wherein the image data converter further includes a first link component and a second link component; wherein the image data converter is operable to convert the set of image data for display on the single display device; and wherein the first link component is operable to transport the first set of color channel data to the single display device and wherein the second link component is operable to transport the second set of color channel data to the single display device in parallel with the first link component.
2. The system of claim 1, wherein the single display device is operable to display the primary color system based on the set of image data, wherein the primary color system displayed on the single display device is based on the set of image data.
3. The system of claim 1, wherein the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 493nm, a third primary at approximately 540nm, and a fourth primary at approximately 640nm.
4. The system of claim 1, wherein the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 485nm, a third primary at approximately 510nm, a fourth primary at approximately 535nm, and a fifth primary at approximately 640nm.
5. The system of claim 1, wherein the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 490nm, a third primary at approximately 506nm, a fourth primary at approximately 520nm, a fifth primary at approximately 545nm, and a sixth primary at approximately 640nm.
6. The system of claim 1, wherein the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 508nm, a fifth primary at approximately 520nm, a sixth primary at approximately 540nm, and a seventh primary at approximately 640nm.
7. The system of claim 1, wherein the at least four primary color values include a first primary at approximately 460nm, a second primary at approximately 480nm, a third primary at approximately 495nm, a fourth primary at approximately 500nm, a fifth primary at approximately 511nm, a sixth primary at approximately 521nm, a seventh primary at approximately 545nm, and an eighth primary at approximately 640nm.
8. The system of claim 1, wherein the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 502nm, a sixth primary at approximately 512nm, a seventh primary at approximately 520nm, an eighth primary at approximately 535nm, a ninth primary at approximately 550nm, and a tenth primary at approximately 660nm.
9. The system of claim 1, wherein the at least four primary color values include a first primary at approximately 440nm, a second primary at approximately 470nm, a third primary at approximately 485nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 505nm, a seventh primary at approximately 511nm, an eighth primary at approximately 517nm, a ninth primary at approximately 523nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 550nm, and a twelfth primary at approximately 670nm.
10. The system of claim 1, wherein the at least four primaries include a first primary at approximately 400nm, a second primary at approximately 468nm, a third primary at approximately 484nm, a fourth primary at approximately 493nm, a fifth primary at approximately 500nm, a sixth primary at approximately 506nm, a seventh primary at approximately 512nm, an eighth primary at approximately 518nm, a ninth primary at approximately 524nm, a tenth primary at approximately 535nm, an eleventh primary at approximately 556nm, and a twelfth primary at approximately 700nm.
11. The system of claim 1, wherein the at least four primaries include a magenta primary, a blue-magenta primary, a blue primary, a cyan-blue primary, a cyan primary, a green-cyan primary, a green primary, a yellow-green primary, a yellow primary, a red-yellow primary, a red primary, and a magenta-red primary.
12. The system of claim 1, wherein the set of SDP parameters is modifiable.
166
13. The system of claim 1, wherein the mid-Kelvin white emitter is modified to include a green bias.
14. The system of claim 1, wherein the first set of color channel data is converted by the first link component and the second set of color channel data is converted by the second link component, and wherein the first set of color channel data and the second set of color channel data are combined to form the set of image data for display on the single display device.
15. The system of claim 1, further including a standardized transport format, wherein the first link component includes a first standardized transport format link and wherein the second link component includes a second standardized transport format link, wherein the standardized transport format is operable to receive the first set of image data and the second set of image data using the first standardized transport format link and the second standardized transport format link, and wherein the first standardized transport format link and the second standardized transport format link are operable to combine the first set of image data and the second set of image data into a combined set of image data.
16. A system for displaying a primary color system, comprising: a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data; an image data converter, wherein the image data converter includes a digital interface, wherein the digital interface is operable to encode and decode the set of image data; a set of Session Description Protocol (SDP) parameters, wherein the set of SDP parameters is modifiable; and a display device, wherein the display device is a single display device; wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter;
167 wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter; wherein the single display device and the image data converter are in network communication; wherein the image data converter further includes a first link component and a second link component; wherein the image data converter is operable to convert the set of image data for display on the single display device; wherein the first link component is operable to transport the first set of color channel data to the single display device and wherein the second link component is operable to transport the second set of color channel data to the single display device in parallel with the first link component; and wherein once the set of image data has been converted by the image data converter for the single display device, the set of SDP parameters is modified based on the conversion.
17. The system of claim 16, wherein the mid-Kelvin white emitter is modified to include a green bias.
18. The system of claim 16, wherein the at least one white emitter includes a white emitter matching a white point of the primary color system.
19. A method for displaying a multi -primary color system, comprising: providing a set of image data, wherein the set of image data includes a first set of color channel data and a second set of color channel data, wherein the set of image data further includes primary color data for at least four primary color values, wherein the at least four primary color values include at least one white emitter;
168 encoding the set of image data using a digital interface of an image data converter, wherein the image data converter is in network communication with a single display device, and wherein the image data converter further includes a first link component and a second link component; decoding the set of image data using the digital interface of the image data converter; and converting the set of image data for display on the single display device using the image data converter; transporting the first set of color channel data to the single display device using the first link component; and transporting the second set of color channel data to the single display device using the first link component in parallel with the first link component; wherein the at least one white emitter includes at least three white emitters, wherein the at least three white emitters each have a different color temperature, and wherein the at least three white emitters include a mid-Kelvin white emitter.
20. The method of claim 19, wherein the at least one white emitter includes a white emitter matching a white point of the primary color system.
EP21876192.2A 2020-10-01 2021-08-31 System and method for a multi-primary wide gamut color system Pending EP4222731A1 (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US17/060,917 US11030934B2 (en) 2018-10-25 2020-10-01 System and method for a multi-primary wide gamut color system
US17/076,383 US11069279B2 (en) 2018-10-25 2020-10-21 System and method for a multi-primary wide gamut color system
US17/082,741 US11069280B2 (en) 2018-10-25 2020-10-28 System and method for a multi-primary wide gamut color system
US17/209,959 US11373575B2 (en) 2018-10-25 2021-03-23 System and method for a multi-primary wide gamut color system
PCT/US2021/048361 WO2022072102A1 (en) 2020-10-01 2021-08-31 System and method for a multi-primary wide gamut color system

Publications (1)

Publication Number Publication Date
EP4222731A1 true EP4222731A1 (en) 2023-08-09

Family

ID=80950808

Family Applications (1)

Application Number Title Priority Date Filing Date
EP21876192.2A Pending EP4222731A1 (en) 2020-10-01 2021-08-31 System and method for a multi-primary wide gamut color system

Country Status (5)

Country Link
EP (1) EP4222731A1 (en)
KR (1) KR20230097030A (en)
AU (1) AU2021351632A1 (en)
CA (1) CA3197643A1 (en)
WO (1) WO2022072102A1 (en)

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4197788B2 (en) * 1999-02-25 2008-12-17 オリンパス株式会社 Color reproduction system
JP5074101B2 (en) * 2007-06-04 2012-11-14 オリンパス株式会社 Multispectral image processing apparatus and color reproduction system using the same
US9544560B2 (en) * 2015-01-09 2017-01-10 Vixs Systems, Inc. Dynamic range converter with generic architecture and methods for use therewith
US10263951B2 (en) * 2017-01-09 2019-04-16 Star2Star Communications, LLC Network address family translation method and system
CN107863081B (en) * 2017-12-13 2019-08-23 京东方科技集团股份有限公司 Multiple bases conversion method and its converter, display control method, display device
US11373575B2 (en) * 2018-10-25 2022-06-28 Baylor University System and method for a multi-primary wide gamut color system
US11030934B2 (en) * 2018-10-25 2021-06-08 Baylor University System and method for a multi-primary wide gamut color system
US11069279B2 (en) * 2018-10-25 2021-07-20 Baylor University System and method for a multi-primary wide gamut color system

Also Published As

Publication number Publication date
CA3197643A1 (en) 2022-04-07
KR20230097030A (en) 2023-06-30
WO2022072102A1 (en) 2022-04-07
AU2021351632A1 (en) 2023-05-18

Similar Documents

Publication Publication Date Title
US11893924B2 (en) System and method for a multi-primary wide gamut color system
US11069280B2 (en) System and method for a multi-primary wide gamut color system
US11315466B2 (en) System and method for a multi-primary wide gamut color system
US11189214B2 (en) System and method for a multi-primary wide gamut color system
US11436967B2 (en) System and method for a multi-primary wide gamut color system
US11373575B2 (en) System and method for a multi-primary wide gamut color system
US11289003B2 (en) System and method for a multi-primary wide gamut color system
US11651718B2 (en) System and method for a multi-primary wide gamut color system
US11721266B2 (en) System and method for a multi-primary wide gamut color system
US11341890B2 (en) System and method for a multi-primary wide gamut color system
AU2021351632A1 (en) System and method for a multi-primary wide gamut color system
US11984055B2 (en) System and method for a multi-primary wide gamut color system
US20220343822A1 (en) System and method for a multi-primary wide gamut color system
US20230070395A1 (en) System and method for a multi-primary wide gamut color system

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20230428

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)