US20120308160A1 - Method and system for producing formatted information related to defects of appliances - Google Patents

Method and system for producing formatted information related to defects of appliances Download PDF

Info

Publication number
US20120308160A1
US20120308160A1 US13489892 US201213489892A US2012308160A1 US 20120308160 A1 US20120308160 A1 US 20120308160A1 US 13489892 US13489892 US 13489892 US 201213489892 A US201213489892 A US 201213489892A US 2012308160 A1 US2012308160 A1 US 2012308160A1
Authority
US
Grant status
Application
Patent type
Prior art keywords
image
restitution
appliance
information
formatted
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13489892
Inventor
Bruno Liege
Frederic Guichard
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
DxO Labs
Original Assignee
DxO Labs
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration, e.g. from bit-mapped to bit-mapped creating a similar image
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • G06T1/0007Image acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image, e.g. from bit-mapped to bit-mapped creating a different image
    • G06T3/0056Geometric image transformation in the plane of the image, e.g. from bit-mapped to bit-mapped creating a different image the transformation method being selected according to the characteristics of the input image
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration, e.g. from bit-mapped to bit-mapped creating a similar image
    • G06T5/001Image restoration
    • G06T5/002Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration, e.g. from bit-mapped to bit-mapped creating a similar image
    • G06T5/001Image restoration
    • G06T5/003Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration, e.g. from bit-mapped to bit-mapped creating a similar image
    • G06T5/006Geometric correction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/00002Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for
    • H04N1/00007Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for relating to particular apparatus or devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/00002Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for
    • H04N1/00026Methods therefor
    • H04N1/00045Methods therefor using a reference pattern designed for the purpose, e.g. a test chart
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/00002Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for
    • H04N1/00071Diagnosis, testing or measuring; Detecting, analysing or monitoring not otherwise provided for characterised by the action taken
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/387Composing, repositioning or otherwise geometrically modifying originals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/40Picture signal circuits
    • H04N1/40093Modification of content of picture, e.g. retouching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/46Colour picture communication systems
    • H04N1/56Processing of colour picture signals
    • H04N1/58Edge or detail enhancement; Noise or error suppression, e.g. colour misregistration correction

Abstract

A method for producing formatted information related to at least one defect of at least one device within an imaging appliance chain. A processor may retrieve at least one variable characteristic of the at least one device. The at least one variable characteristic of the at least one device may be configured with a first value representing at least one setting of the at least one device. A first formatted information representing the at least one defect of the at least one device may be produced. The first formatted information may be based upon measured results related to the configured at least one variable characteristic. A second formatted information related to at least one variable characteristic of the at least one device having a second value may be produced based upon the first formatted information.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • [0001]
    This application is a division of application Ser. No. 10/483,322 filed Jun. 28, 2004, the entire content of which is incorporated herein in its entirety by reference. U.S. application Ser. No. 10/483,322 is the national stage of PCT Application No. PCT/FR02/01914, filed Jun. 5, 2002, and claims the benefit of priority under 35 U.S.C. §119 from French Application 01-09291, filed Jul. 12, 2001 and French Application No. 01-09292, filed Jul. 12, 2001.
  • PREAMBLE OF THE DESCRIPTION
  • [0002]
    Field in Question, Problem Posed
  • BACKGROUND
  • [0003]
    The present invention relates to a method and a system for producing formatted information related to defects of appliances of an appliance chain. The present invention also relates to the formatted information obtained in this way and addressed to image-processing means.
  • Solution Method
  • [0004]
    The invention relates to a method for producing formatted information related to the defects of appliances of an appliance chain. The formatted information is addressed to image-processing means, especially software, for the purpose of modifying the quality of images processed by the image-processing means. The appliance chain includes in particular at least one image-capture appliance and/or at least one restitution means and/or at least one observer. The method includes the stage of producing data characterizing the defects of appliances of the appliance chain. The data are referred to hereinafter as the formatted information.
  • Image-Capture Appliance
  • [0005]
    The appliance chain includes at least one image-capture appliance, especially a scanner. In a preferred embodiment according to the invention, the method is more particularly designed in this case to produce formatted information related to defects of the image-capture appliance. The method additionally includes the stage of producing data characterizing the defects of the image-capture appliance, especially the sharpness characteristics. The formatted information also includes the data characterizing the defects of the image-capture appliance.
  • [0006]
    The appliance chain includes at least one image-capture appliance. To produce the formatted information related to defects of the image-capture appliance, especially the distortion characteristics, preferably, according to the invention, the method includes a first algorithm for calculating the formatted information. By means of this first calculation algorithm, it is possible to make a choice:
      • within a set of parameterizable transformation models,
      • within a set of parameterizable reverse transformation models, within a set of synthetic images,
      • within a set of reference scenes and within a set of transformed images.
  • [0010]
    This choice is based on:
      • a reference scene, and/or
      • a transformed image, and/or
      • a parameterizable transformation model with which the reference image obtained by capturing the reference scene by means of the image-capture appliance can be transformed to the transformed image, and/or
      • a parameterizable reverse transformation model with which the transformed image can be transformed to the reference image, and/or
      • a synthetic image obtained from the reference scene and/or obtained from the reference image.
  • [0016]
    The choice is made in such a way that the transformed image is close to the synthetic image. The transformed image exhibits a deviation compared with the synthetic image.
  • [0017]
    The formatted information is at least partly composed of the parameters of the chosen parameterizable transformation model and/or of the parameters of the chosen parameterizable reverse transformation model.
  • [0018]
    Preferably, according to the invention, it is possible by means of the first calculation algorithm to choose, within a set of mathematical projections, one mathematical projection with which the synthetic image can be constructed from the reference scene.
  • [0019]
    The appliance chain includes at least one image-capture appliance. Preferably, according to another alternative embodiment of the invention, to produce the formatted information related to defects of the image-capture appliance, especially the distortion characteristics, the method includes a second algorithm for calculating the formatted information. This second algorithm includes the stages of:
      • choosing at least one reference scene,
      • capturing at least one reference image of each reference scene by means of the image-capture appliance.
  • [0022]
    This second algorithm additionally includes the stage of choosing, within a set of parameterizable transformation models and within a set of synthetic images:
      • a parameterizable transformation model with which the reference image can be transformed to a transformed image, and/or
      • a synthetic image obtained from the reference scene and/or obtained from the reference image,
  • [0025]
    such that the synthetic image is close to the transformed image. The transformed image exhibits a deviation compared with the synthetic image. The formatted information is at least partly composed of the parameters of the chosen parameterizable transformation model and/or of the parameters of the parameterizable reverse transformation model. By means of the parameterizable reverse transformation model, the transformed image can be transformed to the reference image.
  • [0026]
    Preferably, according to the invention, it is possible by means of the second calculation algorithm to choose, within a set of mathematical projections, one mathematical projection with which the synthetic image can be constructed from the reference scene.
  • [0027]
    Preferably, according to the invention, the method includes a third algorithm for calculating the deviations. This third algorithm for calculating the deviations includes the stages of:
      • calculating the deviations between the transformed image and the synthetic image,
      • associating the deviations with the formatted information.
  • [0030]
    The appliance chain includes at least one image-capture appliance. The image or images are images of scenes, especially of animated scenes captured by a user by means of the image-capture appliance. Preferably, according to this other alternative embodiment of the invention, to produce the formatted information related to defects of the image-capture appliance, especially the distortion characteristics, the method includes a fourth algorithm for calculating the formatted information. This fourth algorithm includes the stages of:
      • constructing a synthetic-image class by specified mathematical projections of at least one reference scene onto a surface,
      • capturing at least one reference image of at least one reference scene by means of the image-capture appliance,
      • choosing, within a set of parameterizable transformation models, that with which the reference image can be transformed to a transformed image close to the synthetic-image class of the reference scene.
  • [0034]
    The transformed image exhibits a deviation compared with the synthetic-image class. The formatted information is at least partly composed of the parameters of the chosen parameterizable transformation models.
  • [0035]
    Preferably, according to the invention, the method includes a fifth algorithm for calculating the deviations. This fifth algorithm includes the stages of:
      • calculating the deviations between the transformed image and the synthetic-image class,
      • associating the deviations with the formatted information.
    Variable Focal Length—Image-Capture Appliance
  • [0038]
    The image-capture appliance of the appliance chain is provided with at least one variable characteristic depending on the image, especially the focal length and/or the aperture. A fraction of the formatted information is related to the defects of the image-capture appliance provided with the variable characteristic or characteristics. Each variable characteristic can be associated with a value to form a combination composed of the set of variable characteristics and values. Preferably in this case, according to the invention, the method additionally includes the stage of producing the fraction of the said formatted information for a predetermined selection of the said combinations. The formatted information obtained in this way in the said fraction of formatted information is entered in the field of formatted information. This formatted information is referred to hereinafter as the measured formatted information.
  • [0039]
    Preferably, according to the invention, the method includes the stage of obtaining the fraction of formatted information related to an arbitrary combination from measured formatted information. The fraction of formatted information related to an arbitrary combination is referred to hereinafter as extended formatted information.
  • [0040]
    The formatted information can include the extended formatted information instead of and in place of measured formatted information.
  • [0041]
    The measured formatted information and the said extended formatted information may exhibit an interpolation deviation. Preferably, according to the invention, the method additionally includes the stage of selecting zero or one or more variable characteristics, in such a way that the interpolation deviation for the extended formatted information obtained for the variable characteristics selected in this way is below a predetermined interpolation threshold. It results from the combination of technical features that the modification of image quality employs simple calculations. It also results from the combination of technical features that the extended formatted information is compact. It also results from the combination of technical features that the eliminated variable characteristics have the least influence on the defect.
  • Image-Restitution Means
  • [0042]
    The appliance chain can include at least one image-restitution means, especially a printer or a visual display screen. The method is designed additionally to provide, to image-processing means, especially software, formatted information related to the defects of the image-restitution means. Preferably in this case, according to the invention, the method additionally includes the stage of producing data characterizing the defects of the image-restitution means, especially the distortion characteristics. The formatted information also includes data characterizing the defects of the restitution means.
  • [0043]
    The appliance chain includes at least one image-restitution means. Preferably, according to the invention, to produce the formatted information related to defects of the image-restitution means, especially the distortion characteristics, the method includes a sixth algorithm for calculating the formatted information. By means of this sixth algorithm, it is possible to make a choice:
      • within a set of parameterizable restitution transformation models,
      • within a set of parameterizable reverse restitution transformation models,
      • within a set of mathematical restitution projections,
      • within a set of restitution references and within a set of corrected reference restitution images.
  • [0048]
    The choice made by this sixth algorithm is based on:
      • a restitution reference, and/or
      • a corrected reference restitution image, and/or
      • a parameterizable restitution transformation model with which the restitution reference can be transformed to the corrected reference restitution image, and/or
      • a parameterizable reverse restitution transformation model with which the corrected reference restitution image can be transformed to the restitution reference, and/or
      • a mathematical restitution projection with which a synthetic restitution image can be constructed from the corrected reference restitution image.
  • [0054]
    The choice is made by this sixth algorithm in such a way that the synthetic restitution image is close to the restituted reference obtained by restitution of the restitution reference by means of the image-restitution means. The restituted reference exhibits a restitution deviation compared with the synthetic restitution image. The formatted information is at least partly composed of the parameters of the chosen parameterizable restitution transformation model and/or of the parameters of the chosen parameterizable reverse restitution transformation model.
  • [0055]
    The appliance chain includes at least one image-restitution means. Preferably in this case, according to an alternative embodiment of the invention, to produce the formatted information related to the defects of the image-restitution means, especially the distortion characteristics, the method includes a seventh algorithm for calculating the formatted information. This seventh algorithm includes the stages of:
      • choosing at least one restitution reference,
      • restituting the restitution reference to a restituted reference by means of the image-restitution means.
  • [0058]
    By means of this seventh algorithm it is also possible to choose, within a set of parameterizable restitution transformation models and within a set of mathematical restitution projections:
      • a parameterizable restitution transformation model with which the restitution reference can be transformed to a corrected reference restitution image, and
      • a mathematical restitution projection with which a synthetic restitution image can be constructed from the corrected reference restitution image.
  • [0061]
    The choice is made by the seventh algorithm in such a way that the synthetic restitution image is close to the restituted reference. The restituted reference exhibits a restitution deviation compared with the synthetic restitution image. The formatted information is at least partly composed of the parameters of the chosen parameterizable restitution transformation model and/or of the parameters of the parameterizable reverse restitution transformation model. By means of the parameterizable reverse restitution transformation model, it is possible to transform the corrected reference restitution image to the restitution reference.
  • [0062]
    The appliance chain includes at least one image-restitution means. Preferably, according to this other alternative embodiment of the invention, to produce the formatted information related to the defects of the image-restitution means, especially the distortion characteristics, the method includes an eighth algorithm for calculating the formatted information. This eighth algorithm includes the stage of choosing a corrected reference restitution image. This eighth algorithm also includes the stage of making a choice within a set of parameterizable restitution transformation models, within a set of mathematical restitution projections and within a set of restitution references. This choice is based on:
      • a restitution reference, and/or
      • a parameterizable restitution transformation model with which the restitution reference can be transformed to the corrected reference restitution image, and/or
      • a parameterizable reverse restitution transformation model with which the corrected reference restitution image can be transformed to the restitution reference, and/or
      • a mathematical restitution projection with which a synthetic restitution image can be constructed from the corrected reference restitution image.
  • [0067]
    The eighth algorithm makes this choice in such a way that the synthetic restitution image is close to the restituted reference obtained by restitution of the restitution reference by means of the image-restitution means. The restituted reference exhibits a restitution deviation compared with the synthetic restitution image. The formatted information is at least partly composed of the parameters of the chosen parameterizable restitution transformation model and/or of the parameters of the chosen parameterizable reverse restitution transformation model.
  • [0068]
    Preferably, according to the invention, the method includes a ninth algorithm for calculating the restitution deviations. This ninth algorithm includes the stages of:
      • calculating the restitution deviations between the restituted reference and the synthetic restitution image,
      • associating the restitution deviations with the said formatted information.
  • [0071]
    It results from the combination of technical features that it is possible to verify automatically, for example during manufacture of the appliance, that the method has produced formatted information within acceptable tolerances.
  • Variable Focal Length—Image-Restitution Means
  • [0072]
    The image-restitution means of the appliance chain can be provided with at least one variable restitution characteristic depending on the image, especially the focal length and/or the aperture. A fraction of the formatted information is related to the defects of the image-restitution means provided with the variable restitution characteristic or characteristics. Each variable restitution characteristic can be associated with a value to form a restitution combination composed of the set of variable restitution characteristics and values. Preferably in this case, according to the invention, the method additionally includes the stage of producing the fraction of the formatted information for a predetermined selection of the restitution combinations. The formatted information obtained in this way in the fraction and entered in the field of formatted information is referred to hereinafter as the measured formatted restitution information.
  • [0073]
    Preferably, according to the invention, the method includes the stage of obtaining the fraction of formatted information related to an arbitrary combination from measured formatted restitution information. The fraction of formatted information related to an arbitrary restitution combination is referred to hereinafter as extended formatted restitution information.
  • [0074]
    The formatted information can include the extended formatted restitution information instead of and in place of measured formatted restitution information.
  • [0075]
    The measured formatted restitution information and the said extended formatted restitution information can exhibit a restitution interpolation deviation. Preferably in this case, according to the invention, the method additionally includes the stage of selecting zero or one or more variable restitution characteristics, in such a way that the restitution interpolation deviation for the extended formatted restitution information obtained for the variable restitution characteristics selected in this way is below a predetermined restitution interpolation threshold. It results from the combination of technical features that the modification of image quality employs simple calculations. It also results from the combination of technical features that the extended formatted restitution information is compact. It also results from the combination of technical features that the eliminated variable restitution characteristics have the least influence on the defect.
  • [0076]
    Preferably, according to the invention, the restitution means is associated with an image-capture appliance to restitute, in digital form, the restituted reference from the restitution reference. The method is such that, to produce the formatted information related to the defects of the restitution means, there is used the formatted information related to the image-capture appliance associated with the restitution means.
  • [0077]
    The appliance chain includes at least two appliances. Preferably in this case, according to the invention, to produce formatted information related to the defects of the appliances of the appliance chain, the method additionally includes the following stages:
      • the stage of producing formatted information related to each of the appliances of the appliance chain,
      • the stage of combining the formatted information related to each of the appliances of the appliance chain, in order to obtain formatted information related to the appliance chain.
    System
  • [0080]
    The invention relates to a system for producing formatted information related to the defects of appliances of an appliance chain. The formatted information is addressed to image-processing means, especially software, for the purpose of modifying the quality of images processed by the image-processing means. The appliance chain includes in particular at least one image-capture appliance and/or at least one restitution means and/or at least one observer. The system includes data-processing means for producing data characterizing the defects of appliances of the appliance chain. The data are referred to hereinafter as the formatted information.
  • [0081]
    Image-Capture Appliance
  • [0082]
    The appliance chain includes at least one image-capture appliance, especially a scanner. Preferably in this case, according to the invention, the system is more particularly designed to produce formatted information related to defects of the image-capture appliance. The system includes data-processing means for producing data characterizing the defects of the image-capture appliance, especially the sharpness characteristics. The formatted information also includes the data characterizing the defects of the image-capture appliance.
  • [0083]
    The appliance chain includes at least one image-capture appliance. Preferably, according to the invention, to produce the formatted information related to defects of the image-capture appliance, especially the distortion characteristics, the data-processing means are capable of employing a first algorithm for calculating the formatted information. By means of the data-processing means employing the first calculation algorithm, it is possible to make a choice within a set of parameterizable transformation models, within a set of parameterizable reverse transformation models, within a set of synthetic images, within a set of reference scenes and within a set of transformed images.
  • [0084]
    This choice is based on:
      • a reference scene, and/or
      • a transformed image, and/or
      • a parameterizable transformation model with which the reference image obtained by capturing the reference scene by means of the image-capture appliance can be transformed to the transformed image, and/or
      • a parameterizable reverse transformation model with which the transformed image can be transformed to the reference image, and/or
      • a synthetic image obtained from the reference scene and/or obtained from the reference image.
  • [0090]
    The data-processing means make this choice in such a way that the transformed image is close to the synthetic image. The transformed image exhibits a deviation compared with the synthetic image. The formatted information is at least partly composed of the parameters of the chosen parameterizable transformation model and/or of the parameters of the chosen parameterizable reverse transformation model.
  • [0091]
    Preferably, according to the invention, the data-processing means employing the first calculation algorithm include calculating means for choosing, within a set of mathematical projections, one mathematical projection with which the synthetic image can be constructed from the reference scene.
  • [0092]
    The appliance chain includes at least one image-capture appliance. Preferably, according to an alternative embodiment of the invention, to produce the formatted information related to defects of the image-capture appliance, especially the distortion characteristics, the data-processing means employ a second algorithm for calculating the formatted information. The data-processing means employing the second calculation algorithm include calculating means for choosing at least one reference scene. By means of the image-capture appliance, it is possible to capture at least one reference image of each reference scene. By means of the calculating means it is additionally possible to make a choice, within a set of parameterizable transformation models and within a set of synthetic images:
      • of a parameterizable transformation model with which the reference image can be transformed to a transformed image, and/or
      • of a synthetic image obtained from the reference scene and/or obtained from the reference image.
  • [0095]
    The calculating means make this choice in such a way that the synthetic image is close to the transformed image. The transformed image exhibits a deviation compared with the synthetic image. The formatted information is at least partly composed of the parameters of the chosen parameterizable transformation model and/or of the parameters of the parameterizable reverse transformation model. By means of the parameterizable reverse transformation model, the transformed image can be transformed to the reference image.
  • [0096]
    Preferably, according to the invention, the calculating means with which the second calculation algorithm can be employed include specific calculating means for choosing, within a set of mathematical projections, one mathematical projection with which the synthetic image can be constructed from the reference scene.
  • [0097]
    Preferably, according to the invention, the data-processing means are capable of employing a third algorithm for calculating the deviations. For this purpose, the data-processing means include calculating means appropriate for:
      • calculating the deviations between the transformed image and the synthetic image,
      • associating the deviations with the said formatted information.
  • [0100]
    Preferably, according to another alternative embodiment of the invention, the system is such that, to calculate the formatted information related to an image-capture appliance of the appliance chain, it additionally uses at least one reference scene. The image-capture appliance constructs a reference image for each reference scene. To calculate the formatted information relative to the image-capture appliance, the system additionally includes calculating and processing means for:
      • constructing a synthetic-image class by specified mathematical projections of at least one reference scene onto a surface,
      • choosing, within a set of parameterizable transformation models, that with which the reference image can be transformed to a transformed image close to the synthetic-image class of the reference scene.
  • [0103]
    The transformed image exhibits a deviation compared with the synthetic-image class. The formatted information is at least partly composed of the parameters of the chosen parameterizable transformation models.
  • [0104]
    Preferably, according to the invention, the system includes calculating means for:
      • calculating the deviations between the transformed image and the synthetic-image class,
      • associating the deviations with the said formatted information.
    Variable Focal Length—Image-Capture Appliance
  • [0107]
    Preferably, according to the invention, the system is such that the image-capture appliance of the appliance chain is provided with at least one variable characteristic depending on the image, especially the focal length and/or the aperture. A fraction of the formatted information is related to the defects of the image-capture appliance provided with the variable characteristic or characteristics. Each variable characteristic can be associated with a value to form a combination composed of the set of variable characteristics and values. The system additionally includes data-processing means for producing the fraction of the formatted information for a predetermined selection of the combinations. The formatted information obtained in this way in the fraction of formatted information and entered in the field of formatted information is referred to hereinafter as the measured formatted information.
  • [0108]
    Preferably, according to the invention, the data-processing means are capable of obtaining the fraction of formatted information related to an arbitrary combination from measured formatted information. The fraction of formatted information related to an arbitrary combination is referred to hereinafter as extended formatted information.
  • [0109]
    The formatted information can include the extended formatted information instead of and in place of measured formatted information.
  • [0110]
    The measured formatted information and the said extended formatted information exhibit an interpolation deviation. Preferably, according to the invention, the system additionally includes selection means for selecting zero or one or more variable characteristics, in such a way that the interpolation deviation for the extended formatted information obtained for the variable characteristics selected in this way is below a predetermined interpolation threshold.
  • Image-Restitution Means
  • [0111]
    The appliance chain can include at least one image-restitution means, especially a printer or a visual display screen. The system is designed to provide, to image-processing means, especially software, formatted information related to the defects of the image-restitution means. Preferably in this case, according to the invention, the system additionally includes data-processing means for producing data characterizing the defects of the image-restitution means, especially the distortion characteristics. The formatted information also includes data characterizing the defects of the restitution means.
  • [0112]
    Preferably, in the case in which the appliance chain includes at least one image-restitution means, to produce the formatted information related to defects of the image-restitution means, especially the distortion characteristics, the system includes calculating means for making a choice:
      • within a set of parameterizable restitution transformation models,
      • within a set of parameterizable reverse restitution transformation models,
      • within a set of mathematical restitution projections,
      • within a set of restitution references and within a set of corrected reference restitution images.
  • [0117]
    The choice made by the calculating means is based on:
      • a restitution reference, and/or
      • a corrected reference restitution image, and/or
      • a parameterizable restitution transformation model with which the restitution reference can be transformed to the corrected reference restitution image, and/or
      • a parameterizable reverse restitution transformation model with which the corrected reference restitution image can be transformed to the restitution reference, and/or
      • a mathematical restitution projection with which a synthetic restitution image can be constructed from the corrected reference restitution image.
  • [0123]
    The calculating means make this choice in such a way that the synthetic restitution image is close to the restituted reference obtained by restitution of the restitution reference by means of the image-restitution means. The restituted reference exhibits a restitution deviation compared with the synthetic restitution image. The formatted information is at least partly composed of the parameters of the chosen parameterizable restitution transformation model and/or of the parameters of the chosen parameterizable reverse restitution transformation model.
  • [0124]
    The appliance chain can include at least one image-restitution means. Preferably in the case of this alternative embodiment according to the invention, to produce the formatted information related to the defects of the image-restitution means, especially the distortion characteristics, the system includes calculating means for choosing at least one restitution reference. By means of the image-restitution means, it is possible to restitute the restitution reference to a restituted reference. By means of the calculating means it is additionally possible to make a choice within a set of parameterizable restitution transformation models and within a set of mathematical restitution projections:
      • of a parameterizable restitution transformation model with which the restitution reference can be transformed to a corrected reference restitution image, and
      • of a mathematical restitution projection with which a synthetic restitution image can be constructed from the corrected reference restitution image.
  • [0127]
    The calculating means make this choice in such a way that the synthetic restitution image is close to the restituted reference. The restituted reference exhibits a restitution deviation compared with the synthetic restitution image. The formatted information is at least partly composed of the parameters of the chosen parameterizable restitution transformation model and/or of the parameters of the parameterizable reverse restitution transformation model. By means of the parameterizable reverse restitution transformation model, it is possible to transform the corrected reference restitution image to the restitution reference.
  • [0128]
    The appliance chain can include at least one image-restitution means. Preferably in the case of this alternative embodiment according to the invention, to produce the formatted information related to the defects of the image-restitution means, especially the distortion characteristics, the system includes calculating means for choosing a corrected reference restitution image. By means of the calculating means, it is additionally possible to make a choice within a set of parameterizable restitution transformation models, within a set of mathematical restitution projections and within a set of restitution references:
      • of a parameterizable restitution transformation model with which the specific restitution reference can be transformed to the corrected reference restitution image, and/or
      • of a parameterizable reverse restitution transformation model with which the corrected reference restitution image can be transformed to the restitution reference, and/or
      • of a mathematical restitution projection with which a synthetic restitution image can be constructed from the corrected reference restitution image.
  • [0132]
    The calculating means make this choice in such a way that the synthetic restitution image is close to the restituted reference obtained by restitution of the restitution reference by means of the image-restitution means. The restituted reference exhibits a restitution deviation compared with the synthetic restitution image. The formatted information is at least partly composed of the parameters of the chosen parameterizable restitution transformation model and/or of the parameters of the chosen parameterizable reverse restitution transformation model.
  • [0133]
    Preferably, according to the invention, the system includes calculating means for:
      • calculating the restitution deviations between the restituted reference and the synthetic restitution image,
      • associating the restitution deviations with the said formatted information.
    Variable Focal Length—Image-Restitution Means
  • [0136]
    The image-restitution means of the appliance chain can be provided with at least one variable restitution characteristic depending on the image, especially the focal length and/or the aperture. A fraction of the formatted information is related to the defects of the image-restitution means provided with the variable restitution characteristic or characteristics. Each variable restitution characteristic can be associated with a value to form a restitution combination composed of the set of variable restitution characteristics and values. Preferably in the case of this alternative embodiment according to the invention, the system additionally includes in this case data-processing means for producing the fraction of the formatted information for a predetermined selection of the restitution combinations. The formatted information obtained in this way in the fraction and entered in the field of formatted information is referred to hereinafter as the measured formatted restitution information.
  • [0137]
    Preferably, according to the invention, the system includes data-processing means for obtaining the fraction of formatted information related to an arbitrary combination from measured formatted restitution information. The formatted information making up the fraction of formatted information related to an arbitrary restitution combination is referred to hereinafter as extended formatted restitution information.
  • [0138]
    The formatted information can include the extended formatted restitution information instead of and in place of measured formatted restitution information.
  • [0139]
    The measured formatted restitution information and the said extended formatted restitution information exhibit a restitution interpolation deviation. Preferably in this case, according to the invention, the system additionally includes selection means for selecting zero or one or more variable restitution characteristics, in such a way that the restitution interpolation deviation for the extended formatted restitution information obtained for the variable restitution characteristics selected in this way is below a predetermined restitution interpolation threshold.
  • [0140]
    According to an alternative embodiment of the invention, the restitution means is associated with an image-capture appliance to restitute, in digital form, the restituted reference from the restitution reference. Preferably in this case, the system includes data-processing means for producing the formatted information related to the defects of the restitution means, by using the formatted information related to the image-capture appliance associated with the restitution means.
  • [0141]
    The appliance chain can include at least two appliances. Preferably in the case of this alternative embodiment of the invention, to produce formatted information related to the defects of the appliances of the appliance chain, the system additionally includes data-processing means for producing formatted information related to each of the appliances of the appliance chain and for combining the formatted information related to each of the appliances of the appliance chain, in such a way as to obtain formatted information related to the appliance chain.
  • BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS Detailed Description
  • [0142]
    Other characteristics and advantages of the invention will become apparent upon reading of the description of alternative embodiments of the invention, provided by way of indicative and non-limitative examples, and of the figures, wherein respectively:
  • [0143]
    FIG. 1 illustrates a schematic view of image capture,
  • [0144]
    FIG. 2 illustrates a schematic view of image restitution,
  • [0145]
    FIG. 3 illustrates a schematic view of the pixels of an image,
  • [0146]
    FIGS. 4 a and 4 b illustrate two schematic views of a reference scene,
  • [0147]
    FIG. 5 illustrates the organizational diagram of the method with which the difference between the mathematical image and the corrected image can be calculated,
  • [0148]
    FIG. 6 illustrates the organizational diagram of the method with which the best restitution transformation for an image-restitution means can be obtained,
  • [0149]
    FIG. 7 illustrates a schematic view of the elements composing the system according to the invention,
  • [0150]
    FIG. 8 illustrates a schematic view of fields of formatted information,
  • [0151]
    FIG. 9 a illustrates a schematic front view of a mathematical point,
  • [0152]
    FIG. 9 b illustrates a schematic front view of a real point of an image,
  • [0153]
    FIG. 9 c illustrates a schematic side view of a mathematical point,
  • [0154]
    FIG. 9 d illustrates a schematic profile view of a real point of an image,
  • [0155]
    FIG. 10 illustrates a schematic view of an array of characteristic points,
  • [0156]
    FIG. 11 illustrates the organizational diagram of the method with which the formatted information can be obtained,
  • [0157]
    FIG. 12 illustrates the organizational diagram of the method with which the best transformation for an image-capture appliance can be obtained,
  • [0158]
    FIG. 13 illustrates the organizational diagram of the method with which the quality of an image derived from or addressed to a chain of appliances can be modified,
  • [0159]
    FIG. 14 illustrates an example of a file containing formatted information,
  • [0160]
    FIG. 15 illustrates an example of formatted information,
  • [0161]
    FIG. 16 illustrates a representation of parameters of parameterizable models,
  • [0162]
    FIG. 17 illustrates an organizational diagram of the method with which the best transformation for an image-restitution appliance can be obtained.
  • DETAILED DESCRIPTION
  • [0163]
    FIG. 1 illustrates a scene 3 containing an object 107, a sensor 101 and sensor surface 110, an optical center 111, an observation point 105 on a sensor surface 110, an observation direction 106 passing through observation point 105, optical center 111, scene 3, and a surface 10 geometrically associated with sensor surface 110.
  • [0164]
    FIG. 2 illustrates an image 103, an image-restitution means 19 and a restituted image 191 obtained on the restitution medium 190.
  • [0165]
    FIG. 3 illustrates a scene 3, an image-capture appliance 1 and an image 103 composed of pixels 104.
  • [0166]
    FIGS. 4 a and 4 b illustrate two alternative versions of a reference scene 9.
  • [0167]
    FIG. 5 illustrates an organizational diagram employing a scene 3, a mathematical projection 8 giving a mathematical image 70 of scene 3, a real projection 72 giving an image 103 of scene 3 for the characteristics 74 used, a parameterizable transformation model 12 giving a corrected image 71 of image 103, the corrected image 71 exhibiting a difference 73 compared with mathematical image 70.
  • [0168]
    FIG. 6 illustrates an organizational diagram employing an image 103, a real restitution projection 90 giving a restituted image 191 of image 103 for the restitution characteristics 95 used, a parameterizable restitution transformation model 97 giving a corrected restitution image 94 of image 103, a mathematical restitution projection 96 giving a mathematical restitution image 92 of corrected restitution image 94 and exhibiting a restitution difference 93 compared with restituted image 191.
  • [0169]
    FIG. 7 illustrates a system comprising an image-capture appliance 1 composed of an optical system 100, of a sensor 101 and of an electronic unit 102. FIG. 7 also illustrates a memory zone 16 containing an image 103, a database 22 containing formatted information 15, and means 18 for transmission of completed image 120 composed of image 103 and formatted information 15 to calculating means 17 containing image-processing software 4.
  • [0170]
    FIG. 8 illustrates formatted information 15 composed of fields 91.
  • [0171]
    FIGS. 9 a to 9 d illustrate a mathematical image 70, an image 103, the mathematical position 40 of a point, and the mathematical shape 41 of a point, compared with the real position 50 and the real shape 51 of the corresponding point of the image.
  • [0172]
    FIG. 10 illustrates an array 80 of characteristic points.
  • [0173]
    FIG. 11 illustrates an organizational diagram employing an image 103, the characteristics 74 used, and a database 22 of characteristic. The formatted information 15 is obtained from the characteristics 74 used and stored in database 22. The completed image 120 is obtained from image 103 and formatted information 15.
  • [0174]
    FIG. 12 illustrates an organizational diagram employing a reference scene 9, a mathematical projection 8 giving a synthetic image class 7 of reference scene 9, and a real projection 72 giving a reference image 11 of reference scene 9 for the characteristics 74 used. This organizational diagram also employs a parameterizable transformation model 12 giving a transformed image 13 of reference image 11. Transformed image 13 exhibits a deviation compared with synthetic image class 7.
  • [0175]
    FIG. 17 illustrates an organizational diagram employing a restitution reference 209, a real restitution projection 90 giving a restituted reference 211 of the said restitution reference 209 for the restitution characteristics 95 used, a parameterizable restitution transformation model 97 giving a corrected reference restitution image 213 of the said restitution reference 209, a parameterizable reverse restitution transformation model 297 producing the said restitution reference 209 from the said corrected reference restitution image 213. This organizational diagram also employs a mathematical restitution projection 96 giving a synthetic restitution image 307 of the corrected reference restitution image 213. The said synthetic restitution image 307 exhibits a restitution deviation 214 compared with the restituted reference 211.
  • DEFINITIONS AND DETAILED DESCRIPTION
  • [0176]
    Other characteristics and advantages of the invention will become apparent on reading:
      • of the definitions explained hereinafter of the employed technical terms, referring to the indicative and non-limitative examples of FIGS. 1 to 17,
      • of the description of FIGS. 1 to 17.
    Scene
  • [0179]
    Scene 3 is defined as a place in three-dimensional space, containing objects 107 illuminated by light sources.
  • Image-Capture Appliance, Image, Image Capture
  • [0180]
    Referring to FIGS. 3 and 7, a description will now be given of what is understood by image-capture appliance 1 and image 103. Image-capture appliance 1 is defined as an appliance composed of an optical system 100, of one or more sensors 101, of an electronic unit 102 and of a memory zone 16. By means of the said image-capture appliance 1, it is possible to obtain, from a scene 3, fixed or animated digital images 103 recorded in memory zone 16 or transmitted to an external device. Animated images are composed of a succession of fixed images 103 in time. The said image-capture appliance 1 can have the form in particular of a photographic appliance, of a video camera, of a camera connected to or integrated in a PC, of a camera connected to or integrated in a personal digital assistant, of a camera connected to or integrated in a telephone, of a videoconferencing appliance or of a measuring camera or appliance sensitive to wavelengths other than those of visible light, such as a thermal camera.
  • [0181]
    Image capture is defined as the method by which image 103 is calculated by image-capture appliance 1.
  • [0182]
    In the case in which an appliance is equipped with a plurality of interchangeable subassemblies, especially an optical system 100, image-capture appliance 1 is defined as a special configuration of the appliance.
  • Image-Restitution Means, Restituted Image, Image Restitution
  • [0183]
    Referring to FIG. 2, a description will now be given of what is understood by image-restitution means 19. Such an image-restitution means 19 can have the form in particular of a visual display screen, of a television screen, of a flat screen, of a projector, of virtual reality goggles, of a printer.
  • [0184]
    Such an image-restitution means 19 is composed of:
      • an electronic unit,
      • one or more sources of light, of electrons or of ink,
      • one or more modulators: devices for modulation of light, of electrons or of ink,
      • a focusing device, having in particular the form of an optical system in the case of a light projector or the form of electron-beam focusing coils in the case of a CRT screen, or the form of filters in the case of a flat screen,
      • a restitution medium 190 having in particular the form of a screen in the case of a CRT screen, of a flat screen or of a projector, the form of a print medium on which printing is performed in the case of a printer, or the form of a virtual surface in space in the case of a virtual-image projector.
  • [0190]
    By means of the said image-restitution means 19, it is possible to obtain, from an image 103, a restituted image 191 on restitution medium 190.
  • [0191]
    Animated images are composed of a succession of fixed images in time.
  • [0192]
    Image restitution is defined as the method by which the image is displayed or printed by means of image restitution means 19.
  • [0193]
    In the case in which a restitution means 19 is equipped with a plurality of interchangeable subassemblies or of subassemblies that can be shifted relative to one another, especially restitution medium 190, image-restitution means 19 is defined as a special configuration.
  • Sensor Surface, Optical Center, Focal Distance
  • [0194]
    Referring to FIG. 1, a description will now be given of what is defined as sensor surface 110.
  • [0195]
    Sensor surface 110 is defined as the shape in space drawn by the sensitive surface of sensor 101 of image-capture appliance 1 at the moment of image capture. This surface is generally plane.
  • [0196]
    An optical center 111 is defined as a point in space associated with image 103 at the moment of image capture. A focal distance is defined as the distance between this point 111 and plane 110, in the case in which sensor surface 110 is plane.
  • Pixel, Pixel Value, Exposure Time
  • [0197]
    Referring to FIG. 3, a description will now be given of what is understood by pixel 104 and pixel value.
  • [0198]
    A pixel 104 is defined as an elemental zone of sensor surface 110 obtained by creating a grid, generally regular, of the said sensor surface 110. Pixel value is defined as a number associated with this pixel 104.
  • [0199]
    Image capture is defined as determining the value of each pixel 104. The set of these values constitutes image 103.
  • [0200]
    During image capture, the pixel value is obtained by integration, over the surface of pixel 104, during a time period defined as exposure time, of part of the light flux derived from scene 3 via optical system 100, and by converting the result of this integration to a digital value. The integration of the light flux and/or the conversion of the result of this integration to a digital value are performed by means of electronic unit 102.
  • [0201]
    This definition of the concept of pixel value is applicable to the case of black-and-white or color images 103, whether they be fixed or animated.
  • [0202]
    Depending on the cases, however, the part in question of the light flux is obtained in various ways:
  • [0203]
    a) In the case of a color image 103, sensor surface 110 is generally composed of a plurality of types of pixels 104, associated respectively with light fluxes of different wavelengths, examples being red, green and blue pixels.
  • [0204]
    b) In the case of a color image 103, there may also be a plurality of sensors 101 disposed side-by-side, each receiving part of the light flux.
  • [0205]
    c) In the case of a color image 103, the colors used may be different from red, green and blue, such as for North American NTSC television, and they may exceed three in number.
  • [0206]
    d) Finally, in the case of an interlaced television scanning camera, the animated images produced are composed of an alternation of images 103 containing even-numbered lines and of images 103 containing odd-numbered lines.
  • Configuration Used, Adjustments Used, Characteristics Used
  • [0207]
    The configuration used is defined as the list of removable subassemblies of image-capture appliance 1, such as optical system 100 which, if it is interchangeable, is mounted on image-capture appliance 1. The configuration used is characterized in particular by:
      • the type of optical system 100,
      • the serial number of optical system 100 or any other designation.
  • [0210]
    Adjustments used are defined as:
      • the configuration used as defined hereinabove, as well as
      • the value of the manual or automatic adjustments available in the configuration used and having an impact on the content of image 103. These adjustments may be made by the user, especially by means of pushbuttons, or may be calculated by image-capture appliance 1. These adjustments may be stored in the appliance, especially on a removable medium, or on any device connected to the appliance. These adjustments may include in particular the adjustments of focusing, diaphragm and focal length of optical system 100, the adjustments of exposure time, the adjustments of white balance, and the integrated image-processing adjustments, such as digital zoom, compression and contrast.
  • [0213]
    Characteristics 74 used or set of characteristics 74 used are defined as:
  • [0214]
    a) Parameters related to the intrinsic technical characteristics of image-capture appliance 1, determined during the phase of design of image-capture appliance 1. For example, these parameters may include the formula of optical system 100 of the configuration used, which impacts the geometric defects and the sharpness of the captured images; the formula of optical system 100 of the configuration used includes in particular the shape, the arrangement and the material of the lenses of optical system 100.
  • [0215]
    These parameters may additionally include:
      • the geometry of sensor 101, or in other words sensor surface 110 as well as the shape and relative arrangement of pixels 104 on this surface,
      • the noise generated by electronic unit 102,
      • the equation for conversion of light flux to pixel value.
  • [0219]
    b) Parameters associated with the intrinsic technical characteristics of image-capture appliance 1, determined during the phase of manufacture of image-capture appliance 1 and, in particular:
      • the exact positioning of the lenses in optical system 100 of the configuration used,
      • the exact positioning of optical system 100 relative to sensor 101.
  • [0222]
    c) Parameters associated with the technical characteristics of image-capture appliance 1, determined at the moment of capture of image 103 and, in particular:
      • the position and orientation of sensor surface 110 relative to scene 3,
      • the adjustments used,
      • the external factors, such as temperature, if they have an influence.
  • [0226]
    d) The user's preferences, especially the color temperature to be used for image restitution. For example, these preferences are selected by the user by means of pushbuttons.
  • Observation Point, Observation Direction
  • [0227]
    Referring to FIG. 1, a description will now be given of what is understood by observation point 105 and observation direction 106.
  • [0228]
    Mathematical surface 10 is defined as a surface that is geometrically associated with sensor surface 110. For example, if the sensor surface is plane, it will be possible for mathematical surface 10 to coincide with the sensor surface.
  • [0229]
    Observation direction 106 is defined as a line passing through at least one point of scene 3 and through optical center 111. Observation point 105 is defined as the intersection of observation direction 106 and surface 10.
  • Observed Color, Observed Intensity
  • [0230]
    Referring to FIG. 1, a description will now be given of what is understood by observed color and observed intensity. Observed color is defined as the color of the light emitted, transmitted or reflected by the said scene 3 in the said observation direction 106 at a given instant, and observed from the said observation point 105. Observed intensity is defined as the intensity of the light emitted by the said scene 3 in the said observation direction 106 at the same instant, and observed from the said observation point 105.
  • [0231]
    The color can be characterized in particular by a light intensity that is a function of wavelength, or else by two values as measured by a colorimeter. The intensity can be characterized by a value such as measured with a photometer.
  • [0232]
    The said observed color and the said observed intensity depend in particular on the relative position of objects 107 in scene 3 and on the illumination sources present as well as on the transparency and reflection characteristics of objects 107 at the moment of observation.
  • Mathematical Projection, Mathematical Image, Mathematical Point, Mathematical Color of a Point, Mathematical Intensity of a Point, Mathematical Shape of a Point, Mathematical Position of a Point
  • [0233]
    Referring in particular to FIGS. 1, 5, 9 a, 9 b, 9 c and 9 d, a description will be given of the concepts of mathematical projection 8, mathematical image 70, mathematical point, mathematical color of a point, mathematical intensity of a point, mathematical shape 41 of a point, and mathematical position 40 of a point.
  • [0234]
    Referring to FIG. 5, a description will now be given of how a mathematical image 70 is constructed by specified mathematical projection 8 of at least one scene 3 on mathematical surface 10.
  • [0235]
    Firstly, a description will be given of what is understood by specified mathematical projection 8.
  • [0236]
    A specified mathematical projection 8 associates a mathematical image 70 with:
      • a scene 3 at the moment of capture of an image,
      • and with the characteristics 74 used.
  • [0239]
    A specified mathematical projection 8 is a transformation with which the characteristics of each point of mathematical image 70 can be determined from scene 3 at the moment of image capture and from the characteristics 74 used.
  • [0240]
    Mathematical projection 8 is preferentially defined in the manner to be described hereinafter.
  • [0241]
    Mathematical position 40 of the point is defined as the position of observation point 105 on mathematical surface 10.
  • [0242]
    Mathematical shape 41 of the point is defined as the geometric, punctiform shape of observation point 105.
  • [0243]
    Mathematical color of the point is defined as the observed color.
  • [0244]
    Mathematical intensity of the point is defined as the observed intensity.
  • [0245]
    Mathematical point is defined as the association of mathematical position 40, mathematical shape 41, mathematical color and mathematical intensity for the observation point 105 under consideration. Mathematical image 70 is composed of the set of said mathematical points.
  • [0246]
    The mathematical projection 8 of scene 3 is mathematical image 70.
  • Real Projection, Real Point, Real Color of a Point, Real Intensity of a Point, Real Shape of a Point, Real Position of a Point
  • [0247]
    Referring in particular to FIGS. 3, 5, 9 a, 9 b, 9 c and 9 d, a description will be given hereinafter of the concepts of real projection 72, real point, real color of a point, real intensity of a point, real shape 51 of a point, and real position 50 of a point.
  • [0248]
    During image capture, image-capture appliance 1 associates an image 103 of scene 3 with the characteristics 74 used. The light originating from scene 3 in an observation direction 106 passes through optical system 100 and arrives at sensor surface 110.
  • [0249]
    For the said observation direction, there is then obtained what is defined as a real point, which exhibits differences compared with the mathematical point.
  • [0250]
    Referring to FIGS. 9 a to 9 d, a description will now be given of the differences between the real point and the mathematical point.
  • [0251]
    The real shape 51 associated with the said observation direction 106 is not a point on the sensor surface, but it has the form of a cloud in three-dimensional space, where it has an intersection with one or more pixels 104. These differences are due in particular to coma, spherical aberration, astigmatism, grouping into pixels 104, chromatic aberration, depth of field, diffraction, parasitic reflections and field curvature of image-capture appliance 1. They give an impression of blurring, or of lack of sharpness of image 103.
  • [0252]
    In addition, real position 50 associated with the said observation direction 106 exhibits a difference compared with mathematical position 40 of a point. This difference is due in particular to the geometric distortion, which gives an impression of deformation: for example, vertical walls appear to be curved. It is also due to the fact that the number of pixels 104 is limited, and that consequently the real position 50 can have only a finite number of values.
  • [0253]
    In addition, the real intensity associated with the said observation direction 106 exhibits differences compared with the mathematical intensity of a point. These differences are due in particular to gamma and vignetting: for example, the edges of image 103 appear to be darker. Furthermore, noise may be added to the signal.
  • [0254]
    Finally, the real color associated with the said observation direction 106 exhibits differences compared with the mathematical color of a point. These differences are due in particular to gamma and the color cast. Furthermore, noise may be added to the signal.
  • [0255]
    A real point is defined as the association of the real position 50, the real shape 51, the real color and the real intensity for the observation direction 106 under consideration.
  • [0256]
    The real projection 72 of scene 3 is composed of the set of real points.
  • [0257]
    Parameterizable Transformation Model, Parameters, Corrected Image
  • [0258]
    A parameterizable transformation model 12 (or parameterizable transformation 12 for short) is defined as a mathematical transformation in which a corrected image 71 can be obtained from an image 103 and from the value of parameters. As indicated hereinbelow, the said parameters can in particular be calculated from the characteristics 74 used.
  • [0259]
    By means of the said parameterizable transformation 12, it is possible in particular to determine, for each real point of image 103, the corrected position of the said real point, the corrected color of the said real point, the corrected intensity of the said real point, and the corrected shape of the said real point, from the value of the parameters, from the real position of the said real point and from the values of the pixels of image 103. As an example, the corrected position can be calculated by means of polynomials of fixed degree as a function of the real position, the coefficients of the polynomials depending on the value of the parameters. The corrected color and the corrected intensity can be, for example, weighted sums of the values of the pixels, the coefficients depending on the value of the parameters and on the real position, or else can be nonlinear functions of the values of the pixels of image 103.
  • [0260]
    A parameterizable reverse transformation model 212 (or parameterizable reverse transformation 212 for short) is defined as a mathematical transformation in which an image 103 can be obtained from a corrected image 71 and from the value of parameters. The said parameters can be calculated in particular from the characteristics 74 used as indicated hereinbelow.
  • [0261]
    By means of the said parameterizable reverse transformation 212, it is possible in particular to determine, for each point of the corrected image 71, the real point of image 103 corresponding to the said point of corrected image 71, and in particular the position of the said real point, the color of the said real point, the intensity of the said real point, and the shape of the said real point, from the value of the parameters and from corrected image 71. As an example, the position of the real point can be calculated by means of polynomials of fixed degree as a function of the position of the point of the corrected image 71, the coefficients of the polynomials depending on the value of the parameters.
  • [0262]
    The parameters can include in particular: the focal length of optical system 100 of the configuration used, or a related value such as the position of a group of lenses, the focusing of optical system 100 of the configuration used, or a related value such as the position of a group of lenses, the aperture of optical system 100 of the configuration used, or a related value such as the position of the diaphragm.
  • Difference Between the Mathematical Image and the Corrected Image
  • [0263]
    Referring to FIG. 5, the difference 73 between mathematical image 70 and corrected image 71 for a given scene 3 and given characteristics 74 used is defined as one or more values determined from numbers characterizing the position, color, intensity, and shape of all or part of the corrected points and of all or part of the mathematical points.
  • [0264]
    For example, the difference 73 between mathematical image 70 and corrected image 71 for a given scene 3 and given characteristics 74 used can be determined as follows:
      • There are chosen characteristic points which, for example, may be the points of an orthogonal array 80 of regularly disposed points, as illustrated in FIG. 10.
      • The difference 73 is calculated, for example, by taking, for each characteristic point, the sum of the absolute values of the differences between each number characterizing the corrected position, the corrected color, the corrected intensity and the corrected shape respectively for the real point and for the mathematical point. The sum function of the absolute values of the differences may be replaced by another function such as the mean, the sum of the squares or any other function with which the numbers can be combined.
    Reference Scene
  • [0267]
    A reference scene 9 is defined as a scene 3 for which certain characteristics are known. As an example, FIG. 4 a shows a reference scene 9 composed of a paper sheet bearing regularly disposed, solid black circles. FIG. 4 b shows another paper sheet bearing the same circles, with the addition of colored lines and areas. The circles are used to measure the real position 50 of a point, the lines to measure the real shape 51 of a point, and the colored areas to measure the real color of a point and the real intensity of a point. This reference scene 9 may be composed of a material other than paper.
  • Reference Image
  • [0268]
    Referring to FIG. 12, a definition will now be given of the concept of reference image 11. A reference image 11 is defined as an image of reference scene 9 obtained with image-capture appliance 1.
  • Synthetic Image, Synthetic-Image Class
  • [0269]
    Referring to FIG. 12, a definition will now be given of the concept of synthetic image 207 and of synthetic-image class 7. A synthetic image 207 is defined as a mathematical image 70 obtained by mathematical projection 8 of a reference scene 9. A synthetic-image class 7 is defined as a set of mathematical images 70 obtained by mathematical projection 8 of one or more reference scenes 9 for one or more sets of characteristics 74 used. In the case in which there is only one reference scene 9 and only one set of characteristics 74 used, the synthetic-image class 7 comprises only one synthetic image 207.
  • Transformed Image
  • [0270]
    Referring to FIG. 12, a definition will now be given of the concept of transformed image 13. A transformed image 13 is defined as the corrected image obtained by application of a parameterizable transformation model 12 to a reference image 11.
  • Transformed Image Close to a Synthetic-Image Class, Deviation
  • [0271]
    Referring to FIG. 12, a description will now be given of the concept of transformed image 13 close to a synthetic-image class 7 and of the concept of deviation 14.
  • [0272]
    The difference between a transformed image 13 and a synthetic-image class 7 is defined as the smallest difference between the said transformed image 13 and any one of the synthetic images 207 of the said synthetic-image class 7.
  • [0273]
    Referring to FIG. 12, a description will next be given of a fourth algorithm with which it is possible to choose, among the parameterizable transformation models 12, that with which each reference image 11 can be transformed to a transformed image 13 close to the synthetic-image class 7 of the reference scene 9 corresponding to the said reference image 11, in different cases of reference scenes 9 and characteristics 74 used.
      • In the case of a given reference scene 9 associated with a set of given characteristics 74 used, there is chosen the parameterizable transformation 12 (and its parameters) with which the reference image 11 can be transformed to the transformed image 13 that exhibits the smallest difference compared with synthetic-image class 7. Synthetic-image class 7 and transformed image 13 are then said to be close. Deviation 14 is defined as the said difference.
      • In the case of a group of given reference scenes associated with sets of given characteristics 74 used, the parameterizable transformation 12 (and its parameters) is chosen as a function of the differences between the transformed image 13 of each reference scene 9 and the synthetic-image class 7 of each reference scene 9 under consideration. There is chosen the parameterizable transformation 12 (and its parameters) with which the reference images 11 can be transformed to transformed images 13 such that the sum of the said differences is minimized. The sum function may be replaced by another function such as the product. Synthetic-image class 7 and transformed images 13 are then said to be close. Deviation 14 is defined as a value obtained from the said differences, for example by calculating the mean thereof.
      • In the case in which certain characteristics 74 used are unknown, it is possible to determine them from the capture of a plurality of reference images 11 of at least one reference scene 9. In this case, there are simultaneously determined the unknown characteristics and the parameterizable transformation 12 (and its parameters) with which the reference images 11 can be transformed to transformed images 13, such that the sum of the said differences is minimized, in particular by iterative calculation or by solving equations concerning the sum of the said differences and/or their product and/or any other appropriate combination of the said differences. Synthetic-image class 7 and transformed images 13 are then said to be close. The unknown characteristics may be, for example, the relative positions and orientations of sensor surface 110 and of each reference scene 9 under consideration. Deviation 14 is defined as a value obtained from the said differences, for example by calculating the mean thereof. Referring to FIG. 12, a description will next be given of a first calculation algorithm with which a choice can be made:
      • within a set of parameterizable transformation models,
      • within a set of parameterizable reverse transformation models,
      • within a set of synthetic images,
      • within a set of reference scenes and within a set of transformed images.
  • [0281]
    This choice is based on:
      • a reference scene 9, and/or
      • a transformed image 13, and/or
      • a parameterizable transformation model 12 with which the reference image 11 obtained by capturing reference scene 9 by means of image-capture appliance 1 can be transformed to transformed image 13, and/or
      • a parameterizable reverse transformation model 212 with which transformed image 13 can be transformed to reference image 11, and/or
      • a synthetic image 207 obtained from reference scene 9 and/or obtained from reference image 11.
  • [0287]
    The choice adopted is that which minimizes the difference between transformed image 13 and synthetic image 207. Synthetic image 207 and transformed image 13 are then said to be close. Deviation 14 is defined as the said difference.
  • [0288]
    Preferably, according to the invention, it is possible by means of the first calculation algorithm to choose, within a set of mathematical projections, one mathematical projection 8 with which synthetic image 207 can be constructed from reference scene 9.
  • [0289]
    Referring to FIG. 12, a description will next be given of a second calculation algorithm that includes the stages of:
      • choosing at least one reference scene 9,
      • capturing at least one reference image 11 of each reference scene 9 by means of image-capture appliance 1.
  • [0292]
    This second algorithm additionally includes the stage of choosing, within a set of parameterizable transformation models and within a set of synthetic images:
      • a parameterizable transformation model 12 with which reference image 11 can be transformed to a transformed image 13, and/or
      • a synthetic image 207 obtained from reference scene 9 and/or obtained from reference image 11,
  • [0295]
    The choice adopted is that which minimizes the difference between transformed image 13 and synthetic image 207. Synthetic image 207 and transformed image 13 are then said to be close. Deviation 14 is defined as the said difference.
  • [0296]
    Preferably, according to the invention, it is possible by means of the second calculation algorithm to choose, within a set of mathematical projections, one mathematical projection 8 with which synthetic image 207 can be constructed from reference scene 9.
  • Best Transformation
  • [0297]
    The best transformation is defined as:
      • the transformation with which, among the parameterizable transformation models 12, each reference image 11 can be transformed to a transformed image 13 close to synthetic-image class 7 of the reference scene 9 corresponding to the said reference image 11, and/or,
      • the parameterizable transformation models 12 among which the parameterizable transformation models, such as the transformed image 13, are close to synthetic image 207, and/or
      • the parameterizable reverse transformation models 212 among which the parameterizable reverse models, such as the transformed image 13, are close to the synthetic image 207.
    Calibration
  • [0301]
    Calibration is defined as a method with which data related to the intrinsic characteristics of image-capture appliance 1 can be obtained, for one or more configurations used, each composed of an optical system 100 associated with an image-capture appliance 1.
  • [0302]
    Case 1: in the case in which there is only one configuration, the said method includes the following stages:
      • the stage of mounting the said optical system 100 on the said image-capture appliance 1,
      • the stage of choosing one or more reference scenes 9,
      • the stage of choosing several characteristics 74 used,
      • the stage of capturing images of the said reference scenes 9 for the said characteristics used,
      • the stage of calculating the best transformation for each group of reference scenes 9 corresponding to the same characteristics 74 used.
  • [0308]
    Case 2: in the case in which all the configurations corresponding to a given image-capture appliance 1 and to all optical systems 100 of the same type are taken into consideration, the said method includes the following stages:
      • the stage of choosing one or more reference scenes 9,
      • the stage of choosing several characteristics 74 used,
      • the stage of calculating images 103 from characteristics 74 used and in particular from formulas for optical system 100 of the configuration used and from values of parameters, by means, for example, of software for calculating the optical system by ray tracing,
      • the stage of calculating the best transformation for each group of reference scenes 9 corresponding to the same characteristics used.
  • [0313]
    Case 3: in the case in which all the configurations corresponding to a given optical system 100 and to all the image-capture appliances 1 of the same type are taken into consideration, the said method includes the following stages:
      • the stage of mounting the said optical system 100 on an image-capture appliance 1 of the type under consideration,
      • the stage of choosing one or more reference scenes 9,
      • the stage of choosing several characteristics 74 used,
      • the stage of capturing images of the said reference scenes 9 for the said characteristics used,
      • the stage of calculating the best transformation for each group of reference scenes 9 corresponding to the same characteristics used.
  • [0319]
    Calibration can be performed preferentially by the manufacturer of image-capture appliance 1, for each appliance and configuration in case 1. This method is more precise but imposes more limitations and is highly suitable in the case in which optical system 100 is not interchangeable.
  • [0320]
    Alternatively, calibration can be performed by the manufacturer of image-capture appliance 1, for each appliance type and configuration in case 2. This method is less precise but is simpler.
  • [0321]
    Alternatively, calibration can be performed by the manufacturer of image-capture appliance 1 or by a third party, for each optical system 100 and type of appliance in case 3. This method is a compromise in which one optical system 100 can be used on all image-capture appliances 1 of one type, without repeating the calibration for each combination of image-capture appliance 1 and optical system 100. In the case in which an image-capture appliance has a non-interchangeable optical system, the method permits the calibration to be performed only one time for a given type of appliance.
  • [0322]
    Alternatively, calibration can be performed by the appliance seller or installer, for each image-capture appliance 1 and configuration in case 1.
  • [0323]
    Alternatively, calibration can be performed by the appliance seller or installer, for each optical system 100 and type of appliance in case 3.
  • [0324]
    Alternatively, calibration can be performed by the appliance user, for each appliance and configuration in case 1.
  • [0325]
    Alternatively, calibration can be performed by the appliance user, for each optical system 100 and type of appliance in case 3.
  • Design of the Digital Optical System
  • [0326]
    Design of the digital optical system is defined as a method for reducing the cost of optical system 100, by:
      • designing an optical system 100 having defects, especially in positioning of real points, or choosing the same from a catalog,
      • reducing the number of lenses, and/or
      • simplifying the shape of the lenses, and/or
      • using less expensive materials, processing operations or manufacturing processes.
  • [0331]
    The said method includes the following stages:
      • the stage of choosing an acceptable difference (within the meaning defined hereinabove),
      • the stage of choosing one or more reference scenes 9,
      • the stage of choosing several characteristics 74 used.
  • [0335]
    The said method also includes iteration of the following stages:
      • the stage of choosing an optical formula that includes in particular the shape, material and arrangement of the lenses,
      • the stage of calculating images 103 from the characteristics 74 used and in particular from the formulas for optical system 100 of the configuration used, by employing, for example, software for calculating the optical system by ray tracing, or by making measurements on a prototype,
      • the stage of calculating the best transformation for each group of reference scenes 9 corresponding to the same characteristics 74 used,
      • the stage of verifying if the difference is acceptable, until the difference is acceptable.
    Formatted Information
  • [0340]
    Formatted information 15 associated with image 103, or formatted information 15, is defined as all or part of the following data:
      • data related to the intrinsic technical characteristics of image-capture appliance 1, especially the distortion characteristics, and/or
      • data related to the technical characteristics of image-capture appliance 1 at the moment of image capture, especially the exposure time, and/or
      • data related to the preferences of the said user, especially the color temperature, and/or
      • data related to the deviations 14.
  • [0345]
    Database of Characteristics
  • [0346]
    A database 22 of characteristics is defined as a database containing formatted information 15 for one or more image-capture appliances 1 and for one or more images 103.
  • [0347]
    The said database 22 of characteristics can be stored in centralized or distributed manner, and in particular can be:
      • integrated into image-capture appliance 1,
      • integrated into optical system 100,
      • integrated into a removable storage device,
      • integrated into a PC or other computer connected to the other elements during image capture,
      • integrated into a PC or other computer connected to the other elements after image capture,
      • integrated into a PC or other computer capable of reading a storage medium shared with image-capture appliance 1,
      • integrated into a remote server connected to a PC or other computer, itself connected to the other image-capture elements.
    Fields
  • [0355]
    Referring to FIG. 8, a definition will now be given of the concept of fields 91. The formatted information 15 associated with image 103 can be recorded in several forms and structured into one or more tables, but it corresponds logically to all or part of fields 91, comprising:
  • [0356]
    (a) the focal distance,
  • [0357]
    (b) the depth of field
  • [0358]
    (c) the geometric defects.
  • [0359]
    The said geometric defects include geometric defects of image 103 characterized by the parameters associated with the filming characteristics 74 and a parameterizable transformation representing the characteristics of image-capture appliance 1 at the moment of filming. By means of the said parameters and of the said parameterizable transformation, it is possible to calculate the corrected position of a point of image 103.
  • [0360]
    The said geometric defects also include the vignetting characterized by the parameters associated with filming characteristics 74 and a parameterizable transformation representing the characteristics of image-capture appliance 1 at the moment of filming. By means of the said parameters and the said parameterizable transformation, it is possible to calculate the corrected intensity of a point of image 103.
  • [0361]
    The said geometric defects also include the color cast characterized by the parameters associated with filming characteristics 74 and a parameterizable transformation representing the characteristics of image-capture appliance 1 at the moment of filming. By means of the said parameters and the said parameterizable transformation, it is possible to calculate the corrected color of a point of image 103.
  • [0362]
    The said fields 91 also include (d) the sharpness of image 103.
  • [0363]
    The said sharpness includes the blurring in resolution of image 103 characterized by the parameters associated with filming characteristics 74 and a parameterizable transformation representing the characteristics of image-capture appliance 1 at the moment of filming. By means of the said parameters and the said parameterizable transformation, it is possible to calculate the corrected shape of a point of image 103. Blurring covers in particular coma, spherical aberration, astigmatism, grouping into pixels 104, chromatic aberration, depth of field, diffraction, parasitic reflections and field curvature.
  • [0364]
    The said sharpness also includes the blurring in depth of field, in particular spherical aberrations, coma and astigmatism. The said blurring depends on the distance of the points of scene 3 relative to image-capture appliance 1, and it is characterized by the parameters associated with filming characteristics 74 and a parameterizable transformation representing the characteristics of image-capture appliance 1 at the moment of filming. By means of the said parameters and of the said parameterizable transformation, it is possible to calculate the corrected shape of a point of image 103.
  • [0365]
    The said fields 91 also include (e) parameters of the quantization method. The said parameters depend on the geometry and physics of sensor 101, on the architecture of electronic unit 102 and on any processing software that may be used.
  • [0366]
    The said parameters include a function that represents the variations of intensity of a pixel 104 as a function of wavelength and light flux derived from the said scene 3. The said function includes in particular gamma information.
  • [0367]
    The said parameters also include:
      • the geometry of the said sensor 101, especially the shape, the relative position and the number of sensitive elements of the said sensor 101,
      • a function representative of the spatial and temporal distribution of noise of image-capture appliance 1,
      • a value representative of the exposure time for image capture.
  • [0371]
    The said fields 91 also include (f) parameters of the digital-processing operations performed by image-capture appliance 1, especially digital zoom and compression. These parameters depend on the processing software of image-capture appliance 1 and on the user's adjustments.
  • [0372]
    The said fields 91 also include:
  • [0373]
    (g) parameters representative of the user's preferences, especially as regards the degree of blurring and the resolution of image 103.
  • [0374]
    (h) the deviations 14.
  • Calculation of Formatted Information
  • [0375]
    The formatted information 15 can be calculated and recorded in database 22 in several stages.
  • [0376]
    a) A stage at the end of design of image-capture appliance 1.
  • [0377]
    By means of this stage it is possible to obtain intrinsic technical characteristics of image-capture appliance 1, and in particular:
      • the spatial and temporal distribution of the noise generated by electronic unit 102,
      • the formula for conversion of light flux to pixel value,
      • the geometry of sensor 101.
  • [0381]
    b) A stage at the end of calibration or design of the digital optical system.
  • [0382]
    By means of this stage it is possible to obtain other intrinsic technical characteristics of image-capture appliance 1, and in particular, for a certain number of values of characteristics used, the best associated transformation and the associated deviation 14.
  • [0383]
    c) A stage in which the user's preferences are chosen by means of pushbuttons, menus or removable media, or of connection to another device.
  • [0384]
    d) An image capture stage.
  • [0385]
    By means of this stage (d) it is possible to obtain technical characteristics of image-capture appliance 1 at the moment of image capture, and in particular the exposure time, which is determined by the manual or automatic adjustments made.
  • [0386]
    By means of stage (d) it is also possible to obtain the focal distance. The focal distance is calculated from:
      • a measurement of the position of the group of lenses of variable focal length of optical system 100 of the configuration used, or
      • a set value input to the positioning motor, or
      • a manufacturer's value if the focal length is fixed.
  • [0390]
    The said focal distance can then be determined by analysis of the content of image 103.
  • [0391]
    By means of stage (d) it is also possible to obtain the depth of field. The depth of field is calculated from:
      • a measurement of the position of the group of focusing lenses of optical system 100 of the configuration used, or
      • a set value input to the positioning motor, or
      • a manufacturer's value if the depth of field is fixed.
  • [0395]
    By means of stage (d) it is also possible to obtain the defects of geometry and of sharpness. The defects of geometry and of sharpness correspond to a transformation calculated by means of a combination of transformations of the database 22 of characteristics obtained at the end of stage (b). This combination is chosen to represent the values of parameters corresponding to the characteristics 74 used, especially the focal distance.
  • [0396]
    By means of stage (d) it is also possible to obtain the parameters of digital processing performed by image-capture appliance 1. These parameters are determined by the manual or automatic adjustments made.
  • [0397]
    The calculation of formatted information 15 according to stages (a) to (d) can be performed by:
      • a device or software integrated into image-capture appliance 1, and/or
      • driver software in a PC or other computer, and/or
      • software in a PC or other computer, and/or
      • a combination of the three.
  • [0402]
    The foregoing transformations in stage (b) and stage (d) can be stored in the form of:
      • a general mathematical formula,
      • a mathematical formula for each point,
      • a mathematical formula for certain characteristic points.
  • [0406]
    The mathematical formulas can be described by:
      • a list of coefficients,
      • a list of coefficients and coordinates.
  • [0409]
    By means of these different methods it is possible to reach a compromise between the size of the memory available for storage of the formulas and the calculating power available for calculation of the corrected images 71.
  • [0410]
    In addition, in order to retrieve the data, identifiers associated with the data are recorded in database 22. These identifiers include in particular:
      • an identifier of the type and of the reference of image-capture appliance 1,
      • an identifier of the type and of the reference of optical system 100, if it is removable,
      • an identifier of the type and of the reference of any other removable element having a link to the stored information,
      • an identifier of image 103,
      • an identifier of the formatted information 15.
    Completed Image
  • [0416]
    As described by FIG. 11, a completed image 120 is defined as the image 103 associated with the formatted information 15. This completed image 120 can preferentially have the form of a file P100, as described by FIG. 14. Completed image 120 can also be distributed into a plurality of files.
  • [0417]
    Completed image 120 can be calculated by image-capture appliance 1. It can also be calculated by an external calculating device, such as a computer.
  • Image-Processing Software
  • [0418]
    Image-processing software 4 is defined as software that accepts one or more completed images 120 as input and that performs processing operations on these images. These processing operations can include in particular:
      • calculating a corrected image 71,
      • performing measurements in the real world,
      • combining several images,
      • improving the fidelity of the images relative to the real world,
      • improving the subjective quality of images,
      • detecting objects or persons 107 in a scene 3,
      • adding objects or persons 107 to a scene 103,
      • replacing or modifying objects or persons 107 in a scene 3,
      • removing shadows from a scene 3,
      • adding shadows to a scene 3,
      • searching for objects in an image base.
  • [0430]
    The said image-processing software can be:
      • integrated into image-capture appliance 1,
      • run on calculating means 17 connected to image-capture appliance 1 by transmission means 18.
    Digital Optical System
  • [0433]
    A digital optical system is defined as the combination of an image-capture appliance 1, a database 22 of characteristics and a calculating means 17 that permits:
      • image capture of an image 103,
      • calculation of the completed image,
      • calculation of the corrected image 71.
  • [0437]
    Preferentially, the user obtains corrected image 71 directly. If he wishes, the user may demand suppression of automatic correction.
  • [0438]
    The database 22 of characteristics may be:
      • integrated into image-capture appliance 1,
      • integrated into a PC or other computer connected to the other elements during image capture,
      • integrated into a PC or other computer connected to the other elements after image capture,
      • integrated into a PC or other computer capable of reading a storage medium shared with image-capture appliance 1,
  • [0443]
    integrated into a remote server connected to a PC or other computer, itself connected to the other image-capture elements.
  • [0444]
    Calculating means 17 may be:
      • integrated onto a component together with sensor 101,
      • integrated onto a component together with part of electronics unit 102,
      • integrated into image-capture appliance 1,
      • integrated into a PC or other computer connected to the other elements during image capture,
      • integrated into a PC or other computer connected to the other elements after image capture,
      • integrated into a PC or other computer capable of reading a storage medium shared with image-capture appliance 1,
      • integrated into a remote server connected to a PC or other computer, itself connected to the other image-capture elements.
    Processing of the Complete Chain
  • [0452]
    The foregoing paragraphs have essentially presented precise details of the concepts and description of the method and system according to the invention for providing, to image-processing software 4, formatted information 15 related to the characteristics of image-capture appliance 1.
  • [0453]
    In the paragraphs to follow, an expanded definition will be given of the concepts and a supplemented description will be given of the method and system according to the invention for providing, to image-processing software 4, formatted information 15 related to the characteristics of image-restitution means 19. In this way the processing of a complete chain will be explained.
  • [0454]
    By means of the processing of the complete chain, it is possible:
      • to improve the quality of image 103 from one end of the chain to the other, to obtain a restituted image 191 while correcting the defects of image-capture appliance 1 and of image-restitution means 19, and/or
      • to use optical systems of lower quality and of lower cost in a video projector in combination with software for improvement of image quality.
    Definitions Associated with the Image-Restitution Means
  • [0457]
    On the basis of FIGS. 2, 17 and 6, a description will now be given of how the characteristics of an image-restitution means 19 such as a printer, a visual display screen or a projector are taken into account in the formatted information 15.
  • [0458]
    The supplements or modifications to be made to the definitions in the case of an image-restitution means 19 may be inferred by analogy by a person skilled in the art by analogy with the definitions provided in the case of an image-capture appliance 1. Nevertheless, in order to illustrate this method, a description with reference in particular to FIG. 6 and FIG. 17 will now be given of the main supplements or modifications.
  • [0459]
    By restitution characteristics 95 used there are designated the intrinsic characteristics of image-restitution means 19, the characteristics of image-restitution means 19 at the moment of image restitution, and the user's preferences at the moment of image restitution. In the case of a projector in particular, the restitution characteristics 95 used include the shape and position of the screen used.
  • [0460]
    By parameterizable restitution transformation model 97 (or parameterizable restitution transformation 97 for short), there is designated a mathematical transformation similar to parameterizable transformation model 12. By parameterizable reverse restitution transformation model 297 (or parameterizable reverse restitution transformation 297 for short), there is designated a mathematical transformation similar to parameterizable reverse transformation model 212.
  • [0461]
    By corrected restitution image 94 there is designated the image obtained by application of parameterizable restitution transformation 97 to image 103.
  • [0462]
    By mathematical restitution projection 96 there is designated a mathematical projection that associates, with a corrected restitution image 94, a mathematical restitution image 92 on the mathematical restitution surface geometrically associated with the surface of restitution medium 190. The mathematical restitution points of the mathematical restitution surface have a shape, position, color and intensity calculated from corrected restitution image 94.
  • [0463]
    By real restitution projection 90 there is designated a projection that associates a restituted image 191 with an image 103. The pixel values of image 103 are converted by the electronic unit of restitution means 19 to a signal that drives the modulator of restitution means 19. Real restitution points are obtained on restitution medium 190. The said real restitution points are characterized by shape, color, intensity and position. The phenomenon of grouping into pixels 104 described hereinabove in the case of an image-capture appliance 1 does not occur in the case of an image-restitution means. On the other hand, an inverse phenomenon occurs, with the result in particular that lines take on a staircase appearance.
  • [0464]
    Restitution difference 93 is designated as the difference between restituted image 191 and mathematical restitution image 92. This restitution difference 93 is obtained by analogy with difference 73.
  • [0465]
    By restitution reference 209 there is designated an image 103 in which the values of pixels 104 are known. By restituted reference 211 there is designated the restituted image 191 obtained by mathematical restitution projection 90 of restitution reference 209. By corrected reference restitution image 213, there is designated the corrected restitution image 94 corresponding to restitution reference 209 for parameterizable restitution transformation model 97 and/or for parameterizable reverse restitution transformation model 297. By synthetic restitution image 307 there is designated the mathematical restitution image 92 obtained by mathematical restitution projection 96 of corrected reference restitution image 213.
  • [0466]
    By best restitution transformation there is designated:
      • for a restitution reference 209 and the restitution characteristics 95 used, that with which image 103 can be transformed to a corrected restitution image 94 such that its mathematical restitution projection 92 exhibits the minimum restitution difference 93 compared with restituted image 191, and/or
      • the parameterizable restitution transformation 97 among the parameterizable restitution transformation models such that restituted reference 211 exhibits the minimum restitution difference 93 compared with the synthetic restitution image 307, and/or
      • the parameterizable reverse restitution transformation 297 among the parameterizable reverse transformation models such that the restituted reference 211 exhibits the minimum restitution difference 93 compared with the synthetic restitution image 307.
  • [0470]
    The restituted reference 211 and the synthetic restitution image 307 are then said to be close.
  • [0471]
    The methods of restitution calibration and of design of the digital optical restitution system are comparable with the methods of calibration and of design of the digital optical system in the case of an image-capture appliance 1. Nevertheless, differences are present in certain stages, and in particular the following stages:
      • the stage of choosing a restitution reference 209;
      • the stage of performing restitution of the said restitution reference;
      • the stage of calculating the best restitution transformation.
  • [0475]
    Preferably, according to the invention, the method includes a sixth algorithm for calculation of the formatted information 15. By means of this sixth algorithm it is possible to make a choice:
      • within a set of parameterizable restitution transformation models,
      • within a set of parameterizable reverse restitution transformation models,
      • within a set of mathematical restitution projections,
      • within a set of restitution references and within a set of corrected reference restitution images.
  • [0480]
    The choice made by this sixth algorithm is based on:
      • a restitution reference 209, and/or
      • a corrected reference restitution image 213, and/or
      • a parameterizable restitution transformation model 97 with which the restitution reference 209 can be transformed to the corrected reference restitution image 213, and/or
      • a parameterizable reverse restitution transformation model 297 with which the corrected reference restitution image 213 can be transformed to the restitution reference 209, and/or
      • a mathematical restitution projection 96 with which a synthetic restitution image 307 can be constructed from the corrected reference restitution image 213.
  • [0486]
    The choice is made by this sixth algorithm in such a way that the synthetic restitution image 307 is close to the restituted reference 211 obtained by restitution of restitution reference 209 by means of image-restitution means 19. Restituted reference 211 exhibits a restitution deviation 214 compared with synthetic restitution image 307.
  • [0487]
    According to an alternative embodiment of the invention, the method includes a seventh algorithm for calculation of the formatted information. This seventh algorithm includes the stages of:
      • choosing at least one restitution reference 209,
      • restituting restitution reference 209 to a restituted reference 211 by means of image-restitution means 19.
  • [0490]
    By means of this seventh algorithm it is also possible to choose, within a set of parameterizable restitution transformation models and within a set of mathematical restitution projections:
      • a parameterizable restitution transformation model 97 with which restitution reference 209 can be transformed to a corrected reference restitution image 213, and
      • a mathematical restitution projection 96 with which a synthetic restitution image 307 can be constructed from corrected reference restitution image 213.
  • [0493]
    The choice is made by the seventh algorithm in such a way that synthetic restitution image 307 is close to restituted reference 211. The restituted reference exhibits a restitution deviation 214 compared with the synthetic restitution image 307. By means of parameterizable reverse restitution transformation model 297, it is possible to transform corrected reference restitution image 213 to restitution reference 209.
  • [0494]
    According to another alternative embodiment of the invention, the method includes an eighth algorithm for calculation of the formatted information. This eighth algorithm includes the stage of choosing a corrected reference restitution image 213. This eighth algorithm also includes the stage of making a choice within a set of parameterizable restitution transformation models, within a set of mathematical restitution projections and within a set of restitution references. This choice is based on:
      • a restitution reference 209, and/or
      • a parameterizable restitution transformation model 97 with which restitution reference 209 can be transformed to corrected reference restitution image 213, and/or
      • a parameterizable reverse restitution transformation model 297 with which the corrected reference restitution image 213 can be transformed to the restitution reference 209, and/or
      • a mathematical restitution projection 96 with which a synthetic restitution image 307 can be constructed from the corrected reference restitution image 213.
  • [0499]
    The eighth algorithm makes this choice in such a way that synthetic restitution image 307 is close to restituted reference 211 obtained by restitution of restitution reference 209, by means of image-restitution means 19. Restituted reference 211 exhibits a restitution deviation compared with synthetic restitution image 307.
  • [0500]
    Preferably, according to the invention, the method includes a ninth algorithm for calculating the restitution deviations 214. This ninth algorithm includes the stages of:
      • calculating the restitution deviations 214 between restituted reference 211 and synthetic restitution image 307,
      • associating restitution deviations 214 with formatted information 15.
  • [0503]
    It results from the combination of technical features that it is possible to verify automatically, for example during manufacture of the appliance, that the method has produced formatted information within acceptable tolerances.
  • [0504]
    The formatted information 15 related to an image-capture appliance 1 and that related to an image-restitution means 19 can be used end-to-end for the same image.
  • [0505]
    It is also possible to combine the formatted information 15 related to each of the appliances to obtain formatted information 15 related to the appliance chain, for example by addition of a vector field, in the case of geometric distortion.
  • [0506]
    In the foregoing, a description was given of the concept of field in the case of an image-capture appliance 1. This concept is also applicable by analogy in the case of image-restitution means 19. Nonetheless the parameters of the quantization method are replaced by the parameters of the signal-reconstitution method, meaning: the geometry of restitution medium 190 and its position, a function representing the spatial and temporal distribution of the noise of image-restitution means 19.
  • [0507]
    In an alternative embodiment according to the invention, restitution means 19 is associated with an image-capture appliance 1 to restitute, in digital form, restituted reference 211 from restitution reference 209. The method is such that, to produce the formatted information 15 related to the defects P5 of restitution means 19, the formatted information 15 related to image-capture appliance 1 associated with the restitution means is used, for example, to correct the defects of image-capture appliance 1 in such a way that restituted reference 211 contains only the defects P5 of restitution means 19.
  • Generalization of the Concepts
  • [0508]
    The technical features of which the invention is composed and which are specified in the claims have been defined, described and illustrated by referring essentially to image-capture appliances of digital type, or in other words appliances that produce digital images. It can be easily understood that the same technical features are applicable in the case of image-capture appliances that would be the combination of an appliance based on silver technology (a photographic or cinematographic appliance using sensitive silver halide films, negatives or reversal films) with a scanner for producing a digital image from the developed sensitive films. Certainly it is appropriate in this case to adapt at least some of the definitions used. Such adaptations are within the capability of the person skilled in the art. In order to demonstrate the obvious character of such adaptations, it is merely necessary to mention that the concepts of pixel and pixel value illustrated by referring to FIG. 3 must, in the case of the combination of an appliance based on silver technology with a scanner, be applied to an elemental zone of the surface of the film after this has been digitized by means of the scanner. Such transpositions of definitions are self-evident and can be extended to the concept of the configuration used. As an example, the list of removable subassemblies of image-capture appliance 1 included in the configuration used can be supplemented by the type of photographic film effectively used in the appliance based on silver technology.
  • [0509]
    Other characteristics and advantages of the invention will become clear upon reading the indicative and non-limitative definitions and examples explained hereinafter with reference to FIGS. 1 to 17.
  • Appliance
  • [0510]
    Referring in particular to FIGS. 2, 3 and 13, a description will be given of the concept of appliance P25. Within the meaning of the invention, an appliance P25 may be in particular:
      • an image-capture appliance 1, such as a disposable photo appliance, a digital photo appliance, a reflex appliance, a scanner, a fax machine, an endoscope, a camcorder, a surveillance camera, a game, a camera integrated into or connected to a telephone, to a personal digital assistant or to a computer, a thermal camera or an echographic appliance,
      • an image-restitution appliance 19 or image-restitution means 19, such as a screen, a projector, a television set, virtual-reality goggles or a printer,
      • an appliance, including its installation, such as a projector, a screen and the manner in which they are positioned,
      • the positioning of an observer relative to an image-restitution appliance 19, which introduces parallax errors in particular,
      • a human being or observer having vision defects, such as astigmatism,
      • an appliance which it is hoped can be emulated, to produce images having, for example, an appearance similar to those produced by an appliance of the Leica brand,
      • an image-processing device, such as zoom software, which has the edge effect of adding blurring,
      • a virtual appliance equivalent to a plurality of appliances P25,
      • a more complex appliance P25, such as a scanner/fax/printer, a photo-printing Minilab, or a videoconferencing appliance can be regarded as an appliance P25 or as a plurality of appliances P25.
    Appliance Chain
  • [0520]
    Referring in particular to FIG. 13, a description will now be given of the concept of appliance chain P3. An appliance chain P3 is defined as a set of appliances P25. The concept of appliance chain P3 may also include a concept of order.
  • [0521]
    The following examples constitute appliance chains P3:
      • a single appliance P25,
      • an image-capture appliance and an image-restitution appliance 19,
      • a photo appliance, a scanner or a printer, for example in a photo-printing Minilab,
      • a digital photo appliance or a printer, for example in a photo-printing Minilab,
      • a scanner, a screen or a printer, for example in a computer,
      • a screen or projector, and the eye of a human being,
      • one appliance and another appliance which it is hoped can be emulated,
      • a photo appliance and a scanner,
      • an image-capture appliance and image-processing software,
      • image-processing software and an image-restitution appliance 19,
      • a combination of the preceding examples,
      • another set of appliances P25.
    Defect
  • [0534]
    Referring in particular to FIG. 13, a description will now be given of the concept of defect P5. A defect P5 of appliance P25 is defined as a defect related to the characteristics of the optical system and/or of the sensor and/or of the electronic unit and/or of the software integrated in an appliance P25; examples of defects P5 include geometric defects, sharpness defects, colorimetry defects, geometric distortion defects, geometric chromatic aberration defects, geometric vignetting defects, contrast defects, colorimetry defects, in particular rendering of colors and color cast, defects of flash uniformity, sensor noise, grain, astigmatism defects and spherical aberration defects.
  • Image
  • [0535]
    Referring in particular to FIGS. 2, 5, 6 and 13, a description will now be given of the concept of image 103. Image 103 is defined as a digital image captured or modified or restituted by an appliance P25. Image 103 may originate from an appliance P25 of appliance chain P3. Image 103 may be addressed to an appliance P25 of appliance chain P3. More generally, image 103 may be derived from or addressed to appliance chain P3. In the case of animated images, such as video images, composed of a time sequence of fixed images, image 103 is defined as one fixed image of the sequence of images.
  • Formatted Information
  • [0536]
    Referring in particular to FIGS. 7, 8, 10 and 13, a description will now be given of the concept of formatted information 15. Formatted information 15 is defined as data related to the defects P5 or characterizing the defects P5 of one or more appliances P25 of appliance chain P3 and enabling image-processing means P1 to modify the quality of images 103 by making allowance for the defects P5 of appliance P25.
  • [0537]
    To produce the formatted information 15, there can be used various methods and systems based on measurements and/or simulations and/or calibrations, such as, for example, the calibration method described hereinabove.
  • [0538]
    To transmit the formatted information 15, there can be used a file P100 containing the completed image 120. As an example, an image-capture appliance 1 such as a digital photo appliance can produce files containing image 103, formatted information 15 copied from an internal memory of the appliance, and data in Exif format containing the adjustments used.
  • [0539]
    To produce the formatted information 15, it is possible, for example, to use the method and the system described in the International Patent Application filed on the same day as the present application in the name of Vision IQ and entitled “Method and system for producing formatted information related to geometric distortions”. That application describes a method for producing formatted information 15 related to the appliances P25 of an appliance chain P3. Appliance chain P3 is composed in particular of at least one image-capture appliance and/or at least one image-restitution appliance. The method includes the stage of producing formatted information 15 related to the geometric distortions of at least one appliance P25 of the chain.
  • [0540]
    Appliance P25 preferably makes it possible to capture or restitute an image on a medium. Appliance P25 contains at least one fixed characteristic and/or one variable characteristic depending on the image. The fixed characteristic and/or variable characteristic can be associated with one or more values of characteristics, especially the focal length and/or the focusing and their values of associated characteristics. The method includes the stage of producing, from a measured field, measured formatted information related to the geometric distortions of the appliance. The formatted information 15 may include the measured formatted information.
  • [0541]
    To produce the formatted information 15, it is possible, for example, to use the method and the system described in the International Patent Application filed on the same day as the present application in the name of Vision IQ and entitled “Method and system for producing formatted information related to the defects of at least one appliance of a chain, especially to blurring”. That application describes a method for producing formatted information 15 related to the appliances P25 of an appliance chain P3. Appliance chain P3 is composed in particular of at least one image-capture appliance and/or at least one image-restitution appliance. The method includes the stage of producing formatted information 15 related to the defects P5 of at least one appliance P25 of the chain. Preferably, appliance P25 with which an image can be captured or restituted contains at least one fixed characteristic and/or one variable characteristic depending on the image (I). The fixed and/or variable characteristics can be associated with one or more values of characteristics, especially the focal length and/or the focusing and their values of associated characteristics. The method includes the stage of producing measured formatted information related to the defects P5 of appliance P25 from a measured field. The formatted information 15 may include the measured formatted information.
  • [0542]
    To produce the formatted information 15, it is possible, for example, to use the method and the system described in the International Patent Application filed on the same day as the present application in the name of Vision IQ and entitled “Method and system for reducing update frequency of image processing means”. That application describes a method for reducing the update frequency of image-processing means P1, in particular software and/or a component. The image-processing means make it possible to modify the quality of the digital images derived from or addressed to an appliance chain P3. Appliance chain P3 is composed in particular of at least one image-capture appliance and/or at least one image-restitution appliance. Image-processing means P1 employ formatted information 15 related to the defects P5 of at least one appliance of appliance chain P5. The formatted information 15 depends on at least one variable. The formatted information 15 makes it possible to establish a correspondence between one part of the variables and of the identifiers. By means of the identifiers it is possible to determine the value of the variable corresponding to the identifier by taking the identifier and the image into account. It results from the combination of technical features that it is possible to determine the value of a variable, especially in the case in which the physical significance and/or the content of the variable are known only after distribution of image-processing means P1. It also results from the combination of technical features that the time between two updates of the correction software can be spaced apart. It also results from the combination of technical features that the various economic players that produce appliances and/or image-processing means can update their products independently of other economic players, even if the latter radically change the characteristics of their product or are unable to force their client to update their products. It also results from the combination of technical features that a new functionality can be deployed progressively by starting with a limited number of economic players and pioneer users.
  • [0543]
    To search for the formatted information 15, it is possible, for example, to use the method and the system described in the International Patent Application filed on the same day as the present application in the name of Vision IQ and entitled “Method and system for modifying the quality of at least one image derived from or addressed to an appliance chain”. That application describes a method for modifying the quality of at least one image 103 derived from or addressed to a specified appliance chain. The specified appliance chain is composed of at least one image-capture appliance and/or at least one image-restitution appliance. The image-capture appliances and/or the image-restitution appliances being progressively introduced on the market by separate economic players belong to an indeterminate set of appliances. The appliances P25 of the set of appliances exhibit defects P5 that can be characterized by formatted information 15. For the image in question, the method includes the following stages:
      • the stage of compiling directories of the sources of formatted information related to the appliances P25 of the set of appliances,
      • the stage of automatically searching for specific formatted information related to the specified appliance chain among the formatted information 15 compiled in this way,
      • the stage of automatically modifying the image by means of image-processing software and/or image-processing components, while taking into account the specific formatted information obtained in this way.
  • [0547]
    To exploit the formatted information 15, it is possible, for example, to use the method and the system described in the International Patent Application filed on the same day as the present application in the name of Vision IQ and entitled “Method and system for calculating a transformed image from a digital image and formatted information related to a geometric transformation”. That application describes a method for calculating a transformed image from a digital image and formatted information 15 related to a geometric transformation, especially formatted information 15 related to the distortions and/or chromatic aberrations of an appliance chain P3. The method includes the stage of calculating the transformed image from an approximation of the geometric transformation. It results therefrom that the calculation is economical in terms of memory resources, in memory bandpass, in calculating power and therefore in electricity consumption. It also results therefrom that the transformed image does not exhibit any visible or annoying defect as regards its subsequent use.
  • [0548]
    To exploit the formatted information 15, it is possible, for example, to use the method and the system described in the International Patent Application filed on the same day as the present application in the name of Vision IQ and entitled “Method and system for modifying a digital image, taking into account its noise”. That application describes a method for calculating a transformed image from a digital image and formatted information 15 related to the defects P5 of an appliance chain P3. Appliance chain P3 includes image-capture appliances and/or image-restitution appliances. Appliance chain P3 contains at least one appliance P25. The method includes the stage of automatically determining the characteristic data from the formatted information 15 and/or the digital image. It results from the combination of technical features that the transformed image does not exhibit any visible or annoying defect, especially defects related to noise, as regards its subsequent use.
  • Image-Processing Means
  • [0549]
    Referring in particular to FIGS. 7 and 13, a description will now be given of the concept of image-processing means P1. Within the meaning of the present invention, image-processing means P1 are defined, for example, as image-processing software 4 and/or a component and/or an equipment item and/or a system capable of modifying the quality of image 103 by employing formatted information 15 in order to produce a modified image, such as a corrected image 71 or a corrected restitution image 97. The modified image may be addressed to a second appliance of appliance chain P3, distinct or not from appliance P25, for example, the following appliance in appliance chain P3.
  • [0550]
    The modification of image quality by image-processing means P1 may consist, for example, in:
      • suppressing or attenuating the defects P5 of one or more appliances P25 of appliance chain P3 in image 103, and/or
      • modifying image 103 to add at least one defect P5 of one or more appliances P25 of appliance chain P3 in such a way that the modified image resembles an image captured by appliance or appliances P25, and/or
      • modifying image 103 to add at least one defect P5 of one or more appliances P25 of appliance chain P3 in such a way that the restitution of the modified image resembles an image restituted by appliance or appliances P25, and/or
      • modifying image 103 by taking into account the formatted information 15 related to the vision defects P5 of the eye P25 of a human being in appliance chain P3 in such a way that restitution of the modified image is perceived by the eye of the human being as corrected for all or part of the defects P5.
  • [0555]
    A correction algorithm is defined as the method employed by an image-processing means P1 to modify image quality depending on the defect P5.
  • [0556]
    Image-processing means P1 may assume various forms depending on the application.
  • [0557]
    Image-processing means P1 may be integrated entirely or partly in appliance P25, as in the following examples:
      • an image-capture appliance that produces modified images, such as a digital photo appliance in which image-processing means P1 are integrated,
      • an image-restitution appliance 19, which displays or prints modified images, such as a video projector in which image-processing means P1 are included,
      • a hybrid appliance, which corrects the defects of its elements, such as a scanner/printer/fax machine in which image-processing means P1 are included,
      • a professional image-capture appliance, which produces modified images, such as an endoscope in which image-processing means P1 are included.
  • [0562]
    In the case in which image-processing means P1 are integrated in appliance P25, appliance P25 in practice corrects its own defects P5, and the appliances P25 of appliance chain P3 can be determined by design, for example in a fax machine: a scanner and a printer; nevertheless, the user is able to use only part of the appliances P25 of appliance chain P3, for example if the fax machine can also be used as a stand-alone printer.
  • [0563]
    Image-processing means P1 can be integrated entirely or partly in a computer, for example in the following manner:
      • in an operating system, such as Windows or the Mac OS, in order to modify automatically the quality of images derived from or addressed to a plurality of appliances P25, which may vary depending on image 103 and/or in time, examples being scanners, photo appliances and printers; the automatic correction may be made, for example, when image 103 is input into the system, or when printing is requested by the user,
      • in an image-processing application, such as Photoshop™, to modify automatically the quality of images derived from or addressed to a plurality of appliances P25, which may vary depending on image 103 and/or in time, examples being scanners, photo appliances and printers; the automatic correction may be made, for example, when the user activates a filter command in Photoshop™,
      • in a photo-printing appliance (such as Photofinishing or Minilab in English), to modify automatically the quality of images derived from a plurality of photo appliances, which may vary depending on the image and/or in time, examples being disposable cameras, digital photo appliances and compact disks, the automatic correction may take into account the photo appliances as well as the integrated scanner and printer, and may be applied at the moment at which the printing jobs are initiated,
      • on a server, for example on the Internet, to modify automatically the quality of images derived from a plurality of photo appliances, which may vary depending on the image and/or in time, examples being disposable cameras and digital photo appliances, the automatic correction may take into account the photo appliances as well as a printer, for example, and may be applied at the moment at which the images 103 are recorded on the server, or at the moment at which the printing jobs are initiated.
  • [0568]
    In the case in which image-processing means P1 are integrated in a computer, image-processing means P1 are for practical purposes compatible with multiple appliances P25, and at least one appliance P25 of appliance chain P3 may vary from one image 103 to another.
  • [0569]
    To provide formatted information 15 in a standard format to image-processing means P1, it is possible, for example, to associate the formatted information 15 with image 103:
      • in a file P100,
      • by using identifiers of appliances P25 of appliance chain P3, such as data in Exif format in file P100, in order to retrieve formatted information 15 in database 22 of characteristics.
    Variable Characteristic
  • [0572]
    On the basis of FIG. 13, a description will now be given of the concept of variable characteristic P6. According to the invention, a variable characteristic P6 is defined as a measurable factor, which is variable from one image 103 to another that has been captured, modified or restituted by the same appliance P25, and which has an influence on defect P5 of the image that has been captured, modified or restituted by appliance P25, especially:
      • a global variable characteristic, which is fixed for a given image 103, an example being a characteristic of appliance P25 at the moment of capture or restitution of the image, related to an adjustment of the user or related to an automatic function of appliance P25, such as the focal length,
      • a local variable characteristic, which is variable within a given image 103, an example being coordinates x, y or rho, theta in the image, permitting image-processing means P1 to apply local processing that differs depending on the zone of the image.
  • [0575]
    A measurable factor which is variable from one appliance P25 to another but which is fixed from one image 103 to another that has been captured, modified or restituted by the same appliance P25 is not generally considered to be a variable characteristic P6. An example is the focal length for an appliance P25 with fixed focal length.
  • [0576]
    The adjustments used as described hereinabove are examples of variable characteristics P6.
  • [0577]
    The formatted information 15 may depend on at least one variable characteristic P6.
  • [0578]
    By variable characteristic P6 there can be understood in particular:
      • the focal length of the optical system,
      • the redimensioning applied to the image (digital zoom factor: enlargement of part of the image; and/or under-sampling: reduction of the number of pixels of the image),
      • the nonlinear brightness correction, such as the gamma correction,
      • the enhancement of contour, such as the level of deblurring applied by appliance P25,
      • the noise of the sensor and of the electronic unit,
      • the aperture of the optical system,
      • the focusing distance,
      • the number of the frame on a film,
      • the underexposure or overexposure,
      • the sensitivity of the film or sensor,
      • the type of paper used in a printer,
      • the position of the center of the sensor in the image,
      • the rotation of the image relative to the sensor,
      • the position of a projector relative to the screen,
      • the white balance used,
      • the activation of a flash and/or its power,
      • the exposure time,
      • the sensor gain,
      • the compression,
      • the contrast,
      • another adjustment applied by the user of appliance P25, such as a mode of operation,
      • another automatic adjustment of appliance P25,
      • another measurement performed by appliance P25.
  • [0602]
    In the case of a restitution means 19, the variable characteristic P6 can also be defined as variable restitution characteristic.
  • Variable Characteristic Value
  • [0603]
    On the basis of FIG. 13, a description will now be given of the concept of variable characteristic value P26. A variable characteristic value P26 is defined as the value of variable characteristic P6 at the moment of capture, modification or restitution of a specified image, such value being obtained, for example, from data in Exif format present in file P100. Image-processing means P1 can then process or modify the quality of image 103 as a function of variable characteristics P6, by using formatted information 15 that depends on variable characteristics P6 and by determining the value P26 of the variable characteristics.
  • [0604]
    In the case of a restitution means 19, the value of variable characteristic P6 can also be defined as a variable restitution characteristic.
  • Measured Formatted Information, Extended Formatted Information
  • [0605]
    As illustrated in FIG. 15, the formatted information 15 or a fraction of the formatted information 15 can include measured formatted information P101 to illustrate a raw measurement, such as a mathematical field related to geometric distortion defects at a certain number of characteristic points of an array 80. As illustrated in FIG. 15, the formatted information 15 or a fraction of the formatted information 15 can include extended formatted information P102, which can be calculated from measured formatted information P101, for example by interpolation for real points other than the characteristic points of array 80. In the foregoing, it has been seen that a formatted information item 15 might depend on variable characteristics P6. According to the invention, a combination P120 is defined as a combination composed of variable characteristics P6 and of values P26 of variable characteristics, an example being a combination P120 composed of the focal length, of the focusing, of the diaphragm aperture, of the capture speed, of the aperture, etc. and of associated values. It is difficult to imagine how the formatted information 15 related to different combinations P120 can be calculated, all the more so because certain characteristics of combination P120, such as the focal length and the distance, can vary continuously.
  • [0606]
    The invention provides for calculating the formatted information 15 in the form of extended formatted information P102 by interpolation from measured formatted information P101 related to a predetermined selection of combinations P120 of known variable characteristics P6.
  • [0607]
    For example, measured formatted information P101 related to the combination P120 of “focal length=2, distance=7, capture speed= 1/100”, to the combination of “focal length=10, distance=7, capture speed= 1/100” and to the combination of “focal length=50, distance=7, capture speed= 1/100” is used to calculate extended formatted information P102 that depends on focal length as the variable characteristic P6. By means of this extended formatted information P102, it is possible in particular to determine formatted information related to the combination of “focal length=25, distance=7 and capture speed= 1/100”.
  • [0608]
    The measured formatted information P101 and the extended formatted information P102 may exhibit an interpolation deviation P121. The invention may include the stage of selecting zero or one or more variable characteristics P6, such that interpolation deviation P121 for the extended formatted information P102 obtained for the variable characteristics P6 selected in this way is smaller than a predetermined interpolation threshold. In fact, certain variable characteristics P6 may have a smaller influence than others on the defect P5, and the error introduced by making the approximation that these are constant may merely be minimum; for example, the focusing adjustment may have merely a slight influence on the vignetting defect, and for this reason may not be part of the variable characteristics P6 selected. The variable characteristics P6 may be selected at the moment of production of the formatted information 15. It results from the combination of technical features that the modification of image quality employs simple calculations. It also results from the combination of technical features that the extended formatted information P102 is compact. It also results from the combination of technical features that the eliminated variable characteristics P6 have the least influence on the defect P5. It results from the combination of technical features that image quality can be modified with specified precision by means of the formatted information 15.
  • [0609]
    In the case of a restitution means 19, the combination 120 may also be defined as a restitution combination.
  • [0610]
    In the case of a restitution means 19, the measured formatted information P101 may also be defined as measured formatted restitution information.
  • [0611]
    In the case of a restitution means 19, the extended formatted information P102 may also be defined as extended formatted restitution information.
  • [0612]
    In the case of a restitution means 19, the interpolation deviations P121 may also be defined as interpolation restitution deviations.
  • Parameterizable Model, Parameters
  • [0613]
    Referring in particular to FIGS. 5, 6 and 16, a description will now be given of the concept of parameters P9 and parameterizable model P10. Within the meaning of the invention, a parameterizable model P10 is defined as a mathematical model that may depend on variables P6 and that may be related to one or more defects P5 of one or more appliances P25; parameterizable transformation model 12, parameterizable reverse transformation model 212, parameterizable restitution transformation model 97 and parameterizable restitution transformation model 297 described hereinabove are examples of parameterizable models P10; for example, a parameterizable model P10 may be related to:
      • sharpness defects or blurring of a digital photo appliance,
      • geometric vignetting defects of a photo appliance which it is hoped can be emulated,
      • geometric distortion defects and geometric chromatic aberration defects of a projector,
      • sharpness or blurring defects of a disposable photo appliance combined with a scanner.
  • [0618]
    The formatted information 15 related to a defect P5 of an appliance P25 may be presented in the form of the parameters P9 of a parameterizable model P10 depending on variable characteristics P6; by means of the parameters P9 of parameterizable model P10, it is possible to identify a mathematical function P16 in a set of mathematical functions, such as multi-variable polynomials; by means of the mathematical functions P16, it is possible to modify image quality as a function of specified values of the variables P6.
  • [0619]
    In such a way that image-processing means P1 can use the parameters P9 of parameterizable transformation model P10 to calculate the modified image, for example to calculate the corrected intensity or the corrected restitution intensity of a point of the image.
  • Color Plane
  • [0620]
    Referring in particular to FIG. 15, a description will now be given of the concept of color plane P20 of a colored image 103. Image 103 can be decomposed into color planes P20 in various ways: number of planes (1, 3 or more), precision (8 bits unsigned, 16 bits signed, floating, etc.) and significance of the planes (relative to a standard color space). Image 103 can then be decomposed in various ways into color planes P20: red color plane composed of red pixels, green color plane, blue color plane (RGB) or brightness, saturation, hue, etc.; on the other hand, color spaces such as PIM exist, or negative pixel values are possible in order to permit representation of subtractive colors, which cannot be represented in positive RGB; finally, it is possible to encode a pixel value on 8 bits or 16 bits, or by using floating values. As an example of how the formatted information 15 may be related to the color planes P20, the sharpness defects can be characterized differently for the planes of red, green and blue color, to permit image-processing means P1 to correct the sharpness defect differently for each color plane P20.
  • Provision of the Formatted Information
  • [0621]
    On the basis in particular of FIGS. 8, 13, 15 and 16, a description will now be given of an alternative embodiment of the invention. To provide formatted information 15 in a standard format to image-processing means P1, the system includes data-processing means and the method includes the stage of filling in at least one field 91 of the standard format with the formatted information 15. Field 91 may then contain in particular:
      • values related to the defects P5, for example in the form of parameters P9, in such a way that image-processing means P1 can use the parameters P9 to modify image quality by taking the defects P5 into account, and/or
      • values related to the sharpness defects, for example in the form of parameters P9, in such a way that image-processing means P1 can use the parameters P9 to modify image quality by taking the sharpness defects into account, and to calculate the corrected shape or the corrected restitution shape of a point of the image, and/or
      • values related to the colorimetry defects, for example in the form of parameters P9, in such a way that image-processing means P1 can use the parameters P9 to modify image quality by taking the colorimetry defects into account, and to calculate the corrected color or the corrected restitution color of a point of the image, and/or
      • values related to the geometric distortion defects and/or to the geometric chromatic aberration defects, for example in the form of parameters P9, in such a way that image-processing means P1 can use the parameters P9 to modify image quality by taking the geometric distortion defects and/or the geometric chromatic aberration defects into account, and to calculate the corrected position or the corrected restitution position of a point of the image, and/or
      • values related to the geometric vignetting defects, for example in the form of parameters P9, in such a way that image-processing means P1 can use the parameters P9 to modify image quality by taking the geometric vignetting defects into account, and to calculate the corrected intensity or the corrected restitution intensity of a point of the image, and/or
      • values related to the deviations 14, and/or
      • values that are functions of variable characteristics P6 depending on image 103, for example the polynomial coefficients and terms which depend on the variable characteristic P6 corresponding to the focal length and with which it is possible to calculate the corrected intensity of a point of the image as a function of its distance from the center, in such a way that the image-processing means can calculate the corrected intensity of a point for any value of focal length of the image-capture appliance at the moment at which image 103 was captured,
      • values related to formatted information related to the color planes P20,
      • values related to formatted information,
      • values related to measured formatted information,
      • values related to extended formatted information.
    Production of Formatted Information
  • [0633]
    On the basis in particular of FIGS. 7, 12 and 17, a description will now be given of an alternative embodiment of the invention. To produce formatted information 15 related to the defects P5 of the appliances P25 of an appliance chain P3, the invention can employ data-processing means and the first algorithm and/or second algorithm and/or third algorithm and/or fourth algorithm and/or fifth algorithm and/or sixth algorithm and/or seventh algorithm and/or eighth algorithm as described hereinabove.
  • Application of the Invention to Cost Reduction
  • [0634]
    Cost reduction is defined as a method and system for lowering the cost of an appliance P25 or of an appliance chain P3, especially the cost of the optical system of an appliance or of an appliance chain, the method consisting in:
      • reducing the number of lenses, and/or
      • simplifying the shape of the lenses, and/or
      • designing an optical system having defects P5 that are larger than those desired for the appliance or the appliance chain, or choosing the same from a catalog, and/or
      • using materials, components, processing operations or manufacturing methods that are less costly for the appliance or the appliance chain and that add defects.
  • [0639]
    The method and system according to the invention can be used to lower the cost of an appliance or of an appliance chain: it is possible to design a digital optical system, to produce formatted information related to the defects of the appliance or of the appliance chain, to use this formatted information to enable image-processing means P1, whether they are integrated or not, to modify the quality of images derived from or addressed to the appliance or to the appliance chain, in such a way that the combination of the appliance or the appliance chain with the image-processing means is capable of capturing, modifying or restituting images of the desired quality at reduced cost.

Claims (1)

  1. 1. A method for producing formatted information related to defects of appliances of at least two appliance chains, addressed to an image-processor for modifying quality of images processed by the image-processor, each appliance chain including at least one image-capture appliance, at least one image-restitution appliance, or a combination thereof, the method comprising:
    producing data of the formatted information according to a standard format, characterizing the defects of the appliances of each appliance chain, and
    storing the formatted information in a database, wherein
    at least one image-capture appliance is provided with at least one variable characteristic depending on the image, fractions of the formatted information being related to the defects of the at least one image-capture appliance, each variable characteristic being configured to be associated with a value to form a combination including the set of the variable characteristics and of the values, the method further including
    producing a first fraction of the formatted information for a predetermined selection of the combinations, the first fraction of the formatted information being entered in a field of the formatted information as measured formatted information, and
    obtaining a second fraction of the formatted information related to an arbitrary combination from the measured formatted information, as extended formatted information, the formatted information including the extended formatted information in place of the measured formatted information.
US13489892 2001-07-12 2012-06-06 Method and system for producing formatted information related to defects of appliances Abandoned US20120308160A1 (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
FR0109291A FR2827459B1 (en) 2001-07-12 2001-07-12 Method and system to provide some image processing software formatted information related to characteristics of the image capture devices and / or image reproduction means
FR01-09292 2001-07-12
FR0109292A FR2827460B1 (en) 2001-07-12 2001-07-12 Method and system for providing, in a standard format, some image processing software information related to characteristics of the image capture devices and / or means of resti
FR01-09291 2001-07-12
PCT/FR2002/001914 WO2003007241A1 (en) 2001-07-12 2002-06-05 Method and system for producing formatted information related to defects of appliances
US10483322 US7760955B2 (en) 2001-07-12 2002-06-05 Method and system for producing formatted information related to defects of appliances
US12838184 US20100278415A1 (en) 2001-07-12 2010-07-16 Method and system for producing formatted information related to defects of appliances
US13489892 US20120308160A1 (en) 2001-07-12 2012-06-06 Method and system for producing formatted information related to defects of appliances

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13489892 US20120308160A1 (en) 2001-07-12 2012-06-06 Method and system for producing formatted information related to defects of appliances

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12838184 Continuation US20100278415A1 (en) 2001-07-12 2010-07-16 Method and system for producing formatted information related to defects of appliances

Publications (1)

Publication Number Publication Date
US20120308160A1 true true US20120308160A1 (en) 2012-12-06

Family

ID=26213095

Family Applications (10)

Application Number Title Priority Date Filing Date
US10483497 Active 2024-02-16 US7724977B2 (en) 2001-07-12 2002-06-05 Method and system for providing formatted data to image processing means in accordance with a standard format
US10483496 Active 2023-07-16 US7343040B2 (en) 2001-07-12 2002-06-05 Method and system for modifying a digital image taking into account it's noise
US10482413 Active 2025-06-13 US8675980B2 (en) 2001-07-12 2002-06-05 Method and system for reducing update frequency of image-processing means
US10483495 Active 2023-07-10 US7346221B2 (en) 2001-07-12 2002-06-05 Method and system for producing formatted data related to defects of at least an appliance of a set, in particular, related to blurring
US10483494 Active 2024-04-21 US7792378B2 (en) 2001-07-12 2002-06-05 Method and system for modifying image quality
US10483322 Active 2025-12-24 US7760955B2 (en) 2001-07-12 2002-06-05 Method and system for producing formatted information related to defects of appliances
US12838198 Active US8559743B2 (en) 2001-07-12 2010-07-16 Method and system for modifying image quality
US12838184 Abandoned US20100278415A1 (en) 2001-07-12 2010-07-16 Method and system for producing formatted information related to defects of appliances
US13489892 Abandoned US20120308160A1 (en) 2001-07-12 2012-06-06 Method and system for producing formatted information related to defects of appliances
US14021235 Active 2022-11-10 US9536284B2 (en) 2001-07-12 2013-09-09 Method and system for modifying image quality of an image

Family Applications Before (8)

Application Number Title Priority Date Filing Date
US10483497 Active 2024-02-16 US7724977B2 (en) 2001-07-12 2002-06-05 Method and system for providing formatted data to image processing means in accordance with a standard format
US10483496 Active 2023-07-16 US7343040B2 (en) 2001-07-12 2002-06-05 Method and system for modifying a digital image taking into account it's noise
US10482413 Active 2025-06-13 US8675980B2 (en) 2001-07-12 2002-06-05 Method and system for reducing update frequency of image-processing means
US10483495 Active 2023-07-10 US7346221B2 (en) 2001-07-12 2002-06-05 Method and system for producing formatted data related to defects of at least an appliance of a set, in particular, related to blurring
US10483494 Active 2024-04-21 US7792378B2 (en) 2001-07-12 2002-06-05 Method and system for modifying image quality
US10483322 Active 2025-12-24 US7760955B2 (en) 2001-07-12 2002-06-05 Method and system for producing formatted information related to defects of appliances
US12838198 Active US8559743B2 (en) 2001-07-12 2010-07-16 Method and system for modifying image quality
US12838184 Abandoned US20100278415A1 (en) 2001-07-12 2010-07-16 Method and system for producing formatted information related to defects of appliances

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14021235 Active 2022-11-10 US9536284B2 (en) 2001-07-12 2013-09-09 Method and system for modifying image quality of an image

Country Status (9)

Country Link
US (10) US7724977B2 (en)
EP (7) EP1410327B1 (en)
JP (6) JP4295612B2 (en)
KR (4) KR100879832B1 (en)
CN (6) CN1316427C (en)
CA (1) CA2453423C (en)
DE (5) DE60234207D1 (en)
ES (2) ES2311061T3 (en)
WO (6) WO2003007241A1 (en)

Families Citing this family (128)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6950211B2 (en) * 2001-07-05 2005-09-27 Corel Corporation Fine moire correction in images
DE60234207D1 (en) * 2001-07-12 2009-12-10 Do Labs Method and system designed to reduce the update frequency
DE60224035D1 (en) * 2002-08-23 2008-01-24 St Microelectronics Srl A method for noise filtering an image sequence numerical
US8294999B2 (en) 2003-01-16 2012-10-23 DigitalOptics Corporation International Optics for an extended depth of field
US7773316B2 (en) * 2003-01-16 2010-08-10 Tessera International, Inc. Optics for an extended depth of field
US7627193B2 (en) * 2003-01-16 2009-12-01 Tessera International, Inc. Camera with image enhancement functions
US7609425B2 (en) * 2003-01-31 2009-10-27 Canon Kabushiki Kaisha Image data processing apparatus, method, storage medium and program
US8471852B1 (en) 2003-05-30 2013-06-25 Nvidia Corporation Method and system for tessellation of subdivision surfaces
JP4096828B2 (en) * 2003-07-15 2008-06-04 セイコーエプソン株式会社 Image processing apparatus
US7369699B1 (en) * 2003-08-29 2008-05-06 Apple Inc. Methods and apparatuses for restoring color and enhancing electronic images
GB0323680D0 (en) * 2003-10-09 2003-11-12 Ta Vision Lab Ltd Deconvolution of a digital image
US7944467B2 (en) * 2003-12-01 2011-05-17 Omnivision Technologies, Inc. Task-based imaging systems
JP2007513427A (en) 2003-12-01 2007-05-24 シーディーエム オプティックス, インコーポレイテッド System and method for optimizing the design of the optical system and a digital system
US7317843B2 (en) * 2004-04-01 2008-01-08 Microsoft Corporation Luminance correction
US7463296B2 (en) 2004-04-01 2008-12-09 Microsoft Corporation Digital cameras with luminance correction
US8285041B2 (en) * 2004-09-14 2012-10-09 Olympus Corporation Image processing apparatus, image recording apparatus, and image processing method
US7461331B2 (en) * 2004-12-21 2008-12-02 Fotomedia Technologies, Llc Automated construction of print order for images capture during a session
EP1679907A1 (en) * 2005-01-05 2006-07-12 Dialog Semiconductor GmbH Hexagonal color pixel structure with white pixels
FR2881011B1 (en) 2005-01-19 2007-06-29 Dxo Labs Sa Process for the realization of a capture device and / or return of images and apparatus obtained by such process
US7683950B2 (en) * 2005-04-26 2010-03-23 Eastman Kodak Company Method and apparatus for correcting a channel dependent color aberration in a digital image
US20060274209A1 (en) * 2005-06-03 2006-12-07 Coretronic Corporation Method and a control device using the same for controlling a display device
US20070294720A1 (en) * 2005-07-01 2007-12-20 Searete Llc Promotional placement in media works
US20090210946A1 (en) * 2005-07-01 2009-08-20 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Media markup for promotional audio content
US20080180539A1 (en) * 2007-01-31 2008-07-31 Searete Llc, A Limited Liability Corporation Image anonymization
US20070005651A1 (en) 2005-07-01 2007-01-04 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Restoring modified assets
US20080013859A1 (en) * 2005-07-01 2008-01-17 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Implementation of media content alteration
US20070266049A1 (en) * 2005-07-01 2007-11-15 Searete Llc, A Limited Liability Corportion Of The State Of Delaware Implementation of media content alteration
US20100154065A1 (en) * 2005-07-01 2010-06-17 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Media markup for user-activated content alteration
US9215512B2 (en) 2007-04-27 2015-12-15 Invention Science Fund I, Llc Implementation of media content alteration
US20090300480A1 (en) * 2005-07-01 2009-12-03 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Media segment alteration with embedded markup identifier
US9426387B2 (en) 2005-07-01 2016-08-23 Invention Science Fund I, Llc Image anonymization
US9065979B2 (en) * 2005-07-01 2015-06-23 The Invention Science Fund I, Llc Promotional placement in media works
US9092928B2 (en) * 2005-07-01 2015-07-28 The Invention Science Fund I, Llc Implementing group content substitution in media works
US20090151004A1 (en) * 2005-07-01 2009-06-11 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Media markup for visual content alteration
US20080052104A1 (en) * 2005-07-01 2008-02-28 Searete Llc Group content substitution in media works
US20080028422A1 (en) * 2005-07-01 2008-01-31 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Implementation of media content alteration
US20090150199A1 (en) * 2005-07-01 2009-06-11 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Visual substitution options in media works
US20090037243A1 (en) * 2005-07-01 2009-02-05 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Audio substitution options in media works
US9230601B2 (en) 2005-07-01 2016-01-05 Invention Science Fund I, Llc Media markup system for content alteration in derivative works
US20070263865A1 (en) * 2005-07-01 2007-11-15 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Authorization rights for substitute media content
US20080086380A1 (en) * 2005-07-01 2008-04-10 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Alteration of promotional content in media works
US20070276757A1 (en) * 2005-07-01 2007-11-29 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Approval technique for media content alteration
US20090150444A1 (en) * 2005-07-01 2009-06-11 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Media markup for audio content alteration
US20090235364A1 (en) * 2005-07-01 2009-09-17 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Media markup for promotional content alteration
US9583141B2 (en) * 2005-07-01 2017-02-28 Invention Science Fund I, Llc Implementing audio substitution options in media works
US20090204475A1 (en) * 2005-07-01 2009-08-13 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Media markup for promotional visual content
US8910033B2 (en) * 2005-07-01 2014-12-09 The Invention Science Fund I, Llc Implementing group content substitution in media works
US20070005422A1 (en) * 2005-07-01 2007-01-04 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Techniques for image generation
US20080052161A1 (en) * 2005-07-01 2008-02-28 Searete Llc Alteration of promotional content in media works
US20080244755A1 (en) * 2007-03-30 2008-10-02 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Authorization for media content alteration
US20080270161A1 (en) * 2007-04-26 2008-10-30 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Authorization rights for substitute media content
CN102147853B (en) * 2005-09-19 2013-01-09 全视技术有限公司 Image outputting method for generating scene
JP2007096405A (en) * 2005-09-27 2007-04-12 Fujifilm Corp Method, device and program for judging direction of camera shake
US8571346B2 (en) 2005-10-26 2013-10-29 Nvidia Corporation Methods and devices for defective pixel detection
US7750956B2 (en) * 2005-11-09 2010-07-06 Nvidia Corporation Using a graphics processing unit to correct video and audio data
US8588542B1 (en) 2005-12-13 2013-11-19 Nvidia Corporation Configurable and compact pixel processing apparatus
FR2895104A1 (en) * 2005-12-19 2007-06-22 Dxo Labs Sa A method for providing data to a digital processing means
FR2895103B1 (en) * 2005-12-19 2008-02-22 Dxo Labs Sa Method and digital data processing system
FR2895102B1 (en) 2005-12-19 2012-12-07 Dxo Labs Method for processing an object in a platform processor (s) and memory (s) and platform using the method
US8295562B2 (en) * 2006-01-13 2012-10-23 Carl Zeiss Microimaging Ais, Inc. Medical image modification to simulate characteristics
US20070165961A1 (en) * 2006-01-13 2007-07-19 Juwei Lu Method And Apparatus For Reducing Motion Blur In An Image
US8737832B1 (en) * 2006-02-10 2014-05-27 Nvidia Corporation Flicker band automated detection system and method
US8310533B2 (en) * 2006-03-27 2012-11-13 GE Sensing & Inspection Technologies, LP Inspection apparatus for inspecting articles
US20070239417A1 (en) * 2006-03-31 2007-10-11 D-Blur Technologies Ltd. Camera performance simulation
US20070269123A1 (en) * 2006-05-16 2007-11-22 Randall Don Briggs Method and apparatus for performing image enhancement in an image processing pipeline
JP4974586B2 (en) * 2006-05-24 2012-07-11 オリンパス株式会社 Microscope imaging apparatus
JP4839148B2 (en) * 2006-07-12 2011-12-21 株式会社リコー Network device, terminal device, a program and a recording medium
US8594441B1 (en) 2006-09-12 2013-11-26 Nvidia Corporation Compressing image-based data using luminance
DE102006057190A1 (en) * 2006-12-05 2008-06-12 Carl Zeiss Meditec Ag A method for producing high-quality images of the front and / or rear portions of the eye
US8203609B2 (en) * 2007-01-31 2012-06-19 The Invention Science Fund I, Llc Anonymization pursuant to a broadcasted policy
US8723969B2 (en) * 2007-03-20 2014-05-13 Nvidia Corporation Compensating for undesirable camera shakes during video capture
US7936915B2 (en) * 2007-05-29 2011-05-03 Microsoft Corporation Focal length estimation for panoramic stitching
US8634103B2 (en) * 2007-06-12 2014-01-21 Qualcomm Incorporated Print image matching parameter extraction and rendering on display devices
US8724895B2 (en) 2007-07-23 2014-05-13 Nvidia Corporation Techniques for reducing color artifacts in digital images
US8570634B2 (en) 2007-10-11 2013-10-29 Nvidia Corporation Image processing of an incoming light field using a spatial light modulator
US8780128B2 (en) 2007-12-17 2014-07-15 Nvidia Corporation Contiguously packed data
US9177368B2 (en) 2007-12-17 2015-11-03 Nvidia Corporation Image distortion correction
US8698908B2 (en) 2008-02-11 2014-04-15 Nvidia Corporation Efficient method for reducing noise and blur in a composite still image from a rolling shutter camera
US9379156B2 (en) * 2008-04-10 2016-06-28 Nvidia Corporation Per-channel image intensity correction
US8280194B2 (en) * 2008-04-29 2012-10-02 Sony Corporation Reduced hardware implementation for a two-picture depth map algorithm
US8194995B2 (en) * 2008-09-30 2012-06-05 Sony Corporation Fast camera auto-focus
US8553093B2 (en) 2008-09-30 2013-10-08 Sony Corporation Method and apparatus for super-resolution imaging using digital imaging devices
US8373718B2 (en) 2008-12-10 2013-02-12 Nvidia Corporation Method and system for color enhancement with color volume adjustment and variable shift along luminance axis
US8290260B2 (en) * 2008-12-15 2012-10-16 Xerox Corporation Method and system for creating integrated remote custom rendering profile
US20100198876A1 (en) * 2009-02-02 2010-08-05 Honeywell International, Inc. Apparatus and method of embedding meta-data in a captured image
DE102009002393A1 (en) * 2009-04-15 2010-11-04 Arnold & Richter Cine Technik Gmbh & Co. Betriebs Kg Method and device for the processing of recorded images to a digital video camera
US8749662B2 (en) 2009-04-16 2014-06-10 Nvidia Corporation System and method for lens shading image correction
CN101551661B (en) 2009-05-12 2013-04-24 广东工业大学 Control method for multi-robot system
US9519814B2 (en) 2009-06-12 2016-12-13 Hand Held Products, Inc. Portable data terminal
FR2948521B1 (en) 2009-07-21 2012-01-27 Dxo Labs Method for estimating a defect of an image capture system and associated systems
US8698918B2 (en) 2009-10-27 2014-04-15 Nvidia Corporation Automatic white balancing for photography
KR20110065997A (en) * 2009-12-10 2011-06-16 삼성전자주식회사 Image processing apparatus and method of processing image
KR101451136B1 (en) * 2010-03-19 2014-10-16 삼성테크윈 주식회사 Method and Apparatus for correcting Vignetting
US8335390B2 (en) * 2010-03-22 2012-12-18 Sony Corporation Blur function modeling for depth of field rendering
WO2011143223A3 (en) * 2010-05-10 2012-02-02 Board Of Regents, The University Of Texas System Determining quality of an image or a video using a distortion classifier
CN102338972A (en) * 2010-07-21 2012-02-01 华晶科技股份有限公司 Assistant focusing method using multiple face blocks
US20120019709A1 (en) * 2010-07-21 2012-01-26 Altek Corporation Assisting focusing method using multiple face blocks
EP2630417A2 (en) 2010-10-24 2013-08-28 Airlight Energy IP SA Absorber tube for a trough collector
EP2447889A1 (en) * 2010-10-29 2012-05-02 Siemens Aktiengesellschaft Method for modeling a defect management in a manufacturing process and for handling the defect during the production process based on said modeled defect management
CN102625043B (en) 2011-01-25 2014-12-10 佳能株式会社 Image processing apparatus, imaging apparatus, and image processing method
US8842931B2 (en) * 2011-02-18 2014-09-23 Nvidia Corporation System, method, and computer program product for reducing noise in an image using depth-based sweeping over image samples
JP5367749B2 (en) * 2011-03-25 2013-12-11 株式会社東芝 Server apparatus, communication method, and program
US20160321316A1 (en) * 2011-06-03 2016-11-03 Gdial Inc. Systems and methods for atomizing and individuating data as data quanta
US8712181B2 (en) * 2011-06-14 2014-04-29 Apteryx, Inc. Real-time application of filters based on image attributes
EP2552099B1 (en) 2011-07-27 2013-08-28 Axis AB Method and camera for providing an estimation of a mean signal to noise ratio value for an image
WO2013073418A1 (en) * 2011-11-16 2013-05-23 オリンパスメディカルシステムズ株式会社 Image signal processor for endoscope
JP2013123812A (en) * 2011-12-13 2013-06-24 Canon Inc Inspecting device, inspecting method, and computer program
US20130329996A1 (en) * 2012-06-10 2013-12-12 Apple Inc. Method and system for auto-enhancing photographs with tonal response curves
JP5656926B2 (en) 2012-06-22 2015-01-21 キヤノン株式会社 Image processing method, image processing apparatus and an image pickup apparatus
US8976271B2 (en) 2012-07-19 2015-03-10 Canon Kabushiki Kaisha Optical system and image pickup apparatus
JP6284937B2 (en) 2012-07-26 2018-02-28 デピュー シンセス プロダクツ,インコーポレーテッドDePuy Synthes Products, Inc. YCbCr pulsed illumination system in an optical insufficient environmental
KR20150037958A (en) 2012-07-26 2015-04-08 올리브 메디컬 코포레이션 Wide dynamic range using monochromatic sensor
US9798698B2 (en) 2012-08-13 2017-10-24 Nvidia Corporation System and method for multi-color dilu preconditioner
US9508318B2 (en) 2012-09-13 2016-11-29 Nvidia Corporation Dynamic color profile management for electronic devices
US8867817B1 (en) * 2012-10-29 2014-10-21 Amazon Technologies, Inc. Display analysis using scanned images
US9307213B2 (en) 2012-11-05 2016-04-05 Nvidia Corporation Robust selection and weighting for gray patch automatic white balancing
GB201219921D0 (en) * 2012-11-05 2012-12-19 British Broadcasting Corp Method and apparatus for focus detection
US9026553B2 (en) * 2012-11-29 2015-05-05 Unisys Corporation Data expanse viewer for database systems
EP2967300A4 (en) 2013-03-15 2016-11-23 Olive Medical Corp Controlling the integral light energy of a laser pulse
WO2014144947A1 (en) 2013-03-15 2014-09-18 Olive Medical Corporation Super resolution and color motion artifact correction in a pulsed color imaging system
CN105283117A (en) 2013-03-15 2016-01-27 奥利弗医疗公司 Calibration using distal cap
US9826208B2 (en) 2013-06-26 2017-11-21 Nvidia Corporation Method and system for generating weights for use in white balancing an image
US9756222B2 (en) 2013-06-26 2017-09-05 Nvidia Corporation Method and system for performing white balancing operations on captured images
US9167706B2 (en) 2013-08-05 2015-10-20 Steven J. Holmstrom Electronic flight bag retention device
US9396409B2 (en) 2014-09-29 2016-07-19 At&T Intellectual Property I, L.P. Object based image processing
WO2016065632A1 (en) * 2014-10-31 2016-05-06 华为技术有限公司 Image processing method and device
JP2016224701A (en) * 2015-05-29 2016-12-28 キヤノン株式会社 Controller, control method, and program
US9838646B2 (en) * 2015-09-24 2017-12-05 Cisco Technology, Inc. Attenuation of loudspeaker in microphone array

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4827333A (en) * 1983-10-11 1989-05-02 Hitachi Denshi Kabushiki Kaisha Dynamic image correction method of offsets of various image pickup characteristics of a television camera system caused by variations in various image pickup conditions
US5047861A (en) * 1990-07-31 1991-09-10 Eastman Kodak Company Method and apparatus for pixel non-uniformity correction
US5157497A (en) * 1991-02-25 1992-10-20 Matsushita Electric Industrial Co., Ltd. Method and apparatus for detecting and compensating for white shading errors in a digitized video signal
US5323204A (en) * 1992-11-03 1994-06-21 Eastman Kodak Company Automatic optimization of photographic exposure parameters for non-standard display sizes and/or different focal length photographing modes through determination and utilization of extra system speed
US5696850A (en) * 1995-12-21 1997-12-09 Eastman Kodak Company Automatic image sharpening in an electronic imaging system
US5818523A (en) * 1995-02-16 1998-10-06 Samsung Electronics Co., Ltd. Method of correcting white shading for video cameras and apparatus adopting the same
US5905530A (en) * 1992-08-24 1999-05-18 Canon Kabushiki Kaisha Image pickup apparatus
US5990935A (en) * 1997-04-04 1999-11-23 Evans & Sutherland Computer Corporation Method for measuring camera and lens properties for camera tracking
US6115104A (en) * 1997-09-02 2000-09-05 Dainippon Screen Mfg, Co., Ltd. Image processing using parameters related to image input and output devices
US6211911B1 (en) * 1994-10-14 2001-04-03 Olympus Optical Co., Ltd. Image processing apparatus
US20010009590A1 (en) * 1997-03-24 2001-07-26 Holm Jack M. Pictorial digital image processing incorporating image and output device modifications
US6273535B1 (en) * 1997-02-14 2001-08-14 Canon Kabushiki Kaisha Image forming system and images forming apparatus
US6603885B1 (en) * 1998-04-30 2003-08-05 Fuji Photo Film Co., Ltd. Image processing method and apparatus
US8559743B2 (en) * 2001-07-12 2013-10-15 Dxo Labs Method and system for modifying image quality

Family Cites Families (64)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2652695B1 (en) * 1989-10-03 1993-04-16 Thomson Csf Method and image display device has automatic correction of defects by cons-reaction.
FR2661061B1 (en) * 1990-04-11 1992-08-07 Multi Media Tech Method and image area changing device.
US5251271A (en) * 1991-10-21 1993-10-05 R. R. Donnelley & Sons Co. Method for automatic registration of digitized multi-plane images
JPH05176166A (en) 1991-12-25 1993-07-13 Hitachi Ltd Color reproducing method
DE69331719T2 (en) * 1992-06-19 2002-10-24 Agfa Gevaert Nv Method and apparatus for noise suppression
US5461440A (en) * 1993-02-10 1995-10-24 Olympus Optical Co., Ltd. Photographing image correction system
US5353362A (en) * 1993-05-17 1994-10-04 Tucci Robert R Method of generation of two electromagnetic modes using squeezers
JPH0715631A (en) * 1993-06-29 1995-01-17 Nippon Telegr & Teleph Corp <Ntt> Noise eliminating method/device for image signal
US5499057A (en) 1993-08-27 1996-03-12 Sony Corporation Apparatus for producing a noise-reducded image signal from an input image signal
US5485568A (en) 1993-10-08 1996-01-16 Xerox Corporation Structured image (Sl) format for describing complex color raster images
DE69524669D1 (en) 1994-05-26 2002-01-31 Canon Kk Image processing method and apparatus
US6334219B1 (en) * 1994-09-26 2001-12-25 Adc Telecommunications Inc. Channel selection for a hybrid fiber coax network
US5606365A (en) * 1995-03-28 1997-02-25 Eastman Kodak Company Interactive camera for network processing of captured images
US5694484A (en) * 1995-05-15 1997-12-02 Polaroid Corporation System and method for automatically processing image data to provide images of optimal perceptual quality
JPH0998299A (en) 1995-10-02 1997-04-08 Canon Inc Image processing device and method
JP3409541B2 (en) 1995-11-14 2003-05-26 三菱電機株式会社 The color correction method and the color correction apparatus and the color correction application device and a color imaging system
JPH09214807A (en) * 1996-01-31 1997-08-15 Canon Inc Device and method for processing image
JP3950188B2 (en) * 1996-02-27 2007-07-25 株式会社リコー Image distortion correction parameter determining method and an imaging apparatus
JPH1083024A (en) 1996-09-09 1998-03-31 Fuji Photo Film Co Ltd Camera and printer
JP3791635B2 (en) * 1996-10-22 2006-06-28 富士写真フイルム株式会社 Image reproducing method, an image reproducing apparatus, an image processing method and image processing apparatus
US6173087B1 (en) * 1996-11-13 2001-01-09 Sarnoff Corporation Multi-view image registration with application to mosaicing and lens distortion correction
US6100925A (en) * 1996-11-27 2000-08-08 Princeton Video Image, Inc. Image insertion in video streams using a combination of physical sensors and pattern recognition
US6094221A (en) * 1997-01-02 2000-07-25 Andersion; Eric C. System and method for using a scripting language to set digital camera device features
US6222583B1 (en) 1997-03-27 2001-04-24 Nippon Telegraph And Telephone Corporation Device and system for labeling sight images
JP3225882B2 (en) 1997-03-27 2001-11-05 日本電信電話株式会社 Scene labeling system
JPH10319929A (en) 1997-05-19 1998-12-04 Matsushita Electric Ind Co Ltd Display device
JPH11146308A (en) 1997-11-13 1999-05-28 Fuji Photo Film Co Ltd Image information recorder and image print system
US6493028B1 (en) 1997-11-26 2002-12-10 Flashpoint Technology, Inc. Method and system for extending the available image file formats in an image capture device
DE19855885A1 (en) * 1997-12-04 1999-08-05 Fuji Photo Film Co Ltd Image processing method for digital photographic copier providing photographic prints
US6069982A (en) * 1997-12-23 2000-05-30 Polaroid Corporation Estimation of frequency dependence and grey-level dependence of noise in an image
JPH11220687A (en) * 1998-01-30 1999-08-10 Fuji Photo Film Co Ltd Method for processing picture and device therefor
US6381375B1 (en) * 1998-02-20 2002-04-30 Cognex Corporation Methods and apparatus for generating a projection of an image
JP3926918B2 (en) 1998-03-20 2007-06-06 富士通株式会社 Image correction processing apparatus and a program recording medium
JP4338155B2 (en) 1998-06-12 2009-10-07 キヤノン株式会社 Image processing apparatus and method, a computer-readable memory
JP4187830B2 (en) 1998-07-03 2008-11-26 東芝医用システムエンジニアリング株式会社 The medical image synthesis apparatus
EP1097431B1 (en) * 1998-07-15 2003-12-10 Kodak Polychrome Graphics LLC Imaging system and method
JP4095184B2 (en) 1998-10-29 2008-06-04 キヤノン株式会社 Image processing apparatus and method
JP2000165647A (en) * 1998-11-26 2000-06-16 Seiko Epson Corp Image data processing method, image data printer, and recording medium storing image data processing program
JP4154053B2 (en) * 1998-12-25 2008-09-24 キヤノン株式会社 Image recording and reproducing system, image recording apparatus and an image reproducing apparatus
US6538691B1 (en) * 1999-01-21 2003-03-25 Intel Corporation Software correction of image distortion in digital cameras
JP4072302B2 (en) 1999-04-13 2008-04-09 キヤノン株式会社 Data processing method and apparatus and a storage medium
CN2567940Y (en) 1999-06-30 2003-08-20 罗技电子股份有限公司 Camera equipment
US6856427B1 (en) * 1999-05-20 2005-02-15 Eastman Kodak Company System for printing correct exposure in a rendered digital image
US6693668B1 (en) * 1999-06-04 2004-02-17 Canon Kabushiki Kaisha Self-diagnostic image sensor
US6470151B1 (en) * 1999-06-22 2002-10-22 Canon Kabushiki Kaisha Camera, image correcting apparatus, image correcting system, image correcting method, and computer program product providing the image correcting method
US6707950B1 (en) * 1999-06-22 2004-03-16 Eastman Kodak Company Method for modification of non-image data in an image processing chain
JP2001016449A (en) 1999-06-25 2001-01-19 Ricoh Co Ltd Image input device
US6633408B1 (en) * 1999-06-29 2003-10-14 Kodak Polychrome Graphics, Llc Spectral modeling of photographic printing based on dye concentration
JP4822571B2 (en) * 1999-08-03 2011-11-24 キヤノン株式会社 Digital x-ray imaging systems and methods
DE19943183A1 (en) * 1999-09-09 2001-03-15 Heimann Systems Gmbh & Co A method for adjusting a color image, in particular an X-ray image
JP2001094848A (en) 1999-09-20 2001-04-06 Canon Inc Camera with monitor
EP1236018A1 (en) * 1999-11-12 2002-09-04 Go Sensors, L.L.C. Robust landmarks for machine vision and methods for detecting same
US6809837B1 (en) * 1999-11-29 2004-10-26 Xerox Corporation On-line model prediction and calibration system for a dynamically varying color reproduction device
KR100414083B1 (en) * 1999-12-18 2004-01-07 엘지전자 주식회사 Method for compensating image distortion and image displaying apparatus using the same
US6816625B2 (en) * 2000-08-16 2004-11-09 Lewis Jr Clarence A Distortion free image capture system and method
JP3429280B2 (en) * 2000-09-05 2003-07-22 理化学研究所 Method of correcting lens distortion of the image
JP4399133B2 (en) * 2000-09-08 2010-01-13 カシオ計算機株式会社 Shooting condition providing unit, imaging condition setting system, shooting condition providing method
US6956966B2 (en) * 2001-04-03 2005-10-18 Electronics For Imaging, Inc. Method and apparatus for automated image correction for digital image acquisition
CN100346633C (en) * 2001-07-12 2007-10-31 杜莱布斯公司 Method and system for correcting chromatic aberrations of a colour image produced by an optical system
FR2827459B1 (en) * 2001-07-12 2004-10-29 Poseidon Method and system to provide some image processing software formatted information related to characteristics of the image capture devices and / or image reproduction means
US6873727B2 (en) * 2001-07-23 2005-03-29 Hewlett-Packard Development Company, L.P. System for setting image characteristics using embedded camera tag information
FR2895103B1 (en) * 2005-12-19 2008-02-22 Dxo Labs Sa Method and digital data processing system
FR2895104A1 (en) * 2005-12-19 2007-06-22 Dxo Labs Sa A method for providing data to a digital processing means
FR2895102B1 (en) * 2005-12-19 2012-12-07 Dxo Labs Method for processing an object in a platform processor (s) and memory (s) and platform using the method

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4827333A (en) * 1983-10-11 1989-05-02 Hitachi Denshi Kabushiki Kaisha Dynamic image correction method of offsets of various image pickup characteristics of a television camera system caused by variations in various image pickup conditions
US5047861A (en) * 1990-07-31 1991-09-10 Eastman Kodak Company Method and apparatus for pixel non-uniformity correction
US5157497A (en) * 1991-02-25 1992-10-20 Matsushita Electric Industrial Co., Ltd. Method and apparatus for detecting and compensating for white shading errors in a digitized video signal
US5905530A (en) * 1992-08-24 1999-05-18 Canon Kabushiki Kaisha Image pickup apparatus
US5323204A (en) * 1992-11-03 1994-06-21 Eastman Kodak Company Automatic optimization of photographic exposure parameters for non-standard display sizes and/or different focal length photographing modes through determination and utilization of extra system speed
US6211911B1 (en) * 1994-10-14 2001-04-03 Olympus Optical Co., Ltd. Image processing apparatus
US5818523A (en) * 1995-02-16 1998-10-06 Samsung Electronics Co., Ltd. Method of correcting white shading for video cameras and apparatus adopting the same
US5696850A (en) * 1995-12-21 1997-12-09 Eastman Kodak Company Automatic image sharpening in an electronic imaging system
US6273535B1 (en) * 1997-02-14 2001-08-14 Canon Kabushiki Kaisha Image forming system and images forming apparatus
US20010009590A1 (en) * 1997-03-24 2001-07-26 Holm Jack M. Pictorial digital image processing incorporating image and output device modifications
US5990935A (en) * 1997-04-04 1999-11-23 Evans & Sutherland Computer Corporation Method for measuring camera and lens properties for camera tracking
US6115104A (en) * 1997-09-02 2000-09-05 Dainippon Screen Mfg, Co., Ltd. Image processing using parameters related to image input and output devices
US6603885B1 (en) * 1998-04-30 2003-08-05 Fuji Photo Film Co., Ltd. Image processing method and apparatus
US8559743B2 (en) * 2001-07-12 2013-10-15 Dxo Labs Method and system for modifying image quality

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Reg G. Willson, "Modeling and Calibration of Automated Zoom Lenses", SPIE Videometrics III, Vol. 2350, 1994, pages 170 - 186 *

Also Published As

Publication number Publication date Type
WO2003007239A1 (en) 2003-01-23 application
US7724977B2 (en) 2010-05-25 grant
CN1316427C (en) 2007-05-16 grant
CN1527989A (en) 2004-09-08 application
KR100879832B1 (en) 2009-01-22 grant
CA2453423C (en) 2014-10-14 grant
CN1526115A (en) 2004-09-01 application
KR20040043157A (en) 2004-05-22 application
EP1444651B1 (en) 2009-10-28 grant
KR20040043155A (en) 2004-05-22 application
JP2004534341A (en) 2004-11-11 application
CN1316426C (en) 2007-05-16 grant
EP1444651A2 (en) 2004-08-11 application
KR20040044187A (en) 2004-05-27 application
WO2003007240A1 (en) 2003-01-23 application
CN1526117A (en) 2004-09-01 application
EP1410327B1 (en) 2005-11-16 grant
DE60207417T2 (en) 2006-08-10 grant
ES2311061T3 (en) 2009-02-01 grant
CN1554074A (en) 2004-12-08 application
WO2003007242A3 (en) 2003-04-10 application
EP2015247A2 (en) 2009-01-14 application
JP2004537791A (en) 2004-12-16 application
JP4295612B2 (en) 2009-07-15 grant
EP2015247B1 (en) 2016-04-13 grant
CN1305010C (en) 2007-03-14 grant
JP2004535128A (en) 2004-11-18 application
WO2003007241A1 (en) 2003-01-23 application
KR100940147B1 (en) 2010-02-03 grant
EP1415275A1 (en) 2004-05-06 application
US20110085740A1 (en) 2011-04-14 application
JP4367757B2 (en) 2009-11-18 grant
CN100361153C (en) 2008-01-09 grant
WO2003007242A2 (en) 2003-01-23 application
WO2003007243A2 (en) 2003-01-23 application
US20100278415A1 (en) 2010-11-04 application
US20040252906A1 (en) 2004-12-16 application
DE60227374D1 (en) 2008-08-14 grant
KR100940148B1 (en) 2010-02-03 grant
EP1410326B1 (en) 2011-01-26 grant
JP2004534342A (en) 2004-11-11 application
US7792378B2 (en) 2010-09-07 grant
CN1535448A (en) 2004-10-06 application
WO2003007236A2 (en) 2003-01-23 application
DE60207417D1 (en) 2005-12-22 grant
US20050008242A1 (en) 2005-01-13 application
EP1410331A2 (en) 2004-04-21 application
EP2015247A3 (en) 2009-03-18 application
KR20040043154A (en) 2004-05-22 application
JP4295613B2 (en) 2009-07-15 grant
ES2253542T3 (en) 2006-06-01 grant
US20050002586A1 (en) 2005-01-06 application
JP4020262B2 (en) 2007-12-12 grant
DE60234207D1 (en) 2009-12-10 grant
DE60239061D1 (en) 2011-03-10 grant
US9536284B2 (en) 2017-01-03 grant
EP1410326A1 (en) 2004-04-21 application
US20040234152A1 (en) 2004-11-25 application
KR100957878B1 (en) 2010-05-13 grant
CN1526116A (en) 2004-09-01 application
US7346221B2 (en) 2008-03-18 grant
EP1415275B1 (en) 2008-07-02 grant
CN1273931C (en) 2006-09-06 grant
WO2003007236A3 (en) 2003-04-24 application
US20040218803A1 (en) 2004-11-04 application
US7343040B2 (en) 2008-03-11 grant
US20140099042A1 (en) 2014-04-10 application
CA2453423A1 (en) 2003-01-23 application
US8675980B2 (en) 2014-03-18 grant
JP2004534491A (en) 2004-11-11 application
JP2005509333A (en) 2005-04-07 application
EP1410327A2 (en) 2004-04-21 application
EP1410331B1 (en) 2015-08-12 grant
US20040247196A1 (en) 2004-12-09 application
EP1442425A1 (en) 2004-08-04 application
US8559743B2 (en) 2013-10-15 grant
JP4452497B2 (en) 2010-04-21 grant
JP4614657B2 (en) 2011-01-19 grant
CN1305006C (en) 2007-03-14 grant
WO2003007243A3 (en) 2004-01-22 application
US7760955B2 (en) 2010-07-20 grant

Similar Documents

Publication Publication Date Title
Martinez et al. Ten years of art imaging research
US5267030A (en) Method and associated apparatus for forming image data metrics which achieve media compatibility for subsequent imaging application
US7053953B2 (en) Method and camera system for blurring portions of a verification image to show out of focus areas in a captured archival image
US6249315B1 (en) Strategy for pictorial digital image processing
US8406562B2 (en) System and method for automated calibration and correction of display geometry and color
US6097470A (en) Digital photofinishing system including scene balance, contrast normalization, and image sharpening digital image processing
US7289664B2 (en) Method of detecting and correcting the red eye
US6301382B1 (en) Extracting a matte of a foreground object from multiple backgrounds by triangulation
US20040169873A1 (en) Automatic determination of custom parameters based on scanned image data
US5420979A (en) Method and apparatus for using composite transforms to form intermediary image data metrics which achieve device/media compatibility for subsequent imaging applications
US20050007477A1 (en) Correction of optical distortion by image processing
US6995793B1 (en) Video tap for a digital motion camera that simulates the look of post processing
US6975437B2 (en) Method, apparatus and recording medium for color correction
US6853400B1 (en) System and method for correcting aberration of lenses through which images are projected
US6646760B1 (en) Image processing method and apparatus
US6097471A (en) Digital photofinishing system including film under-exposure gamma, scene balance, and image sharpening digital image processing
US6864915B1 (en) Method and apparatus for production of an image captured by an electronic motion camera/sensor that emulates the attributes/exposure content produced by a motion camera film system
US7046400B2 (en) Adjusting the color, brightness, and tone scale of rendered digital images
US20050141002A1 (en) Image-processing method, image-processing apparatus and image-recording apparatus
US20050169519A1 (en) Image processing apparatus, image pickup apparatus, image processing method, image data output method, image processing program and image data ouput program
US20020140825A1 (en) Image processing method and apparatus for carrying out image processing on digital image signals under different image processing conditions
US20050134801A1 (en) Method and system for preserving the creative intent within a motion picture production chain
US20020163657A1 (en) Method of digital processing for digital cinema projection of tone scale and color
US6594388B1 (en) Color image reproduction of scenes with preferential color mapping and scene-dependent tone scaling
US20040227978A1 (en) Image processor