US20130021490A1 - Facial Image Processing in an Image Capture Device - Google Patents

Facial Image Processing in an Image Capture Device Download PDF

Info

Publication number
US20130021490A1
US20130021490A1 US13/281,521 US201113281521A US2013021490A1 US 20130021490 A1 US20130021490 A1 US 20130021490A1 US 201113281521 A US201113281521 A US 201113281521A US 2013021490 A1 US2013021490 A1 US 2013021490A1
Authority
US
United States
Prior art keywords
image
capture device
face
image capture
facial
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/281,521
Inventor
Geraint James
Efrat Swissa
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Avago Technologies International Sales Pte Ltd
Original Assignee
Broadcom Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Broadcom Corp filed Critical Broadcom Corp
Priority to US13/281,521 priority Critical patent/US20130021490A1/en
Assigned to BROADCOM CORPORATION reassignment BROADCOM CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SWISSA, EFRAT, JAMES, GERAINT
Publication of US20130021490A1 publication Critical patent/US20130021490A1/en
Assigned to BANK OF AMERICA, N.A., AS COLLATERAL AGENT reassignment BANK OF AMERICA, N.A., AS COLLATERAL AGENT PATENT SECURITY AGREEMENT Assignors: BROADCOM CORPORATION
Assigned to AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD. reassignment AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BROADCOM CORPORATION
Assigned to BROADCOM CORPORATION reassignment BROADCOM CORPORATION TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS Assignors: BANK OF AMERICA, N.A., AS COLLATERAL AGENT
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/73Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • H04N19/139Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/192Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding the adaptation method, adaptation tool or adaptation type being iterative or recursive
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • H04N19/436Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation using parallelised computational arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/537Motion estimation other than block-based
    • H04N19/54Motion estimation other than block-based using feature points or meshes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/56Motion estimation with initialisation of the vector search, e.g. estimating a good candidate to initiate a search
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • H04N23/681Motion detection
    • H04N23/6811Motion detection based on the image signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • H04N23/682Vibration or motion blur correction
    • H04N23/683Vibration or motion blur correction performed by a processor, e.g. controlling the readout of an image memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20201Motion blur correction

Definitions

  • Image capture devices can apply various image processing techniques. These techniques can be applied globally or, in other words, to an entire image. Images captured by an image capture device can often contain various objects and/or subjects such that application of a single image processing technique to the entirety of the image can result in a less than desirable result.
  • FIGS. 1A-1B are drawings of a mobile device incorporating an image capture device according to various embodiments of the disclosure.
  • FIG. 2 is a drawing of an image capture device that can be incorporated into the mobile device of FIG. 1 according to various embodiments of the disclosure.
  • FIGS. 3-4 are drawings of an example image that can be captured and processed by an image capture device of FIG. 2 according to various embodiments of the disclosure.
  • FIGS. 5-7 are drawings of an example image processing techniques that can be performed on an image captured by an image capture device of FIG. 2 according to various embodiments of the disclosure.
  • FIG. 8 is a flowchart depicting one example of a process that can be executed in an image capture device according to various embodiments of the disclosure.
  • Embodiments of the present disclosure relate to systems and methods that can be executed in an image capture device. More specifically, embodiments of the present disclosure further comprise systems and methods that can implement facial beautification techniques to faces and/or facial regions that can be detected in images and/or video captured by an image capture device.
  • An image capture device can include a camera, video camera, a mobile device with an integrated image capture device, or other devices suitable to capturing imagery and/or video as can be appreciated.
  • an image capture device according to an embodiment of the disclosure can include a device such as a smartphone, tablet computing system, laptop computer, desktop computer, or any other computing device that has the capability to receive and/or capture imagery via image capture hardware.
  • image capture device hardware can include components such as lenses, image sensors (e.g., charge coupled devices, CMOS image sensor, etc.), processor(s), image signal processor(s), a main processor, memory, mass storage, or any other hardware or software components that can facilitate capture of imagery and/or video.
  • an image signal processor can be incorporated as a part of a main processor in an image capture device module that is in turn incorporated into a device having its own processor, memory and other components.
  • An image capture device can provide a user interface via a display that is integrated into the image capture device.
  • the display can be integrated with a mobile device, such as a smartphone and/or tablet computing device, and can include a touchscreen input device (e.g., a capacitive touchscreen, etc.) with which a user may interact with the user interface that is presented thereon.
  • the image capture device hardware can also include one or more buttons, dials, toggles, switches, or other input devices with which the user can interact with software executed in the image capture device.
  • FIGS. 1A-1B show a mobile device 102 that can comprise and/or incorporate an image capture device according to various embodiments of the disclosure.
  • the mobile device 102 may comprise, for example, a processor-based system, such as a computer system.
  • a computer system may be embodied in the form of a desktop computer, a laptop computer, a personal digital assistant, a mobile device (e.g., cellular telephone, smart phone, etc.), tablet computing system, set-top box, music players, or other devices with like capability.
  • the mobile device can include, for example, an image capture device 104 , which can further include a lens system as well as other hardware components that can be integrated with the device to facilitate image capture.
  • the mobile device 102 can also include a display device 141 upon which various content and other user interfaces may be rendered.
  • the mobile device 102 can also include one or more input devices with which a user can interact with a user interface rendered on the display device 141 .
  • the mobile device 102 can include or be in communication with a mouse, touch input device (e.g., capacitive and/or resistive touchscreen incorporated with the display device 141 ), keyboard, or other input devices.
  • the mobile device 102 may be configured to execute various applications, such as a camera application that can interact with an image capture module that includes various hardware and/or software components that facilitate capture and/or storage of images and/or video.
  • the camera application can interact with application programming interfaces (API's) and/or other software libraries and/or drivers that are provided for the purpose interacting with image capture hardware, such as the lens system and other image capture hardware.
  • API's application programming interfaces
  • the camera application can be a special purpose application, a plug-in or executable library, one or more API's, image control algorithms, image capture device firmware, or other software that can facilitate communication with image capture hardware in communication with the mobile device 102 .
  • a camera application according to embodiments of the present disclosure can capture imagery and/or video via the various image capture hardware as well as facilitate storage of the captured imagery and/or video in memory and/or mass storage associated with the mobile device 102 .
  • FIG. 2 illustrates an embodiment of the various image capture components, or one example of an image capture device 104 , that can be incorporated in the mobile device 102 illustrated in FIGS. 1A-1B .
  • an image capture device according to an embodiment of the disclosure more generally comprises an image capture device that can provide images in digital form.
  • the image capture device 104 includes a lens system 200 that conveys images of viewed scenes to an image sensor 202 .
  • the image sensor 202 comprises a charge-coupled device (CCD) or a complementary metal oxide semiconductor (CMOS) sensor that is driven by one or more sensor drivers 204 .
  • the analog image signals captured by the sensor 202 are provided to an analog-to-digital (ND) converter 206 for conversion into binary code that can be processed by a processor 208 .
  • the processor can also execute a facial processing application 151 that can identify one or more faces appearing in an image and perform facial beautification image processing techniques on the face.
  • the facial processing application 151 can take the form of API's, firmware, or other software accessible to the image capture device 104 and/or a mobile device 102 or other system in which the image capture device 104 is integrated.
  • Operation of the sensor driver(s) 204 is controlled through a camera controller 210 that is in bi-directional communication with the processor 208 .
  • the controller 210 can control one or more motors 212 that are used to drive the lens system 200 (e.g., to adjust focus, zoom, and/or aperture settings).
  • the controller 210 can also communicate with a flash system, user input devices (e.g., buttons, dials, toggles, etc.) or other components associated with the image capture device 104 . Operation of the camera controller 210 may be adjusted through manipulation of a user interface.
  • a user interface comprises the various components used to enter selections and commands into the image capture device 104 and therefore can include various buttons as well as a menu system that, for example, is displayed to the user in, for example, a camera application executed on a mobile device 102 and/or on a back panel associated with a standalone digital camera.
  • the digital image signals are processed in accordance with instructions from an image signal processor 218 that can be implemented as a standalone processor within the image capture device as well as being a part of the processor 208 . Processed (e.g., compressed) images may then be stored in storage memory, such as that contained within a removable solid-state memory card (e.g., Flash memory card).
  • the embodiment shown in FIG. 2 further includes a device interface 224 through which the image capture device 104 can communicate with a mobile device or other computing system in which it may be integrated.
  • the device interface 224 can allow the image capture device to communicate with a main processor associated with a mobile device as well as memory, mass storage, or other resources associated with the mobile device.
  • the device interface 224 can communicate with a mobile device in various communications protocols, and this communication can be facilitated, at a software level, by various device drivers, libraries, API's or other software associated with the image capture device 104 that is executed in the mobile device.
  • An image capture device can apply various image processing techniques to various regions of an image that can be associated with a particular region type.
  • the identification and characterization of regions within captured imagery as well as application of various image processing techniques can be accomplished via software executed by the processor 208 , the ISP 218 as well as a processor associated with a device in communication with the image capture device 104 . It should be appreciated that the specific implementation and/or embodiments disclosed herein are merely examples.
  • the facial processing application 151 can perform various image processing techniques on faces that are identified in an image captured by the image capture device 104 .
  • a user can specify one or more image processing techniques that can be applied to the faces identified in the image.
  • These image processing techniques can include scar removal algorithms, blemish removal algorithms, color correction techniques, or any other image processing technique that can be employed to modify the appearance of a face in an image captured by the image capture device 104 .
  • the facial processing application 151 can analyze an image captured by the image capture device and determine whether there are objects that correspond to facial features.
  • An image processing technique can also include an image processing operation that alters a dimension of the face and/or facial features within the face.
  • such a technique can include an alteration of the shape and/or size of a nose, forehead, chin, lip, cheeks, or other portion of the face.
  • the technique can include modifying the dimensions of the face, or altering the appearance of facial bone structures.
  • Another example of an image processing technique can include a technique that applies, enhances, or alters color of a region of the face and/or skin.
  • the facial processing application 151 can enhance and/or modify colorization of a subject's cheeks, whiten teeth, or other portions of the face within the image.
  • the image processing technique can include removal of grey hairs that may appear in an image by adjusting a color of a hair region. Therefore, an image processing technique can also include, but is not limited to, the application of one or more signal processing techniques, filters, or any other process that receives as an input image data associated with a region and outputs image data that is altered or modified in some form.
  • the facial processing application 151 can detect whether an image captured by the image capture device 104 contains a face.
  • the facial processing application 151 can employ object detection algorithms, edge detection algorithms, or other image recognition techniques on the captured image to identify whether one or more objects within an image corresponds a head and/or face.
  • the facial processing application 151 can then employ one or more facial recognition algorithms to determine whether the face(s) detected in the image correspond to a facial profile that can specify one or more image processing techniques that can be applied to a region of the image that corresponds to the location of the face within the image.
  • the facial processing application 151 can calculate a facial signature, which can comprise a digital representation of a face depicted within an image that allows the face to be uniquely identified.
  • a facial profile can contain the facial signature as well as references to one or more image processing techniques that the facial processing application 151 can apply to the face.
  • the facial processing application 151 can generate a resultant image that includes the processed face.
  • the facial processing application 151 can generate a user interface that allows the user to specify image processing techniques to apply to a given face depicted within an image.
  • the facial processing application 151 can store a reference to a facial signature as well as the image processing techniques selected by the user in a memory accessible to the image capture device.
  • the facial profile can also store a location within a facial at which a specified image processing technique should be applied.
  • facial profiles can be stored in memory of a mobile device 102 in which the image capture device 104 is integrated. Therefore, the facial processing application 151 can apply the selected image processing techniques to the face in subsequent images that contain the face when it can be identified by its facial signature. Additional examples and variations are provided in the discussion that accompanies the following drawings.
  • FIG. 3 illustrates an example image that can be captured by the image capture device 104 ( FIG. 2 ) according to various embodiments according to the disclosure.
  • the image capture device 104 is incorporated into a mobile device 102 , which can execute a camera application that renders a user interface for display on a display device associated with the mobile device 102 . It should be appreciated that this is only one non-limiting illustrative implementation.
  • FIG. 3 illustrates an example of an image 300 that can be captured by the image capture device.
  • the image 300 can be captured via a camera application executed on a mobile device where the camera application is configured to communicate with API's associated with the image capture device for the purposes of initiating capture of imagery, display of imagery on a display of the mobile device as well as storage of captured imagery in the form of still images and/or video in memory or mass storage associated with the mobile device.
  • FIG. 4 continues the example of FIG. 3 by illustrating an example of a face that can be detected by the facial processing application 151 .
  • the facial processing application 151 can be invoked when an image is captured by the image capture device 104 . Accordingly, the facial processing application 151 can calculate a facial signature to determine whether a facial profile associated with the face is stored in a memory accessible to the facial processing application 151 .
  • a facial profile can specify one or more image processing techniques that can be applied to at least a portion of the region of the image.
  • a facial profile can represent a preconfigured set of one of more image processing techniques that can be applied to the face 302 when the image capture device 104 captures an image that contains the face 302 having a facial signature that matches a signature associated with a facial profile accessible to the facial processing application 151 .
  • the facial processing application 151 can also process frames of a video in which a face corresponding to a facial signature associated with a facial profile by applying the image processing techniques that are specified by the profile. In this way, the facial processing application 151 can apply the image processing techniques specified by a facial profile to a face that appears in a video when the face can be identified as matching a facial signature associated with the profile.
  • FIGS. 5-6 illustrate an example of a region 506 in the image 300 in which an image processing technique can be applied.
  • the facial processing application 151 detects that the face corresponds to a facial profile by calculating a facial signature and determining that the signature matches a profile accessible to the facial processing application 151 , it can apply the specified image processing techniques.
  • the facial profile can specify a blemish removal algorithm that can be applied to the region.
  • the facial processing application 151 can apply the same image processing techniques that are specified by the facial profile.
  • FIG. 7 illustrates how additional image processing techniques can be specified by the facial profile, such as those that can alter the dimensions of a face and/or regions of the face.
  • FIG. 8 shown is a flowchart that provides one example of the operation of a portion of a facial processing application 151 executed by an image capture device 104 , a mobile device 102 or any other device in which an image capture device 104 is integrated according to various embodiments of the disclosure. It is understood that the flowchart of FIG. 8 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the portion of logic employed by the image capture device as described herein. As an alternative, the flowchart of FIG. 8 may be viewed as depicting an example of steps of a method implemented in a computing device, processor, or other circuits according to one or more embodiments.
  • image capture can be initiated in the image capture device so that one or more images are captured by the lens system, image sensor, and other image capture device hardware as discussed above.
  • the facial processing application 151 can determine whether a face is present within the captured image. If, in box 803 , the facial processing application 151 determines that a face is present in the image, a facial signature associated with the face can be calculated in box 805 . In box 807 , the facial processing application 151 can determine whether the face corresponds to a facial profile accessible to the facial processing application 151 . If so, then in box 815 , the facial processing application 151 applies the image processing techniques that are referenced by the corresponding facial profile. If not, then in box 809 , the facial processing application can facilitate generating a user interface that prompts a user to create a facial profile corresponding to the face detected in the captured image.
  • the facial processing application 151 can store the profile in a memory accessible to the image capture device 104 . Subsequently, the facial processing application 151 can apply the image processing techniques specified by the facial profile. In some embodiments, the facial processing application 151 can prompt a user prior to applying the image processing techniques specified by the facial profile. Additionally, in some embodiments, the facial processing application 151 can apply the image processing techniques and modify the image data captured by the image capture device 104 . In other embodiments, the facial processing application 151 can generate a copy of the image and apply the image processing techniques to the copy, retaining the originally captured image data. In yet another embodiment, the image processing application 151 can apply the image processing techniques by including a reference to the image processing technique and the location within the image to which it is applied in meta data associated and stored with the image data.
  • Embodiments of the present disclosure can be implemented in various devices, for example, having a processor, memory as well as image capture hardware that can be coupled to a local interface.
  • the logic described herein can be executable by one or more processors integrated with a device.
  • an application executed in a computing device such as a mobile device, can invoke API's that provide the logic described herein as well as facilitate interaction with image capture hardware.
  • any component discussed herein is implemented in the form of software, any one of a number of programming languages may be employed such as, for example, processor specific assembler languages, C, C++, C#, Objective C, Java, Javascript, Perl, PHP, Visual Basic, Python, Ruby, Delphi, Flash, or other programming languages.
  • executable means a program file that is in a form that can ultimately be run by a processor.
  • executable programs may be, for example, a compiled program that can be translated into machine code in a format that can be loaded into a random access portion of memory and run by a processor, source code that may be expressed in proper format such as object code that is capable of being loaded into a random access portion of the memory and executed by the processor, or source code that may be interpreted by another executable program to generate instructions in a random access portion of the memory to be executed by the processor, etc.
  • An executable program may be stored in any portion or component of the memory including, for example, random access memory (RAM), read-only memory (ROM), hard drive, solid-state drive, USB flash drive, memory card, optical disc such as compact disc (CD) or digital versatile disc (DVD), floppy disk, magnetic tape, or other memory components.
  • RAM random access memory
  • ROM read-only memory
  • hard drive solid-state drive
  • USB flash drive USB flash drive
  • memory card such as compact disc (CD) or digital versatile disc (DVD), floppy disk, magnetic tape, or other memory components.
  • CD compact disc
  • DVD digital versatile disc
  • each block may represent a module, segment, or portion of code that comprises program instructions to implement the specified logical function(s).
  • the program instructions may be embodied in the form of source code that comprises human-readable statements written in a programming language or machine code that comprises numerical instructions recognizable by a suitable execution system such as a processor in a computer system or other system.
  • the machine code may be converted from the source code, etc.
  • each block may represent a circuit or a number of interconnected circuits to implement the specified logical function(s).
  • FIG. 8 shows a specific order of execution, it is understood that the order of execution may differ from that which is depicted. For example, the order of execution of two or more blocks may be scrambled relative to the order shown. Also, two or more blocks shown in succession in FIG. 8 may be executed concurrently or with partial concurrence. Further, in some embodiments, one or more of the blocks shown in FIG. 8 may be skipped or omitted. In addition, any number of counters, state variables, warning semaphores, or messages might be added to the logical flow described herein, for purposes of enhanced utility, accounting, performance measurement, or providing troubleshooting aids, etc. It is understood that all such variations are within the scope of the present disclosure.
  • any logic or application described herein that comprises software or code can be embodied in any non-transitory computer-readable medium for use by or in connection with an instruction execution system such as, for example, a processor in a computer device or other system.
  • the logic may comprise, for example, statements including instructions and declarations that can be fetched from the computer-readable medium and executed by the instruction execution system.
  • a “computer-readable medium” can be any medium that can contain, store, or maintain the logic or application described herein for use by or in connection with the instruction execution system.
  • the computer-readable medium can comprise any one of many physical media such as, for example, magnetic, optical, or semiconductor media.
  • a suitable computer-readable medium would include, but are not limited to, magnetic tapes, magnetic floppy diskettes, magnetic hard drives, memory cards, solid-state drives, USB flash drives, or optical discs.
  • the computer-readable medium may be a random access memory (RAM) including, for example, static random access memory (SRAM) and dynamic random access memory (DRAM), or magnetic random access memory (MRAM).
  • the computer-readable medium may be a read-only memory (ROM), a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or other type of memory device.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Studio Devices (AREA)
  • Image Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Disclosed are various embodiments of applying image processing techniques in an image capture device to faces within an image. Faces and/or facial features can be detected within an image. A facial signature can be calculated that corresponds to the face within the image. A facial profile corresponding to the facial signature can reference one or more image processing techniques. The image processing techniques can be applied to the region of the image in which the face is located.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims priority to co-pending U.S. provisional application entitled, “Image Capture Device Systems and Methods,” having Ser. No. 61/509,747, filed Jul. 20, 2011, which is entirely incorporated herein by reference.
  • BACKGROUND
  • Image capture devices (e.g., still cameras, video cameras, etc.) can apply various image processing techniques. These techniques can be applied globally or, in other words, to an entire image. Images captured by an image capture device can often contain various objects and/or subjects such that application of a single image processing technique to the entirety of the image can result in a less than desirable result.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Many aspects of the invention can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present invention. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views.
  • FIGS. 1A-1B are drawings of a mobile device incorporating an image capture device according to various embodiments of the disclosure.
  • FIG. 2 is a drawing of an image capture device that can be incorporated into the mobile device of FIG. 1 according to various embodiments of the disclosure.
  • FIGS. 3-4 are drawings of an example image that can be captured and processed by an image capture device of FIG. 2 according to various embodiments of the disclosure.
  • FIGS. 5-7 are drawings of an example image processing techniques that can be performed on an image captured by an image capture device of FIG. 2 according to various embodiments of the disclosure.
  • FIG. 8 is a flowchart depicting one example of a process that can be executed in an image capture device according to various embodiments of the disclosure.
  • DETAILED DESCRIPTION
  • Embodiments of the present disclosure relate to systems and methods that can be executed in an image capture device. More specifically, embodiments of the present disclosure further comprise systems and methods that can implement facial beautification techniques to faces and/or facial regions that can be detected in images and/or video captured by an image capture device. An image capture device can include a camera, video camera, a mobile device with an integrated image capture device, or other devices suitable to capturing imagery and/or video as can be appreciated. In some embodiments, an image capture device according to an embodiment of the disclosure can include a device such as a smartphone, tablet computing system, laptop computer, desktop computer, or any other computing device that has the capability to receive and/or capture imagery via image capture hardware.
  • Accordingly, image capture device hardware can include components such as lenses, image sensors (e.g., charge coupled devices, CMOS image sensor, etc.), processor(s), image signal processor(s), a main processor, memory, mass storage, or any other hardware or software components that can facilitate capture of imagery and/or video. In some embodiments, an image signal processor can be incorporated as a part of a main processor in an image capture device module that is in turn incorporated into a device having its own processor, memory and other components.
  • An image capture device according to an embodiment of the disclosure can provide a user interface via a display that is integrated into the image capture device. The display can be integrated with a mobile device, such as a smartphone and/or tablet computing device, and can include a touchscreen input device (e.g., a capacitive touchscreen, etc.) with which a user may interact with the user interface that is presented thereon. The image capture device hardware can also include one or more buttons, dials, toggles, switches, or other input devices with which the user can interact with software executed in the image capture device.
  • Referring now to the drawings, FIGS. 1A-1B show a mobile device 102 that can comprise and/or incorporate an image capture device according to various embodiments of the disclosure. The mobile device 102 may comprise, for example, a processor-based system, such as a computer system. Such a computer system may be embodied in the form of a desktop computer, a laptop computer, a personal digital assistant, a mobile device (e.g., cellular telephone, smart phone, etc.), tablet computing system, set-top box, music players, or other devices with like capability. The mobile device can include, for example, an image capture device 104, which can further include a lens system as well as other hardware components that can be integrated with the device to facilitate image capture. The mobile device 102 can also include a display device 141 upon which various content and other user interfaces may be rendered. The mobile device 102 can also include one or more input devices with which a user can interact with a user interface rendered on the display device 141. For example, the mobile device 102 can include or be in communication with a mouse, touch input device (e.g., capacitive and/or resistive touchscreen incorporated with the display device 141), keyboard, or other input devices.
  • The mobile device 102 may be configured to execute various applications, such as a camera application that can interact with an image capture module that includes various hardware and/or software components that facilitate capture and/or storage of images and/or video. In one embodiment, the camera application can interact with application programming interfaces (API's) and/or other software libraries and/or drivers that are provided for the purpose interacting with image capture hardware, such as the lens system and other image capture hardware. The camera application can be a special purpose application, a plug-in or executable library, one or more API's, image control algorithms, image capture device firmware, or other software that can facilitate communication with image capture hardware in communication with the mobile device 102. Accordingly, a camera application according to embodiments of the present disclosure can capture imagery and/or video via the various image capture hardware as well as facilitate storage of the captured imagery and/or video in memory and/or mass storage associated with the mobile device 102.
  • FIG. 2 illustrates an embodiment of the various image capture components, or one example of an image capture device 104, that can be incorporated in the mobile device 102 illustrated in FIGS. 1A-1B. Although one implementation is shown in FIG. 2 and described herein, an image capture device according to an embodiment of the disclosure more generally comprises an image capture device that can provide images in digital form.
  • The image capture device 104 includes a lens system 200 that conveys images of viewed scenes to an image sensor 202. By way of example, the image sensor 202 comprises a charge-coupled device (CCD) or a complementary metal oxide semiconductor (CMOS) sensor that is driven by one or more sensor drivers 204. The analog image signals captured by the sensor 202 are provided to an analog-to-digital (ND) converter 206 for conversion into binary code that can be processed by a processor 208. The processor can also execute a facial processing application 151 that can identify one or more faces appearing in an image and perform facial beautification image processing techniques on the face. In some embodiments, the facial processing application 151 can take the form of API's, firmware, or other software accessible to the image capture device 104 and/or a mobile device 102 or other system in which the image capture device 104 is integrated.
  • Operation of the sensor driver(s) 204 is controlled through a camera controller 210 that is in bi-directional communication with the processor 208. In some embodiments, the controller 210 can control one or more motors 212 that are used to drive the lens system 200 (e.g., to adjust focus, zoom, and/or aperture settings). The controller 210 can also communicate with a flash system, user input devices (e.g., buttons, dials, toggles, etc.) or other components associated with the image capture device 104. Operation of the camera controller 210 may be adjusted through manipulation of a user interface. A user interface comprises the various components used to enter selections and commands into the image capture device 104 and therefore can include various buttons as well as a menu system that, for example, is displayed to the user in, for example, a camera application executed on a mobile device 102 and/or on a back panel associated with a standalone digital camera.
  • The digital image signals are processed in accordance with instructions from an image signal processor 218 that can be implemented as a standalone processor within the image capture device as well as being a part of the processor 208. Processed (e.g., compressed) images may then be stored in storage memory, such as that contained within a removable solid-state memory card (e.g., Flash memory card). The embodiment shown in FIG. 2 further includes a device interface 224 through which the image capture device 104 can communicate with a mobile device or other computing system in which it may be integrated. For example, the device interface 224 can allow the image capture device to communicate with a main processor associated with a mobile device as well as memory, mass storage, or other resources associated with the mobile device. The device interface 224 can communicate with a mobile device in various communications protocols, and this communication can be facilitated, at a software level, by various device drivers, libraries, API's or other software associated with the image capture device 104 that is executed in the mobile device.
  • An image capture device according to an embodiment of the disclosure can apply various image processing techniques to various regions of an image that can be associated with a particular region type. The identification and characterization of regions within captured imagery as well as application of various image processing techniques can be accomplished via software executed by the processor 208, the ISP 218 as well as a processor associated with a device in communication with the image capture device 104. It should be appreciated that the specific implementation and/or embodiments disclosed herein are merely examples.
  • As noted above, the facial processing application 151 can perform various image processing techniques on faces that are identified in an image captured by the image capture device 104. A user can specify one or more image processing techniques that can be applied to the faces identified in the image. These image processing techniques can include scar removal algorithms, blemish removal algorithms, color correction techniques, or any other image processing technique that can be employed to modify the appearance of a face in an image captured by the image capture device 104. In one embodiment, the facial processing application 151 can analyze an image captured by the image capture device and determine whether there are objects that correspond to facial features. An image processing technique can also include an image processing operation that alters a dimension of the face and/or facial features within the face. For example, such a technique can include an alteration of the shape and/or size of a nose, forehead, chin, lip, cheeks, or other portion of the face. For example, the technique can include modifying the dimensions of the face, or altering the appearance of facial bone structures. Another example of an image processing technique can include a technique that applies, enhances, or alters color of a region of the face and/or skin. For example, the facial processing application 151 can enhance and/or modify colorization of a subject's cheeks, whiten teeth, or other portions of the face within the image. As yet another example, the image processing technique can include removal of grey hairs that may appear in an image by adjusting a color of a hair region. Therefore, an image processing technique can also include, but is not limited to, the application of one or more signal processing techniques, filters, or any other process that receives as an input image data associated with a region and outputs image data that is altered or modified in some form.
  • Accordingly, the facial processing application 151 can detect whether an image captured by the image capture device 104 contains a face. In some embodiments, the facial processing application 151 can employ object detection algorithms, edge detection algorithms, or other image recognition techniques on the captured image to identify whether one or more objects within an image corresponds a head and/or face. The facial processing application 151 can then employ one or more facial recognition algorithms to determine whether the face(s) detected in the image correspond to a facial profile that can specify one or more image processing techniques that can be applied to a region of the image that corresponds to the location of the face within the image. In some embodiments, the facial processing application 151 can calculate a facial signature, which can comprise a digital representation of a face depicted within an image that allows the face to be uniquely identified.
  • Accordingly, a facial profile can contain the facial signature as well as references to one or more image processing techniques that the facial processing application 151 can apply to the face. In this way, the facial processing application 151 can generate a resultant image that includes the processed face. To facilitate the above functionality, the facial processing application 151 can generate a user interface that allows the user to specify image processing techniques to apply to a given face depicted within an image. The facial processing application 151 can store a reference to a facial signature as well as the image processing techniques selected by the user in a memory accessible to the image capture device. The facial profile can also store a location within a facial at which a specified image processing technique should be applied. In some embodiments, facial profiles can be stored in memory of a mobile device 102 in which the image capture device 104 is integrated. Therefore, the facial processing application 151 can apply the selected image processing techniques to the face in subsequent images that contain the face when it can be identified by its facial signature. Additional examples and variations are provided in the discussion that accompanies the following drawings.
  • Accordingly, reference is now made to FIG. 3, which illustrates an example image that can be captured by the image capture device 104 (FIG. 2) according to various embodiments according to the disclosure. In the depicted non-limiting examples of FIGS. 3-4, the image capture device 104 is incorporated into a mobile device 102, which can execute a camera application that renders a user interface for display on a display device associated with the mobile device 102. It should be appreciated that this is only one non-limiting illustrative implementation.
  • Therefore, FIG. 3 illustrates an example of an image 300 that can be captured by the image capture device. As one example, the image 300 can be captured via a camera application executed on a mobile device where the camera application is configured to communicate with API's associated with the image capture device for the purposes of initiating capture of imagery, display of imagery on a display of the mobile device as well as storage of captured imagery in the form of still images and/or video in memory or mass storage associated with the mobile device.
  • FIG. 4 continues the example of FIG. 3 by illustrating an example of a face that can be detected by the facial processing application 151. In one embodiment, the facial processing application 151 can be invoked when an image is captured by the image capture device 104. Accordingly, the facial processing application 151 can calculate a facial signature to determine whether a facial profile associated with the face is stored in a memory accessible to the facial processing application 151. As noted above, a facial profile can specify one or more image processing techniques that can be applied to at least a portion of the region of the image. In this sense, a facial profile can represent a preconfigured set of one of more image processing techniques that can be applied to the face 302 when the image capture device 104 captures an image that contains the face 302 having a facial signature that matches a signature associated with a facial profile accessible to the facial processing application 151. Additionally, the facial processing application 151 can also process frames of a video in which a face corresponding to a facial signature associated with a facial profile by applying the image processing techniques that are specified by the profile. In this way, the facial processing application 151 can apply the image processing techniques specified by a facial profile to a face that appears in a video when the face can be identified as matching a facial signature associated with the profile.
  • Continuing the example of FIG. 4, reference is now made to FIGS. 5-6, which illustrate an example of a region 506 in the image 300 in which an image processing technique can be applied. In the depicted example, when the facial processing application 151 detects that the face corresponds to a facial profile by calculating a facial signature and determining that the signature matches a profile accessible to the facial processing application 151, it can apply the specified image processing techniques. In FIGS. 5-6, the facial profile can specify a blemish removal algorithm that can be applied to the region. Additionally, in subsequent images captured by the image capture device 104 that contain a face with a facial signature that corresponds to the facial profile, the facial processing application 151 can apply the same image processing techniques that are specified by the facial profile. FIG. 7 illustrates how additional image processing techniques can be specified by the facial profile, such as those that can alter the dimensions of a face and/or regions of the face.
  • Referring next to FIG. 8, shown is a flowchart that provides one example of the operation of a portion of a facial processing application 151 executed by an image capture device 104, a mobile device 102 or any other device in which an image capture device 104 is integrated according to various embodiments of the disclosure. It is understood that the flowchart of FIG. 8 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the portion of logic employed by the image capture device as described herein. As an alternative, the flowchart of FIG. 8 may be viewed as depicting an example of steps of a method implemented in a computing device, processor, or other circuits according to one or more embodiments.
  • First, in box 801, image capture can be initiated in the image capture device so that one or more images are captured by the lens system, image sensor, and other image capture device hardware as discussed above. In box 803, the facial processing application 151 can determine whether a face is present within the captured image. If, in box 803, the facial processing application 151 determines that a face is present in the image, a facial signature associated with the face can be calculated in box 805. In box 807, the facial processing application 151 can determine whether the face corresponds to a facial profile accessible to the facial processing application 151. If so, then in box 815, the facial processing application 151 applies the image processing techniques that are referenced by the corresponding facial profile. If not, then in box 809, the facial processing application can facilitate generating a user interface that prompts a user to create a facial profile corresponding to the face detected in the captured image.
  • If, in box 811, the user opts to create such a profile, then in box 813 the facial processing application 151 can store the profile in a memory accessible to the image capture device 104. Subsequently, the facial processing application 151 can apply the image processing techniques specified by the facial profile. In some embodiments, the facial processing application 151 can prompt a user prior to applying the image processing techniques specified by the facial profile. Additionally, in some embodiments, the facial processing application 151 can apply the image processing techniques and modify the image data captured by the image capture device 104. In other embodiments, the facial processing application 151 can generate a copy of the image and apply the image processing techniques to the copy, retaining the originally captured image data. In yet another embodiment, the image processing application 151 can apply the image processing techniques by including a reference to the image processing technique and the location within the image to which it is applied in meta data associated and stored with the image data.
  • Embodiments of the present disclosure can be implemented in various devices, for example, having a processor, memory as well as image capture hardware that can be coupled to a local interface. The logic described herein can be executable by one or more processors integrated with a device. In one embodiment, an application executed in a computing device, such as a mobile device, can invoke API's that provide the logic described herein as well as facilitate interaction with image capture hardware. Where any component discussed herein is implemented in the form of software, any one of a number of programming languages may be employed such as, for example, processor specific assembler languages, C, C++, C#, Objective C, Java, Javascript, Perl, PHP, Visual Basic, Python, Ruby, Delphi, Flash, or other programming languages.
  • As such, these software components can be executable by one or more processors in various devices. In this respect, the term “executable” means a program file that is in a form that can ultimately be run by a processor. Examples of executable programs may be, for example, a compiled program that can be translated into machine code in a format that can be loaded into a random access portion of memory and run by a processor, source code that may be expressed in proper format such as object code that is capable of being loaded into a random access portion of the memory and executed by the processor, or source code that may be interpreted by another executable program to generate instructions in a random access portion of the memory to be executed by the processor, etc. An executable program may be stored in any portion or component of the memory including, for example, random access memory (RAM), read-only memory (ROM), hard drive, solid-state drive, USB flash drive, memory card, optical disc such as compact disc (CD) or digital versatile disc (DVD), floppy disk, magnetic tape, or other memory components.
  • Although various logic described herein may be embodied in software or code executed by general purpose hardware as discussed above, as an alternative the same may also be embodied in dedicated hardware or a combination of software/general purpose hardware and dedicated hardware. If embodied in dedicated hardware, each can be implemented as a circuit or state machine that employs any one of or a combination of a number of technologies. These technologies may include, but are not limited to, discrete logic circuits having logic gates for implementing various logic functions upon an application of one or more data signals, application specific integrated circuits having appropriate logic gates, or other components, etc. Such technologies are generally well known by those skilled in the art and, consequently, are not described in detail herein.
  • The flowchart of FIG. 8 shows the functionality and operation of an implementation of portions of an image capture device according to embodiments of the disclosure. If embodied in software, each block may represent a module, segment, or portion of code that comprises program instructions to implement the specified logical function(s). The program instructions may be embodied in the form of source code that comprises human-readable statements written in a programming language or machine code that comprises numerical instructions recognizable by a suitable execution system such as a processor in a computer system or other system. The machine code may be converted from the source code, etc. If embodied in hardware, each block may represent a circuit or a number of interconnected circuits to implement the specified logical function(s).
  • Although the flowchart of FIG. 8 shows a specific order of execution, it is understood that the order of execution may differ from that which is depicted. For example, the order of execution of two or more blocks may be scrambled relative to the order shown. Also, two or more blocks shown in succession in FIG. 8 may be executed concurrently or with partial concurrence. Further, in some embodiments, one or more of the blocks shown in FIG. 8 may be skipped or omitted. In addition, any number of counters, state variables, warning semaphores, or messages might be added to the logical flow described herein, for purposes of enhanced utility, accounting, performance measurement, or providing troubleshooting aids, etc. It is understood that all such variations are within the scope of the present disclosure.
  • Also, any logic or application described herein that comprises software or code can be embodied in any non-transitory computer-readable medium for use by or in connection with an instruction execution system such as, for example, a processor in a computer device or other system. In this sense, the logic may comprise, for example, statements including instructions and declarations that can be fetched from the computer-readable medium and executed by the instruction execution system. In the context of the present disclosure, a “computer-readable medium” can be any medium that can contain, store, or maintain the logic or application described herein for use by or in connection with the instruction execution system. The computer-readable medium can comprise any one of many physical media such as, for example, magnetic, optical, or semiconductor media. More specific examples of a suitable computer-readable medium would include, but are not limited to, magnetic tapes, magnetic floppy diskettes, magnetic hard drives, memory cards, solid-state drives, USB flash drives, or optical discs. Also, the computer-readable medium may be a random access memory (RAM) including, for example, static random access memory (SRAM) and dynamic random access memory (DRAM), or magnetic random access memory (MRAM). In addition, the computer-readable medium may be a read-only memory (ROM), a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or other type of memory device.
  • It should be emphasized that the above-described embodiments of the present disclosure are merely possible examples of implementations set forth for a clear understanding of the principles of the disclosure. Many variations and modifications may be made to the above-described embodiment(s) without departing substantially from the spirit and principles of the disclosure. All such modifications and variations are intended to be included herein within the scope of this disclosure and protected by the following claims.

Claims (20)

1. An image capture device, comprising:
at least one image sensor; and
an application executed in the image capture device, the application comprising:
logic that initiates capture of at least one image via the at least one image sensor associated with the image capture device;
logic that detects a face within the image;
logic that calculates a facial signature associated with the face;
logic that identifies a facial profile accessible to the image capture device corresponding to the facial signature, the facial profile specifying at least one image processing technique; and
logic that applies the at least one image processing technique to at least a portion of the face within the image.
2. The image capture device of claim 1, wherein the image capture device is configured to capture a plurality of images associated with a plurality of frames of a video, and the application further comprises:
logic that tracks the face in subsequent frames of the video; and
logic that applies the image processing technique associated with the face in the subsequent frames.
3. The image capture device of claim 1, wherein the at least one image processing technique further comprises at least one of: a wrinkle reduction algorithm, a blemish removal algorithm, whitening of teeth appearing in the image, adjusting colorization of hair appearing in the image, altering dimensions of the face, and a skin tone correction algorithm.
4. The image capture device of claim 1, wherein the at least one image processing technique further comprises adjusting at least one dimension associated with the face.
5. The image capture device of claim 4, wherein the at least one dimension is associated with at least one sub-region of the face, the sub-region being one of: a nose, a forehead, a cheek, a lip, and a chin.
6. The image capture device of claim 1, wherein the at least one image processing technique further comprises increasing at least one color level associated with a portion of the face.
7. The image capture device of claim 6, wherein the application further comprises:
logic that identifies at least one cheek region associated with the face; and
the image processing technique further comprises increasing the at least one color level in the at least one cheek region.
8. The image capture device of claim 1, wherein the application further comprises:
logic that generates a user interface allowing a user to specify the at least one image processing technique associated with the face;
logic that associates the facial signature and a reference corresponding to the at least one image processing technique with the facial profile; and
logic that stores the facial profile in a memory accessible to the image capture device.
9. The image capture device of claim 8, wherein the logic that stores the facial profile further comprises logic that stores the facial profile in a mobile device memory associated with a mobile device in which the image capture device is integrated.
10. The image capture device of claim 1, wherein the logic that applies the at least one image processing technique to at least a portion of the face within the image further comprises logic that stores a reference associated with the at least one image processing technique in meta data associated with the image captured by the image capture device.
11. A method executed in an image capture device, comprising the steps of:
initiating capture of at least one image via at least one image sensor associated with the image capture device;
detecting an object corresponding to a face within the image captured by the image capture device;
calculating a facial signature associated with the face;
identifying a facial profile accessible to the image capture device corresponding to the facial signature, the facial profile specifying at least one image processing technique; and
applying the at least one image processing technique to at least a portion of the face within the image.
12. The method of claim 11, wherein the image capture device is configured to capture a plurality of images associated with a plurality of frames of a video, and the method further comprises the steps of:
tracking the face in subsequent frames of the video; and
applying the image processing technique associated with the face in the subsequent frames.
13. The method of claim 11, wherein the at least one image processing technique further comprises at least one of: a wrinkle reduction algorithm, a blemish removal algorithm and a skin tone correction algorithm.
14. The method of claim 11, wherein the at least one image processing technique further comprises adjusting at least one dimension associated with the face.
15. The method of claim 14, wherein the at least one dimension is associated with at least one sub-region of the face, the sub-region being one of: a nose, a forehead, a cheek, a lip, and a chin.
16. The method of claim 11, wherein the at least one image processing technique further comprises increasing at least one color level associated with a portion of the face.
17. The image capture device of claim 16, further comprising the step of identifying at least one cheek region associated with the face, wherein the image processing technique further comprises increasing the at least one color level in the at least one cheek region.
18. The method of claim 11, further comprising the steps of:
generating a user interface allowing a user to specify the at least one image processing technique associated with the face;
associating the facial signature and a reference corresponding to the at least one image processing technique with the facial profile; and
storing the facial profile in a memory accessible to the image capture device.
19. The method of claim 18, wherein the step of storing the facial profile further comprises the step of storing the facial profile in a mobile device memory associated with a mobile device in which the image capture device is integrated.
20. A system, comprising:
means for capturing at least one image via an image sensor means;
means for detecting an object corresponding to a face within the image;
means for calculating a facial signature associated with the face;
means for identifying a facial profile accessible to the image capture device corresponding to the facial signature, the facial profile specifying at least one image processing technique; and
means for applying the at least one image processing technique to at least a portion of the face within the image.
US13/281,521 2011-07-20 2011-10-26 Facial Image Processing in an Image Capture Device Abandoned US20130021490A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/281,521 US20130021490A1 (en) 2011-07-20 2011-10-26 Facial Image Processing in an Image Capture Device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201161509747P 2011-07-20 2011-07-20
US13/281,521 US20130021490A1 (en) 2011-07-20 2011-10-26 Facial Image Processing in an Image Capture Device

Publications (1)

Publication Number Publication Date
US20130021490A1 true US20130021490A1 (en) 2013-01-24

Family

ID=47555520

Family Applications (9)

Application Number Title Priority Date Filing Date
US13/232,045 Abandoned US20130021488A1 (en) 2011-07-20 2011-09-14 Adjusting Image Capture Device Settings
US13/232,052 Abandoned US20130021512A1 (en) 2011-07-20 2011-09-14 Framing of Images in an Image Capture Device
US13/235,975 Abandoned US20130021504A1 (en) 2011-07-20 2011-09-19 Multiple image processing
US13/245,941 Abandoned US20130021489A1 (en) 2011-07-20 2011-09-27 Regional Image Processing in an Image Capture Device
US13/281,521 Abandoned US20130021490A1 (en) 2011-07-20 2011-10-26 Facial Image Processing in an Image Capture Device
US13/313,345 Abandoned US20130022116A1 (en) 2011-07-20 2011-12-07 Camera tap transcoder architecture with feed forward encode data
US13/313,352 Active 2032-01-11 US9092861B2 (en) 2011-07-20 2011-12-07 Using motion information to assist in image processing
US13/330,047 Abandoned US20130021484A1 (en) 2011-07-20 2011-12-19 Dynamic computation of lens shading
US13/413,863 Abandoned US20130021491A1 (en) 2011-07-20 2012-03-07 Camera Device Systems and Methods

Family Applications Before (4)

Application Number Title Priority Date Filing Date
US13/232,045 Abandoned US20130021488A1 (en) 2011-07-20 2011-09-14 Adjusting Image Capture Device Settings
US13/232,052 Abandoned US20130021512A1 (en) 2011-07-20 2011-09-14 Framing of Images in an Image Capture Device
US13/235,975 Abandoned US20130021504A1 (en) 2011-07-20 2011-09-19 Multiple image processing
US13/245,941 Abandoned US20130021489A1 (en) 2011-07-20 2011-09-27 Regional Image Processing in an Image Capture Device

Family Applications After (4)

Application Number Title Priority Date Filing Date
US13/313,345 Abandoned US20130022116A1 (en) 2011-07-20 2011-12-07 Camera tap transcoder architecture with feed forward encode data
US13/313,352 Active 2032-01-11 US9092861B2 (en) 2011-07-20 2011-12-07 Using motion information to assist in image processing
US13/330,047 Abandoned US20130021484A1 (en) 2011-07-20 2011-12-19 Dynamic computation of lens shading
US13/413,863 Abandoned US20130021491A1 (en) 2011-07-20 2012-03-07 Camera Device Systems and Methods

Country Status (1)

Country Link
US (9) US20130021488A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130258118A1 (en) * 2012-03-30 2013-10-03 Verizon Patent And Licensing Inc. Automatic skin tone calibration for camera images
US20150227609A1 (en) * 2014-02-13 2015-08-13 Yahoo! Inc. Automatic group formation and group detection through media recognition
WO2016183380A1 (en) * 2015-05-12 2016-11-17 Mine One Gmbh Facial signature methods, systems and software
US20170140596A1 (en) * 2014-07-03 2017-05-18 Brady Worldwide, Inc. Lockout/tagout device with non-volatile memory and related system
US10551913B2 (en) 2015-03-21 2020-02-04 Mine One Gmbh Virtual 3D methods, systems and software
US10853625B2 (en) 2015-03-21 2020-12-01 Mine One Gmbh Facial signature methods, systems and software
US11995902B2 (en) 2020-11-30 2024-05-28 Mine One Gmbh Facial signature methods, systems and software

Families Citing this family (79)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE69735727T2 (en) 1997-01-27 2007-01-04 Peter D. Louisville Haaland PROCESS FOR REDUCING THE REFLECTION OF OPTICAL SUBSTRATES
US10116839B2 (en) 2014-08-14 2018-10-30 Atheer Labs, Inc. Methods for camera movement compensation for gesture detection and object recognition
JP5781351B2 (en) * 2011-03-30 2015-09-24 日本アビオニクス株式会社 Imaging apparatus, pixel output level correction method thereof, infrared camera system, and interchangeable lens system
JP5778469B2 (en) 2011-04-28 2015-09-16 日本アビオニクス株式会社 Imaging apparatus, image generation method, infrared camera system, and interchangeable lens system
KR101796481B1 (en) * 2011-11-28 2017-12-04 삼성전자주식회사 Method of eliminating shutter-lags with low power consumption, camera module, and mobile device having the same
US9462255B1 (en) 2012-04-18 2016-10-04 Amazon Technologies, Inc. Projection and camera system for augmented reality environment
US9619036B2 (en) 2012-05-11 2017-04-11 Comcast Cable Communications, Llc System and methods for controlling a user experience
US9438805B2 (en) * 2012-06-08 2016-09-06 Sony Corporation Terminal device and image capturing method
US8957973B2 (en) * 2012-06-11 2015-02-17 Omnivision Technologies, Inc. Shutter release using secondary camera
US20130335587A1 (en) * 2012-06-14 2013-12-19 Sony Mobile Communications, Inc. Terminal device and image capturing method
TWI498771B (en) * 2012-07-06 2015-09-01 Pixart Imaging Inc Gesture recognition system and glasses with gesture recognition function
KR101917650B1 (en) * 2012-08-03 2019-01-29 삼성전자 주식회사 Method and apparatus for processing a image in camera device
US9554042B2 (en) * 2012-09-24 2017-01-24 Google Technology Holdings LLC Preventing motion artifacts by intelligently disabling video stabilization
US9286509B1 (en) * 2012-10-19 2016-03-15 Google Inc. Image optimization during facial recognition
JP2014086849A (en) * 2012-10-23 2014-05-12 Sony Corp Content acquisition device and program
US9060127B2 (en) * 2013-01-23 2015-06-16 Orcam Technologies Ltd. Apparatus for adjusting image capture settings
JP2014176034A (en) * 2013-03-12 2014-09-22 Ricoh Co Ltd Video transmission device
US9552630B2 (en) * 2013-04-09 2017-01-24 Honeywell International Inc. Motion deblurring
US9595083B1 (en) * 2013-04-16 2017-03-14 Lockheed Martin Corporation Method and apparatus for image producing with predictions of future positions
US9916367B2 (en) 2013-05-03 2018-03-13 Splunk Inc. Processing system search requests from multiple data stores with overlapping data
US8738629B1 (en) 2013-05-03 2014-05-27 Splunk Inc. External Result Provided process for retrieving data stored using a different configuration or protocol
US10003792B2 (en) 2013-05-27 2018-06-19 Microsoft Technology Licensing, Llc Video encoder for images
US10796617B2 (en) * 2013-06-12 2020-10-06 Infineon Technologies Ag Device, method and system for processing an image data stream
US9529513B2 (en) * 2013-08-05 2016-12-27 Microsoft Technology Licensing, Llc Two-hand interaction with natural user interface
US9270959B2 (en) 2013-08-07 2016-02-23 Qualcomm Incorporated Dynamic color shading correction
DE112014004664T5 (en) * 2013-10-09 2016-08-18 Magna Closures Inc. DISPLAY CONTROL FOR VEHICLE WINDOW
US9973672B2 (en) 2013-12-06 2018-05-15 Huawei Device (Dongguan) Co., Ltd. Photographing for dual-lens device using photographing environment determined using depth estimation
US10931866B2 (en) 2014-01-05 2021-02-23 Light Labs Inc. Methods and apparatus for receiving and storing in a camera a user controllable setting that is used to control composite image generation performed after image capture
US9245347B2 (en) * 2014-01-30 2016-01-26 Adobe Systems Incorporated Image Cropping suggestion
US9251594B2 (en) * 2014-01-30 2016-02-02 Adobe Systems Incorporated Cropping boundary simplicity
KR102128468B1 (en) * 2014-02-19 2020-06-30 삼성전자주식회사 Image Processing Device and Method including a plurality of image signal processors
CN103841328B (en) * 2014-02-27 2015-03-11 深圳市中兴移动通信有限公司 Low-speed shutter shooting method and device
WO2015139165A1 (en) 2014-03-17 2015-09-24 Microsoft Technology Licensing, Llc Encoder-side decisions for screen content encoding
US20150297986A1 (en) * 2014-04-18 2015-10-22 Aquifi, Inc. Systems and methods for interactive video games with motion dependent gesture inputs
WO2015170503A1 (en) * 2014-05-08 2015-11-12 ソニー株式会社 Information processing apparatus and information processing method
US10051196B2 (en) * 2014-05-20 2018-08-14 Lenovo (Singapore) Pte. Ltd. Projecting light at angle corresponding to the field of view of a camera
CN106796383A (en) * 2014-08-06 2017-05-31 凯文·J·沃琳 The orientation system of image recorder
KR102225947B1 (en) * 2014-10-24 2021-03-10 엘지전자 주식회사 Mobile terminal and method for controlling the same
CN105549302B (en) 2014-10-31 2018-05-08 国际商业机器公司 The coverage suggestion device of photography and vedio recording equipment
US10334158B2 (en) * 2014-11-03 2019-06-25 Robert John Gove Autonomous media capturing
US20160148648A1 (en) * 2014-11-20 2016-05-26 Facebook, Inc. Systems and methods for improving stabilization in time-lapse media content
WO2016123792A1 (en) 2015-02-06 2016-08-11 Microsoft Technology Licensing, Llc Skipping evaluation stages during media encoding
US11721414B2 (en) 2015-03-12 2023-08-08 Walmart Apollo, Llc Importing structured prescription records from a prescription label on a medication package
US20160316220A1 (en) * 2015-04-21 2016-10-27 Microsoft Technology Licensing, Llc Video encoder management strategies
US10447926B1 (en) 2015-06-19 2019-10-15 Amazon Technologies, Inc. Motion estimation based video compression and encoding
US10165186B1 (en) * 2015-06-19 2018-12-25 Amazon Technologies, Inc. Motion estimation based video stabilization for panoramic video from multi-camera capture device
US10136132B2 (en) 2015-07-21 2018-11-20 Microsoft Technology Licensing, Llc Adaptive skip or zero block detection combined with transform size decision
EP3136726B1 (en) * 2015-08-27 2018-03-07 Axis AB Pre-processing of digital images
US9648223B2 (en) * 2015-09-04 2017-05-09 Microvision, Inc. Laser beam scanning assisted autofocus
US9456195B1 (en) * 2015-10-08 2016-09-27 Dual Aperture International Co. Ltd. Application programming interface for multi-aperture imaging systems
US9578221B1 (en) * 2016-01-05 2017-02-21 International Business Machines Corporation Camera field of view visualizer
JP6514140B2 (en) * 2016-03-17 2019-05-15 株式会社東芝 Imaging support apparatus, method and program
EP3466051A1 (en) 2016-05-25 2019-04-10 GoPro, Inc. Three-dimensional noise reduction
US9639935B1 (en) 2016-05-25 2017-05-02 Gopro, Inc. Apparatus and methods for camera alignment model calibration
WO2017205597A1 (en) * 2016-05-25 2017-11-30 Gopro, Inc. Image signal processing-based encoding hints for motion estimation
US10140776B2 (en) * 2016-06-13 2018-11-27 Microsoft Technology Licensing, Llc Altering properties of rendered objects via control points
US9851842B1 (en) * 2016-08-10 2017-12-26 Rovi Guides, Inc. Systems and methods for adjusting display characteristics
US10366122B2 (en) * 2016-09-14 2019-07-30 Ants Technology (Hk) Limited. Methods circuits devices systems and functionally associated machine executable code for generating a searchable real-scene database
CN109154979A (en) * 2016-10-26 2019-01-04 奥康科技有限公司 For analyzing image and providing the wearable device and method of feedback
CN106550227B (en) * 2016-10-27 2019-02-22 成都西纬科技有限公司 A kind of image saturation method of adjustment and device
US10477064B2 (en) 2017-08-21 2019-11-12 Gopro, Inc. Image stitching with electronic rolling shutter correction
US10791265B1 (en) 2017-10-13 2020-09-29 State Farm Mutual Automobile Insurance Company Systems and methods for model-based analysis of damage to a vehicle
US11587046B1 (en) 2017-10-25 2023-02-21 State Farm Mutual Automobile Insurance Company Systems and methods for performing repairs to a vehicle
CN111345036A (en) * 2017-10-26 2020-06-26 京瓷株式会社 Image processing apparatus, imaging apparatus, driving assistance apparatus, moving object, and image processing method
KR20190087977A (en) * 2017-12-25 2019-07-25 저텍 테크놀로지 컴퍼니 리미티드 Laser beam scanning display and augmented reality glasses
KR20200029270A (en) 2018-09-10 2020-03-18 삼성전자주식회사 Electronic apparatus for recognizing an object and controlling method thereof
JP7456385B2 (en) * 2018-10-25 2024-03-27 ソニーグループ株式会社 Image processing device, image processing method, and program
US10771696B2 (en) * 2018-11-26 2020-09-08 Sony Corporation Physically based camera motion compensation
WO2020142471A1 (en) * 2018-12-30 2020-07-09 Sang Chul Kwon Foldable mobile phone
US11289078B2 (en) * 2019-06-28 2022-03-29 Intel Corporation Voice controlled camera with AI scene detection for precise focusing
US10861127B1 (en) 2019-09-17 2020-12-08 Gopro, Inc. Image and video processing using multiple pipelines
US11064118B1 (en) * 2019-12-18 2021-07-13 Gopro, Inc. Systems and methods for dynamic stabilization adjustment
US11006044B1 (en) * 2020-03-03 2021-05-11 Qualcomm Incorporated Power-efficient dynamic electronic image stabilization
US11284157B2 (en) * 2020-06-11 2022-03-22 Rovi Guides, Inc. Methods and systems facilitating adjustment of multiple variables via a content guidance application
TWI774039B (en) * 2020-08-12 2022-08-11 瑞昱半導體股份有限公司 System for compensating image with fixed pattern noise
US11563899B2 (en) * 2020-08-14 2023-01-24 Raytheon Company Parallelization technique for gain map generation using overlapping sub-images
CN114079735B (en) * 2020-08-19 2024-02-23 瑞昱半导体股份有限公司 Image compensation system for fixed image noise
US11902671B2 (en) * 2021-12-09 2024-02-13 Fotonation Limited Vehicle occupant monitoring system including an image acquisition device with a rolling shutter image sensor
EP4344480A1 (en) * 2022-02-07 2024-04-03 GoPro, Inc. Methods and apparatus for real-time guided encoding

Family Cites Families (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100325253B1 (en) * 1998-05-19 2002-03-04 미야즈 준이치롯 Motion vector search method and apparatus
US6486908B1 (en) * 1998-05-27 2002-11-26 Industrial Technology Research Institute Image-based method and system for building spherical panoramas
US20010047517A1 (en) * 2000-02-10 2001-11-29 Charilaos Christopoulos Method and apparatus for intelligent transcoding of multimedia data
JP2001245303A (en) * 2000-02-29 2001-09-07 Toshiba Corp Moving picture coder and moving picture coding method
US6407680B1 (en) * 2000-12-22 2002-06-18 Generic Media, Inc. Distributed on-demand media transcoding system and method
US7034848B2 (en) * 2001-01-05 2006-04-25 Hewlett-Packard Development Company, L.P. System and method for automatically cropping graphical images
WO2002098130A2 (en) * 2001-05-31 2002-12-05 Canon Kabushiki Kaisha Information storing apparatus and method therefor
US7801215B2 (en) * 2001-07-24 2010-09-21 Sasken Communication Technologies Limited Motion estimation technique for digital video encoding applications
US20030126622A1 (en) * 2001-12-27 2003-07-03 Koninklijke Philips Electronics N.V. Method for efficiently storing the trajectory of tracked objects in video
KR100850705B1 (en) * 2002-03-09 2008-08-06 삼성전자주식회사 Method for adaptive encoding motion image based on the temperal and spatial complexity and apparatus thereof
JP4275358B2 (en) * 2002-06-11 2009-06-10 株式会社日立製作所 Image information conversion apparatus, bit stream converter, and image information conversion transmission method
US7259784B2 (en) * 2002-06-21 2007-08-21 Microsoft Corporation System and method for camera color calibration and image stitching
US20040131276A1 (en) * 2002-12-23 2004-07-08 John Hudson Region-based image processor
EP3404479A1 (en) * 2002-12-25 2018-11-21 Nikon Corporation Blur correction camera system
US20130107938A9 (en) * 2003-05-28 2013-05-02 Chad Fogg Method And Apparatus For Scalable Video Decoder Using An Enhancement Stream
KR100566290B1 (en) * 2003-09-18 2006-03-30 삼성전자주식회사 Image Scanning Method By Using Scan Table and Discrete Cosine Transform Apparatus adapted it
JP4123171B2 (en) * 2004-03-08 2008-07-23 ソニー株式会社 Method for manufacturing vibration type gyro sensor element, vibration type gyro sensor element, and method for adjusting vibration direction
WO2005094270A2 (en) * 2004-03-24 2005-10-13 Sharp Laboratories Of America, Inc. Methods and systems for a/v input device to diplay networking
US8315307B2 (en) * 2004-04-07 2012-11-20 Qualcomm Incorporated Method and apparatus for frame prediction in hybrid video compression to enable temporal scalability
US20060109900A1 (en) * 2004-11-23 2006-05-25 Bo Shen Image data transcoding
JP2006203682A (en) * 2005-01-21 2006-08-03 Nec Corp Converting device of compression encoding bit stream for moving image at syntax level and moving image communication system
TW200816798A (en) * 2006-09-22 2008-04-01 Altek Corp Method of automatic shooting by using an image recognition technology
US7843824B2 (en) * 2007-01-08 2010-11-30 General Instrument Corporation Method and apparatus for statistically multiplexing services
US7924316B2 (en) * 2007-03-14 2011-04-12 Aptina Imaging Corporation Image feature identification and motion compensation apparatus, systems, and methods
WO2008142948A1 (en) * 2007-05-23 2008-11-27 Nec Corporation Dynamic image distribution system, conversion device, and dynamic image distribution method
KR20100031755A (en) * 2007-07-30 2010-03-24 닛본 덴끼 가부시끼가이샤 Connection terminal, distribution system, conversion method, and program
US20090060039A1 (en) * 2007-09-05 2009-03-05 Yasuharu Tanaka Method and apparatus for compression-encoding moving image
US8098732B2 (en) * 2007-10-10 2012-01-17 Sony Corporation System for and method of transcoding video sequences from a first format to a second format
US8063942B2 (en) * 2007-10-19 2011-11-22 Qualcomm Incorporated Motion assisted image sensor configuration
US8170342B2 (en) * 2007-11-07 2012-05-01 Microsoft Corporation Image recognition of content
JP2009152672A (en) * 2007-12-18 2009-07-09 Samsung Techwin Co Ltd Recording apparatus, reproducing apparatus, recording method, reproducing method, and program
JP5242151B2 (en) * 2007-12-21 2013-07-24 セミコンダクター・コンポーネンツ・インダストリーズ・リミテッド・ライアビリティ・カンパニー Vibration correction control circuit and imaging apparatus including the same
JP2009159359A (en) * 2007-12-27 2009-07-16 Samsung Techwin Co Ltd Moving image data encoding apparatus, moving image data decoding apparatus, moving image data encoding method, moving image data decoding method and program
US20090217338A1 (en) * 2008-02-25 2009-08-27 Broadcom Corporation Reception verification/non-reception verification of base/enhancement video layers
US20090323810A1 (en) * 2008-06-26 2009-12-31 Mediatek Inc. Video encoding apparatuses and methods with decoupled data dependency
US7990421B2 (en) * 2008-07-18 2011-08-02 Sony Ericsson Mobile Communications Ab Arrangement and method relating to an image recording device
JP2010039788A (en) * 2008-08-05 2010-02-18 Toshiba Corp Image processing apparatus and method thereof, and image processing program
JP2010147808A (en) * 2008-12-18 2010-07-01 Olympus Imaging Corp Imaging apparatus and image processing method in same
US8311115B2 (en) * 2009-01-29 2012-11-13 Microsoft Corporation Video encoding using previously calculated motion information
US20100194851A1 (en) * 2009-02-03 2010-08-05 Aricent Inc. Panorama image stitching
US20100229206A1 (en) * 2009-03-03 2010-09-09 Viasat, Inc. Space shifting over forward satellite communication channels
US8520083B2 (en) * 2009-03-27 2013-08-27 Canon Kabushiki Kaisha Method of removing an artefact from an image
US20100309987A1 (en) * 2009-06-05 2010-12-09 Apple Inc. Image acquisition and encoding system
JP5473536B2 (en) * 2009-10-28 2014-04-16 京セラ株式会社 Portable imaging device with projector function
US20110170608A1 (en) * 2010-01-08 2011-07-14 Xun Shi Method and device for video transcoding using quad-tree based mode selection
US8681255B2 (en) * 2010-09-28 2014-03-25 Microsoft Corporation Integrated low power depth camera and projection device
US9007428B2 (en) * 2011-06-01 2015-04-14 Apple Inc. Motion-based image stitching
US8554011B2 (en) * 2011-06-07 2013-10-08 Microsoft Corporation Automatic exposure correction of images

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130258118A1 (en) * 2012-03-30 2013-10-03 Verizon Patent And Licensing Inc. Automatic skin tone calibration for camera images
US9118876B2 (en) * 2012-03-30 2015-08-25 Verizon Patent And Licensing Inc. Automatic skin tone calibration for camera images
US20150227609A1 (en) * 2014-02-13 2015-08-13 Yahoo! Inc. Automatic group formation and group detection through media recognition
US10121060B2 (en) * 2014-02-13 2018-11-06 Oath Inc. Automatic group formation and group detection through media recognition
US20170140596A1 (en) * 2014-07-03 2017-05-18 Brady Worldwide, Inc. Lockout/tagout device with non-volatile memory and related system
US10551913B2 (en) 2015-03-21 2020-02-04 Mine One Gmbh Virtual 3D methods, systems and software
US10853625B2 (en) 2015-03-21 2020-12-01 Mine One Gmbh Facial signature methods, systems and software
US11960639B2 (en) 2015-03-21 2024-04-16 Mine One Gmbh Virtual 3D methods, systems and software
WO2016183380A1 (en) * 2015-05-12 2016-11-17 Mine One Gmbh Facial signature methods, systems and software
US11995902B2 (en) 2020-11-30 2024-05-28 Mine One Gmbh Facial signature methods, systems and software

Also Published As

Publication number Publication date
US20130021489A1 (en) 2013-01-24
US20130021488A1 (en) 2013-01-24
US20130021483A1 (en) 2013-01-24
US20130021491A1 (en) 2013-01-24
US20130021484A1 (en) 2013-01-24
US20130021504A1 (en) 2013-01-24
US20130021512A1 (en) 2013-01-24
US9092861B2 (en) 2015-07-28
US20130022116A1 (en) 2013-01-24

Similar Documents

Publication Publication Date Title
US20130021490A1 (en) Facial Image Processing in an Image Capture Device
KR101772177B1 (en) Method and apparatus for obtaining photograph
KR101929077B1 (en) Image identificaiton method and image identification device
CN109644231B (en) Method, system, and medium providing improved video stability for mobile devices
EP3480784B1 (en) Image processing method, and device
KR20190101693A (en) Electronic device displaying a interface for editing video data and method for controlling thereof
KR101349699B1 (en) Apparatus and method for extracting and synthesizing image
EP3644599A1 (en) Video processing method and apparatus, electronic device, and storage medium
US9973687B2 (en) Capturing apparatus and method for capturing images without moire pattern
EP4055812B1 (en) A system for performing ambient light image correction
WO2015189343A1 (en) Methods and systems for color processing of digital images
KR102655625B1 (en) Method and photographing device for controlling the photographing device according to proximity of a user
WO2018019068A1 (en) Photographing method and device, and mobile terminal
JP2008033718A (en) Imaging device and regional enlargement display method
KR20150141059A (en) Apparatus and method for providing thumbnail image of moving picture
US9225906B2 (en) Electronic device having efficient mechanisms for self-portrait image capturing and method for controlling the same
EP3629572A1 (en) Image capture apparatus capable of performing hdr combination, method of controlling same, and storage medium
US20160127651A1 (en) Electronic device and method for capturing image using assistant icon
US20140285649A1 (en) Image acquisition apparatus that stops acquisition of images
KR102562720B1 (en) Detection apparatus, image processing apparatus, detection method, and image processing method
JP6373446B2 (en) Program, system, apparatus and method for selecting video frame
JP2022120681A (en) Image processing device and image processing method
US9600735B2 (en) Image processing device, image processing method, program recording medium
KR20140134844A (en) Method and device for photographing based on objects
US20170372495A1 (en) Methods and systems for color processing of digital images

Legal Events

Date Code Title Description
AS Assignment

Owner name: BROADCOM CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JAMES, GERAINT;SWISSA, EFRAT;SIGNING DATES FROM 20111020 TO 20111025;REEL/FRAME:027365/0367

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH CAROLINA

Free format text: PATENT SECURITY AGREEMENT;ASSIGNOR:BROADCOM CORPORATION;REEL/FRAME:037806/0001

Effective date: 20160201

Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH

Free format text: PATENT SECURITY AGREEMENT;ASSIGNOR:BROADCOM CORPORATION;REEL/FRAME:037806/0001

Effective date: 20160201

AS Assignment

Owner name: AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD., SINGAPORE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BROADCOM CORPORATION;REEL/FRAME:041706/0001

Effective date: 20170120

Owner name: AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BROADCOM CORPORATION;REEL/FRAME:041706/0001

Effective date: 20170120

AS Assignment

Owner name: BROADCOM CORPORATION, CALIFORNIA

Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:BANK OF AMERICA, N.A., AS COLLATERAL AGENT;REEL/FRAME:041712/0001

Effective date: 20170119