EP3123226A1 - Camera to capture multiple sub-images for generation of an image - Google Patents

Camera to capture multiple sub-images for generation of an image

Info

Publication number
EP3123226A1
EP3123226A1 EP14887071.0A EP14887071A EP3123226A1 EP 3123226 A1 EP3123226 A1 EP 3123226A1 EP 14887071 A EP14887071 A EP 14887071A EP 3123226 A1 EP3123226 A1 EP 3123226A1
Authority
EP
European Patent Office
Prior art keywords
sub
image
images
image sensor
lenses
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP14887071.0A
Other languages
German (de)
French (fr)
Other versions
EP3123226B1 (en
EP3123226A4 (en
Inventor
Hong W. Wong
Wah Yiu Kwong
Jiancheng TAO
Xiaoguo Liang
Andy Keates
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intel Corp
Original Assignee
Intel Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corp filed Critical Intel Corp
Publication of EP3123226A1 publication Critical patent/EP3123226A1/en
Publication of EP3123226A4 publication Critical patent/EP3123226A4/en
Application granted granted Critical
Publication of EP3123226B1 publication Critical patent/EP3123226B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/272Means for inserting a foreground image in a background image, i.e. inlay, outlay
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B13/00Optical objectives specially designed for the purposes specified below
    • G02B13/001Miniaturised objectives for electronic devices, e.g. portable telephones, webcams, PDAs, small digital cameras
    • G02B13/0015Miniaturised objectives for electronic devices, e.g. portable telephones, webcams, PDAs, small digital cameras characterised by the lens design
    • G02B13/002Miniaturised objectives for electronic devices, e.g. portable telephones, webcams, PDAs, small digital cameras characterised by the lens design having at least one aspherical surface
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/45Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/56Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/58Means for changing the camera field of view without moving the camera body, e.g. nutating or panning of optics or image sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/69Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders

Definitions

  • Embodiments described herein generally relate to the field of electronic devices and, more particularly, a camera to capture multiple sub-images for the generation of an image.
  • Figure l is an illustration of mobile devices to include an embodiment of a camera assembly
  • Figures 2A and 2B illustrates examples of conventional lenses and image sensors
  • Figure 3 A is an illustration of elements of an embodiment of a camera assembly including multiple lenses
  • Figure 3B is an illustration of elements of an embodiment of a camera assembly including an aspheric lens
  • Figure 4 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including multiple lenses
  • Figure 5 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including an aspheric lens
  • Figure 6 is an illustration of stitching sub-images together to generate a combined image according to an embodiment
  • Figure 7 is an illustration of an embodiment of a mobile device including a camera assembly
  • Figure 8 is a flow chart to illustratean embodiment of a process to generate a combined image from captured sub-images.
  • Figure 9 illustraterates an apparatus or system to generate combined images from captured sub-images.
  • Embodiments described herein are generally directed to a camera to capture multiple sub-images for the generation of an image.
  • Mobile device means a smartphone, smartwatch, tablet computer, handheld computer, mobile Internet device, wearable technology, or other mobile apparatus that includes processing ability and communication ability.
  • Camera assembly means an apparatus or subsystem for capturing images, including still photographs and video images.
  • a camera assembly may further include processing of images.
  • a camera having a lens on a first side of the mobile device,the first side includes a main display, may be referred to as a front- facing camera and a camera having a lens on a second side of the mobile device, the second side not including a main display, may be referred to as a rear-facing camera.
  • a camera includes, but is not limited to, a lens and an image sensor.
  • a camera may further include a processing unit.
  • Lens means one or more optical elements for transmission of light from a scene to an image sensor.
  • Each optical element of a lens may be a simple lens (a single element to refract light), a mirror, or other element to transmit or otherwise affect light.
  • a lens including multiple lens elements may be referred to as a compound lens.
  • an aspheric lens means a lens including at least one simple lens with surface profiles that are not portions of a sphere or cylinder.
  • an aspheric lens may include a lens in an oval shape with one dimension that is larger than another, or, stated in another way, with a diameter through the semi-major axis (the largest axis of the oval) that is larger than a diameter through the semi-minor axis (the smallest axis of the oval).
  • Cameras on mobile devices are increasing used for functions that require high performance.
  • a camera on a mobile device is requiredto provide good low light performance in order to be certified for certain video communicationoperations, including SkypeTM and MicrosoftLync® certification. Further, it may be expected that certification requirements will expand andbecome more stringent in the future in order to provide improved user experience in video communications.
  • Another aspect of improved camera performance is the increasing resolution of mobile device cameras.
  • the increase in camera resolution provides improved image clarity, butmay have a negative impact on low light performance.
  • a larger image sensor is needed to maintain the low light performance as the number of pixelsfor the image sensor increases.
  • Other technologies are also driving the need for low light performance for cameras, such as facial recognition of the end user of a device to effect computer login authentication.
  • the location of a camera in an apparatus is limited by the physical dimensions of the apparatus.
  • the bezel width of a forward side of a mobile device (where "bezel” refers to the portion of the device around or beside a display screen or input elements on a front side of the device), where a camera lens would generally be installed, limits the size of the lens and optical sensor in at least one dimension (referred to herein as the Y dimension of the mobile device).
  • a lens for animage circle becomes increasingly difficult as the image circle increases in size.
  • the larger lens typically uses more lens elements and may require more expensive lens elements because of the need for more aspheric or special glasses to keep the image sharp and evenly lit over the larger image circle.
  • the lens for a larger image sensor is generally larger, heavier and more expensive.
  • creating a lens for a smaller sensor is easier and less expensive to produce than a lens for a larger sensor at the same level of lens performance. This factor is true for lenses for medium and large format cameras as well as lenses for small format cameras, such as mobile device camera assemblies.
  • an apparatus, system, or process includes wide image sensing to simultaneously capture multiple sub-images.
  • a camera provides enhanced low light performance while reducing impact to the overall system's Y-dimension through use of wide image sensing to capture multiple sub-images.
  • a camera includes wide aspect ratio image sensing together with modified lens design and image processing to enable both a compact form factor and enhanced low light performance foruse in, for example, mobile devices or other devices with limited physical size.
  • a camera assembly includes one or more lenses to transmit multiple portions of an image (referred to herein as sub-images) of a sceneon one or more image sensors, the one or more image sensors having a width sufficient to receive the multiplesub-images.
  • the sub-images may overlap.
  • a processor or other image processing element provides for combining at least a first sub-image and a second sub-image to generate a combined image.
  • the one or more lenses transmit at least a first sub- image and a second sub-image that differ in a first dimension, such as a vertical differentiation of the first and second sub-images, in order that the first sub-image and the second sub-image are arranged along a second dimension, such as a horizontal arrangement, onto the one or more image sensors.
  • a processing element processes the sub-images including stitching the sub-images together to generate a combined image of the original scene. While the examples described and illustrated herein illustrate two sub-images, embodiments are not limited to two sub-images, and may include a larger number of sub-images.
  • a mobile device may include, but is not limited to, a smart phone 110 or a tablet computer 130.
  • the smart phone 110 includes a display screen 115 on a first side of the smart phone, and the tablet computer 130 includes a display screen 135 on a first side of the tablet computer.
  • Each mobile device 110, 130 includes a bezel, or Y-dimension 120, 140,of a certain size.
  • the available Y- dimension 120 for the smart phone is 12 mm (millimeters) and the available Y- dimension 140 for the tablet computer 130 is 18.5 mm.
  • the usable Y-dimension of each mobile device 120, 140 is actually smaller than the indicated sizes because of the width of the cover material.
  • the shape of the cover may further limit the usable Y-dimension.
  • the mobile devices 110, 130 include camera assemblies that provide improved low-light performance though use of wide image sensing to simultaneously capture multiple sub-images of a scene, with two or more sub-images being combined to generate a combined image of the scene.
  • Figures 2A and 2B illustrates examples of conventional lenses and image sensors.
  • Figure 2A illustrates front and side views of a first conventional camera assembly 200 including a small aspect lens 210, which in this example is 7.5 mm in diameter, and an image sensor 220.
  • the camera assembly 200 may be utilized in, for example, a mobile device.
  • an image sensor of this size may not provide sufficient low light performance for certain purposes, including video communications.
  • Figure 2B illustrates front and side views of a second conventional camera assembly 250, the camera assembly 250 including a larger lens 260 and image sensor 270 to allow for the collection of a greater quantity of light.
  • the diameter of the lens 260 has been increased to 14 mm.
  • the second camera assembly 250 may provide improved low light performance in comparison with the first camera assembly 200, such camera assembly 250 may be too large to fit within the physical constraints of a device.
  • the lens and image sensor may be too large to fit within the bezel area (the available Y-dimension) of the front side (the display side) of a mobile device. For this reason, the camera assembly 250 may not be usable as a front- facing camera of such a device.
  • FIG 3 A is an illustration of elements of an embodiment of a camera assembly including multiple lenses.
  • an embodiment of a camera assembly 300 includes multiple lenses to simultaneously capture multiple sub-images, such as a first lens 310 to transmit light for a first sub-image and a second lens 315 to transmit light for a second sub-image.
  • the camera assembly 300 includes a wide image sensor 320, wherein the sub-image from the first lens 310 is directed to a first portion of the image sensor 320 and the second sub-image from the second lens 315 is directed to a second portion of the image sensor 320.
  • the wide image sensor 320 includes a single sensor element.
  • the wide image sensor 320 includes multiple sensor elements, such as a first sensor element for the capture of the first sub-image and a second sensor element for the capture of the second sub-image.
  • the multiple lenses 310, 315 reduce the impact in a first dimension bythe arrangement side by side along a second dimension.
  • each of the lenses 310, 315 is 7.5 mm in diameter, and thus the total width in the Y-dimension is 7.5 mm because the multiple lenses are arranged in the X-dimension.
  • the light transmitted has essentially been doubled, but the effective increase in captured light may vary depending on the specific implementation.
  • the image sensor 320 is increased in size only in the X-dimension, thus reducing any impact in the Y- dimension for the camera assembly while improving the low light performance of the camera assembly 300.
  • Figure 3B is an illustration of elements of an embodiment of a camera assembly including an aspheric lens.
  • an embodiment of a camera assembly includes an aspheric lens 360 to simultaneously capture multiple sub-images.
  • the aspheric lens 360 is formed to generate multiple sub-images from a scene, the aspheric lens 360 being shaped such that the lens is substantially larger in a second dimension than the lens is in a first dimension.
  • camera assembly includes a wide image sensor 370, wherein the a first sub-image transmitted by the aspheric lens 360 is directed to a first portion of the image sensor 370 and a second sub-image transmitted by the aspheric lens 370 is directed to a second portion of the image sensor.
  • the wide image sensor 370 includes a single sensor element.
  • the wide image sensor 370 includes multiple sensor elements, such as a first sensor element for the capture of the first sub-image and a second sensor element for the capture of the second sub-image.
  • the aspheric lens 360 reduces the impact in a first dimension by the shape and arrangement of the lens. As illustrated in Figure 3B, the aspheric lens 360 is 7.5 mm in width in the Y-dimension, while such lens is substantially larger in the X-dimension. However, the lens captures additional light in comparison with a spherical lens. Further, the image sensor 370 is increased in size in the X-dimension, thus reducing any impact in the Y-dimension for the camera assembly 350 while improving the low light performance of the camera assembly.
  • Figure 4 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including multiple lenses.
  • a camera assembly 400 includes multiple lenses, including a first lens 410 and a second lens 415, to simultaneously capture multiple sub-images, such as a first sub- image 430 transmitted via the first lens 410 and a second sub-image 435 transmitted via the second lens 415.
  • the first lens and the second lens are arranged in a second dimension (X-dimension, or horizontal in Figure 4) to reduce width in a first dimension (Y-dimension, or vertical in Figure 4).
  • the camera assembly 400 further includes a wide aspect ratio image sensor 420, wherein the image sensor 420 is sufficiently sized to allow the capture of the first sub-image 430 adjacent to the second sub-image 435 in the X-dimension, or horizontally in the illustration.
  • the wide aspect ratio image sensor 420 includes a single sensor element, and in some embodiments the image sensor includes multiple sensor elements.
  • the first sub- image 430 and second sub-image 435 are portions of the full image arranged vertically (along the first dimension), such as the first sub-image 430 being a top portion of a full image of the scene and the second sub-image 435 being a bottom portion of the full image of the scene.
  • the multiple sub- images are captured in a horizontal arrangement (along the second dimension) on the image sensor.
  • the first sub-image 430 being a first vertically arranged sub-image (the top sub-image in Figure 4), is arranged in a first horizontal position (a left position in Figure 4)
  • the second sub-image 435 being a second vertically arranged sub-image (the bottom sub-image in Figure 4) is arranged in a second horizontal position (a right position in Figure 4).
  • the multiple sub-images 430, 435 are processed to generate a combined image 450 of the original scene.
  • the sub-images 430, 435 are rearranged in processing to return such sub-images to their original position in the combined image 450 of the scene.
  • the first sub-image 430 captured in the left portion of the image sensor 420 is processed to become a top portion of the combined image 450
  • the second sub-image 435 captured in the right portion of the image sensor 420 is processed to become a top portion of the combined image 450.
  • processing of sub-images includes image stabilization, correction of sensor/lens spatial separation, and merging ofsub-images into the combined image 450.
  • Figure 5 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including an aspheric lens.
  • a camera assembly 500 includes an aspheric lens 510 to
  • the aspheric lens is an oval lens or other lens shape that is larger in one dimension than in a second dimension arranged such that the larger dimension (such as the semi- major axis of an oval-shaped lens) is aligned in a second dimension (X-dimension, or horizontal in Figure 5) and the smaller dimension (such as the semi-minor axis of an oval-shaped lens) is aligned in a first dimension (Y-dimension, or vertical in Figure 5) to reduce width in the first dimension.
  • the camera assembly 500 further includes a wide aspect ratio image sensor 520, wherein the image sensor 520 is sufficiently wide to allow the capture of the first sub-image 530 adjacent to the second sub-image 535 in the X-dimension, or horizontally in the illustration.
  • the wide aspect ratio image sensor 520 includes a single sensor element, and in some embodiments the image sensor includes multiple sensor elements.
  • the first sub- image 530 and second sub-image 535 are portions of the full image arranged vertically (along the first dimension), such as the first sub-image 530 being a top portion of a full image of the scene and the second sub-image 535 being a bottom portion of the full image of the scene.
  • the multiple sub- images are captured in a horizontal arrangement (along the second dimension) on the image sensor.
  • the first sub-image 530 being a first vertically arranged sub-image (the top sub-image in Figure 5) is arranged in a first horizontal position (a left position in Figure 5)
  • the second sub-image 535 being a second vertically arranged sub-image (the bottom sub-image in Figure 5) is arranged in a second horizontal position (a right position in Figure 5).
  • the multiple sub-images 530, 535 are processed to generate a combined image 550 of the original scene.
  • the sub-images 530, 535 are rearranged in processing to return such sub-images to their original position in the combined image 550 of the scene.
  • the first sub-image 530 captured in the left portion of the image sensor 520 is processed to become a top portion of the combined image 550
  • the second sub-image 535 captured in the right portion of the image sensor 520 is processed to become a lower(or bottom)portion of the combined image 550.
  • Figure 6 is an illustration of combining sub-images to generate a combined image according to an embodiment.
  • an image sensor 620 is a wide aspect ratio with a width (Y-dimension) that is significantly shorter than a length (X-dimension).
  • a first sub-image 630 is captured in a first portion of the image sensor 620 and a second sub-image 635 is captured in a second portion of the image sensor 620.
  • the first sub-image 630 and the second sub-image 635 may have spatial separation 640 because of imperfect operation of the one or more lenses or the image sensor.
  • the multiple sub-images 630, 635 are processed, wherein the processing may include image stabilization, correction of sensor or lens spatial separation, such as cropping one or more of the sub-images to remove over-lapping between the sub-images and aligning the cropped sub-images,and combining the sub-images generate a combined image 650.
  • FIG. 7 is ablock diagram of an embodiment of a mobile device including a camera assembly.
  • a mobile device 700 includes a camera assembly including one or more lenses 710, wherein the one or more lenses may include multiple lenses, such as the multiple lenses 410, 415 illustrated in Figure 4 or the aspheric lens 510 illustrated in Figure 5, and a wide aspect ratio image sensor 715, such as the wide aspect image 420 or 520 illustrated in Figure 4 or Figure 5 respectively.
  • the camera assembly 705 is a front- facing camera with the one or more lenses of the camera assembly on a same side as a display screen 760, and thus the camera assembly is located in a limited Y- dimension space of the mobile device.
  • the one or more lenses 710 and the wide aspect ratio image sensor 715 are arranged such that a longer dimension of such elements is arranged in an X-dimension of the mobile device 700 and a shorter dimension of such elements is arranged in the Y-dimension of the mobile device 700.
  • the camera assembly 705 is operable to simultaneously capture multiple sub-images 730 on the image sensor, wherein the sub-images are captured in an arrangement along the X-dimension.
  • a processor 770 which may be a general-purpose processor, a dedicated graphics processor, or other image processing element of the mobile device 700, receives the multiple sub-images 730 for processing.
  • the processing of the sub-images may be as illustrated in Figure 6, wherein the processing of the sub-images includes image stabilization, correction of lens or image sensor spatial separation, and merging the two images into a combined image 750.
  • the processor may provide the combined image for display on the screen 760, for transmission using a transceiver 780 to another device, including transmission of the image in a video stream for a video communication, for storage in a memory 790, or a combination of such operations.
  • Figure 8 is a flow chart to illustrate an embodiment of a process to generate a combined image from captured sub-images.
  • a process includes enabling a camera function 800, and may further include enabling a video communication service (such as Skype or Lync) or other function 801.
  • the process includes capturing an image 802, which may be a single image or one image in a series of images, such as a video stream.
  • the capture of the image includes simultaneously capturing multiple sub-images, including capturing a first sub-image using a first portion of an image sensor and capturing a second sub-image using a second portion of the image sensor, wherein the image sensor is a wide aspect ratio image sensor.
  • the first sub-image may be an upper portion of a full image of a scene and the second sub- image may be a lower portion of the full image of the scene.
  • the process further includes processing the sub-images 810, wherein the processing may include providing image stabilization 812, correcting sensor or lens spatial separation (which may include eliminating overlap between the sub-images and aligning the sub-images for combination) 814, and combining the multiple sub-images, including the first sub-image and the second sub-image, to generate a combined image of the original scene 816.
  • the process may further include displaying the combined image, transmitting the combined image to another user (such as, for example, if the combined image is one image in a series of images for a video stream), storing the combined image in a memory, or a combination of such operations 820. If there is an additional new image for capture 822, such as if a video communicating session is continuing, then the process returns to the capturing of the next image 802.
  • Figure 9 Illustrates an apparatus or system to generate combined images from captured sub-images.In this illustration, certain standard and well- known components that are not germane to the present description are not shown. Elements shown as separate elements may be combined, including, for example, an SoC (System on Chip) combining multiple elements on a single chip.
  • the apparatus or system may include, but is not limited to, a mobile device.
  • theapparatus or system900 (referred to generally herein as an apparatus)includes an interconnect or crossbar 905 or other communication means for transmission of data.
  • the interconnect 905 is illustrated as a single interconnect for simplicity, but may represent multiple different
  • interconnects or buses and the component connections to such interconnects may vary.
  • the interconnect905 shown in Figure 9 is an abstraction that represents any one or more separate physical buses, point-to-point connections, or both connected by appropriate bridges, adapters, or controllers.
  • the apparatus 900 includes a camera assembly 970, wherein the camera assembly includes one or more lenses 972 to collect light for multiple sub-images of a scene and a wide aspect ratio image sensor 974 to capture the plurality of sub-images.
  • the camera assembly 970 may include the camera assembly 400 illustrated in Figure 4 or the camera assembly 500 illustrated in Figure 5.
  • the apparatus900 may include a processing means such as the one or more processors 910 coupled to the interconnect 905 for processing information.
  • the processors 910 may comprise one or more physical processors and one or more logical processors.
  • the processors may include a general-purpose processor.
  • the processors 910 may include an additional image processing element.
  • functions of the processors 910 include the processing of the sub-images captured by the camera assembly 970 to generate a combined image.
  • the apparatus900 includes one or more transmitters or receivers 940coupled to the interconnect 905.
  • the apparatus900 may include one or more antennas 942 for the transmission and reception of data via wireless communication.
  • the apparatus 900 includes one or more ports 945 for the transmission and reception of data via wired communications.
  • the data that is transmitted and received includes image data, including the transmission of combined images generated from collected sub-images by the camera assembly 970, where the transmitted combined images may be a stream of video images.
  • the apparatus900further comprises a random access memory (RAM) or other dynamic storage device or element as a main memory 915 for storing information and instructions to be executed by the processors 910.
  • the apparatus900 may include one or more non- volatile memory elements 925, including, for example, flash memory, for the storage of certain elements.
  • the apparatus 900 also may comprise a read only memory (ROM) 930 or other static storage device for storing static information and instructions for the processors 910, and data storage 935, such as a solid state drive, for the storage of data.
  • ROM read only memory
  • memory of the apparatus900 may include storage for combined images generated from the sub-images collected by the camera assembly 970.
  • the apparatus 900 includes one or more input devices 950 for the input of data, including hard and soft buttons, a joy stick, a mouse or other pointing device, voice command system, or gesture recognition system.
  • the apparatus900 includes an output display 955, where the display 955 may include a liquid crystal display (LCD) or any other display technology, for displaying information or content to a user.
  • the display 955 may include a touch-screen that is also utilized as at least a part of an input device950.
  • the display 955 may display combined images generated from the sub-images collected by the camera assembly 970.
  • the apparatus900 may also comprise a batteryor other power source 960, which may include a solar cell, a fuel cell, a charged capacitor, near field inductive coupling, or other system or device for providing or generating power in the apparatus 900.
  • the power provided by the power source 960 may be distributed as required to elements of the apparatus 900.
  • Various embodiments may include various processes. These processes may be performed by hardware components or may be embodied in computer program or machine-executable instructions, which may be used to cause a general-purpose or special-purpose processor or logic circuits programmed with the instructions to perform the processes. Alternatively, the processes may be performed by a combination of hardware and software.
  • Portions of various embodiments may be provided as a computer program product, which may include a computer-readable medium having stored thereon computer program instructions, which may be used to program a computer (or other electronic devices) for execution by one or more processors to perform a process according to confidencembodiments.
  • the computer-readable medium may include, but is not limited to, magnetic disks, optical disks, compact disk read-only memory (CD-ROM), and magneto-optical disks, read-only memory (ROM), random access memory (RAM), erasable programmable read-only memory
  • EPROM electrically-erasable programmable read-only memory
  • EEPROM electrically-erasable programmable read-only memory
  • magnet or optical cards flash memory, or other type of computer-readable medium suitable for storing electronic instructions.
  • embodiments may also be downloaded as a computer program product, wherein the program may be transferred from a remote computer to a requesting computer.
  • element A may be directly coupled to element B or be indirectly coupled through, for example, element C.
  • a component, feature, structure, process, or characteristic A “causes” a component, feature, structure, process, or characteristic B, it means that "A” is at least a partial cause of "B” but that there may also be at least one other component, feature, structure, process, or characteristic that assists in causing "B.”
  • the specification indicates that a component, feature, structure, process, or characteristic "may”, “might”, or “could” be included, that particular component, feature, structure, process, or characteristic is not required to be included. If the specification or claim refers to "a” or “an” element, this does not mean there is only one of the described elements.
  • An embodiment is an implementation or example.
  • Reference in the specification to "an embodiment,” “one embodiment,” “some embodiments,” or “other embodiments” means that a particular feature, structure, or characteristic described in connection with the embodiments is included in at least some embodiments, but not necessarily all embodiments.
  • the various appearances of "an embodiment,” “one embodiment,” or “some embodiments” are not necessarily all referring to the same embodiments. It should be appreciated that in the foregoing description of exemplary embodiments, various features are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various novel aspects.
  • a camera assembly includes one or more lenses to transmit light from a scene; and an image sensor to simultaneously capture multiple sub-images of the scene via the one or more lenses, the sub-images including a first sub-image and a second sub-image.
  • a processing element is to process the sub-images sensed by the image sensor, the processing of the sub-images includes combining at least the first sub-image and the second sub-image to generate a combined image of the scene.
  • the image sensor is a wide aspect ratio image sensor that is smaller in a first dimension than in a second dimension.
  • the sub-images are captured on portions of the image sensor along the second dimension, including a first portion of the image sensor to capture the first sub-image and a second portion of the image sensor to capture the second sub-image.
  • the one or more lenses of the camera assembly include multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image.
  • the one or more lenses of the camera assembly include an aspheric lens to transmit light for the first sub-image and the second sub-image.
  • the image sensor of the camera assembly includes a single sensor element. In some embodiments, the image sensor of the camera assembly includes multiple sensor elements.
  • the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.
  • an apparatus includes a camera assembly including one or more lenses to transmit light from a scene, andan image sensor to simultaneously capture multiple sub-images of the scene from the one or more lenses, the sub-images including a first sub-image and a second sub-image; a processing element, the processing element to process the sub-images sensed by the image sensor; and a transceiver to transmit one or more images.
  • the processing of the sub-images includes combining the sub-images to generate a combined image of the scene.
  • the image sensor is a wide aspect ratio image sensor that is smaller in a first dimension than in a second dimension, wherein the sub-images are captured on portions of the image sensor along the second dimension, including a first portion of the image sensor to capture the first sub-image and a second portion of the image sensor to capture the second sub- image.
  • the first sub-image is a sub-image of an upper portion of the scene and the second sub-image is sub-image of a lower portion of the scene.
  • the one or more lenses of the camera assembly include multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image.
  • the one or more lenses of the camera assembly include an aspheric lens to transmit light for the first sub-image and the second sub-image.
  • the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.
  • the apparatus is a mobile device.
  • the mobile device includes a display screen in a front side of the mobile device, the one or more lenses also being in the front side of the mobile device.
  • the combined image is one of multiple images in a data stream, the transceiver being operable to transmit the data stream.
  • a non-transitory computer-readable storage medium having stored thereon data representing sequences of instructions that, when executed by a processor, cause the processor to perform operations including enabling operation of a camera, the camera including one or more lenses and an image sensor; simultaneously capturing by the image sensor multiple sub- images of a scene, the sub-images including a first sub-image and a second sub- image; and processing the sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the sub-images to generate a combined image of the scene.
  • capturing the sub-images includes capturing each sub-image on a different portion of the image sensor, including capturing the first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
  • the one or more lenses include: multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image; oran aspheric lens to transmit light for the first sub-image and the second sub-image.
  • the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.
  • the processing of the sub-images further includes image stabilization for the sub-images.
  • a method includes: enabling operation of a camera assembly, the camera assembly including one or more lenses and an image sensor;enabling a video communication; simultaneously capturing by the image sensor a multiple sub-images of a scene, the sub-images including a first sub-image and a second sub-image;processing the sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the sub-images to generate a combined image of the scene; andtransmitting the image as a part of a data stream for the video communication.
  • capturing the sub-images includes capturing each sub-image on a different portion of the image sensor, including capturing the first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
  • the one or more lenses include:multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image; oran aspheric lens to transmit light for the first sub-image and the second sub-image.
  • the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image. In some embodiments, the processing of the sub-images further includes image stabilization for the plurality of sub-images.
  • an apparatus includes a means for enabling operation of a camera, the camera including one or more lenses and an image sensor; a means for simultaneously capturing by the image sensor a plurality of sub-images of a scene, the plurality of sub-images including a first sub-image and a second sub-image; and a means for processing the plurality of sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the plurality of sub-images to generate a combined image of the scene.
  • capturing the plurality of sub-images includes capturing each sub-image on a different portion of the image sensor, including capturing the first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
  • the one or more lenses of the apparatus include: a plurality of lenses, including a first lens to transmit light for the first sub- image and a second lens to transmit light for the second sub-image; oran aspheric lens to transmit light for the first sub-image and the second sub-image.
  • the processing of the sub-images by the means for processing further includes correction of spatial separation of the first sub-image and the second sub-image. In some embodiments, the processing of the sub-images by the means for processing further includes image stabilization for the plurality of sub-images.

Abstract

A camera assembly (400,500) includes one or more lenses (410,415,510) to transmit light from a scene, and an image sensor (420,520) to simultaneously capture a plurality of sub-images of the scene via the one or more lenses (410,415,510). The plurality of sub-images including a first sub-image (430,530) and a second sub-image (435,535). A processing element is to process the sub-image sensed by the image sensor (320,370), the processing of the sub-images includes combining at least the first sub-image (430,530) and the second sub-image (435,535) to generate a combined image of the scene.

Description

CAMERA TO CAPTURE MULTIPLE SUB-IMAGES FOR GENERATION OF AN IMAGE
TECHNICAL FIELD
[0001] Embodiments described herein generally relate to the field of electronic devices and, more particularly, a camera to capture multiple sub-images for the generation of an image.
BACKGROUND
[0002] As the use of mobile devices has expanded, the demands on the image capture capabilities of such devices have increased. Where low-resolution cameras for occasional photos were once acceptable in mobile devices, modern devices require cameras that are capable of high-resolution photography in difficult conditions, such as low light environments. These requirements are being applied to the front-facing cameras of such devices, which face the user of the devices, as well as the rear-facing cameras as devices are increasingly being used for video communications, including use to capture images of users for Skype™ and
MicrosoftLync® communications.
[0003] At the same time, in order to provide better user experience in using mobile devices, newer mobile devices have larger displays than older devices, with the display covering more of the front face of mobile devices than was true in older mobile devices.
[0004] However, the improvements in mobile device displays have had the side effect of limiting the space for placement of a camera on the front side of a mobile device because the bezel of such devices has been reduced to allow for an increased display sizes for a particular size of mobile device.
BRIEF DESCRIPTION OF THE DRAWINGS
[0005] Embodiments described here are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings in which like reference numerals refer to similar elements.
[0006] Figure lis an illustration of mobile devices to include an embodiment of a camera assembly; [0007] Figures 2A and 2B illustrates examples of conventional lenses and image sensors;
[0008] Figure 3 A is an illustration of elements of an embodiment of a camera assembly including multiple lenses;
[0009] Figure 3B is an illustration of elements of an embodiment of a camera assembly including an aspheric lens;
[0010] Figure 4 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including multiple lenses;
[0011] Figure 5 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including an aspheric lens;
[0012] Figure 6 is an illustration of stitching sub-images together to generate a combined image according to an embodiment;
[0013] Figure 7 is an illustration of an embodiment of a mobile device including a camera assembly;
[0014] Figure 8 is a flow chart to illustratean embodiment of a process to generate a combined image from captured sub-images; and
[0015] Figure 9illustrates an apparatus or system to generate combined images from captured sub-images.
DETAILED DESCRIPTION
[0016] Embodiments described herein are generally directed toa camera to capture multiple sub-images for the generation of an image.
[0017] For the purposes of this description:
[0018] "Mobile device" means a smartphone, smartwatch, tablet computer, handheld computer, mobile Internet device, wearable technology, or other mobile apparatus that includes processing ability and communication ability.
[0019] "Camera" or "camera assembly" means an apparatus or subsystem for capturing images, including still photographs and video images. A camera assembly may further include processing of images.For purposes of a mobile device, a camera having a lens on a first side of the mobile device,the first side includes a main display, may be referred to as a front- facing camera and a camera having a lens on a second side of the mobile device, the second side not including a main display, may be referred to as a rear-facing camera. A camera includes, but is not limited to, a lens and an image sensor. A camera may further include a processing unit.
[0020] "Lens" means one or more optical elements for transmission of light from a scene to an image sensor. Each optical element of a lens may be a simple lens (a single element to refract light), a mirror, or other element to transmit or otherwise affect light. A lens including multiple lens elements may be referred to as a compound lens.
[0021] "Aspheric lens" means a lens including at least one simple lens with surface profiles that are not portions of a sphere or cylinder. In a particular example, an aspheric lens may include a lens in an oval shape with one dimension that is larger than another, or, stated in another way, with a diameter through the semi-major axis (the largest axis of the oval) that is larger than a diameter through the semi-minor axis (the smallest axis of the oval).
[0022] Cameras on mobile devices are increasing used for functions that require high performance. For example, a camera on a mobile device is requiredto provide good low light performance in order to be certified for certain video communicationoperations, including Skype™ and MicrosoftLync® certification. Further, it may be expected that certification requirements will expand andbecome more stringent in the future in order to provide improved user experience in video communications.
[0023] Another aspect of improved camera performance is the increasing resolution of mobile device cameras. The increase in camera resolution provides improved image clarity, butmay havea negative impact on low light performance. In general, a larger image sensor is needed to maintain the low light performance as the number of pixelsfor the image sensor increases. Other technologies are also driving the need for low light performance for cameras, such as facial recognition of the end user of a device to effect computer login authentication.
[0024] However, the location of a camera in an apparatus, in particular a mobile device, is limited by the physical dimensions of the apparatus. In particular, the bezel width of a forward side ofa mobile device (where "bezel" refers to the portion of the device around or beside a display screen or input elements on a front side of the device), where a camera lens would generally be installed, limits the size of the lens and optical sensor in at least one dimension (referred to herein as the Y dimension of the mobile device).
[0025] In camera assemblies, an increase in the size of the camera image sensor is a commonmodification toenhance the low light performanceof the camera. However, with the physical limitations of a mobile device, a simple increase in size of the image sensor may not be practical. Other possible conventional solutions include better and faster lenses and BSI (backside illumination). However, these conventional solutions may not be sufficient to provide acceptable low light performance in a device such as mobile device having a limited physical dimension for the camera.
[0026] In general, designing a lens for animage circle becomes increasingly difficult as the image circle increases in size. The larger lens typically uses more lens elements and may require more expensive lens elements because of the need for more aspheric or special glasses to keep the image sharp and evenly lit over the larger image circle. For these reasons, the lens for a larger image sensor is generally larger, heavier and more expensive. Thus, in general creating a lens for a smaller sensor is easier and less expensive to produce than a lens for a larger sensor at the same level of lens performance. This factor is true for lenses for medium and large format cameras as well as lenses for small format cameras, such as mobile device camera assemblies.
[0027] In some embodiments, an apparatus, system, or process includes wide image sensing to simultaneously capture multiple sub-images. In some embodiments, a camera provides enhanced low light performance while reducing impact to the overall system's Y-dimension through use of wide image sensing to capture multiple sub-images. In some embodiments, a camera includes wide aspect ratio image sensing together with modified lens design and image processing to enable both a compact form factor and enhanced low light performance foruse in, for example, mobile devices or other devices with limited physical size.
[0028] In some embodiments, a camera assembly includes one or more lenses to transmit multiple portions of an image (referred to herein as sub-images) of a sceneon one or more image sensors, the one or more image sensors having a width sufficient to receive the multiplesub-images. The sub-images may overlap.In some embodiments, a processor or other image processing element provides for combining at least a first sub-image and a second sub-image to generate a combined image. In some embodiments, the one or more lenses transmit at least a first sub- image and a second sub-image that differ in a first dimension, such as a vertical differentiation of the first and second sub-images, in order that the first sub-image and the second sub-image are arranged along a second dimension, such as a horizontal arrangement, onto the one or more image sensors. In some embodiments, a processing element processes the sub-images including stitching the sub-images together to generate a combined image of the original scene. While the examples described and illustrated herein illustrate two sub-images, embodiments are not limited to two sub-images, and may include a larger number of sub-images.
[0029] Figure lis an illustration of mobile devices to include an embodiment of a camera assembly. As illustrated in Figure 1, a mobile device may include, but is not limited to, a smart phone 110 or a tablet computer 130. The smart phone 110 includes a display screen 115 on a first side of the smart phone, and the tablet computer 130 includes a display screen 135 on a first side of the tablet computer.Each mobile device 110, 130 includes a bezel, or Y-dimension 120, 140,of a certain size. In the particular examples shown in Figure 1, the available Y- dimension 120 for the smart phone is 12 mm (millimeters) and the available Y- dimension 140 for the tablet computer 130 is 18.5 mm.
[0030] However, the usable Y-dimension of each mobile device 120, 140 is actually smaller than the indicated sizes because of the width of the cover material. Depending on the curvature of the mobile device cover, the shape of the cover may further limit the usable Y-dimension.
[0031] One potential solution to the installation of a larger image sensor is to increase the Y-dimension of a mobile device, but this choice reduces the space available for the display screen 115, 135 of the mobile device, thereby reducing the functionality and attractiveness of the device for users.
[0032] In some embodiments, the mobile devices 110, 130 include camera assemblies that provide improved low-light performance though use of wide image sensing to simultaneously capture multiple sub-images of a scene, with two or more sub-images being combined to generate a combined image of the scene.
[0033] Figures 2A and 2B illustrates examples of conventional lenses and image sensors. Figure 2A illustrates front and side views of a first conventional camera assembly 200 including a small aspect lens 210, which in this example is 7.5 mm in diameter, and an image sensor 220. The camera assembly 200 may be utilized in, for example, a mobile device. However, an image sensor of this size may not provide sufficient low light performance for certain purposes, including video communications.
[0034] To improve low light performance, Figure 2B illustrates front and side views of a second conventional camera assembly 250, the camera assembly 250 including a larger lens 260 and image sensor 270 to allow for the collection of a greater quantity of light. In this example, the diameter of the lens 260 has been increased to 14 mm.
[0035] However, while the second camera assembly 250 may provide improved low light performance in comparison with the first camera assembly 200, such camera assembly 250 may be too large to fit within the physical constraints of a device. In particular, the lens and image sensor may be too large to fit within the bezel area (the available Y-dimension) of the front side (the display side) of a mobile device. For this reason, the camera assembly 250 may not be usable as a front- facing camera of such a device.
[0036] Figure 3 A is an illustration of elements of an embodiment of a camera assembly including multiple lenses.As illustrated in a front view in Figure 3 A, an embodiment of a camera assembly 300 includes multiple lenses to simultaneously capture multiple sub-images, such as a first lens 310 to transmit light for a first sub-image and a second lens 315 to transmit light for a second sub-image. In some embodiments, the camera assembly 300 includes a wide image sensor 320, wherein the sub-image from the first lens 310 is directed to a first portion of the image sensor 320 and the second sub-image from the second lens 315 is directed to a second portion of the image sensor 320. In some embodiments, the wide image sensor 320 includes a single sensor element. In some embodiments, the wide image sensor 320 includes multiple sensor elements, such as a first sensor element for the capture of the first sub-image and a second sensor element for the capture of the second sub-image.
[0037] In some embodiments, the multiple lenses 310, 315 reduce the impact in a first dimension bythe arrangement side by side along a second dimension. As illustrated in Figure 3A each of the lenses 310, 315 is 7.5 mm in diameter, and thus the total width in the Y-dimension is 7.5 mm because the multiple lenses are arranged in the X-dimension. However, with two lenses the amount of light being transmitted is increased in comparison with a single lens. The light transmittedhas essentially been doubled, but the effective increase in captured light may vary depending on the specific implementation. Further, the image sensor 320 is increased in size only in the X-dimension, thus reducing any impact in the Y- dimension for the camera assembly while improving the low light performance of the camera assembly 300.
[0038] Figure 3B is an illustration of elements of an embodiment of a camera assembly including an aspheric lens. As illustrated in a front view in Figure 3B, an embodiment of a camera assembly includes an aspheric lens 360 to simultaneously capture multiple sub-images. In some embodiments, the aspheric lens 360 is formed to generate multiple sub-images from a scene, the aspheric lens 360 being shaped such that the lens is substantially larger in a second dimension than the lens is in a first dimension. In some embodiments, camera assembly includes a wide image sensor 370, wherein the a first sub-image transmitted by the aspheric lens 360 is directed to a first portion of the image sensor 370 and a second sub-image transmitted by the aspheric lens 370 is directed to a second portion of the image sensor. In some embodiments, the wide image sensor 370 includes a single sensor element. In some embodiments, the wide image sensor 370 includes multiple sensor elements, such as a first sensor element for the capture of the first sub-image and a second sensor element for the capture of the second sub-image.
[0039] In some embodiments, the aspheric lens 360 reduces the impact in a first dimension by the shape and arrangement of the lens. As illustrated in Figure 3B, the aspheric lens 360 is 7.5 mm in width in the Y-dimension, while such lens is substantially larger in the X-dimension. However, the lens captures additional light in comparison with a spherical lens. Further, the image sensor 370 is increased in size in the X-dimension, thus reducing any impact in the Y-dimension for the camera assembly 350 while improving the low light performance of the camera assembly.
[0040] Figure 4 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including multiple lenses. In some embodiments, a camera assembly 400 includes multiple lenses, including a first lens 410 and a second lens 415, to simultaneously capture multiple sub-images, such as a first sub- image 430 transmitted via the first lens 410 and a second sub-image 435 transmitted via the second lens 415. In some embodiments, the first lens and the second lens are arranged in a second dimension (X-dimension, or horizontal in Figure 4) to reduce width in a first dimension (Y-dimension, or vertical in Figure 4). In some embodiments, the camera assembly 400 further includes a wide aspect ratio image sensor 420, wherein the image sensor 420 is sufficiently sized to allow the capture of the first sub-image 430 adjacent to the second sub-image 435 in the X-dimension, or horizontally in the illustration. In some embodiments, the wide aspect ratio image sensor 420 includes a single sensor element, and in some embodiments the image sensor includes multiple sensor elements.
[0041] In some embodiments, as illustrated in Figure 4 the first sub- image 430 and second sub-image 435 are portions of the full image arranged vertically (along the first dimension), such as the first sub-image 430 being a top portion of a full image of the scene and the second sub-image 435 being a bottom portion of the full image of the scene. In some embodiments, the multiple sub- images are captured in a horizontal arrangement (along the second dimension) on the image sensor. Thus, the first sub-image 430, being a first vertically arranged sub-image (the top sub-image in Figure 4), is arranged in a first horizontal position (a left position in Figure 4), and the second sub-image 435, being a second vertically arranged sub-image (the bottom sub-image in Figure 4), is arranged in a second horizontal position (a right position in Figure 4).
[0042] In some embodiments, the multiple sub-images 430, 435 are processed to generate a combined image 450 of the original scene. In some embodiments, the sub-images 430, 435 are rearranged in processing to return such sub-images to their original position in the combined image 450 of the scene. In the example illustrated in Figure 4, the first sub-image 430 captured in the left portion of the image sensor 420 is processed to become a top portion of the combined image 450, and the second sub-image 435 captured in the right portion of the image sensor 420 is processed to become a top portion of the combined image 450. In some embodiments, processing of sub-images includes image stabilization, correction of sensor/lens spatial separation, and merging ofsub-images into the combined image 450.
[0043] Figure 5 is an illustration of the capture of an image utilizing an embodiment of a camera assembly including an aspheric lens. In some
embodiments, a camera assembly 500 includes an aspheric lens 510 to
simultaneously capture multiple sub-images, such as a first sub-image 530 and a second sub-image 535 captured via the aspheric lens. In some embodiments, the aspheric lens is an oval lens or other lens shape that is larger in one dimension than in a second dimension arranged such that the larger dimension (such as the semi- major axis of an oval-shaped lens) is aligned in a second dimension (X-dimension, or horizontal in Figure 5) and the smaller dimension (such as the semi-minor axis of an oval-shaped lens) is aligned in a first dimension (Y-dimension, or vertical in Figure 5) to reduce width in the first dimension. In some embodiments, the camera assembly 500 further includes a wide aspect ratio image sensor 520, wherein the image sensor 520 is sufficiently wide to allow the capture of the first sub-image 530 adjacent to the second sub-image 535 in the X-dimension, or horizontally in the illustration. In some embodiments, the wide aspect ratio image sensor 520 includes a single sensor element, and in some embodiments the image sensor includes multiple sensor elements.
[0044] In some embodiments, as illustrated in Figure 5 the first sub- image 530 and second sub-image 535 are portions of the full image arranged vertically (along the first dimension), such as the first sub-image 530 being a top portion of a full image of the scene and the second sub-image 535 being a bottom portion of the full image of the scene. In some embodiments, the multiple sub- images are captured in a horizontal arrangement (along the second dimension) on the image sensor. Thus, the first sub-image 530, being a first vertically arranged sub-image (the top sub-image in Figure 5), is arranged in a first horizontal position (a left position in Figure 5), and the second sub-image 535, being a second vertically arranged sub-image (the bottom sub-image in Figure 5), is arranged in a second horizontal position (a right position in Figure 5).
[0045] In some embodiments, the multiple sub-images 530, 535 are processed to generate a combined image 550 of the original scene. In some embodiments, the sub-images 530, 535 are rearranged in processing to return such sub-images to their original position in the combined image 550 of the scene. In the example illustrated in Figure 5, the first sub-image 530 captured in the left portion of the image sensor 520 is processed to become a top portion of the combined image 550, and the second sub-image 535 captured in the right portion of the image sensor 520 is processed to become a lower(or bottom)portion of the combined image 550.
[0046] Figure 6 is an illustration of combining sub-images to generate a combined image according to an embodiment. In some embodiments, an image sensor 620 is a wide aspect ratio with a width (Y-dimension) that is significantly shorter than a length (X-dimension). In some embodiments, a first sub-image 630 is captured in a first portion of the image sensor 620 and a second sub-image 635 is captured in a second portion of the image sensor 620.
[0047] In some embodiments, the first sub-image 630 and the second sub-image 635 may have spatial separation 640 because of imperfect operation of the one or more lenses or the image sensor. In some embodiments, the multiple sub-images 630, 635 are processed, wherein the processing may include image stabilization, correction of sensor or lens spatial separation, such as cropping one or more of the sub-images to remove over-lapping between the sub-images and aligning the cropped sub-images,and combining the sub-images generate a combined image 650.
[0048] Figure 7 is ablock diagram of an embodiment of a mobile device including a camera assembly. In some embodiments, a mobile device 700 includes a camera assembly including one or more lenses 710, wherein the one or more lenses may include multiple lenses, such as the multiple lenses 410, 415 illustrated in Figure 4 or the aspheric lens 510 illustrated in Figure 5, and a wide aspect ratio image sensor 715, such as the wide aspect image 420 or 520 illustrated in Figure 4 or Figure 5 respectively. In some embodiments, the camera assembly 705 is a front- facing camera with the one or more lenses of the camera assembly on a same side as a display screen 760, and thus the camera assembly is located in a limited Y- dimension space of the mobile device. The one or more lenses 710 and the wide aspect ratio image sensor 715 are arranged such that a longer dimension of such elements is arranged in an X-dimension of the mobile device 700 and a shorter dimension of such elements is arranged in the Y-dimension of the mobile device 700.
[0049] In some embodiments, the camera assembly 705 is operable to simultaneously capture multiple sub-images 730 on the image sensor, wherein the sub-images are captured in an arrangement along the X-dimension. In some embodiments, a processor 770, which may be a general-purpose processor, a dedicated graphics processor, or other image processing element of the mobile device 700, receives the multiple sub-images 730 for processing. In some embodiments, the processing of the sub-images may be as illustrated in Figure 6, wherein the processing of the sub-images includes image stabilization, correction of lens or image sensor spatial separation, and merging the two images into a combined image 750. In some embodiments, the processor may provide the combined image for display on the screen 760, for transmission using a transceiver 780 to another device, including transmission of the image in a video stream for a video communication, for storage in a memory 790, or a combination of such operations.
[0050] Figure 8 is a flow chart to illustrate an embodiment of a process to generate a combined image from captured sub-images. In some embodiments, a process includes enabling a camera function 800, and may further include enabling a video communication service (such as Skype or Lync) or other function 801. In some embodiments, the process includes capturing an image 802, which may be a single image or one image in a series of images, such as a video stream. In some embodiments, the capture of the image includes simultaneously capturing multiple sub-images, including capturing a first sub-image using a first portion of an image sensor and capturing a second sub-image using a second portion of the image sensor, wherein the image sensor is a wide aspect ratio image sensor. For example, the first sub-image may be an upper portion of a full image of a scene and the second sub- image may be a lower portion of the full image of the scene.
[0051] In some embodiments, the process further includes processing the sub-images 810, wherein the processing may include providing image stabilization 812, correcting sensor or lens spatial separation (which may include eliminating overlap between the sub-images and aligning the sub-images for combination) 814, and combining the multiple sub-images, including the first sub-image and the second sub-image, to generate a combined image of the original scene 816. In some embodiments, the process may further include displaying the combined image, transmitting the combined image to another user (such as, for example, if the combined image is one image in a series of images for a video stream), storing the combined image in a memory, or a combination of such operations 820. If there is an additional new image for capture 822, such as if a video communicating session is continuing, then the process returns to the capturing of the next image 802.
[0052] Figure 9illustrates an apparatus or system to generate combined images from captured sub-images.In this illustration, certain standard and well- known components that are not germane to the present description are not shown. Elements shown as separate elements may be combined, including, for example, an SoC (System on Chip) combining multiple elements on a single chip. The apparatus or system may include, but is not limited to, a mobile device.
[0053] Under some embodiments, theapparatus or system900 (referred to generally herein as an apparatus)includes an interconnect or crossbar 905 or other communication means for transmission of data.The interconnect 905 is illustrated as a single interconnect for simplicity, but may represent multiple different
interconnects or buses and the component connections to such interconnects may vary. The interconnect905 shown in Figure 9 is an abstraction that represents any one or more separate physical buses, point-to-point connections, or both connected by appropriate bridges, adapters, or controllers.
[0054] In some embodiments, the apparatus 900 includes a camera assembly 970, wherein the camera assembly includes one or more lenses 972 to collect light for multiple sub-images of a scene and a wide aspect ratio image sensor 974 to capture the plurality of sub-images. The camera assembly 970 may include the camera assembly 400 illustrated in Figure 4 or the camera assembly 500 illustrated in Figure 5.
[0055] The apparatus900may include a processing means such as the one or more processors 910 coupled to the interconnect 905 for processing information. The processors 910 may comprise one or more physical processors and one or more logical processors. In some embodiments, the processors may include a general-purpose processor. In some embodiments, the processors 910 may include an additional image processing element. In some embodiments, functions of the processors 910 include the processing of the sub-images captured by the camera assembly 970 to generate a combined image.
[0056] In some embodiments, the apparatus900includes one or more transmitters or receivers 940coupled to the interconnect 905. In some embodiments, the apparatus900may include one or more antennas 942 for the transmission and reception of data via wireless communication. In some embodiments, the apparatus 900 includes one or more ports 945 for the transmission and reception of data via wired communications. In some embodiments, the data that is transmitted and received includes image data, including the transmission of combined images generated from collected sub-images by the camera assembly 970, where the transmitted combined images may be a stream of video images.
[0057] In someembodiments, the apparatus900further comprises a random access memory (RAM) or other dynamic storage device or element as a main memory 915 for storing information and instructions to be executed by the processors 910. The apparatus900may include one or more non- volatile memory elements 925, including, for example, flash memory, for the storage of certain elements. The apparatus 900 also may comprise a read only memory (ROM) 930 or other static storage device for storing static information and instructions for the processors 910, and data storage 935, such as a solid state drive, for the storage of data. In some embodiments, memory of the apparatus900 may include storage for combined images generated from the sub-images collected by the camera assembly 970.
[0058] In some embodiments, the apparatus 900includes one or more input devices 950 for the input of data, including hard and soft buttons, a joy stick, a mouse or other pointing device, voice command system, or gesture recognition system. In some embodiments, the apparatus900 includes an output display 955, where the display 955 may include a liquid crystal display (LCD) or any other display technology, for displaying information or content to a user. In some environments, the display 955 may include a touch-screen that is also utilized as at least a part of an input device950. In some embodiments, the display 955 may display combined images generated from the sub-images collected by the camera assembly 970.
[0059] The apparatus900may also comprise a batteryor other power source 960, which may include a solar cell, a fuel cell, a charged capacitor, near field inductive coupling, or other system or device for providing or generating power in the apparatus 900. The power provided by the power source 960 may be distributed as required to elements of the apparatus 900.
[0060] In the description above, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the described embodiments. It will be apparent, however, to one skilled in the art that embodiments may be practiced without some of these specific details. In other instances, well-known structures and devices are shown in block diagram form. There may be intermediate structure between illustrated components.The components described or illustrated herein may have additional inputs or outputs that are not illustrated or described.
[0061] Various embodiments may include various processes. These processes may be performed by hardware components or may be embodied in computer program or machine-executable instructions, which may be used to cause a general-purpose or special-purpose processor or logic circuits programmed with the instructions to perform the processes. Alternatively, the processes may be performed by a combination of hardware and software.
[0062] Portions of various embodiments may be provided as a computer program product, which may include a computer-readable medium having stored thereon computer program instructions, which may be used to program a computer (or other electronic devices) for execution by one or more processors to perform a process according to certainembodiments. The computer-readable medium may include, but is not limited to, magnetic disks, optical disks, compact disk read-only memory (CD-ROM), and magneto-optical disks, read-only memory (ROM), random access memory (RAM), erasable programmable read-only memory
(EPROM), electrically-erasable programmable read-only memory (EEPROM), magnet or optical cards, flash memory, or other type of computer-readable medium suitable for storing electronic instructions. Moreover, embodiments may also be downloaded as a computer program product, wherein the program may be transferred from a remote computer to a requesting computer.
[0063] Many of the methods are described in their most basic form, but processes can be added to or deleted from any of the methods and information can be added or subtracted from any of the described messages without departing from the basic scope of the present embodiments. It will be apparent to those skilled in the art that many further modifications and adaptations can be made. The particular embodiments are not provided to limit the concept but to illustrate it. The scope of the embodiments is not to be determined by the specific examples provided above but only by the claims below.
[0064] If it is said that an element "A" is coupled to or with element "B," element A may be directly coupled to element B or be indirectly coupled through, for example, element C. When the specification or claims state that a component, feature, structure, process, or characteristic A "causes" a component, feature, structure, process, or characteristic B, it means that "A" is at least a partial cause of "B" but that there may also be at least one other component, feature, structure, process, or characteristic that assists in causing "B." If the specification indicates that a component, feature, structure, process, or characteristic "may", "might", or "could" be included, that particular component, feature, structure, process, or characteristic is not required to be included. If the specification or claim refers to "a" or "an" element, this does not mean there is only one of the described elements.
[0065] An embodiment is an implementation or example. Reference in the specification to "an embodiment," "one embodiment," "some embodiments," or "other embodiments" means that a particular feature, structure, or characteristic described in connection with the embodiments is included in at least some embodiments, but not necessarily all embodiments. The various appearances of "an embodiment," "one embodiment," or "some embodiments" are not necessarily all referring to the same embodiments. It should be appreciated that in the foregoing description of exemplary embodiments, various features are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various novel aspects. This method of disclosure, however, is not to be interpreted as reflecting an intention that the claimed embodiments requires more features than are expressly recited in each claim. Rather, as the following claims reflect, novel aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the claims are hereby expressly incorporated into this description, with each claim standing on its own as a separate embodiment.
[0066] In some embodiments, a camera assembly includes one or more lenses to transmit light from a scene; and an image sensor to simultaneously capture multiple sub-images of the scene via the one or more lenses, the sub-images including a first sub-image and a second sub-image. In some embodiments, a processing element is to process the sub-images sensed by the image sensor, the processing of the sub-images includes combining at least the first sub-image and the second sub-image to generate a combined image of the scene.
[0067] In some embodiments, the image sensor is a wide aspect ratio image sensor that is smaller in a first dimension than in a second dimension. In some embodiments, the sub-images are captured on portions of the image sensor along the second dimension, including a first portion of the image sensor to capture the first sub-image and a second portion of the image sensor to capture the second sub-image.
[0068] In some embodiments, the one or more lenses of the camera assembly include multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image. [0069] In some embodiments, the one or more lenses of the camera assembly include an aspheric lens to transmit light for the first sub-image and the second sub-image.
[0070] In some embodiments, the image sensor of the camera assembly includes a single sensor element.In some embodiments, the image sensor of the camera assembly includes multiple sensor elements.
[0071] In some embodiments, the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.
[0072] In some embodiments, an apparatus includes a camera assembly including one or more lenses to transmit light from a scene, andan image sensor to simultaneously capture multiple sub-images of the scene from the one or more lenses, the sub-images including a first sub-image and a second sub-image; a processing element, the processing element to process the sub-images sensed by the image sensor; and a transceiver to transmit one or more images. In some embodiments, the processing of the sub-images includes combining the sub-images to generate a combined image of the scene.
[0073] In some embodiments, the image sensor is a wide aspect ratio image sensor that is smaller in a first dimension than in a second dimension, wherein the sub-images are captured on portions of the image sensor along the second dimension, including a first portion of the image sensor to capture the first sub-image and a second portion of the image sensor to capture the second sub- image.
[0074] In some embodiments, the first sub-image is a sub-image of an upper portion of the scene and the second sub-image is sub-image of a lower portion of the scene.
[0075] In some embodiments, the one or more lenses of the camera assembly include multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image.
[0076] In some embodiments, the one or more lenses of the camera assembly include an aspheric lens to transmit light for the first sub-image and the second sub-image. [0077] In some embodiments, the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.
[0078] In some embodiments, the apparatus is a mobile device. In some embodiments, the mobile device includes a display screen in a front side of the mobile device, the one or more lenses also being in the front side of the mobile device.
[0079] In some embodiments, the combined image is one of multiple images in a data stream, the transceiver being operable to transmit the data stream.
[0080] In some embodiments, a non-transitory computer-readable storage medium having stored thereon data representing sequences of instructions that, when executed by a processor, cause the processor to perform operations including enabling operation of a camera, the camera including one or more lenses and an image sensor; simultaneously capturing by the image sensor multiple sub- images of a scene, the sub-images including a first sub-image and a second sub- image; and processing the sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the sub-images to generate a combined image of the scene.
[0081] In some embodiments, capturing the sub-images includes capturing each sub-image on a different portion of the image sensor, including capturing the first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
[0082] In some embodiments, the one or more lenses include: multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image; oran aspheric lens to transmit light for the first sub-image and the second sub-image.
[0083] In some embodiments, the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.In some embodiments, the processing of the sub-images further includes image stabilization for the sub-images.
[0084] In some embodiments, a method includes: enabling operation of a camera assembly, the camera assembly including one or more lenses and an image sensor;enabling a video communication; simultaneously capturing by the image sensor a multiple sub-images of a scene, the sub-images including a first sub-image and a second sub-image;processing the sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the sub-images to generate a combined image of the scene; andtransmitting the image as a part of a data stream for the video communication.
[0085] In some embodiments, capturing the sub-images includes capturing each sub-image on a different portion of the image sensor, including capturing the first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
[0086] In some embodiments, the one or more lenses include:multiple lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image; oran aspheric lens to transmit light for the first sub-image and the second sub-image.
[0087] In some embodiments, the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image. In some embodiments, the processing of the sub-images further includes image stabilization for the plurality of sub-images.
[0088] In some embodiments, an apparatus includes a means for enabling operation of a camera, the camera including one or more lenses and an image sensor;a means for simultaneously capturing by the image sensor a plurality of sub-images of a scene, the plurality of sub-images including a first sub-image and a second sub-image; anda means for processing the plurality of sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the plurality of sub-images to generate a combined image of the scene.
[0089] In some embodiments, capturing the plurality of sub-images includes capturing each sub-image on a different portion of the image sensor, including capturing the first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
[0090] In some embodiments, the one or more lenses of the apparatus include: a plurality of lenses, including a first lens to transmit light for the first sub- image and a second lens to transmit light for the second sub-image; oran aspheric lens to transmit light for the first sub-image and the second sub-image.
[0091] In some embodiments, the processing of the sub-images by the means for processing further includes correction of spatial separation of the first sub-image and the second sub-image. In some embodiments, the processing of the sub-images by the means for processing further includes image stabilization for the plurality of sub-images.

Claims

CLAIMS What is claimed is:
1. A camera assembly comprising:
one or more lenses to transmit light from a scene; and
an image sensor to simultaneously capture a plurality of sub-images of the scene via the one or more lenses, the plurality of sub-images including a first sub- image and a second sub-image;
wherein aprocessing element is to process the sub-images sensed by the image sensor, the processingof the sub-images includes combining at least the first sub-image and the second sub-image to generate a combined image of the scene.
2. The camera assembly of claim 1, wherein the image sensor is a wide aspect ratio image sensor that is smaller in a first dimension than in a second dimension.
3. The camera assembly of claim 2, wherein the plurality of sub-images are captured on portions of the image sensor along the second dimension, including a first portion of the image sensor to capture the first sub-image and a second portion of the image sensor to capture the second sub-image.
4. The camera assembly of claim 1, wherein the one or more lenses include a plurality of lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image.
5. The camera assembly of claim 1, wherein the one or more lenses include an aspheric lens to transmit light forthe first sub-image and the second sub-image.
6. The camera assembly of claim 1, wherein the image sensor includes a single sensor element.
7. The camera assembly of claim 1, wherein the image sensor includes a plurality of sensor elements.
8. The camera assembly of claim 1, wherein the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub-image.
9. An apparatus comprising:
a camera assembly including:
one or more lenses to transmit light from a scene, and an image sensor to simultaneously capture a plurality of sub-images of the scene from the one or more lenses, the plurality of sub-images including a first sub-image and a second sub-image;
a processing element, the processing element to process the sub-images sensed by the image sensor; and
a transceiver to transmit one or more images;
wherein the processing of the sub-images includes combining the plurality of sub-images to generate a combined image of the scene.
10. The apparatus of claim 9, wherein the image sensor is a wide aspect ratio image sensor that is smaller in a first dimension than in a second dimension, wherein the plurality of sub-images are captured on portions of the image sensor along the second dimension, including a first portion of the image sensor to capture the first sub-image and a second portion of the image sensor to capture the second sub-image.
11. The apparatus of claim 10, wherein the first sub-image is a sub-image of an upper portion of the scene and the second sub-image is sub-image of a lower portion of the scene.
12. The apparatus of claim 9, wherein the one or more lenses include a plurality of lenses, including a first lens to transmit light for the first sub-image and a second lens to transmit light for the second sub-image.
13. The apparatus of claim 9, wherein the one or more lenses include an aspheric lens to transmit light for the first sub-image and the second sub-image.
14. The apparatus of claim 9, wherein the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.
15. The apparatus of claim 9, wherein the apparatus is a mobile device.
16. The apparatus of claim 15, wherein the mobile device includes a display screen in a front side of the mobile device, the one or more lenses also being in the front side of the mobile device.
17. The apparatus of claim 9, wherein the combined image is one of a plurality of images in a data stream, the transceiver being operable to transmit the data stream.
18. A non-transitory computer-readable storage medium having stored thereon data representing sequences of instructions that, when executed by a processor, cause the processor to perform operations comprising:
enabling operation of a camera, the camera including one or more lenses and an image sensor;
simultaneously capturing by the image sensor a plurality of sub-images of a scene, the plurality of sub-images including a first sub-image and a second sub- image; and
processingthe plurality of sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the plurality of sub-images to generate a combined image of the scene.
19. The article of claim 18, wherein capturing the plurality of sub-images includes capturing each sub-image on a different portion of the image sensor, including capturing the first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
20. The article of claim 18, wherein the one or more lenses include:
a plurality of lenses, including a first lens to transmit light for the first sub- image and a second lens to transmit light for the second sub-image; or an aspheric lens to transmit light for the first sub-image and the second sub- image.
21. The article of claim 18, wherein the processing of the sub-images further includes correction of spatial separation of the first sub-image and the second sub- image.
22. The article of claim 18, wherein the processing of the sub-images further includes image stabilization for the plurality of sub-images.
23. A method comprising:
enabling operation of a camera assembly, the camera assembly includingone or more lenses andan image sensor;
enabling a video communication;
simultaneously capturing by the image sensor a plurality of sub-images of a scene, the plurality of sub-images including a first sub-image and a second sub- image;
processing the plurality of sub-images captured by the image sensor, wherein the processing of the sub-images includes combining the plurality of sub- images to generate a combined image of the scene; and
transmitting the image as a part of a data stream for the video
communication.
24. The method of claim 23, wherein capturing the plurality of sub-images includes capturing each sub-image on a different portion of the image sensor, including capturingthe first sub-image on a first portion of the image sensor and capturing the second sub-image on a second portion of the image sensor.
25. The method of claim 23, wherein the one or more lenses include:
a plurality of lenses, including a first lens to transmit light for the first sub- image and a second lens to transmit light for the second sub-image; or
an aspheric lens to transmit light forthe first sub-image and the second sub- image.
EP14887071.0A 2014-03-27 2014-03-27 Apparatus and method for capturing multiple sub-images to generate an image Active EP3123226B1 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2014/074195 WO2015143670A1 (en) 2014-03-27 2014-03-27 Camera to capture multiple sub-images for generation of an image

Publications (3)

Publication Number Publication Date
EP3123226A1 true EP3123226A1 (en) 2017-02-01
EP3123226A4 EP3123226A4 (en) 2017-11-29
EP3123226B1 EP3123226B1 (en) 2019-02-20

Family

ID=54193908

Family Applications (1)

Application Number Title Priority Date Filing Date
EP14887071.0A Active EP3123226B1 (en) 2014-03-27 2014-03-27 Apparatus and method for capturing multiple sub-images to generate an image

Country Status (7)

Country Link
US (1) US9762815B2 (en)
EP (1) EP3123226B1 (en)
JP (1) JP6416925B2 (en)
KR (1) KR101815164B1 (en)
CN (1) CN106062610B (en)
TW (1) TWI551138B (en)
WO (1) WO2015143670A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR3052252B1 (en) * 2016-06-07 2019-05-10 Thales OPTRONIC VISION EQUIPMENT FOR A TERRESTRIAL VEHICLE
GB2573593B (en) 2018-05-10 2021-11-03 Samsung Electronics Co Ltd Augmented reality rendering method and apparatus
WO2022045072A1 (en) 2020-08-28 2022-03-03 富士フイルム株式会社 Imaging device, imaging instruction method, and imaging instruction program

Family Cites Families (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5581409A (en) * 1994-09-30 1996-12-03 Republic Lens Co., Inc. Imaging system to combine disparate fields of view
JPWO2003015400A1 (en) * 2001-08-07 2004-12-02 日立マクセル株式会社 The camera module
EP1677518A4 (en) * 2003-10-22 2009-12-02 Panasonic Corp Imaging device and method of producing the device, portable apparatus, and imaging element and method of producing the element
JP2005303694A (en) * 2004-04-13 2005-10-27 Konica Minolta Holdings Inc Compound eye imaging device
KR100716829B1 (en) * 2005-08-10 2007-05-09 삼성전기주식회사 Mobile camera optical system and method for producing image thereof
KR100796849B1 (en) * 2006-09-04 2008-01-22 삼성전자주식회사 Method for photographing panorama mosaics picture in mobile device
CN101207784A (en) 2006-12-22 2008-06-25 北京上行逶式信息公司 Mobile phone video device
US7683962B2 (en) 2007-03-09 2010-03-23 Eastman Kodak Company Camera using multiple lenses and image sensors in a rangefinder configuration to provide a range map
CN101378559A (en) * 2007-08-30 2009-03-04 联想移动通信科技有限公司 Mobile phone capable of panorama shooting by pick-up head combination
CN101400002A (en) * 2007-09-24 2009-04-01 鸿富锦精密工业(深圳)有限公司 Stereo video apparatus
US7956924B2 (en) * 2007-10-18 2011-06-07 Adobe Systems Incorporated Fast computational camera based on two arrays of lenses
US8355042B2 (en) 2008-10-16 2013-01-15 Spatial Cam Llc Controller in a camera for creating a panoramic image
JP2011523538A (en) * 2008-05-20 2011-08-11 ペリカン イメージング コーポレイション Image capture and processing using monolithic camera arrays with different types of imagers
JP2010268343A (en) * 2009-05-18 2010-11-25 Olympus Imaging Corp Photographing device and photographing method
JP5423235B2 (en) * 2009-08-20 2014-02-19 ソニー株式会社 Imaging device
KR101649635B1 (en) 2009-11-06 2016-08-19 엘지전자 주식회사 Portable terminal
CN102131044B (en) * 2010-01-20 2014-03-26 鸿富锦精密工业(深圳)有限公司 Camera module
CN102263926A (en) * 2010-05-31 2011-11-30 鸿富锦精密工业(深圳)有限公司 Electronic equipment and image processing method thereof
US20130265459A1 (en) * 2011-06-28 2013-10-10 Pelican Imaging Corporation Optical arrangements for use with an array camera
CN202143153U (en) * 2011-07-27 2012-02-08 天津三星光电子有限公司 Digital camera
US10412367B2 (en) * 2011-08-05 2019-09-10 3D Media Ltd Multi-lens camera with a single image sensor
CN103167223A (en) * 2011-12-09 2013-06-19 富泰华工业(深圳)有限公司 Mobile device with wide-angle shooting function and image acquisition method thereof
JP6123274B2 (en) * 2012-03-08 2017-05-10 株式会社リコー Imaging device
US20130258044A1 (en) * 2012-03-30 2013-10-03 Zetta Research And Development Llc - Forc Series Multi-lens camera
JP2014123896A (en) * 2012-12-21 2014-07-03 Olympus Imaging Corp Imaging apparatus, imaging method and program

Also Published As

Publication number Publication date
JP6416925B2 (en) 2018-10-31
KR101815164B1 (en) 2018-01-04
US20160044251A1 (en) 2016-02-11
KR20160113682A (en) 2016-09-30
TWI551138B (en) 2016-09-21
TW201543890A (en) 2015-11-16
CN106062610A (en) 2016-10-26
US9762815B2 (en) 2017-09-12
EP3123226B1 (en) 2019-02-20
EP3123226A4 (en) 2017-11-29
CN106062610B (en) 2019-09-17
WO2015143670A1 (en) 2015-10-01
JP2017514326A (en) 2017-06-01

Similar Documents

Publication Publication Date Title
CN111033569B (en) Apparatus for editing image using depth map and method thereof
CN112889266B (en) Electronic device including camera module in display and method for compensating image around camera module
EP3435655B1 (en) Electronic device for acquiring image using plurality of cameras and method for processing image using the same
EP3067746B1 (en) Photographing method for dual-camera device and dual-camera device
US10645278B2 (en) Imaging control apparatus and control method therefor
KR102344104B1 (en) The Electronic Device Controlling the Effect for Displaying of the Image and Method for Displaying the Image
CN112840634B (en) Electronic device and method for obtaining image
CN103369246B (en) Camera resolution conversion device
US11641524B2 (en) Electronic device and method for displaying image in electronic device
KR20200094500A (en) Electronic device and method for processing line data included in image frame data into multiple intervals
US9762815B2 (en) Camera to capture multiple sub-images for generation of an image
WO2021185374A1 (en) Image capturing method and electronic device
CN116711316A (en) Electronic device and operation method thereof
CN109302600A (en) A kind of stereo scene filming apparatus
US11689816B2 (en) Imaging apparatus, operation method of imaging apparatus, and program
US11636708B2 (en) Face detection in spherical images
CN116530090A (en) Method for taking pictures by using multiple cameras and device thereof
CN117729320A (en) Image display method, device and storage medium
CN114979458A (en) Image shooting method and electronic equipment
CN115552399A (en) Biometric authentication system and electronic device thereof
CN102955350A (en) Stereoscopic image acquisition device
CN102213911A (en) Three-dimensional image capture device and method

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20160822

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

DAX Request for extension of the european patent (deleted)
A4 Supplementary search report drawn up and despatched

Effective date: 20171027

RIC1 Information provided on ipc code assigned before grant

Ipc: H04N 5/225 20060101ALI20171023BHEP

Ipc: G02B 13/06 20060101AFI20171023BHEP

Ipc: G02B 13/18 20060101ALI20171023BHEP

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

INTG Intention to grant announced

Effective date: 20180917

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602014041593

Country of ref document: DE

REG Reference to a national code

Ref country code: AT

Ref legal event code: REF

Ref document number: 1098969

Country of ref document: AT

Kind code of ref document: T

Effective date: 20190315

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: NL

Ref legal event code: FP

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190620

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190520

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190520

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190620

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190521

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 1098969

Country of ref document: AT

Kind code of ref document: T

Effective date: 20190220

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602014041593

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190327

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20190331

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

26N No opposition filed

Effective date: 20191121

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190331

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190331

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190327

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190331

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190327

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20140327

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190220

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20230216

Year of fee payment: 10

Ref country code: DE

Payment date: 20230222

Year of fee payment: 10

P01 Opt-out of the competence of the unified patent court (upc) registered

Effective date: 20230518

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: NL

Payment date: 20231215

Year of fee payment: 11

Ref country code: FR

Payment date: 20231212

Year of fee payment: 11