EP3168681B1 - Verfahren zur automatischen fokussierung einer digitalkamera - Google Patents
Verfahren zur automatischen fokussierung einer digitalkamera Download PDFInfo
- Publication number
- EP3168681B1 EP3168681B1 EP16206486.9A EP16206486A EP3168681B1 EP 3168681 B1 EP3168681 B1 EP 3168681B1 EP 16206486 A EP16206486 A EP 16206486A EP 3168681 B1 EP3168681 B1 EP 3168681B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- focal point
- image
- component
- user input
- electronic device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims description 15
- 230000004044 response Effects 0.000 claims description 5
- 238000004590 computer program Methods 0.000 claims 2
- 230000015654 memory Effects 0.000 description 46
- 238000004891 communication Methods 0.000 description 22
- 230000000007 visual effect Effects 0.000 description 9
- 230000001413 cellular effect Effects 0.000 description 8
- 238000012545 processing Methods 0.000 description 8
- 230000008859 change Effects 0.000 description 4
- 230000005674 electromagnetic induction Effects 0.000 description 4
- 238000012552 review Methods 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 3
- 239000003086 colorant Substances 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000009977 dual effect Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000005291 magnetic effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 229920000642 polymer Polymers 0.000 description 2
- 238000003825 pressing Methods 0.000 description 2
- 239000010409 thin film Substances 0.000 description 2
- HBBGRARXTFLTSG-UHFFFAOYSA-N Lithium ion Chemical compound [Li+] HBBGRARXTFLTSG-UHFFFAOYSA-N 0.000 description 1
- VVQNEPGJFQJSBK-UHFFFAOYSA-N Methyl methacrylate Chemical compound COC(=O)C(C)=C VVQNEPGJFQJSBK-UHFFFAOYSA-N 0.000 description 1
- 229920005372 Plexiglas® Polymers 0.000 description 1
- NIXOWILDQLNWCW-UHFFFAOYSA-N acrylic acid group Chemical group C(C=C)(=O)O NIXOWILDQLNWCW-UHFFFAOYSA-N 0.000 description 1
- 239000008186 active pharmaceutical agent Substances 0.000 description 1
- 229910017052 cobalt Inorganic materials 0.000 description 1
- 239000010941 cobalt Substances 0.000 description 1
- GUTLYIVDDKVIGB-UHFFFAOYSA-N cobalt atom Chemical compound [Co] GUTLYIVDDKVIGB-UHFFFAOYSA-N 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000010408 film Substances 0.000 description 1
- 210000003811 finger Anatomy 0.000 description 1
- 238000009432 framing Methods 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 229910001416 lithium ion Inorganic materials 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000007935 neutral effect Effects 0.000 description 1
- 229920003023 plastic Polymers 0.000 description 1
- 229920003229 poly(methyl methacrylate) Polymers 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 238000010079 rubber tapping Methods 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 239000012780 transparent material Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B13/00—Viewfinders; Focusing aids for cameras; Means for focusing for cameras; Autofocus systems for cameras
- G03B13/32—Means for focusing
- G03B13/34—Power focusing
- G03B13/36—Autofocus systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B7/00—Mountings, adjusting means, or light-tight connections, for optical elements
- G02B7/28—Systems for automatic generation of focusing signals
- G02B7/285—Systems for automatic generation of focusing signals including two or more different focus detection devices, e.g. both an active and a passive focus detecting device
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B17/00—Details of cameras or camera bodies; Accessories therefor
- G03B17/56—Accessories
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B31/00—Associated working of cameras or projectors with sound-recording or sound-reproducing means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
- H04N23/632—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/675—Focus control based on electronic image sensor signals comprising setting of focusing regions
Definitions
- a digital camera is an electronic device used to capture and store images electronically in a digital format, instead of using photographic film like conventional cameras, or recording images in an analog format to magnetic tape like many video cameras.
- Modern compact digital cameras are typically multifunctional, with some devices capable of recording sound and/or video as well as still images.
- Autofocus is a feature that allows a user of a digital camera to obtain the correct focus on a subject rather than requiring the operator to adjust focus manually.
- a user may assist the camera by determining which area of the photograph to focus on by performing a button half press or other convoluted user interaction to achieve the desired focus.
- JP 2004 242010 A describes a digital camera having an automatic focusing function in which the position focused can be confirmed in the image through the display.
- Minolta DiMAGE 7Hi (Available from: http://www.dcresource.com/reviews/minolta/dimage_7hi-review/ ) [Accessed 15 March 2017] provides a review of the Minolta DiMAGE 7Hi digital camera which uses a "double flash” (known as TTL Flash Metering) to ensure proper exposure.
- the first flash is used to illuminate the subject, while the camera judges the correct exposure to use.
- the second flash is the one that actually lights up the subject for the picture.
- the camera also uses information from the autofocus system to ensure proper flash exposure.
- JP 2004 117490 A describes a touch panel provided on a monitor for displaying a photographic image according to a video signal from a camera main body.
- a photographer touches two points diagonally located in the focus area to be set on the image plane of the monitor, the focus area is given from a focus area setting part to an autofocus circuit, and processing for autofocus control with the focus area as an object is performed by the autofocus circuit.
- Various embodiments are generally directed to techniques to automatically focus a digital camera. Some embodiments are particularly directed to automatically focusing a digital camera using a display device.
- the display device may be integrated with, or separate from, the digital camera.
- a mobile electronics device may comprise a digital camera having a lens component and lens position component.
- a display may be coupled to the digital camera to reproduce an image with a first focal point.
- the digital camera may also include a focal point selection module coupled to the display to select a second focal point for the image.
- the digital camera may also include a focus control module coupled to the focal point selection module and the lens position component to provide focus control signals to the lens position component to focus the lens component on the second focal point.
- Other embodiments are described and claimed.
- FIG. 1 illustrates a mobile electronic device 100 in accordance with one or more embodiments.
- the mobile electronic device 100 may be implemented as a combination handheld computer and mobile telephone, sometimes referred to as a smart phone.
- smart phones include, for example, Palm® products such as Palm® TrcoTM smart phones.
- Palm® products such as Palm® TrcoTM smart phones.
- the mobile electronic device 100 may comprise, or be implemented as, any type of wireless device, mobile station, or portable computing device with a self-contained power source (e.g., battery) such as a laptop computer, ultra-laptop computer, personal digital assistant (PDA), cellular telephone, combination cellular telephone/PDA, mobile unit, subscriber station, user terminal, portable computer, handheld computer, palmtop computer, wearable computer, media player, pager, messaging device, data communication device, and so forth.
- a self-contained power source e.g., battery
- a self-contained power source e.g., battery
- a self-contained power source e.g., battery
- a self-contained power source e.g., battery
- a self-contained power source e.g., battery
- the mobile electronic device 100 may provide voice communications functionality in accordance with different types of cellular radiotelephone systems.
- cellular radiotelephone systems may include Code Division Multiple Access (CDMA) systems, Global System for Mobile Communications (GSM) systems, North American Digital Cellular (NADC) systems, Time Division Multiple Access (TDMA) systems, Extended-TDMA (E-TDMA) systems, Narrowband Advanced Mobile Phone Service (NAMPS) systems, third generation (3G) systems such as Wide-band CDMA (WCDMA), CDMA-2000, Universal Mobile Telephone System (UMTS) systems, and so forth.
- CDMA Code Division Multiple Access
- GSM Global System for Mobile Communications
- NADC North American Digital Cellular
- TDMA Time Division Multiple Access
- E-TDMA Extended-TDMA
- NAMPS Narrowband Advanced Mobile Phone Service
- WCDMA Wide-band CDMA
- CDMA-2000 Code Division Multiple Access
- UMTS Universal Mobile Telephone System
- the mobile electronic device 100 may be arranged to provide data communications functionality in accordance with different types of cellular radiotelephone systems.
- Examples of cellular radiotelephone systems offering data communications services may include GSM with General Packet Radio Service (GPRS) systems (GSM/GPRS), CDMA/1xRTT systems, Enhanced Data Rates for Global Evolution (EDGE) systems, Evolution Data Only or Evolution Data Optimized (EV-DO) systems, Evolution For Data and Voice (EV-DV) systems, High Speed Downlink Packet Access (HSDPA) systems, High Speed Uplink Packet Access (HSUPA), and so forth.
- GSM General Packet Radio Service
- EDGE Enhanced Data Rates for Global Evolution
- EV-DO Evolution Data Only or Evolution Data Optimized
- EV-DV Evolution For Data and Voice
- HSDPA High Speed Downlink Packet Access
- HSUPA High Speed Uplink Packet Access
- the mobile electronic device 100 may be arranged to provide voice and/or data communications functionality in accordance with different types of wireless network systems.
- wireless network systems may include a wireless local area network (WLAN) system, wireless metropolitan area network (WMAN) system, wireless wide area network (WWAN) system, and so forth.
- suitable wireless network systems offering data communication services may include the Institute of Electrical and Electronics Engineers (IEEE) 802.xx series of protocols, such as the IEEE 802.11a/b/g/n series of standard protocols and variants (also referred to as "WiFi”), the IEEE 802.16 series of standard protocols and variants (also referred to as "WiMAX”), the IEEE 802.20 series of standard protocols and variants, and so forth.
- IEEE 802.xx series of protocols such as the IEEE 802.11a/b/g/n series of standard protocols and variants (also referred to as "WiFi”), the IEEE 802.16 series of standard protocols and variants (also referred to as "WiMAX”), the IEEE 802.20 series of standard
- the mobile electronic device 100 may be arranged to perform data communications in accordance with different types of shorter range wireless systems, such as a wireless personal area network (PAN) system.
- PAN personal area network
- a wireless PAN system offering data communication services may include a Bluetooth system operating in accordance with the Bluetooth Special Interest Group (SIG) series of protocols, including Bluetooth Specification versions v1.0, v1.1, v1.2, v2.0, v2.0 with Enhanced Data Rate (EDR), as well as one or more Bluetooth Profiles, and so forth.
- SIG Bluetooth Special Interest Group
- EDR Enhanced Data Rate
- Other examples may include systems using infrared techniques or near-field communication techniques and protocols, such as electro-magnetic induction (EMI) techniques.
- EMI techniques may include passive or active radiofrequency identification (RFID) protocols and devices.
- RFID radiofrequency identification
- the mobile electronic device 100 may comprise a dual processor architecture including a host processor 102 and a radio processor 104.
- the host processor 102 and the radio processor 104 may be arranged to communicate with each other using interfaces 106 such as one or more universal serial bus (USB) interfaces, micro-USB interfaces, universal asynchronous receiver-transmitter (UART) interfaces, general purpose input/output (GPIO) interfaces, control/status lines, control/data lines, audio lines, and so forth.
- USB universal serial bus
- micro-USB interfaces micro-USB interfaces
- UART universal asynchronous receiver-transmitter
- GPIO general purpose input/output
- the host processor 102 may be responsible for executing various software programs such as system programs and applications programs to provide computing and processing operations for the mobile electronic device 100.
- the radio processor 104 may be responsible for performing various voice and data communications operations for the mobile electronic device 100 such as transmitting and receiving voice and data information over one or more wireless communications channels.
- voice and data communications operations for the mobile electronic device 100 such as transmitting and receiving voice and data information over one or more wireless communications channels.
- the host processor 102 may be implemented as a host central processing unit (CPU) using any suitable processor or logic device, such as a general purpose processor. Although some embodiments may be described with the host processor 102 implemented as a CPU or general purpose processor by way of example, it may be appreciated that the embodiments are not limited in this context.
- the host processor 102 may comprise, or be implemented as, a chip multiprocessor (CMP), dedicated processor, embedded processor, media processor, input/output (I/O) processor, co-processor, microprocessor, controller, microcontroller, application specific integrated circuit (ASIC), field programmable gate array (FPGA), programmable logic device (PLD), or other processing device in accordance with the described embodiments.
- CMP chip multiprocessor
- dedicated processor dedicated processor
- embedded processor media processor
- I/O input/output
- co-processor co-processor
- microprocessor controller
- microcontroller application specific integrated circuit
- FPGA field programmable gate array
- PLD programmable logic device
- the host processor 102 may be coupled through a memory bus 108 to a memory 110.
- the memory bus 108 may comprise any suitable interface and/or bus architecture for allowing the host processor 102 to access the memory 110.
- the memory 110 may be shown as being separate from the host processor 102 for purposes of illustration, it is worthy to note that in various embodiments some portion or the entire memory 110 may be included on the same integrated circuit as the host processor 102. Alternatively, some portion or the entire memory 110 may be disposed on an integrated circuit or other medium (e.g., hard disk drive) external to the integrated circuit of host processor 102.
- the mobile electronic device 100 may comprise an expansion slot to support a multimedia and/or memory card, for example.
- the memory 110 may be implemented using any machine-readable or computer-readable media capable of storing data such as volatile memory or non-volatile memory, removable or non-removable memory, erasable or non-erasable memory, writeable or re-writeable memory, and so forth.
- Examples of machine-readable storage media may include, without limitation, random-access memory (RAM), dynamic RAM (DRAM), Double-Data-Rate DRAM (DDRAM), synchronous DRAM (SDRAM), static RAM (SRAM), read-only memory (ROM), programmable ROM (PROM), erasable programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), flash memory (e.g., NOR or NAND flash memory), content addressable memory (CAM), polymer memory (e.g., ferroelectric polymer memory), phase-change memory, ovonic memory, ferroelectric memory, silicon-oxide-nitride-oxide-silicon (SONOS) memory, magnetic or optical cards, or any other type of media suitable for storing information.
- RAM random-access memory
- DRAM dynamic RAM
- DDRAM Double-Data-Rate DRAM
- SDRAM synchronous DRAM
- SRAM static RAM
- ROM read-only memory
- PROM programmable ROM
- EPROM eras
- the host processor 102 may also be coupled to various components, such as a keypad 112, a display 114, an input/output (I/O) interface 116, various audio/visual (A/V) devices 118, a power supply 120, and so forth.
- the host processor 102 may further implement a digital interface to communicate digital signals between the host processor 102 and the various coupled components. Examples of such digital signals may include, without limitation, digital clock signals such as from an integrated circuit (IC) clock source.
- the digital clock signal is typically a signal used to coordinate the actions of two or more circuits.
- the clock signal oscillates between a high and low state, normally with a fifty percent duty cycle, and is usually a square waveform.
- Circuits using the clock signal for synchronization may become active at either the rising edge or falling edge, or both, of the oscillating clock signal. Consequently, the host processor 102 typically delivers a clock signal to the various components coupled to the host processor 102 in part to synchronize and control the coupled components.
- the mobile electronic device 100 may comprise an alphanumeric keypad 112 coupled to the host processor 102.
- the keypad 112 may comprise, for example, a QWERTY key layout and an integrated number dial pad. In some cases, the keypad 112 may comprise a thumbboard arranged for operation using the thumbs of a user.
- the mobile electronic device 100 also may comprise various keys, buttons, and switches such as, for example, input keys, preset and programmable hot keys, left and right action buttons, a navigation button such as a multidirectional navigation button, phone/send and power/end buttons, preset and programmable shortcut buttons, a volume rocker switch, a ringer on/off switch having a vibrate mode, and so forth.
- the mobile electronic device 100 may comprise a display 114 coupled to the host processor 102.
- the display 114 may comprise any suitable visual interface for displaying content to a user of the mobile electronic device 100.
- the display 114 may be implemented by a LCD such as a touch-sensitive color (e.g., 16-bit color) thin-film transistor (TFT) LCD screen.
- the touch-sensitive LCD may be used with a stylus and/or a handwriting recognizer program.
- the mobile electronic device 100 may comprise an I/O interface 116 coupled to the host processor 102.
- the I/O interface 116 may comprise one or more I/O devices such as a serial connection port, an infrared port, integrated Bluetooth® wireless capability, and/or integrated 802.11x (WiFi) wireless capability, to enable wired (e.g., USB cable) and/or wireless connection to a local computer system, such as a local personal computer (PC).
- a local computer system such as a local personal computer (PC).
- PC personal computer
- mobile electronic device 100 may be arranged to transfer and/or synchronize information with the local computer system.
- the host processor 102 may be coupled to various A/V devices 118 that support A/V capability of the mobile electronic device 100.
- A/V devices 114 may include, for example, a microphone, one or more speakers, an audio port to connect an audio headset, an audio coder/decoder (codec), an audio player, a digital camera, a video camera, a video codec, a video player, and so forth.
- the host processor 102 may be coupled to a power supply 120 arranged to supply and manage power to the elements of the mobile electronic device 100.
- the power supply 120 may be implemented by a rechargeable battery, such as a removable and rechargeable lithium ion battery to provide direct current (DC) power, and/or an alternating current (AC) adapter to draw power from a standard AC main power supply.
- the radio processor 104 may perform voice and/or data communication operations for the mobile electronic device 100.
- the radio processor 104 may be arranged to communicate voice information and/or data information over one or more assigned frequency bands of a wireless communication channel.
- the radio processor 104 may be implemented as a communications processor using any suitable processor or logic device, such as a modem processor or baseband processor. Although some embodiments may be described with the radio processor 104 implemented as a modem processor or baseband processor by way of example, it may be appreciated that the embodiments are not limited in this context.
- the radio processor 104 may comprise, or be implemented as, a digital signal processor (DSP), media access control (MAC) processor, or any other type of communications processor in accordance with the described embodiments.
- DSP digital signal processor
- MAC media access control
- the radio processor 104 may perform analog and/or digital baseband operations for the mobile electronic device 100.
- the radio processor 104 may perform digital-to-analog conversion (DAC), analog-to-digital conversion (ADC), modulation, demodulation, encoding, decoding, encryption, decryption, and so forth.
- DAC digital-to-analog conversion
- ADC analog-to-digital conversion
- modulation demodulation
- encoding decoding
- decoding encryption, decryption, and so forth.
- the mobile electronic device 100 may comprise a memory 122 coupled to the radio processor 104.
- the memory 122 may be implemented using one or more types of machine-readable or computer-readable media capable of storing data such as volatile memory or non-volatile memory, removable or non-removable memory, erasable or non-erasable memory, writeable or re-writeable memory, and so forth.
- the memory 122 may comprise, for example, flash memory and secure digital (SD) RAM.
- SD secure digital
- the mobile electronic device 100 may comprise a transceiver module 124 coupled to the radio processor 104.
- the transceiver module 124 may comprise one or more radio transceivers arranged to communicate using different types of protocols, communication ranges, operating power requirements, RF sub-bands, information types (e.g., voice or data), use scenarios, applications, and so forth.
- the transceiver module 124 may comprise one or more transceivers arranged to support voice communication for a cellular radiotelephone system such as a GSM, UMTS, and/or CDMA system.
- the transceiver module 124 also may comprise one or more transceivers arranged to perform data communications in accordance with one or more wireless communications protocols such as WWAN protocols (e.g., GSM/GPRS protocols, CDMA/1xRTT protocols, EDGE protocols, EV-DO protocols, EV-DV protocols, HSDPA protocols, etc.), WLAN protocols (e.g., IEEE 802.11a/b/g/n, IEEE 802.16, IEEE 802.20, etc.), PAN protocols, Infrared protocols, Bluetooth protocols, EMI protocols including passive or active RFID protocols, and so forth.
- the transceiver module 124 may comprise a Global Positioning System (GPS) transceiver to support position determination and/or location-based services.
- GPS Global Positioning System
- the transceiver module 124 generally may be implemented using one or more chips as desired for a given implementation. Although the transceiver module 124 may be shown as being separate from and external to the radio processor 104 for purposes of illustration, it is worthy to note that in various embodiments some portion or the entire transceiver module 124 may be included on the same integrated circuit as the radio processor 104. The embodiments are not limited in this context.
- the mobile electronic device 100 may comprise an antenna system 126 for transmitting and/or receiving electrical signals. As shown, the antenna system 126 may be coupled to the radio processor 104 through the transceiver module 124. The antenna system 126 may comprise or be implemented as one or more internal antennas and/or external antennas.
- the mobile electronic device 100 may comprise a subscriber identity module (SIM) 128 coupled to the radio processor 104.
- SIM subscriber identity module
- the SIM 128 may comprise, for example, a removable or non-removable smart card arranged to encrypt voice and data transmissions and to store user-specific data for allowing a voice or data communications network to identify and authenticate the user.
- the SIM 128 also may store data such as personal settings specific to the user.
- the host processor 102 may be arranged to provide processing or computing resources to the mobile electronic device 100.
- the host processor 102 may be responsible for executing various software programs such as system programs and application programs to provide computing and processing operations for the mobile electronic device 100.
- System programs generally may assist in the running of the mobile electronic device 100 and may be directly responsible for controlling, integrating, and managing the individual hardware components of the computer system.
- Examples of system programs may include, without limitation, an operating system (OS), device drivers, programming tools, utility programs, software libraries, application programming interfaces (APIs), and so forth.
- the mobile electronic device 100 may utilize any suitable OS in accordance with the described embodiments such as a Palm OS®, Palm OS® Cobalt, Microsoft® Windows OS, Microsoft Windows® CE, Microsoft Pocket PC, Microsoft Mobile, Symbian OSTM, Embedix OS, Linux, Binary Run-time Environment for Wireless (BREW) OS, JavaOS, a Wireless Application Protocol (WAP) OS, Mac OS X, and so forth.
- OS operating system
- device drivers such as a Palm OS® Cobalt, Microsoft® Windows OS, Microsoft Windows® CE, Microsoft Pocket PC, Microsoft Mobile, Symbian OSTM, Embedix OS, Linux, Binary Run-time Environment for Wireless (BREW) OS, JavaOS, a Wireless Application Protocol (WAP
- Application programs generally may allow a user to accomplish one or more specific tasks.
- Examples of application programs may include, without limitation, one or more messaging applications (e.g., telephone, voicemail, facsimile, e-mail, IM, SMS, MMS, video conferencing), a web browser application, personal information management (PIM) applications (e.g., contacts, calendar, scheduling, tasks), word processing applications, spreadsheet applications, database applications, media applications (e.g., video player, audio player, multimedia player, digital camera, video camera, media management), gaming applications, and so forth.
- the application programs may provide one or more graphical user interfaces (GUIs) to communicate information between the mobile electronic device 100 and a user.
- GUIs graphical user interfaces
- application programs may comprise upper layer programs running on top of the OS of the host processor 102 that operate in conjunction with the functions and protocols of lower layers including, for example, a transport layer such as a Transmission Control Protocol (TCP) layer, a network layer such as an Internet Protocol (IP) layer, and a link layer such as a Point-to-Point (PPP) layer used to translate and format data for communication.
- TCP Transmission Control Protocol
- IP Internet Protocol
- PPP Point-to-Point
- the mobile electronic device 100 may comprise or implement a focus management module 130.
- Focus management module 130 may be arranged to provide automatic focus control to a user.
- focus management module 130 may be responsible for providing control signals to control one or more of A/V devices 118 of mobile electronic device 100.
- the focus management module 130 may be shown as being implemented on the same integrated circuit as the host processor 102 for purposes of illustration, it is worthy to note that in various embodiments some portion or the entire focus management module 130 may be included separate from the host processor 102. The embodiments are not limited in this context.
- FIG. 2 illustrates one embodiment of a digital camera system 200.
- the digital camera system 200 may be representative of, for example, a digital camera system of mobile electronics device 100.
- FIG. 2 illustrates one embodiment of a digital camera system 200.
- the digital camera system 200 may be representative of, for example, a digital camera system of mobile electronics device 100.
- FIG. 2 illustrates one embodiment of a digital camera system 200.
- the digital camera system 200 may be representative of, for example, a digital camera system of mobile electronics device 100.
- FIG. 2 illustrates one embodiment of a digital camera system 200.
- the digital camera system 200 may be representative of, for example, a digital camera system of mobile electronics device 100.
- digital camera system 200 may comprise or implement focus management module 130, pointing component 214, display 114, digital camera 202 and bus 240.
- the bus 240 may comprise any suitable interface and/or bus architecture for allowing the digital camera 202 to communicate with focus management module 130.
- focus management module 130 may be shown as being separate from digital camera 202 for purposes of illustration, it is worthy to note that in various embodiments some portion or the entire focus management module 130 may be included on the same integrated circuit as the digital camera 202. Alternatively, some portion or the entire focus management module 130 may be disposed on an integrated circuit or other medium (e.g., hard disk drive) external to the integrated circuit of digital camera 202.
- Focus management module 130 may comprise or implement focal point selection module 232, focus control module 234 and white balance control module 236.
- Digital camera 202 may comprise or implement controller 220, lens component 204, lens position component 206, microphone position component 208, directional microphone 210, memory 212, flash control module 216 and flash component 218. It can be appreciated that a limited number of elements arc shown for purposes of illustration and not limitation.
- Display 114 of digital camera system 200 may be representative of, for example, display 114 of mobile electronic device 100.
- the display 114 may comprise any suitable visual interface for displaying content to a user of digital camera system 200.
- the display 114 may be implemented by a LCD such as a touch-sensitive color (e.g., 16-bit color) thin-film transistor (TFT) LCD screen.
- the touch-sensitive LCD may be used with a stylus or other pointing device. The embodiments are not limited in this context.
- Digital camera system 200 may be representative of, for example, one of A/V devices 118 of mobile electronics device 100.
- digital camera system 200 may be implemented as an electronic device used to capture and store images electronically in a digital format. Additionally, in some embodiments digital camera system 200 may be capable of recording sound and/or video in addition to still images. The embodiments are not limited in this context.
- Digital camera 202 of digital camera system 200 may comprise controller 220.
- Controller 220 may provide control signals to components of digital camera 202, including lens position component 206, microphone position component 208 and flash control module 216, to provide functionality for digital camera system 200.
- controller 220 may also provide control signals to focus management module 130 via bus 240.
- controller 220 may be implemented as, for example, host processor 102 of mobile clcctronics device 100. Alternatively, controller 220 may be implemented as a separate processor from host processor 102. The embodiments arc not limited in this context.
- Digital camera 202 may further comprise memory 212.
- Memory 212 may be implemented as, for example, memory 110 of mobile electronics device 100. Although the memory 212 may be shown as being separate from controller 220 for purposes of illustration, it is worthy to note that in various embodiments some portion or the entire memory 212 may be included on the same integrated circuit as controller 220. Alternatively, some portion or the entire memory 212 may be disposed on an integrated circuit or other medium (e.g., hard disk drive) external to the integrated circuit of controller 220.
- an integrated circuit or other medium e.g., hard disk drive
- Focus management module 130 of digital camera system 200 may be representative of, for example, focus management module 130 of mobile electronics device 100. Focus management module 130 may be arranged to provide automatic focus control to a user by, for example, providing control signals to control one or more of lens position component 206, microphone position component 208 and flash control module 216, for example. In some embodiments the focus management module 130 may be implemented on the same integrated circuit as controller 220. In various embodiments some portion or the entire focus management module 130 may be included separate from controller 220. In various embodiments, focus management module 130 may be implemented by one or more hardware components, software components, and/or combination thereof. Focus management module 130 may, for example, be implemented as computer executable instructions stored on memory 212 to be executed by controller 220. The embodiments are not limited in this context.
- Digital camera 202 may comprise a lens component 204 and a lens position component 206.
- Lens component 204 may consist of a photographic or optical lens or assembly of lenses made of a transparent material such as glass, plastic, acrylic or Plexiglass, for example.
- the one or more lens elements of lens component 204 may reproduce an image of an object and allow for zooming in or out on the object by mechanically changing the focal length of the lens elements.
- digital zoom may be employed in digital camera system 200 to zoom in or out on an image.
- the one or more lens elements of lens component 204 may be used to focus on different portions of the image by varying the focal length of the lens elements. The desired focus can be obtained with an autofocus feature of digital camera system 200 or by manually focusing on the desired portion of the image, for example.
- Lens position component 206 may be configured to receive control signals from controller 220 and focus control module 234 and use this control information to adjust the lens component 204.
- the lens position component 206 may be implemented by one or more hardware components, software components, and/or combination thereof. The embodiments are not limited in this context.
- lens position component may receive control information from controller 220 indicating that a user wishes to focus on a particular portion of an object. Lens position component 206 may use this control information to configure the lens elements of lens component 204 into the proper configuration to deliver the desired focus.
- display 114 of digital camera system 200 may be coupled to digital camera 202 and may be arranged to capture and rcproducc an image with a first focal point.
- the image may be viewed through lens component 204 and reproduced on display 114.
- the image on display 114 may comprise a substantially real-time preview of the image to allow for framing and previewing before capturing a photograph.
- the focal point refers to the intended center of interest of a photograph and is typically the point where the image will be in the clearest focus.
- Focal point selection module 232 may be coupled to the display 114 and may allow for the selection of a second focal point for the image.
- the second focal point may be selected, for example, by an operator using the display 114 to indicate that a second focal point is desired.
- display 114 may comprise a touchscreen display and may be configured to send coordinates for the second focal point to the focal point selection module 232.
- These coordinates which may in one example comprise X, Y coordinates, may correspond to pressure applied to an area of the display 114. Pressure may be applied to the display 114 by any suitable input device, including a stylus or a finger, for example, and this pressure may be digitized to represent the coordinates.
- the embodiments are not limited in this context.
- Focus control module 234 may be coupled to the focal point selection module 232 and the lens position component 206 to provide focus control signals to the lens position component 206 to focus the lens component 204 on the second focal point.
- the focus control signals may represent a focal length for the lens component 204.
- focus control module 234 may determine the focal length for lens component 204 based on the coordinates for the second focal point determined by focal point selection module 232.
- Lens position component 206 may use the control signals from focus control module 234 to adjust the lens elements of lens component 204 to achieve the desired focus on the second focal point.
- the digital camera system 200 may further comprise a pointing component 214 coupled to the display 114 and the focal point selection module 232.
- the pointing component 214 may be configured to control a pointer on the display 114 in response to user or operator commands.
- the pointing component 214 may send coordinates for the second focal point to the focal point selection module 232.
- the coordinates may correspond to a position for the pointer on the display 114.
- the pointer may comprise any overlay image on display 114 that is capable of indicated the currently selected area or focal point.
- the pointer may include an arrow, box, circle, oval or any other suitable image.
- the overlay image may have varying degrees of translucency as desired for a given implementation.
- the digital camera system 200 may comprise a directional microphone 210 and a microphone position component 208 coupled to the focus control module 234.
- the focus control module 234 may provide focus control signals to the microphone position component 208 to focus the directional microphone 210 on the second focal point.
- focus control module 234 may provide coordinates to microphone position component 208.
- Microphone position component 208 may use the coordinates to focus directional microphone 210 on the sound emanating from the focal point defined by the coordinates.
- Directional microphone 210 may consist of one or more microphones whose response varies with the direction of sound incidence.
- directional microphone 210 may be implemented as a digital directional microphone capable of creating null points at different frequencies to cancel various sounds around a desired focal point.
- the microphone position component 208 may be implemented by one or more hardware components, software components, and/or combination thereof. The embodiments are not limited in this context.
- the digital camera system 200 may comprise a white balance control module 236 coupled to the focal point selection module 232.
- the white balance control module 236 may be configured to modify a white balance setting value for the image based on the second focal point.
- white balance control module 236 may be implemented by one or more hardware components, software components, and/or combination thereof.
- White balance control may refer to the adjustment of the relative amounts of red, green, and blue primary colors in an image such that neutral colors are reproduced correctly. Changing the white balance may change the overall mixture of colors in an image and can be used for generalized color correction.
- Flash component 218 and flash control module 216 of digital camera system 200 may be coupled to the focal point selection module 232 and/or controller 220.
- the flash control module 216 may be configured to select a flash level value representing flash intensity for the flash component 218 based on the second focal point.
- the flash control module 216 may be implemented by one or more hardware components, software components, and/or combination thereof.
- Flash component 218 may consist of a device or group of devices that produce an instantaneous flash of light, typically around 1/3000 of a second at a color temperature of about 5500K, to help illuminate a sccnc. Flash component 218 may be used, for example, to aid in capturing images of quickly moving objects, to create a different temperature light than the ambient light, and to illuminate scenes that do not have enough available light to adequately expose the image.
- Operations for the mobile electronic device 100 may be further described with reference to one or more logic flows. It may be appreciated that the representative logic flows do not necessarily have to be executed in the order presented, or in any particular order, unless otherwise indicated. Moreover, various activities described with respect to the logic flows can be executed in serial or parallel fashion. The logic flows may be implemented using any desired hardware elements, software elements, or combination of both.
- FIG. 3 illustrates a logic flow 300.
- Logic flow 300 may be representative of the operations executed by one or more embodiments described herein.
- the logic flow 300 may be performed by various systems and/or devices and may be implemented as hardware, software, and/or any combination thereof, as desired for a given set of design parameters or performance constraints.
- the logic flow 300 may be implemented by a logic device (e.g., processor) and/or logic comprising instructions, data, and/or code to be executed by a logic device.
- the logic flow 300 is described with reference to FIGS. 1 and 2 . The embodiments are not limited in this context.
- the logic flow 300 may comprise displaying an image with a first focal point from a digital camera at block 302.
- the image may be displayed with a touchscreen display such as, for example, display 114 of mobile electronics device 100.
- the logic flow 300 may further comprise selecting a second focal point for the image at block 304.
- the selection of the second focal point for the image may be performed using a touchscreen display.
- the touchscreen display may be, for example, display 114 of mobile electronic device 100.
- logic flow 300 may comprise focusing the digital camera on the second focal point. Focusing the digital camera on the second focal point may further comprise generating a set of coordinates for the second focal point and focusing a lens for the digital camera on the second focal point.
- a directional microphone may be focused on the second focal point and/or a flash level value representing flash intensity for a flash component may be selected based on the second focal point.
- the image may be captured using a touchscreen display.
- Capturing the image using a touchscreen display may comprise a user applying pressure to the touchscreen display to indicate that the image on the display is ready to be committed to memory. Capturing the image, for example, may require a double tap on the touchscreen display or pressure applied to the touchscreen display in the area of the selected second focal point.
- the digital camera may be capable of zooming in or out on the image displayed on the touchscreen display in various embodiments.
- a user may desire to zoom in on an object to focus a photograph on a specific element of the object.
- the user may accomplish this goal by applying pressure to the touchscreen display in the area where the user wishes to zoom, similar to the way in which a second focal point is selected.
- a user may slide a pointer across the touchscreen display to zoom in or out on the object.
- GUI graphic user interface
- a user may edit or crop the image on the touchscreen display by manipulating the image using the touchscreen display.
- a user who wishes to crop an image may use the touchscrccn display to indicate the desired size of the cropped image by sliding the corners of the image to the desired position.
- a GUI may be implemented for the cropping operations.
- FIG. 4 illustrates one embodiment of a digital camera system 400.
- the digital camera system 400 may be representative of, for example, digital camera system 200.
- digital camera system 400 may comprise or implement lens component 204, directional microphone 210, visual focal point 402 and audio focal point 404. It can be appreciated that a limited number of elements are shown for purposes of illustration and not limitation.
- a mobile electronic device 100 may comprise a digital camera 200 having a lens component 204 and lens position component 206, and a directional microphone 210 and a microphone position component 208.
- a display 114 may be coupled to the digital camera 200 to reproduce an image from the digital camera 200 with a first focal point.
- a focal point selection module 232 may be coupled to the display 114 to select a second and a third focal point for the image.
- a focus control module 234 may be coupled to the focal point selection module 232 and the lens position component 206 and microphone position component 208 to provide focus control signals to the lens position component 206 and the microphone position component 208.
- the focus control signals may be used to focus the lens component 204 on the second focal point and the directional microphone 210 on the third focal point, for example.
- the second focal point may comprise visual focal point 402.
- Visual focal point 402 may comprise, for example, an object that a user desires to focus on when capturing a digital image.
- the third focal point may comprise audio focal point 404.
- Audio focal point 404 may comprise a sound or series of sounds, such as the voice of an individual that a user desires to focus on when capturing a digital video, for example.
- the visual focal point 402 and the audio focal point 404 may be located remotely from each other.
- directional microphone 210 may be configured to track audio focal point 404 as the position of audio focal point 404 changes. For example, if the selected audio focal point 404 comprises the voice of an individual located in a first position, the individual may move to a second, remote position, and the directional microphone 210 may be configured to follow the voice of the individual to the second position without requiring the user to physically change the position of the digital camera. Tracking an audio focal point 404 may be implemented by any suitable means known in the art, such as sound source localization, for example.
- the embodiment described in Fig. 4 may be applicable if a user desires to record a digital video with different audio and video focal points. For example, a user may desire to record a digital video of a waterfall while another individual standing nearby describes the waterfall. In this example, the user may wish to select the waterfall as the visual focal point 402 and the individual describing the scene as the audio focal point 404. In this embodiment, the quality of the recorded digital video may be increased as a result of increased options presented to the user.
- a user of a mobile electronic device may desire to capture a digital photograph of a friend standing near a waterfall with a digital camera contained in a mobile electronic device.
- the user points the lens of the camera of the mobile electronic device in the direction of the friend and an image is displayed on the touchscreen display of the mobile electronic device.
- the intended focal point of the image is the face of the friend.
- the digital camera may originally set the default focal point as the waterfall, for example.
- the user need only tap the portion of the touchscreen display of the mobile electronic device to set the focal point on a different, desired location.
- the user may tap the touchscreen display in the area of the face of their friend to change the focal point away from the waterfall.
- the user may also capture the image by tapping the touchscreen a second time or by performing a double tap on the touchscreen.
- the image may also be captured using the various buttons on the mobile electronic device, including a five way center button press, for example.
- the touchscreen display could be used to continually change the focal point of the video in a manner similar to that described above. Additionally, the touchscreen display could be utilized to control a directional microphone of the mobile electronic device. For example, if the intended video target is the waterfall that is several yards away from the user, and a group of individuals carrying on a conversation arc standing near the user, the user could use the touchscreen display to indicate that the sound of the waterfall, and not the conversation of the members of the nearby group, arc the intended audio target. As described above, a user may also wish to select different audio and visual focal points. The above described embodiments may provide a user with enhanced case of operation and may help to ensure better picture quality and a better photographic experience.
- Various embodiments may comprise one or more elements.
- An element may comprise any structure arranged to perform certain operations.
- Each element may be implemented as hardware, software, or any combination thereof, as desired for a given set of design and/or performance constraints.
- an embodiment may be described with a limited number of elements in a certain topology by way of example, the embodiment may include more or less elements in alternate topologies as desired for a given implementation.
- any reference to "one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment.
- the appearances of the phrase “in one embodiment” in the specification arc not necessarily all referring to the same embodiment.
- exemplary functional components or modules may be implemented by one or more hardware components, software components, and/or combination thereof.
- the functional components and/or modules may be implemented, for example, by logic (e.g., instructions, data, and/or code) to be executed by a logic device (e.g., processor).
- logic e.g., instructions, data, and/or code
- Such logic may be stored internally or externally to a logic device on one or more types of computer-readable storage media.
- processing refers to the action and/or processes of a computer or computing system, or similar electronic computing device, that manipulates and/or transforms data represented as physical quantities (e.g., electronic) within registers and/or memories into other data similarly represented as physical quantities within the memories, registers or other such information storage, transmission or display devices.
- physical quantities e.g., electronic
- Coupled and “connected” along with their derivatives. These terms are not intended as synonyms for each other.
- some embodiments may be described using the terms “connected” and/or “coupled” to indicate that two or more elements arc in direct physical or electrical contact with each other.
- the term “coupled,” however, may also mean that two or more elements arc not in direct contact with each other, but yet still co-operate or interact with each other.
- the term “coupled” may refer to interfaces, message interfaces, API, exchanging messages, and so forth.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Optics & Photonics (AREA)
- Human Computer Interaction (AREA)
- Studio Devices (AREA)
- Telephone Function (AREA)
Claims (12)
- Mobiles elektronisches Gerät (100), das Folgendes umfasst:eine Linsenkomponente (204), die mit einer Linsenpositionskomponente (206) gekoppelt ist;ein Touchscreen-Display (114) mit der Aufgabe, ein Bild mit einem ersten Brennpunkt anzuzeigen;ein mit dem Display gekoppeltes Brennpunktauswahlmodul (232), wobei das Brennpunktauswahlmodul auf einen ersten Benutzereingabetyp auf dem Touchscreen-Display anspricht, um einen gewünschten zweiten Brennpunkt für das Bild auszuwählen, wobei der zweite Brennpunkt einer Stelle auf dem Bild entspricht;ein mit dem Brennpunktauswahlmodul und der Linsenpositionskomponente gekoppeltes Brennpunktsteuermodul (234), wobei das Brennpunktsteuermodul die Aufgabe hat, Brennpunktsteuersignale an die Linsenpositionskomponente anzulegen, um es zuzulassen, dass die Linsenpositionskomponente die Linsenkomponente auf den zweiten Brennpunkt fokussiert, während das Bild angezeigt wird; undeine Blitzlichtkomponente (218), die mit einem Blitzlichtsteuermodul (216) gekoppelt ist, wobei das Blitzlichtsteuermodul zum Auswählen eines Blitzlichtstärkewertes konfiguriert ist, der eine Blitzlichtintensität für die Blitzlichtkomponente auf der Basis des zweiten Brennpunkts repräsentiert, undwobei die ein oder mehreren Prozessoren das Bild auf der Basis des Blitzlichtstärkewertes als Reaktion auf einen zweiten Benutzereingabetyp auf dem Touchscreen-Display erfasst, wobei sich der zweite Benutzereingabetyp vom ersten Benutzereingabetyp unterscheidet.
- Mobiles elektronisches Gerät nach Anspruch 1, wobei der erste Benutzereingabetyp ein einfaches Antippen beinhaltet und wobei der zweite Benutzereingabetyp ein doppeltes Antippen beinhaltet.
- Mobiles elektronisches Gerät nach Anspruch 1 oder 2, wobei das Touchscreen-Display die Aufgabe hat, Koordinaten für den zweiten Brennpunkt zum Brennpunktauswahlmodul zu senden, wobei die Koordinaten der Stelle des ersten Benutzereingabetyps auf dem Touchscreen-Display entsprechen.
- Mobiles elektronisches Gerät nach Anspruch 1 bis 3, das eine Zeigekomponente (214) umfasst, die mit dem Display und dem Brennpunktauswahlmodul gekoppelt ist, wobei die Zeigekomponente die Aufgabe hat, einen Zeiger auf dem Display zu steuern, wobei die Zeigekomponente ferner die Aufgabe hat, Koordinaten für den zweiten Brennpunkt zum Brennpunktauswahlmodul zu senden, wobei die Koordinaten einer Position für den Zeiger auf dem Display entsprechen.
- Mobiles elektronisches Gerät nach einem der Ansprüche 1 bis 4, das eine Richtmikrofonkomponente (210) umfasst, wobei das Brennpunktsteuermodul zum Bereitstellen von Brennpunktsteuersignalen zum Fokussieren des Richtmikrofons auf den zweiten Brennpunkt konfiguriert ist.
- Mobiles elektronisches Gerät nach einem der Ansprüche 1 bis 5, das ein Weißabgleichsteuermodul (236) umfasst, das mit dem Brennpunktauswahlmodul gekoppelt ist, wobei das Weißabgleichsteuermodul zum Modifizieren eines Weißabgleicheinstellwertes für das Bild auf der Basis des zweiten Brennpunkts konfiguriert ist.
- Verfahren, das die folgenden Schritte beinhaltet:Anzeigen eines Bildes mit einem ersten Brennpunkt (302) auf einem Touchscreen-Display (114), wobei das Bild von einer Linsenkomponente (204) bereitgestellt wird;Auswählen eines zweiten Brennpunkts (304) für das Bild als Reaktion auf einen ersten Benutzereingabetyp auf dem Touchscreen-Display, wobei der zweite Brennpunkt einer Stelle auf dem auf dem Touchscreen-Display angezeigten Bild entspricht;Fokussieren der Linsenkomponente vom ersten Brennpunkt auf den zweiten Brennpunkt (306), während das Bild angezeigt wird;Auswählen eines Blitzlichtstärkewertes, der eine Blitzlichtintensität für eine Blitzlichtkomponente (218) auf der Basis des zweiten Brennpunkts repräsentiert; undErfassen des Bildes, auf der Basis des Blitzlichtstärkewertes, als Reaktion auf einen zweiten Benutzereingabetyp auf dem Touchscreen-Display, wobei sich der zweite Benutzereingabetyp von dem ersten Benutzereingabetyp unterscheidet.
- Verfahren nach Anspruch 7, wobei der erste Benutzereingabetyp ein einzelnes Antippen beinhaltet und wobei der zweite Benutzereingabetyp ein doppeltes Antippen beinhaltet.
- Verfahren nach Anspruch 7 oder 8, das ferner die folgenden Schritte beinhaltet:Erzeugen eines Satzes von Koordinaten für den zweiten Brennpunkt; undBestimmen einer Brennweite für eine Linsenkomponente der digitalen Kamera auf der Basis der Koordinaten für den durch das Brennpunktauswahlmodul ausgewählten zweiten Brennpunkt, wobei das Fokussieren der Linsenkomponente vom ersten Brennpunkt auf den zweiten Brennpunkt auf die Erzeugung eines Satzes von Koordinaten für den zweiten Brennpunkt anspricht.
- Verfahren nach den Ansprüchen 7 bis 9, wobei der Schritt des Fokussierens der Linsenkomponente auf den zweiten Brennpunkt den Schritt des Fokussierens eines Richtmikrofons (210) auf den zweiten Brennpunkt beinhaltet.
- Computerprogramm, das computerlesbare Programmelemente mit der Aufgabe umfasst, in dem mobilen elektronischen Gerät der Ansprüche 1-6 das Verfahren nach einem der Ansprüche 7-10 zu implementieren.
- Computerlesbares Speichermedium, das das Computerprogramm von Anspruch 11 umfasst.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SI200832050T SI3168681T1 (sl) | 2007-07-31 | 2008-07-29 | Tehnike za samodejno fokusiranje digitalne kamere |
EP19154330.5A EP3506006B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP20198227.9A EP3779589A1 (de) | 2007-07-31 | 2008-07-29 | Techniken zur automatischen fokussierung einer digitalkamera |
PL16206486T PL3168681T3 (pl) | 2007-07-31 | 2008-07-29 | Techniki automatycznego ustawiania ostrości w cyfrowym aparacie fotograficznym |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/831,051 US8497928B2 (en) | 2007-07-31 | 2007-07-31 | Techniques to automatically focus a digital camera |
PCT/US2008/071481 WO2009018279A1 (en) | 2007-07-31 | 2008-07-29 | Techniques to automatically focus a digital camera |
EP08782499.1A EP2174186B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
Related Parent Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP08782499.1A Division-Into EP2174186B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP08782499.1A Division EP2174186B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20198227.9A Division EP3779589A1 (de) | 2007-07-31 | 2008-07-29 | Techniken zur automatischen fokussierung einer digitalkamera |
EP19154330.5A Division EP3506006B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
Publications (2)
Publication Number | Publication Date |
---|---|
EP3168681A1 EP3168681A1 (de) | 2017-05-17 |
EP3168681B1 true EP3168681B1 (de) | 2019-01-30 |
Family
ID=40304819
Family Applications (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20158892.8A Active EP3712697B1 (de) | 2007-07-31 | 2008-07-29 | Techniken zur automatischen fokussierung einer digitalkamera |
EP16206486.9A Active EP3168681B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP19154330.5A Active EP3506006B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP16190939.5A Active EP3136167B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP08782499.1A Active EP2174186B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP20198227.9A Pending EP3779589A1 (de) | 2007-07-31 | 2008-07-29 | Techniken zur automatischen fokussierung einer digitalkamera |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20158892.8A Active EP3712697B1 (de) | 2007-07-31 | 2008-07-29 | Techniken zur automatischen fokussierung einer digitalkamera |
Family Applications After (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP19154330.5A Active EP3506006B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP16190939.5A Active EP3136167B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP08782499.1A Active EP2174186B1 (de) | 2007-07-31 | 2008-07-29 | Verfahren zur automatischen fokussierung einer digitalkamera |
EP20198227.9A Pending EP3779589A1 (de) | 2007-07-31 | 2008-07-29 | Techniken zur automatischen fokussierung einer digitalkamera |
Country Status (11)
Country | Link |
---|---|
US (2) | US8497928B2 (de) |
EP (6) | EP3712697B1 (de) |
DK (2) | DK3168681T3 (de) |
ES (4) | ES2647068T3 (de) |
HU (3) | HUE049065T2 (de) |
NO (1) | NO2174186T3 (de) |
PL (3) | PL3168681T3 (de) |
PT (2) | PT3168681T (de) |
SI (2) | SI3168681T1 (de) |
TR (1) | TR201904084T4 (de) |
WO (1) | WO2009018279A1 (de) |
Families Citing this family (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100896711B1 (ko) | 2007-02-08 | 2009-05-11 | 삼성전자주식회사 | 터치 스크린을 갖는 휴대 단말기의 탭을 통한 기능 실행방법 |
US8497928B2 (en) | 2007-07-31 | 2013-07-30 | Palm, Inc. | Techniques to automatically focus a digital camera |
US8917985B2 (en) * | 2007-08-29 | 2014-12-23 | Nintendo Co., Ltd. | Imaging apparatus |
JP4260215B1 (ja) * | 2007-08-29 | 2009-04-30 | 任天堂株式会社 | 撮像装置 |
KR101491594B1 (ko) | 2007-11-05 | 2015-02-09 | 삼성전자주식회사 | 터치 스크린을 갖는 휴대 단말기 및 그의 이미지 처리 방법 |
JP4181211B1 (ja) | 2008-06-13 | 2008-11-12 | 任天堂株式会社 | 情報処理装置およびそれにおいて実行される起動用プログラム |
US8130275B2 (en) | 2008-06-13 | 2012-03-06 | Nintendo Co., Ltd. | Information-processing apparatus, and storage medium storing a photographing application launch program executed by information-processing apparatus |
US8237807B2 (en) * | 2008-07-24 | 2012-08-07 | Apple Inc. | Image capturing device with touch screen for adjusting camera settings |
WO2010038296A1 (ja) | 2008-10-01 | 2010-04-08 | 任天堂株式会社 | 情報処理装置、情報処理システム、ならびに起動用プログラムおよびそれを記憶した記憶媒体 |
KR20110061063A (ko) * | 2009-12-01 | 2011-06-09 | 삼성전자주식회사 | 휴대 단말기의 이미지 촬영 방법 및 장치 |
TW201120741A (en) * | 2009-12-04 | 2011-06-16 | Alcor Micro Corp | Vidoe/audio data detecting module and video/audio data detecting method |
US8379134B2 (en) | 2010-02-26 | 2013-02-19 | Research In Motion Limited | Object detection and selection using gesture recognition |
JP5459031B2 (ja) * | 2010-04-13 | 2014-04-02 | ソニー株式会社 | 情報処理装置、情報処理方法及びプログラム |
JP5848561B2 (ja) * | 2011-09-20 | 2016-01-27 | キヤノン株式会社 | 撮像装置及びその制御方法、プログラム、並びに記憶媒体 |
EP2827577A4 (de) * | 2012-03-16 | 2015-10-07 | Ntt Docomo Inc | Informationsendgerät und bildgebungssteuerungsverfahren |
US9041844B2 (en) * | 2012-04-27 | 2015-05-26 | Blackberry Limited | Camera device with a dynamic touch screen shutter and dynamic focal control area |
US9319636B2 (en) * | 2012-12-31 | 2016-04-19 | Karl Storz Imaging, Inc. | Video imaging system with multiple camera white balance capability |
US9200915B2 (en) * | 2013-06-08 | 2015-12-01 | Apple Inc. | Mapping application with several user interfaces |
CN104519116A (zh) * | 2013-10-08 | 2015-04-15 | 深圳市王菱科技开发有限公司 | 支持交互关联系统与wifi无隙连接的平面书刊产品 |
KR101974514B1 (ko) * | 2013-12-27 | 2019-05-02 | 인텔 코포레이션 | 데이터를 처리하기 위한 두 개의 프로세서를 갖는 전자 장치 |
US10429608B1 (en) * | 2016-09-23 | 2019-10-01 | Apple Inc. | Primary-subordinate camera focus based on lens position sensing |
US10846561B1 (en) | 2020-04-01 | 2020-11-24 | Scandit Ag | Recognition and selection of discrete patterns within a scene or image |
US11087105B1 (en) * | 2020-05-15 | 2021-08-10 | Scandit Ag | Performance improvements for recognition of optical patterns in images |
US11216628B2 (en) | 2020-04-01 | 2022-01-04 | Scandit Ag | High-speed scanning of optical patterns using a digital camera |
US11514665B2 (en) | 2020-04-01 | 2022-11-29 | Scandit Ag | Mapping optical-code images to an overview image |
US11295163B1 (en) | 2020-04-01 | 2022-04-05 | Scandit Ag | Recognition of optical patterns in images acquired by a robotic device |
US11290643B1 (en) | 2020-06-26 | 2022-03-29 | Scandit Ag | Efficient digital camera image acquisition and analysis |
US11403477B1 (en) | 2020-05-15 | 2022-08-02 | Scandit Ag | Image exposure performance improvements for recognition of optical patterns |
US11922271B1 (en) | 2020-05-15 | 2024-03-05 | Scandit Ag | Virtual screen standby mode for mobile device camera |
US11244147B2 (en) | 2020-05-15 | 2022-02-08 | Scandit Ag | Optical data exchange while preserving social distancing |
JP7542994B2 (ja) * | 2020-05-26 | 2024-09-02 | キヤノン株式会社 | 電子機器 |
US11417001B1 (en) | 2020-07-27 | 2022-08-16 | Scandit Ag | Detecting discrete optical patterns using depth estimation |
US11495036B1 (en) | 2021-01-29 | 2022-11-08 | Scandit Ag | Segmenting images for optical character recognition |
US11880738B1 (en) | 2021-08-17 | 2024-01-23 | Scandit Ag | Visual odometry for optical pattern scanning in a real scene |
US11557136B1 (en) | 2022-02-23 | 2023-01-17 | Scandit Ag | Identity document verification based on barcode structure |
Family Cites Families (72)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3050348B2 (ja) * | 1992-04-17 | 2000-06-12 | インターナショナル・ビジネス・マシーンズ・コーポレイション | プロセス制御システムにおけるユーザ制御のための方法と装置 |
US5612719A (en) | 1992-12-03 | 1997-03-18 | Apple Computer, Inc. | Gesture sensitive buttons for graphical user interfaces |
US5852667A (en) * | 1995-07-03 | 1998-12-22 | Pan; Jianhua | Digital feed-forward active noise control system |
JPH09116792A (ja) * | 1995-10-19 | 1997-05-02 | Sony Corp | 撮像装置 |
US6151073A (en) | 1996-03-28 | 2000-11-21 | Fotonation, Inc. | Intelligent camera flash system |
US6529234B2 (en) * | 1996-10-15 | 2003-03-04 | Canon Kabushiki Kaisha | Camera control system, camera server, camera client, control method, and storage medium |
US6441854B2 (en) | 1997-02-20 | 2002-08-27 | Eastman Kodak Company | Electronic camera with quick review of last captured image |
KR19990030882A (ko) * | 1997-10-07 | 1999-05-06 | 이해규 | 초점 위치 조절이 가능한 디지탈 스틸 카메라 및 그 제어 방법 |
US5923908A (en) | 1997-10-30 | 1999-07-13 | Eastman Kodak Company | Camera with touch sensitive control |
JPH11136568A (ja) * | 1997-10-31 | 1999-05-21 | Fuji Photo Film Co Ltd | タッチパネル操作式カメラ |
US8479122B2 (en) | 2004-07-30 | 2013-07-02 | Apple Inc. | Gestures for touch sensitive input devices |
US6775361B1 (en) * | 1998-05-01 | 2004-08-10 | Canon Kabushiki Kaisha | Recording/playback apparatus with telephone and its control method, video camera with telephone and its control method, image communication apparatus, and storage medium |
JPH11355617A (ja) * | 1998-06-05 | 1999-12-24 | Fuji Photo Film Co Ltd | 画像表示器付きカメラ |
JP2000002833A (ja) | 1998-06-16 | 2000-01-07 | Minolta Co Ltd | カメラ |
US7253836B1 (en) | 1998-06-30 | 2007-08-07 | Nikon Corporation | Digital camera, storage medium for image signal processing, carrier wave and electronic camera |
US6812968B1 (en) * | 1998-12-31 | 2004-11-02 | Lucent Technologies Inc. | Camera with configurable focus area |
US6684087B1 (en) * | 1999-05-07 | 2004-01-27 | Openwave Systems Inc. | Method and apparatus for displaying images on mobile devices |
JP2001159730A (ja) | 1999-12-02 | 2001-06-12 | Ricoh Co Ltd | 電子カメラ |
JP4560173B2 (ja) * | 2000-05-30 | 2010-10-13 | キヤノン株式会社 | 焦点検出装置 |
US20020080257A1 (en) * | 2000-09-27 | 2002-06-27 | Benjamin Blank | Focus control system and process |
JP4096541B2 (ja) * | 2001-10-01 | 2008-06-04 | 株式会社日立製作所 | 画面表示方法 |
JP2003179678A (ja) * | 2001-10-03 | 2003-06-27 | Nec Corp | 携帯電話機 |
JP3970046B2 (ja) * | 2002-02-19 | 2007-09-05 | オリンパス株式会社 | カメラ |
US20030193600A1 (en) * | 2002-03-28 | 2003-10-16 | Minolta Co., Ltd | Image capturing apparatus |
JP2004086160A (ja) * | 2002-07-03 | 2004-03-18 | Canon Inc | 光学装置 |
JP2004117490A (ja) * | 2002-09-24 | 2004-04-15 | Fuji Photo Optical Co Ltd | オートフォーカスシステム |
US20040085360A1 (en) * | 2002-10-31 | 2004-05-06 | Hallmark Interactive, Llc | Icon-based graphical user interface for text messaging |
US7551223B2 (en) * | 2002-12-26 | 2009-06-23 | Sony Corporation | Apparatus, method, and computer program for imaging and automatic focusing |
JP3956861B2 (ja) * | 2003-02-05 | 2007-08-08 | コニカミノルタフォトイメージング株式会社 | 撮像装置 |
JP2004297751A (ja) * | 2003-02-07 | 2004-10-21 | Sharp Corp | 合焦状態表示装置及び合焦状態表示方法 |
US7417683B2 (en) * | 2003-03-26 | 2008-08-26 | Canon Kabushiki Kaisha | Auto-focusing optical apparatus with focus detection area setting capability |
US7248301B2 (en) | 2003-05-02 | 2007-07-24 | Hewlett-Packard Development Company, L.P. | System and method for providing camera focus feedback |
US20040246360A1 (en) | 2003-06-05 | 2004-12-09 | Stavely Donald J. | System and method for displaying preview images to a camera user |
US7471846B2 (en) * | 2003-06-26 | 2008-12-30 | Fotonation Vision Limited | Perfecting the effect of flash within an image acquisition devices using face detection |
US8896725B2 (en) | 2007-06-21 | 2014-11-25 | Fotonation Limited | Image capture device with contemporaneous reference image capture mechanism |
JP4218446B2 (ja) | 2003-07-03 | 2009-02-04 | 株式会社ニコン | 電子カメラ |
US20050036708A1 (en) | 2003-08-11 | 2005-02-17 | David Boll | Systems and methods for cropping captured images |
US7231231B2 (en) * | 2003-10-14 | 2007-06-12 | Nokia Corporation | Method and apparatus for locking a mobile telephone touch screen |
JP4269883B2 (ja) * | 2003-10-20 | 2009-05-27 | ソニー株式会社 | マイクロホン装置、再生装置及び撮像装置 |
JP2005159856A (ja) * | 2003-11-27 | 2005-06-16 | Nikon Corp | デジタルカメラ |
JP4513317B2 (ja) * | 2003-12-10 | 2010-07-28 | カシオ計算機株式会社 | 複合機器、合焦制御方法及び合焦制御プログラム |
US7705908B2 (en) | 2003-12-16 | 2010-04-27 | Eastman Kodak Company | Imaging method and system for determining camera operating parameter |
JP4374574B2 (ja) | 2004-03-30 | 2009-12-02 | 富士フイルム株式会社 | マニュアルフォーカス調整装置及び合焦アシスト・プログラム |
JP2006023384A (ja) | 2004-07-06 | 2006-01-26 | Konica Minolta Photo Imaging Inc | 撮像装置 |
DE102004037296A1 (de) * | 2004-07-27 | 2006-03-23 | Arnold & Richter Cine Technik Gmbh & Co Betriebs Kg | Verfahren zum Fokussieren des Aufnahmeobjektivs einer Laufbild- oder Videokamera |
JP4344299B2 (ja) * | 2004-09-16 | 2009-10-14 | 富士通マイクロエレクトロニクス株式会社 | 撮像装置及びオートフォーカス合焦時間報知方法 |
JP4443361B2 (ja) * | 2004-09-22 | 2010-03-31 | 株式会社リコー | 撮像装置およびカメラ |
JP3829144B2 (ja) * | 2004-11-25 | 2006-10-04 | シャープ株式会社 | 合焦エリア調節カメラ付携帯端末 |
US7492958B2 (en) | 2005-01-05 | 2009-02-17 | Nokia Corporation | Digital imaging with autofocus |
TWI274952B (en) | 2005-01-26 | 2007-03-01 | Asia Optical Co Inc | Method and system for automatically adjusting the flash intensity of camera |
JP2006259113A (ja) | 2005-03-16 | 2006-09-28 | Konica Minolta Photo Imaging Inc | レンズ交換式デジタルカメラ |
US8525892B2 (en) | 2005-04-21 | 2013-09-03 | Canon Kabushiki Kaisha | Imaging apparatus and control method therefor |
JP4441879B2 (ja) * | 2005-06-28 | 2010-03-31 | ソニー株式会社 | 信号処理装置および方法、プログラム、並びに記録媒体 |
JP4759332B2 (ja) | 2005-07-01 | 2011-08-31 | Hoya株式会社 | 画像信号処理装置 |
JP4929630B2 (ja) | 2005-07-06 | 2012-05-09 | ソニー株式会社 | 撮像装置、制御方法、およびプログラム |
JP4528235B2 (ja) | 2005-09-29 | 2010-08-18 | 富士フイルム株式会社 | デジタルカメラ |
KR101114407B1 (ko) | 2005-10-11 | 2012-02-22 | 삼성전자주식회사 | 터치 스크린을 구비한 휴대용 단말기의 촬영 방법 |
US7697827B2 (en) * | 2005-10-17 | 2010-04-13 | Konicek Jeffrey C | User-friendlier interfaces for a camera |
EP2448247A1 (de) * | 2005-11-02 | 2012-05-02 | Olympus Corporation | Bildprozessor für elektronische Kamera |
RU2384968C1 (ru) * | 2005-12-06 | 2010-03-20 | Панасоник Корпорэйшн | Цифровой фотоаппарат |
WO2007079425A2 (en) | 2005-12-30 | 2007-07-12 | Apple Inc. | Portable electronic device with multi-touch input |
KR101143164B1 (ko) * | 2006-01-09 | 2012-05-08 | 삼성전자주식회사 | 휴대 단말기에서 동영상 촬영 시 오디오 입력 신호 처리 방법 및 장치 |
US20070222859A1 (en) * | 2006-03-23 | 2007-09-27 | Coban Research And Technologies, Inc. | Method for digital video/audio recording with backlight compensation using a touch screen control panel |
TWI320132B (en) | 2006-05-30 | 2010-02-01 | Method for triggering strobe of image-capturing device | |
JP2008011497A (ja) * | 2006-06-01 | 2008-01-17 | Canon Inc | カメラ装置 |
US8169421B2 (en) | 2006-06-19 | 2012-05-01 | Cypress Semiconductor Corporation | Apparatus and method for detecting a touch-sensor pad gesture |
US7860382B2 (en) | 2006-10-02 | 2010-12-28 | Sony Ericsson Mobile Communications Ab | Selecting autofocus area in an image |
US7664384B2 (en) | 2006-11-07 | 2010-02-16 | Sony Ericsson Mobile Communications Ab | User defined autofocus area |
US20080278589A1 (en) | 2007-05-11 | 2008-11-13 | Karl Ola Thorn | Methods for identifying a target subject to automatically focus a digital camera and related systems, and computer program products |
US20080303922A1 (en) * | 2007-06-08 | 2008-12-11 | Imran Chaudhri | Image capture |
US8497928B2 (en) | 2007-07-31 | 2013-07-30 | Palm, Inc. | Techniques to automatically focus a digital camera |
US7844174B2 (en) * | 2008-07-31 | 2010-11-30 | Fuji Xerox Co., Ltd. | System and method for manual selection of multiple evaluation points for camera control |
-
2007
- 2007-07-31 US US11/831,051 patent/US8497928B2/en not_active Ceased
-
2008
- 2008-07-29 PT PT16206486T patent/PT3168681T/pt unknown
- 2008-07-29 EP EP20158892.8A patent/EP3712697B1/de active Active
- 2008-07-29 DK DK16206486.9T patent/DK3168681T3/da active
- 2008-07-29 TR TR2019/04084T patent/TR201904084T4/tr unknown
- 2008-07-29 EP EP16206486.9A patent/EP3168681B1/de active Active
- 2008-07-29 ES ES08782499.1T patent/ES2647068T3/es active Active
- 2008-07-29 EP EP19154330.5A patent/EP3506006B1/de active Active
- 2008-07-29 NO NO08782499A patent/NO2174186T3/no unknown
- 2008-07-29 EP EP16190939.5A patent/EP3136167B1/de active Active
- 2008-07-29 ES ES20158892T patent/ES2970284T3/es active Active
- 2008-07-29 PL PL16206486T patent/PL3168681T3/pl unknown
- 2008-07-29 ES ES16190939T patent/ES2796279T3/es active Active
- 2008-07-29 HU HUE16190939A patent/HUE049065T2/hu unknown
- 2008-07-29 PT PT87824991T patent/PT2174186T/pt unknown
- 2008-07-29 PL PL08782499T patent/PL2174186T3/pl unknown
- 2008-07-29 HU HUE08782499A patent/HUE034427T2/en unknown
- 2008-07-29 SI SI200832050T patent/SI3168681T1/sl unknown
- 2008-07-29 DK DK08782499.1T patent/DK2174186T3/da active
- 2008-07-29 SI SI200831869T patent/SI2174186T1/sl unknown
- 2008-07-29 WO PCT/US2008/071481 patent/WO2009018279A1/en active Application Filing
- 2008-07-29 HU HUE16206486A patent/HUE042961T2/hu unknown
- 2008-07-29 EP EP08782499.1A patent/EP2174186B1/de active Active
- 2008-07-29 EP EP20198227.9A patent/EP3779589A1/de active Pending
- 2008-07-29 ES ES16206486T patent/ES2716773T3/es active Active
- 2008-07-29 PL PL20158892.8T patent/PL3712697T3/pl unknown
-
2020
- 2020-06-22 US US16/908,579 patent/USRE49039E1/en active Active
Non-Patent Citations (1)
Title |
---|
None * |
Also Published As
Publication number | Publication date |
---|---|
US20090033786A1 (en) | 2009-02-05 |
PL2174186T3 (pl) | 2018-01-31 |
EP2174186A1 (de) | 2010-04-14 |
ES2716773T3 (es) | 2019-06-14 |
EP2174186B1 (de) | 2017-08-30 |
EP3712697B1 (de) | 2024-01-17 |
EP3712697C0 (de) | 2024-01-17 |
WO2009018279A1 (en) | 2009-02-05 |
PL3712697T3 (pl) | 2024-05-27 |
EP3506006B1 (de) | 2020-09-30 |
SI3168681T1 (sl) | 2019-04-30 |
EP3136167B1 (de) | 2020-03-18 |
PT3168681T (pt) | 2019-03-04 |
DK2174186T3 (da) | 2017-11-27 |
EP3168681A1 (de) | 2017-05-17 |
HUE034427T2 (en) | 2018-05-28 |
ES2796279T3 (es) | 2020-11-26 |
HUE042961T2 (hu) | 2019-07-29 |
EP3779589A1 (de) | 2021-02-17 |
USRE49039E1 (en) | 2022-04-19 |
TR201904084T4 (tr) | 2019-05-21 |
US8497928B2 (en) | 2013-07-30 |
EP3506006A1 (de) | 2019-07-03 |
NO2174186T3 (de) | 2018-01-27 |
EP2174186A4 (de) | 2010-08-04 |
ES2970284T3 (es) | 2024-05-27 |
EP3136167A1 (de) | 2017-03-01 |
EP3712697A1 (de) | 2020-09-23 |
HUE049065T2 (hu) | 2020-09-28 |
ES2647068T3 (es) | 2017-12-19 |
SI2174186T1 (sl) | 2017-11-30 |
DK3168681T3 (da) | 2019-05-13 |
PT2174186T (pt) | 2017-09-29 |
PL3168681T3 (pl) | 2019-09-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
USRE49039E1 (en) | Techniques to automatically focus a digital camera | |
JP7355941B2 (ja) | 長焦点シナリオにおける撮影方法および端末 | |
JP5358733B2 (ja) | タッチスクリーン機能を変更するためのシステム及び方法 | |
US8203640B2 (en) | Portable terminal having touch sensing based image capture function and image capture method therefor | |
CN113542581A (zh) | 多路录像的取景方法、图形用户界面及电子设备 | |
CN115191110A (zh) | 一种视频拍摄方法与电子设备 | |
US20100066751A1 (en) | Adjusting the display orientation of an image on a mobile terminal | |
CN112019895B (zh) | 功能操作控制方法、功能操作控制装置及存储介质 | |
US20130235233A1 (en) | Methods and devices for capturing images | |
KR100605803B1 (ko) | 휴대용단말을 이용한 멀티분할 촬영 장치 및 방법 | |
WO2022206783A1 (zh) | 拍摄方法、装置、电子设备及可读存储介质 | |
CA2804594A1 (en) | Methods and devices for capturing images | |
CN109005576A (zh) | 一种终端相机的切换方法及终端设备 | |
WO2023202431A1 (zh) | 一种定向拾音方法及设备 | |
CN106470319B (zh) | 一种信息处理方法及设备 | |
CN116339881A (zh) | 内容显示方法、电子设备及计算机可读存储介质 | |
CN116027886A (zh) | 一种折叠屏的显示方法及电子设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN PUBLISHED |
|
AC | Divisional application: reference to earlier application |
Ref document number: 2174186 Country of ref document: EP Kind code of ref document: P |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MT NL NO PL PT RO SE SI SK TR |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20171114 |
|
RBV | Designated contracting states (corrected) |
Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MT NL NO PL PT RO SE SI SK TR |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G03B 13/36 20060101ALI20180118BHEP Ipc: G02B 7/28 20060101ALI20180118BHEP Ipc: G03B 17/00 20060101AFI20180118BHEP Ipc: H04N 5/232 20060101ALI20180118BHEP |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
INTG | Intention to grant announced |
Effective date: 20180226 |
|
GRAJ | Information related to disapproval of communication of intention to grant by the applicant or resumption of examination proceedings by the epo deleted |
Free format text: ORIGINAL CODE: EPIDOSDIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
INTC | Intention to grant announced (deleted) | ||
INTG | Intention to grant announced |
Effective date: 20180806 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
AC | Divisional application: reference to earlier application |
Ref document number: 2174186 Country of ref document: EP Kind code of ref document: P |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MT NL NO PL PT RO SE SI SK TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: REF Ref document number: 1093734 Country of ref document: AT Kind code of ref document: T Effective date: 20190215 |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 602008058942 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: PT Ref legal event code: SC4A Ref document number: 3168681 Country of ref document: PT Date of ref document: 20190304 Kind code of ref document: T Free format text: AVAILABILITY OF NATIONAL TRANSLATION Effective date: 20190222 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: NV Representative=s name: NOVAGRAAF INTERNATIONAL SA, CH |
|
REG | Reference to a national code |
Ref country code: RO Ref legal event code: EPE |
|
REG | Reference to a national code |
Ref country code: SE Ref legal event code: TRGR |
|
REG | Reference to a national code |
Ref country code: DK Ref legal event code: T3 Effective date: 20190506 |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: FP |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG4D |
|
REG | Reference to a national code |
Ref country code: ES Ref legal event code: FG2A Ref document number: 2716773 Country of ref document: ES Kind code of ref document: T3 Effective date: 20190614 |
|
REG | Reference to a national code |
Ref country code: NO Ref legal event code: T2 Effective date: 20190130 |
|
REG | Reference to a national code |
Ref country code: HU Ref legal event code: AG4A Ref document number: E042961 Country of ref document: HU |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 |
|
REG | Reference to a national code |
Ref country code: GR Ref legal event code: EP Ref document number: 20190401319 Country of ref document: GR Effective date: 20190620 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190530 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190430 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 602008058942 Country of ref document: DE |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20191031 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MC Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190729 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190130 |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: UEP Ref document number: 1093734 Country of ref document: AT Kind code of ref document: T Effective date: 20190130 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: RO Payment date: 20230622 Year of fee payment: 16 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: IT Payment date: 20230712 Year of fee payment: 16 Ref country code: ES Payment date: 20230802 Year of fee payment: 16 Ref country code: CH Payment date: 20230801 Year of fee payment: 16 Ref country code: AT Payment date: 20230626 Year of fee payment: 16 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: SI Payment date: 20230622 Year of fee payment: 16 Ref country code: SE Payment date: 20230710 Year of fee payment: 16 Ref country code: HU Payment date: 20230628 Year of fee payment: 16 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: IE Payment date: 20240625 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20240613 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DK Payment date: 20240626 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GR Payment date: 20240625 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: NL Payment date: 20240617 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: CZ Payment date: 20240620 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: NO Payment date: 20240625 Year of fee payment: 17 Ref country code: FR Payment date: 20240613 Year of fee payment: 17 Ref country code: FI Payment date: 20240626 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: PL Payment date: 20240619 Year of fee payment: 17 Ref country code: PT Payment date: 20240614 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: TR Payment date: 20240628 Year of fee payment: 17 Ref country code: BE Payment date: 20240614 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20240613 Year of fee payment: 17 |