WO2023231009A1 - 一种对焦方法、装置及存储介质 - Google Patents

一种对焦方法、装置及存储介质 Download PDF

Info

Publication number
WO2023231009A1
WO2023231009A1 PCT/CN2022/096934 CN2022096934W WO2023231009A1 WO 2023231009 A1 WO2023231009 A1 WO 2023231009A1 CN 2022096934 W CN2022096934 W CN 2022096934W WO 2023231009 A1 WO2023231009 A1 WO 2023231009A1
Authority
WO
WIPO (PCT)
Prior art keywords
target focus
target
area
blocks
focus
Prior art date
Application number
PCT/CN2022/096934
Other languages
English (en)
French (fr)
Inventor
姬向东
Original Assignee
北京小米移动软件有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京小米移动软件有限公司 filed Critical 北京小米移动软件有限公司
Priority to PCT/CN2022/096934 priority Critical patent/WO2023231009A1/zh
Priority to CN202280004614.XA priority patent/CN117652152A/zh
Publication of WO2023231009A1 publication Critical patent/WO2023231009A1/zh

Links

Images

Definitions

  • the present disclosure relates to the field of cameras, and in particular to focusing methods, devices and storage media.
  • the focusing method is to divide the imaging area into multiple windows, and further divide the central window into grid blocks to determine whether the credibility of each grid in the central window meets the focus requirements. If the focus requirements are not met, then Calculate the image data information of the windows around the central window until a window that meets the focus requirements is found as the final focus area.
  • the focusing method may cause focus loss, blurred focus, or the inability to automatically focus on the object that the user wants to focus on.
  • the present disclosure provides a focusing method, device and storage medium.
  • a focusing method including:
  • a target focus area is determined based on the target focus object, and focusing is performed based on the target focus area.
  • performing object recognition in the imaging area and determining a target focus object among the recognized objects includes:
  • an object matching the focus mode is determined as a target focus object.
  • determining the target focus area based on the target focus object includes:
  • the target focus object is a single target focus object, then the block where the single target focus object is located is used as the target focus area;
  • the target focus area is determined based on the blocks to which the at least two target focus objects belong.
  • determining the target focus area based on the blocks to which the at least two target focus objects belong includes:
  • the target focus area is determined based on the pixel ratio of the number of pixels of the target focus objects in the different blocks in the imaging area.
  • determining the target focus area based on the pixel ratio of the number of pixels of the target focus object in the different blocks to the pixels in the imaging area includes:
  • the color dynamic change ranges of the multiple target blocks are determined, and the block with the largest color dynamic change range is selected as the target focus area.
  • focusing based on the target focus area includes:
  • the target focus area is expanded at a set ratio, so that the expanded target focus area includes target focus objects outside the area;
  • a focusing device including:
  • a recognition unit used to perform object recognition in the imaging area and determine the target focus object among the recognized objects
  • a focusing unit is configured to determine a target focus area based on the target focus object, and to focus based on the target focus area.
  • the recognition unit uses the following method to perform object recognition in the imaging area, and determines the target focus object among the recognized objects:
  • an object matching the focus mode is determined as a target focus object.
  • the focusing unit determines the target focus area based on the target focus object in the following manner:
  • the target focus object is a single target focus object, then the block where the single target focus object is located is used as the target focus area;
  • the target focus area is determined based on the blocks to which the at least two target focus objects belong.
  • the focusing unit determines the target focus area based on the blocks to which the at least two target focus objects belong in the following manner:
  • the target focus area is determined based on the pixel ratio of the number of pixels of the target focus objects in the different blocks in the imaging area.
  • the focusing unit uses the following method to determine the target focus area based on the pixel ratio of the number of pixels of the target focus object in the different blocks to the pixels in the imaging area:
  • the color dynamic change ranges of the multiple target blocks are determined, and the block with the largest color dynamic change range is selected as the target focus area.
  • the focusing unit performs focusing based on the target focus area in the following manner:
  • the target focus area is expanded at a set ratio, so that the expanded target focus area includes target focus objects outside the area;
  • a focusing device including:
  • Processor memory used to store instructions executable by the processor
  • the processor is configured to: execute the focusing method described in the first aspect or any implementation of the first aspect.
  • a non-transitory computer-readable storage medium which when instructions in the storage medium are executed by a processor of a mobile terminal, enables the mobile terminal to execute the first aspect or the first The focusing method described in any one of the embodiments.
  • the technical solution provided by the embodiments of the present disclosure may include the following beneficial effects: determine the imaging area, perform object recognition on the imaging area, determine the target focus object among the recognized objects, determine the target focus area based on the target focus object, and in the target focus area to focus within.
  • the focusing method provided by the embodiments of the present disclosure can accurately focus on the physical information that the user is interested in, thereby improving the focusing accuracy, ensuring the clarity of the object that the user wants to focus on, and improving the user's photography experience.
  • FIG. 1 is a flowchart of a focusing method according to an exemplary embodiment.
  • FIG. 2 is a flowchart of a method for object recognition and determining a target focus object according to an exemplary embodiment.
  • FIG. 3 is a flowchart of a method for determining a target focus area according to an exemplary embodiment.
  • FIG. 4 is a flowchart of a method for determining a target focus area according to an exemplary embodiment.
  • FIG. 5 is a flowchart of a method for determining a target focus area according to an exemplary embodiment.
  • FIG. 6 shows a schematic diagram of a focusing method according to an exemplary embodiment of the present disclosure.
  • FIG. 7 shows a schematic diagram of focusing according to an exemplary embodiment of the present disclosure.
  • FIG. 8 is a block diagram of a focusing device according to an exemplary embodiment.
  • FIG. 9 is a block diagram of a device for focusing according to an exemplary embodiment.
  • the focusing method provided by the embodiments of the present disclosure can be applied to different autofocus scenes and can better distinguish the background and the foreground.
  • it can be a scene with multiple depths, or a scene with a flat area in the middle of the image.
  • focusing methods such as area of interest focusing and RGB analysis are used.
  • the imaging area needs to be divided into blocks, and the center window is further divided into grid blocks, that is, a piece of w (width) * h (height) is intercepted from a characteristic place in the camera.
  • the difference between two adjacent pixels in the image is the absolute value, and the same operation is performed by summing all pixels. The larger the value, the clearer the image, which meets the focus requirement. If the focus requirement is not met, the pixels in the surrounding windows are analyzed and the absolute value is taken, until a block that meets the focus requirement appears and is used as the focus area.
  • embodiments of the present disclosure provide a focusing method.
  • the focusing method when receiving the user's focus information requirement, the current area is determined to be the imaging area, and objects in the imaging area are identified. After the identified objects Determine the target focus object, determine the focus area based on the target focus object, and focus based on the target focus area. This reduces focus loss and focus blur, and improves the accuracy of autofocusing on the object the user wants to focus on, ensuring the clarity and accuracy of the focus area to the greatest extent possible, and improving the user's photography experience.
  • FIG. 1 is a flowchart of a focusing method according to an exemplary embodiment. As shown in Figure 1, the focusing method is used in the terminal and includes the following steps.
  • step S11 the imaging area is determined.
  • step S12 object recognition is performed on the imaging area, and a target focus object is determined among the recognized objects.
  • step S13 the target focus area is determined based on the target focus object, and focusing is performed based on the target focus area.
  • the content presented in the viewfinder is determined as the imaging area, and the objects in the imaging area are identified.
  • the target focus object is included in the recognized objects, and the target focus object is determined according to the area where the target focus object is located.
  • Target focus area focus based on the target focus area.
  • the viewing frame of the terminal is the imaging area.
  • the imaging area contains multiple objects. Multiple objects are identified.
  • the target focus object is a tree. The tree is determined among the multiple objects.
  • the position of the tree in the imaging area is the target focus area. Focus on the target focus area, that is, focus on the area where the trees are located.
  • the focusing method provided can achieve precise focusing on the area where the user wants to focus on the object.
  • the focusing method needs to determine the target focus object among the recognized objects.
  • FIG. 2 is a flowchart of a method for object recognition and determining a target focus object according to an exemplary embodiment. As shown in Figure 2, performing object recognition in the imaging area and determining the target focus object among the recognized objects includes the following steps.
  • step S21 the focus mode selected by the user is obtained, and the focus mode is used to represent the target focus object to be focused.
  • the user can select a focus mode on the camera interface.
  • the focus mode is used to represent the target object to be focused.
  • the focus mode can be the sky, land, people, animals, trees, etc. Get the focus mode selected by the user, that is, get the type of object the user wants to focus on.
  • step S22 the imaging area is divided into blocks, and the divided imaging area is semantically segmented to identify all objects included in the imaging area.
  • the user's focus information requirements are received and the imaging area is divided into blocks.
  • the blocking method can be 3*3, 5*5 and other blocking methods, and semantic segmentation is used to identify the divided imaging areas. Information about all object types contained in .
  • the imaging area is divided according to a 3*3 blocking method; of course, the blocking method in this embodiment can be preset in the terminal, or customized by the user according to needs. This embodiment does not specifically limit this.
  • semantic segmentation is a basic task in computer vision.
  • visual input needs to be divided into different semantic interpretable categories.
  • semantic interpretability means that the classification categories are meaningful in the real world. For example, if you need to distinguish trees, distinguish all pixels belonging to trees in the image and color these pixels blue.
  • step S23 the object matching the focus mode is determined as the target focus object among all the recognized objects.
  • the object matching the focus mode selected by the user is determined as the target focus object.
  • the focus mode selected by the user is stone
  • the imaging area contains object information such as the sky, trees, people, and stones.
  • the above object information has been recognized through semantic segmentation.
  • the stone is selected from the identified object information as the target focus. object.
  • the focusing method provided can accurately identify the object that the user wants to focus on, thereby improving the accuracy of focusing and reducing focus blur.
  • the focusing method needs to determine the target focus area according to the target focus object.
  • FIG. 3 is a flowchart of a method for determining a target focus area according to an exemplary embodiment. As shown in Figure 3, determining the target focus area based on the target focus object includes the following steps.
  • step S31 if the target focus object is a single target focus object, the block where the single target focus object is located is used as the target focus area.
  • the target focus area is determined based on the block where the single target focus object is located. For example, if the focus mode selected by the user is trees, the imaging area is divided into blocks, the tree is identified based on semantic segmentation, and the block where the tree is located is determined as the target focus area.
  • step S32 if the number of target focus objects is at least two, the target focus area is determined based on the blocks to which the at least two target focus objects belong.
  • the target focus area when determining the target focus area, if there are two or more target focus objects, the target focus area is determined based on the blocks where the two or more target focus objects are located.
  • the focusing method provided can optimally determine the focus area, making the focus area more suitable for the user's needs.
  • the target focus area needs to be determined based on the blocks to which the multiple focus objects belong.
  • FIG. 4 is a flowchart of a method for determining a target focus area according to an exemplary embodiment. As shown in Figure 4, determining a target focus area based on the blocks to which at least two target focus objects belong includes the following steps.
  • step S41 if the blocks to which at least two target focus objects belong are the same block, then the block is used as the target focus area.
  • the block where the target focus object is located is the same block, then the block is used as the target focus area.
  • the imaging area is divided into blocks, and semantic segmentation is used to identify that the imaging area contains three people, and these three people are all in the same block, then the block is determined as the target focus area.
  • step S42 if the blocks to which at least two target focus objects belong are different blocks, the target focus area is determined based on the pixel ratio of the number of pixels of the target focus objects in the different blocks in the imaging area.
  • the target focus area is determined based on the pixel proportions of the target focus objects contained in different blocks in the imaging area.
  • the pixel ratio refers to the ratio of the pixels of the target focus object to the pixels in the imaging area.
  • the pixel ratio of the object in the imaging area can be used to improve the focusing accuracy, thereby improving the user's photographing experience.
  • the target focus area if there are multiple target focus areas, it is necessary to determine the target focus area with the largest color dynamic range according to the color dynamic range of the target focus object.
  • FIG. 5 is a flowchart of a method for determining a target focus area according to an exemplary embodiment. As shown in Figure 5, determining the target focus area based on the pixel ratio of the pixel number of the target focus object in the imaging area in different blocks includes the following steps.
  • step S51 for each of the different blocks, the pixel ratio of the number of pixels of the target focus object in the imaging area is determined.
  • step S52 target blocks whose pixel ratio is greater than a preset threshold are determined, and the number of target blocks is determined.
  • target blocks with a pixel ratio greater than a preset threshold and the number of target blocks are selected.
  • step S53 if the number of target blocks is one, the target block is used as the target focus area.
  • the preset threshold can be preset in the terminal, or can be set independently by the user according to needs. The embodiments of the present disclosure do not specifically limit this.
  • the imaging area is divided into blocks, the divided imaging area is semantically segmented, and all objects in the imaging area are recognized.
  • the trees occupy three subdivisions. blocks, respectively, are the first block, the second block and the third block.
  • the proportion of pixels of trees in the imaging area in the third block is greater than the preset threshold. The first block and the second block If the number of pixels of the trees in the block accounts for less than the preset threshold in the imaging area, the third block is selected as the target focus area.
  • step S54 if there are multiple target blocks, the color dynamic change ranges of the multiple target blocks are determined, and the block with the largest color dynamic change range is selected as the target focus area.
  • the block with the largest change in color dynamic range will be used as the target focus area.
  • the value of the color dynamic range is the maximum and minimum values of the target focus object.
  • the imaging area is divided into blocks and semantically segmented, and all objects in the imaging area are recognized.
  • people occupy four blocks, namely the first block, The second block, the third block and the fourth block, where the proportion of pixels of people in the first block and the second block is less than the first preset threshold, and the proportion of pixels of the people in the third block and the fourth block is If the pixel proportion is greater than the preset threshold, and the pixel proportion of the person in the first block and the second block is less than the preset threshold, then the color dynamic range of the target focus object contained in the third block and the fourth block is the largest. Determine the target focus area. If the color dynamic range of the target focus object contained in the third block is greater than the target focus object contained in the fourth block, determine the third block as the target focus area.
  • the focusing method provided can identify the object that the user most wants to focus on from at least two target focus objects, so as to achieve precise focusing and reduce the occurrence of focus loss, focus blur and other situations.
  • the target focus area needs to be expanded to a certain extent.
  • the center of the target focus area is used as the center point, and the target focus area is expanded at a set ratio, so that the expanded target is in focus.
  • the area includes target focus objects outside the area, and focus is performed based on the expanded target focus area.
  • FIG. 6 shows a schematic diagram of a focusing method according to an exemplary embodiment of the present disclosure.
  • the pixels of each block are 50*50
  • the imaging area is divided into blocks
  • the fifth block is determined to be the target focus block.
  • there are target focus objects in the third block and the fourth block then Take the center point of the fifth segment as the center, expand outward by a quarter, and focus on the expanded target focus area.
  • FIG. 7 shows a focusing flowchart according to an exemplary embodiment of the present disclosure.
  • the user starts taking pictures and selects the focus mode, divides the imaging area into blocks, that is, multi-window blocking, and performs semantic segmentation on the divided imaging area to identify all objects in the imaging area. If the target in the imaging area is in focus If the number of objects is one, then determine the block where the target focus object is located as the target focus area, and focus; if the number of target focus objects in the imaging area is two or more, then determine the target focus object in the block according to the location of the target focus area.
  • the block is used as the target focus area, where the target block is a block whose pixel ratio is greater than the first preset threshold and the difference in pixel ratio between any two blocks is smaller than the second preset threshold for focusing.
  • the focusing method provided can make full use of the target focus object type, pixel ratio, and color information in the imaging area, reduce focus loss, ensure the clarity of the focus area as much as possible, and improve focus accuracy, thereby Improve users’ photography experience.
  • embodiments of the present disclosure also provide a focusing device.
  • the focusing device provided by the embodiments of the present disclosure includes corresponding hardware structures and/or software modules for performing each function.
  • the embodiments of the present disclosure can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software driving the hardware depends on the specific application and design constraints of the technical solution. Those skilled in the art can use different methods to implement the described functions for each specific application, but such implementation should not be considered to go beyond the scope of the technical solutions of the embodiments of the present disclosure.
  • FIG. 8 is a block diagram of a focusing device according to an exemplary embodiment.
  • the focusing device 100 includes a determining unit 101 , an identifying unit 102 and a focusing unit 103 .
  • Determining unit 101 used to determine the imaging area.
  • the recognition unit 102 is configured to perform object recognition in the imaging area and determine a target focus object among the recognized objects.
  • the focusing unit 103 is configured to determine a target focus area based on the target focus object, and to focus based on the target focus area.
  • the recognition unit 102 performs object recognition in the imaging area and determines the target focus object among the recognized objects in the following manner: obtains the focus mode selected by the user, and the focus mode is used to characterize the target to be focused. Focus on the object; divide the imaging area into blocks, and perform semantic segmentation on the divided imaging area to identify all objects included in the imaging area; determine the object that matches the focus mode among all the identified objects as the target Focus on the object.
  • the focusing unit 103 determines the target focus area based on the target focus object in the following manner: if the target focus object is a single target focus object, then the block where the single target focus object is located is used as the target focus area; if If the number of target focus objects is at least two, the target focus area is determined based on the blocks to which at least two target focus objects belong.
  • the focusing unit 103 determines the target focus area based on the blocks to which at least two target focus objects belong: if the blocks to which the at least two target focus objects belong are the same block, then the block is used as the target. Focus area; if at least two target focus objects belong to different blocks, the target focus area is determined based on the proportion of pixels of the target focus objects in the different blocks to the pixels in the imaging area.
  • the focusing unit 103 determines the target focus area based on the pixel ratio of the number of pixels of the target focus object in the imaging area in different blocks in the following manner:
  • the focusing unit 103 performs focusing based on the target focus area in the following manner: if there is a target focus object within the set range of the target focus area; then the center of the target focus area is used as the center point, and the target focus area is expanded according to the set ratio. Target focus area so that the expanded target focus area includes target focus objects outside the area; focus is performed based on the expanded target focus area.
  • FIG. 9 is a block diagram of a device 200 for focusing according to an exemplary embodiment.
  • the device 200 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, or the like.
  • device 200 may include one or more of the following components: processing component 202, memory 204, power component 206, multimedia component 208, audio component 210, input/output (I/O) interface 212, sensor component 214, and Communication component 216.
  • Processing component 202 generally controls the overall operations of device 200, such as operations associated with display, phone calls, data communications, camera operations, and recording operations.
  • the processing component 202 may include one or more processors 220 to execute instructions to complete all or part of the steps of the above method.
  • processing component 202 may include one or more modules that facilitate interaction between processing component 202 and other components.
  • processing component 202 may include a multimedia module to facilitate interaction between multimedia component 208 and processing component 202.
  • Memory 204 is configured to store various types of data to support operations at device 200 . Examples of such data include instructions for any application or method operating on device 200, contact data, phonebook data, messages, pictures, videos, etc.
  • Memory 204 may be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EEPROM), Programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read-only memory
  • EEPROM erasable programmable read-only memory
  • EPROM Programmable read-only memory
  • PROM programmable read-only memory
  • ROM read-only memory
  • magnetic memory flash memory, magnetic or optical disk.
  • Power component 206 provides power to various components of device 200 .
  • Power components 206 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power to device 200 .
  • Multimedia component 208 includes a screen that provides an output interface between the device 200 and the user.
  • the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touches, swipes, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide action.
  • multimedia component 208 includes a front-facing camera and/or a rear-facing camera.
  • the front camera and/or the rear camera may receive external multimedia data.
  • Each front-facing camera and rear-facing camera can be a fixed optical lens system or have a focal length and optical zoom capabilities.
  • Audio component 210 is configured to output and/or input audio signals.
  • audio component 210 includes a microphone (MIC) configured to receive external audio signals when device 200 is in operating modes, such as call mode, recording mode, and voice recognition mode. The received audio signals may be further stored in memory 204 or sent via communications component 216 .
  • audio component 210 also includes a speaker for outputting audio signals.
  • the I/O interface 212 provides an interface between the processing component 202 and a peripheral interface module, which may be a keyboard, a click wheel, a button, etc. These buttons may include, but are not limited to: Home button, Volume buttons, Start button, and Lock button.
  • Sensor component 214 includes one or more sensors for providing various aspects of status assessment for device 200 .
  • the sensor component 214 can detect the open/closed state of the device 200, the relative positioning of components, such as the display and keypad of the device 200, and the sensor component 214 can also detect a change in position of the device 200 or a component of the device 200. , the presence or absence of user contact with the device 200 , device 200 orientation or acceleration/deceleration and temperature changes of the device 200 .
  • Sensor assembly 214 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • Sensor assembly 214 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 214 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • Communication component 216 is configured to facilitate wired or wireless communication between apparatus 200 and other devices.
  • Device 200 may access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof.
  • the communication component 216 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel.
  • the communications component 216 also includes a near field communications (NFC) module to facilitate short-range communications.
  • NFC near field communications
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • apparatus 200 may be configured by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable Gate array (FPGA), controller, microcontroller, microprocessor or other electronic components are implemented for executing the above method.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field programmable Gate array
  • controller microcontroller, microprocessor or other electronic components are implemented for executing the above method.
  • a non-transitory computer-readable storage medium including instructions such as a memory 204 including instructions, which can be executed by the processor 220 of the device 200 to complete the above method is also provided.
  • the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
  • “plurality” in this disclosure refers to two or more, and other quantifiers are similar.
  • “And/or” describes the relationship between related objects, indicating that there can be three relationships.
  • a and/or B can mean: A exists alone, A and B exist simultaneously, and B exists alone.
  • the character “/” generally indicates that the related objects are in an “or” relationship.
  • the singular forms “a”, “the” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise.
  • first, second, etc. are used to describe various information, but such information should not be limited to these terms. These terms are only used to distinguish information of the same type from each other and do not imply a specific order or importance. In fact, expressions such as “first” and “second” can be used interchangeably.
  • first information may also be called second information, and similarly, the second information may also be called first information.

Landscapes

  • Studio Devices (AREA)

Abstract

本公开是关于一种对焦方法、装置及存储介质。其中,对焦方法包括:确定成像区域;对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体;基于所述目标对焦物体确定目标对焦区域,并基于所述目标对焦区域进行对焦。通过本公开的对焦方法,可以提升对焦准确度,保证用户想要对焦的物体的清晰度,从而提升用户的拍照体验。

Description

一种对焦方法、装置及存储介质 技术领域
本公开涉及相机领域,尤其涉及对焦方法、装置及存储介质。
背景技术
相关技术中,对焦方法是对成像区域划分多个窗口,并对中心窗口进一步进行网格分块,判断中心窗口中的每个网格可信度是否满足对焦要求,若不满足对焦要求,则计算中心窗口周围窗口的图像数据信息,直至找到一个满足对焦要求的窗口作为最终对焦区域。然而,在实际应用中,对焦方法可能会导致丟焦、对焦模糊或者不能自动对焦到用户想要对焦的物体上。
发明内容
为克服相关技术中存在的问题,本公开提供一种对焦方法、装置及存储介质。
根据本公开实施例的第一方面,提供一种对焦方法,包括:
确定成像区域;
对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体;
基于所述目标对焦物体确定目标对焦区域,并基于所述目标对焦区域进行对焦。
在一种实施方式中,所述对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体,包括:
获取用户选择的对焦模式,所述对焦模式用于表征待对焦的目标对焦物体;
对所述成像区域进行分块,并对分块后的成像区域进行语义分割,以识别得到所述成像区域中包括的全部物体;
在识别到的所述全部物体中确定与所述对焦模式匹配的物体,作为目标对焦物体。
在一种实施方式中,所述基于所述目标对焦物体确定目标对焦区域,包括:
若所述目标对焦物体为单一的目标对焦物体,则将所述单一的目标对焦物体所在的分块作为目标对焦区域;
若所述目标对焦物体的数量为至少两个,则基于所述至少两个目标对焦物体所属分块,确定目标对焦区域。
在一种实施方式中,所述基于所述至少两个目标对焦物体所属分块,确定目标对焦区域,包括:
若所述至少两个目标对焦物体所属分块为同一分块,则将所述分块作为目标对焦区域;
若所述至少两个目标对焦物体所属分块为不同分块,则基于所述不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域。
在一种实施方式中,所述基于所述不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域,包括:
针对所述不同分块中各分块,分别确定目标对焦物体的像素数在所述成像区域的像素占比;
确定像素占比大于预设阈值的目标分块,并确定所述目标分块的数量;
若所述目标分块的数量为一个,则将所述目标分块作为目标对焦区域;
若所述目标分块的数量为多个,则确定多个目标分块的颜色动态变化范围,并选择颜色动态变化范围最大的分块作为目标对焦区域。
在一种实施方式中,所述基于所述目标对焦区域进行对焦,包括:
若在所述目标对焦区域的设定范围内存在目标对焦物体;
则以所述目标对焦区域中心为中心点,以设定比例扩展所述目标对焦区域,以使扩展后的目标对焦区域包括区域外的目标对焦物体;
基于扩展后的目标对焦区域进行对焦。
根据本公开实施例的第二方面,提供一种对焦装置,包括:
确定单元,用于确定成像区域;
识别单元,用于对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体;
对焦单元,用于基于所述目标对焦物体确定目标对焦区域,并基于所述目标对焦区域进行对焦。
在一种实施方式中,所述识别单元采用如下方式对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体:
获取用户选择的对焦模式,所述对焦模式用于表征待对焦的目标对焦物体;
对所述成像区域进行分块,并对分块后的成像区域进行语义分割,以识别得到所述成像区域中包括的全部物体;
在识别到的所述全部物体中确定与所述对焦模式匹配的物体,作为目标对焦物体。
在一种实施方式中,所述对焦单元采用如下方式基于所述目标对焦物体确定目标对焦区域:
若所述目标对焦物体为单一的目标对焦物体,则将所述单一的目标对焦物体所在的分块作为目标对焦区域;
若所述目标对焦物体的数量为至少两个,则基于所述至少两个目标对焦物体所属分块,确定目标对焦区域。
在一种实施方式中,所述对焦单元采用如下方式基于所述至少两个目标对焦物体所属分块,确定目标对焦区域:
若所述至少两个目标对焦物体所属分块为同一分块,则将所述分块作为目标对焦区域;
若所述至少两个目标对焦物体所属分块为不同分块,则基于所述不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域。
在一种实施方式中,所述对焦单元采用如下方式基于所述不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域:
针对不同分块中各分块,分别确定目标对焦物体的像素数在所述成像区域的像素占比;
确定像素占比大于预设阈值的目标分块,并确定所述目标分块的数量;
若所述目标分块的数量为一个,则将所述目标分块作为目标对焦区域;
若所述目标分块的数量为多个,则确定多个目标分块的颜色动态变化范围,并选择颜色动态变化范围最大的分块作为目标对焦区域。
在一种实施方式中,所述对焦单元采用如下方式基于所述目标对焦区域进行对焦:
若在所述目标对焦区域的设定范围内存在目标对焦物体;
则以所述目标对焦区域中心为中心点,以设定比例扩展所述目标对焦区域,以使扩展后的目标对焦区域包括区域外的目标对焦物体;
基于扩展后的目标对焦区域进行对焦。
根据本公开实施例的第三方面,提供一种对焦装置,包括:
处理器;用于存储处理器可执行指令的存储器;
其中,所述处理器被配置为:执行第一方面或第一方面任意一种实施方式中所述的对焦方法。
根据本公开实施例的第四方面,提供一种非临时性计算机可读存储介质,当所述存储介质中的指令由移动终端的处理器执行时,使得移动终端能够执行第一方面或第一方面任意一种实施方式中所述的对焦方法。
本公开的实施例提供的技术方案可以包括以下有益效果:确定成像区域,对成像区域进行物体识别,在识别到的物体中确定目标对焦物体,根据目标对焦物体确定目标对焦区域,在目标对焦区域内进行对焦。本公开实施例提供的对焦方法能够精准的对焦用户感兴 趣的物理信息,从而提升对焦准确度,保证用户想要对焦的物体的清晰度,提升用户的拍照体验。
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,并不能限制本公开。
附图说明
此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本公开的实施例,并与说明书一起用于解释本公开的原理。
图1是根据一示例性实施例示出的一种对焦方法的流程图。
图2是根据一示例性实施例示出的一种物体识别和确定目标对焦物体的方法流程图。
图3是根据一示例性实施例示出的一种确定目标对焦区域的方法流程图。
图4是根据一示例性实施例示出的一种确定目标对焦区域的方法流程图。
图5是根据一示例性实施例示出的一种确定目标对焦区域的方法流程图。
图6示出了本公开一示例性实施例示出的一种对焦方法的示意图。
图7示出了本公开一示例性实施例示出的一种对焦的示意图。
图8是根据一示例性实施例示出的一种对焦装置框图。
图9是根据一示例性实施例示出的一种用于对焦的装置的框图。
具体实施方式
这里将详细地对示例性实施例进行说明,其示例表示在附图中。下面的描述涉及附图时,除非另有表示,不同附图中的相同数字表示相同或相似的要素。以下示例性实施例中所描述的实施方式并不代表与本公开相一致的所有实施方式。
本公开实施例提供的对焦方法可应用于不同的自动对焦场景,能够更好的区分背景和前景。例如,可以是多深度场景的拍摄,也可以是图像中间属于平坦区域场景的拍摄。
相关技术中,在对于多深度场景或中间属于平坦区域场景拍摄的对焦时,采用感兴趣区域对焦、RGB分析等对焦方式。
其中,感兴趣区域对焦中为了确定对焦感兴趣区域,需要对成像区域进行分块,进一步对中心窗口进行网格分块,即从相机中有特征的地方截取一片w(宽)*h(高)的矩形图像出来(感兴趣区域),图像中相邻两个像素做差取绝对值,求和所有像素同一操作,值越大,说明图像越清晰,即满足对焦需求。若不满足对焦需求,则分析周围窗口的像素做差取绝对值,直至有满足对焦需求的分块出现,作为对焦区域。
在实际应用中,对焦方式还存在着研究的进步空间,例如,可以减少丢焦、对焦模糊,或者提升自动对焦到用户想要对焦的物体的准确度。
有鉴于此,本公开实施例提供一种对焦方法,在该对焦方法中,接收到用户的对焦信息需求时,确定当前区域为成像区域,对成像区域内的物体进行识别,在经过识别的物体中确定目标对焦物体,根据目标对焦物体确定对焦区域,并基于目标对焦区域进行对焦。从而实现减少丢焦、对焦模糊,并且提升自动对焦到用户想要对焦的物体的准确度,进最大可能保证对焦区域的清晰度与准确度,提升用户的拍照体验。
图1是根据一示例性实施例示出的一种对焦方法的流程图。如图1所示,对焦方法用于终端中,包括以下步骤。
在步骤S11中,确定成像区域。
在步骤S12中,对成像区域进行物体识别,并在识别到的物体中确定目标对焦物体。
在步骤S13中,基于目标对焦物体确定目标对焦区域,并基于目标对焦区域进行对焦。
本公开实施例中,取景框内呈现的内容确定为成像区域,对成像区域内的物体进行识别,目标对焦物体包含在识别到的物体中,确定目标对焦物体,根据目标对焦物体所在的区域确定目标对焦区域,根据目标对焦区域进行对焦。
例如,终端的取景框为成像区域,成像区域内含有多个物体,对多个物体进行识别,目标对焦物体为树木,在多个物体中确定树木,树木在成像区域的位置为目标对焦区域,对目标对焦区域进行对焦,即对树木所在的区域进行对焦。
本公开实施例中,提供的对焦方法能够实现对精准对焦到用户想要对焦物体所在的区域中。
进一步地,本公开实施例中,对焦方法需要在识别到的物体中确定目标对焦物体。
图2是根据一示例性实施例示出的一种物体识别和确定目标对焦物体的方法流程图。如图2所示,对成像区域进行物体识别,并在识别到的物体中确定目标对焦物体,包括以下步骤。
在步骤S21中,获取用户选择的对焦模式,对焦模式用于表征待对焦的目标对焦物体。
本公开实施例中,用户可以在拍照界面选择对焦模式,对焦模式用于表征待对焦的目标对焦物体,例如,对焦模式可以是天空、陆地、人、动物、树木等。获取用户选择的对焦模式,即获取用户想要对焦的物体种类。
在步骤S22中,对成像区域进行分块,并对分块后的成像区域进行语义分割,以识别得到成像区域中包括的全部物体。
本公开实施例中,接收到用户的对焦信息需求,对成像区域进行分块,其中,分块方法可以是3*3、5*5等分块方法,采用语义分割识别分块后的成像区域中包含的所有物体种类的信息。例如,图6中即为按照3*3的分块方法将成像区域进行分割;当然,该实施 例中的分块方法可以是终端中预设的,也可以是用户根据需求进行自定义的,本实施例对此不作具体限制。
其中,语义分割是计算机视觉中的基本任务,在语义分割中需要将视觉输入分为不同的语义可解释类别,“语义的可解释性”即分类类别在真实世界中是有意义的。例如,若需要区分树木,则区分图像中属于树木的所有像素,并把这些像素涂成蓝色。
在步骤S23中,在识别到的全部物体中确定与对焦模式匹配的物体,作为目标对焦物体。
本公开实施例中,根据用户选择的对焦模式,在成像区域采用语义分割得到包含的所有物体种类的信息中,确定与用户选择的对焦模式匹配的物体,作为目标对焦物体。
例如,用户选择的对焦模式为石头,成像区域中包含有天空、树木、人、石头等物体信息,上述物体信息已通过语义分割识别到,在已被识别的物体信息中选择石头,作为目标对焦物体。
本公开实施例中,提供的对焦方法能够准确的识别用户想要对焦的物体,从而提升对焦的准确度,减少对焦模糊的情况。
进一步地,本公开实施例中,对焦方法需要根据目标对焦物体确定目标对焦区域。
图3是根据一示例性实施例示出的一种确定目标对焦区域的方法流程图。如图3所示,基于目标对焦物体确定目标对焦区域,包括以下步骤。
在步骤S31中,若目标对焦物体为单一的目标对焦物体,则将单一的目标对焦物体所在的分块作为目标对焦区域。
本公开实施例中,若目标对焦物体在成像区域内为单一的,则目标对焦区域根据单一的目标对焦物体所在的分块确定。例如,用户选择的对焦模式为树木,将成像区域进行分块,并根据语义分割识别到该树木,将该树木所在的分块确定为目标对焦区域。
在步骤S32中,若目标对焦物体的数量为至少两个,则基于至少两个目标对焦物体所属分块,确定目标对焦区域。
本公开实施例中,在确定目标对焦区域时,存在目标对焦物体的数量两个或大于两个的情况,则根据两个或大于两个目标对焦物体所在的分块确定目标对焦区域。
本公开实施例中,提供的对焦方法能够优选的确定对焦区域,使对焦区域更加贴合用户的需求。
进一步地,本公开实施例中,若存在多个对焦物体,需要根据多个对焦物体所属分块确定目标对焦区域。
图4是根据一示例性实施例示出的一种确定目标对焦区域的方法流程图。如图4所示, 基于至少两个目标对焦物体所属分块,确定目标对焦区域,包括以下步骤。
在步骤S41中,若至少两个目标对焦物体所属分块为同一分块,则将分块作为目标对焦区域。
本公开实施例中,当存在两个或两个以上的目标对焦物体时,若目标对焦物体所在的分块为同一分块,则将该分块作为目标对焦区域。
例如,用户选择的对焦模式为人,对成像区域进行分块,采用语义分割识别到成像区域中包含三个人,这三个人均处在同一分块中,则将该分块确定为目标对焦区域。
在步骤S42中,若至少两个目标对焦物体所属分块为不同分块,则基于不同分块中目标对焦物体的像素数在成像区域内的像素占比,确定目标对焦区域。
本公开实施例中,若两个或两个以上目标对焦物体所在分块不同,则根据成像区域内不同分块中所含目标对焦物体的像素占比,确定目标对焦区域。
其中,像素占比,是指目标对焦物体的像素占成像区域内像素的比例。
本公开实施例中,提供的对焦方法中,能够利用成像区域中物体的像素占比提升对焦准确度,从而提升用户的拍照体验。
进一步地,本公开实施例中,若存在目标对焦区域有多个,需要根据目标对焦物体的颜色动态范围最大的确定目标对焦区域。
图5是根据一示例性实施例示出的一种确定目标对焦区域的方法流程图。如图5所示,基于不同分块中目标对焦物体的像素数在成像区域内的像素占比,确定目标对焦区域,包括以下步骤。
在步骤S51中,针对不同分块中各分块,分别确定目标对焦物体的像素数在成像区域的像素占比。
本公开实施例中,对于包含目标对焦物体的不同分块中的各分块,需要确定每个分块中目标对焦物体的像素数在成像区域的像素占比。
在步骤S52中,确定像素占比大于预设阈值的目标分块,并确定目标分块的数量。
本公开实施例中,选出像素占比大于预设阈值的目标分块以及目标分块的数量。
在步骤S53中,若目标分块数量为一个,则将目标分块作为目标对焦区域。
本公开实施例中,如果目标对焦物体的像素数在成像区域的像素占比大于预设阈值的分块数量为一个,那么将该分块作为目标对焦区域。其中,预设阈值可以为终端中预先设置的,也可以是用户根据需求自主设置的,本公开实施例对此不作具体限制
例如,若用户选择的对焦模式为树木,将成像区域进行分块,对分块后的成像区域进行语义分割,识别到成像区域内所有物体,根据用户选择的对焦模式,树木占据了三个分 块,分别为第一分块、第二分块以及第三分块,其中,第三分块中树木的像素数在成像区域的像素占比大于预设阈值,第一分块和第二分块中树木的像素数在成像区域的像素占比小于预设阈值,则选择第三分块作为目标对焦区域。
在步骤S54中,若目标分块的数量为多个,则确定多个目标分块的颜色动态变化范围,并选择颜色动态变化范围最大的分块作为目标对焦区域。
本公开实施例中,若目标分块的数量至少有两个,即,这些分块中的目标对焦物体的像素占比大于预设阈值,则将颜色动态范围变化最大的分块作为目标对焦区域,其中,颜色动态范围的取值为目标对焦物体的最大值和最小值。
例如,用户选择的对焦模式为人,将成像区域进行分块并进行语义分割,识别到成像区域内所有物体,根据用户选择的对焦模式,人占据了四个分块,分别为第一分块、第二分块、第三分块以及第四分块,其中第一分块与第二分块中人的像素占比小于第一预设阈值,第三分块与第四分块中人的像素占比大于预设阈值,第一分块和第二分块中人的像素占比小于预设阈值,则根据第三分块与第四分块中包含目标对焦物体的颜色动态范围最大的确定目标对焦区域,若第三分块中包含目标对焦物体的颜色动态范围大于第四分块中包含目标对焦物体,确定第三分块为目标对焦区域。
本公开实施例中,提供的对焦方法能够从至少两个目标对焦物体中识别到用户最想对焦的物体,以实现精准对焦,减少丢焦、对焦模糊等情况的出现。
进一步地,本公开实施例中,若存在至少两个目标对焦物体,为了使目标对焦区域尽可能包含更多的目标对焦物体,需要将目标对焦区域进行一定的扩展。
本公开一示例性实施例中,若在目标对焦区域的设定范围内存在目标对焦物体,则以目标对焦区域中心为中心点,以设定比例扩展目标对焦区域,以使扩展后的目标对焦区域包括区域外的目标对焦物体,基于扩展后的目标对焦区域进行对焦。
图6示出了本公开一示例性实施例示出的一种对焦方法的示意图。参阅图6,例如每一分块像素为50*50,对成像区域进行分块,确定第五分块为目标对焦分块,同时,第三分块以及第四分块存在目标对焦物体,则以第五分块的中心点为中心,向外扩展四分之一,对扩展后的目标对焦区域进行对焦。
图7示出了本公开一示例性实施例示出的一种对焦的流程图。参阅图7,用户开始进行拍照并选择对焦模式,对成像区域进行分块,即多窗口分块,对分块的成像区域进行语义分割,以识别成像区域内所有物体,若成像区域内目标对焦物体的数量为一个,则确定该目标对焦物体所在的分块为目标对焦区域,进行对焦;若成像区域内目标对焦物体的数量为两个或两个以上,则根据分块中目标对焦物体在成像区域内的像素占比最大的确定目 标对焦区域,进行对焦;若目标分块的数量为多个,则确定多个目标分块的颜色动态变化范围,并选择颜色动态变化范围最大的目标分块作为目标对焦区域,其中,目标分块为像素占比大于第一预设阈值的像素中,任意两分块之间的像素占比之差小于第二预设阈值的分块,进行对焦。
本公开实施例中,提供的对焦方法,能够充分利用成像区域内的目标对焦物体种类、像素占比、颜色信息,减少丢焦,尽最大可能保证对焦区域的清晰度,提升对焦准确度,从而提升用户的拍照体验。
需要说明的是,本领域内技术人员可以理解,本公开实施例上述涉及的各种实施方式/实施例中可以配合前述的实施例使用,也可以是独立使用。无论是单独使用还是配合前述的实施例一起使用,其实现原理类似。本公开实施中,部分实施例中是以一起使用的实施方式进行说明的。当然,本领域内技术人员可以理解,这样的举例说明并非对本公开实施例的限定。
基于相同的构思,本公开实施例还提供一种对焦装置。
可以理解的是,本公开实施例提供的对焦装置为了实现上述功能,其包含了执行各个功能相应的硬件结构和/或软件模块。结合本公开实施例中所公开的各示例的单元及算法步骤,本公开实施例能够以硬件或硬件和计算机软件的结合形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用和设计约束条件。本领域技术人员可以对每个特定的应用来使用不同的方法来实现所描述的功能,但是这种实现不应认为超出本公开实施例的技术方案的范围。
图8是根据一示例性实施例示出的一种对焦装置框图。参照图7,该对焦装置100包括确定单元101,识别单元102和对焦单元103。
确定单元101,用于确定成像区域.
识别单元102,用于对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体。
对焦单元103,用于基于所述目标对焦物体确定目标对焦区域,并基于所述目标对焦区域进行对焦。
在一种实施方式中,识别单元102采用如下方式对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体:获取用户选择的对焦模式,对焦模式用于表征待对焦的目标对焦物体;对成像区域进行分块,并对分块后的成像区域进行语义分割,以识别得到成像区域中包括的全部物体;在识别到的全部物体中确定与对焦模式匹配的物体,作为目标对焦物体。
在一种实施方式中,对焦单元103采用如下方式基于目标对焦物体确定目标对焦区域:若目标对焦物体为单一的目标对焦物体,则将单一的目标对焦物体所在的分块作为目标对焦区域;若目标对焦物体的数量为至少两个,则基于至少两个目标对焦物体所属分块,确定目标对焦区域。
在一种实施方式中,对焦单元103采用如下方式基于至少两个目标对焦物体所属分块,确定目标对焦区域:若至少两个目标对焦物体所属分块为同一分块,则将分块作为目标对焦区域;若至少两个目标对焦物体所属分块为不同分块,则基于不同分块中目标对焦物体的像素数在成像区域内的像素占比,确定目标对焦区域。
在一种实施方式中,对焦单元103采用如下方式基于不同分块中目标对焦物体的像素数在成像区域内的像素占比,确定目标对焦区域:
针对所述不同分块中各分块,分别确定目标对焦物体的像素数在成像区域的像素占比;确定像素占比大于预设阈值的目标分块,并确定目标分块的数量;若目标分块的数量为一个,则将目标分块作为目标对焦区域;若目标分块的数量为多个,则确定多个目标分块的颜色动态变化范围,并选择颜色动态变化范围最大的分块作为目标对焦区域。
在一种实施方式中,对焦单元103采用如下方式基于目标对焦区域进行对焦:若在目标对焦区域的设定范围内存在目标对焦物体;则以目标对焦区域中心为中心点,以设定比例扩展目标对焦区域,以使扩展后的目标对焦区域包括区域外的目标对焦物体;基于扩展后的目标对焦区域进行对焦。
关于上述实施例中的装置,其中各个模块执行操作的具体方式已经在有关该方法的实施例中进行了详细描述,此处将不做详细阐述说明。
图9是根据一示例性实施例示出的一种用于对焦的装置200的框图。例如,装置200可以是移动电话,计算机,数字广播终端,消息收发设备,游戏控制台,平板设备,医疗设备,健身设备,个人数字助理等。
参照图9,装置200可以包括以下一个或多个组件:处理组件202,存储器204,电力组件206,多媒体组件208,音频组件210,输入/输出(I/O)接口212,传感器组件214,以及通信组件216。
处理组件202通常控制装置200的整体操作,诸如与显示,电话呼叫,数据通信,相机操作和记录操作相关联的操作。处理组件202可以包括一个或多个处理器220来执行指令,以完成上述的方法的全部或部分步骤。此外,处理组件202可以包括一个或多个模块,便于处理组件202和其他组件之间的交互。例如,处理组件202可以包括多媒体模块,以方便多媒体组件208和处理组件202之间的交互。
存储器204被配置为存储各种类型的数据以支持在装置200的操作。这些数据的示例包括用于在装置200上操作的任何应用程序或方法的指令,联系人数据,电话簿数据,消息,图片,视频等。存储器204可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。
电力组件206为装置200的各种组件提供电力。电力组件206可以包括电源管理系统,一个或多个电源,及其他与为装置200生成、管理和分配电力相关联的组件。
多媒体组件208包括在所述装置200和用户之间的提供一个输出接口的屏幕。在一些实施例中,屏幕可以包括液晶显示器(LCD)和触摸面板(TP)。如果屏幕包括触摸面板,屏幕可以被实现为触摸屏,以接收来自用户的输入信号。触摸面板包括一个或多个触摸传感器以感测触摸、滑动和触摸面板上的手势。所述触摸传感器可以不仅感测触摸或滑动动作的边界,而且还检测与所述触摸或滑动操作相关的持续时间和压力。在一些实施例中,多媒体组件208包括一个前置摄像头和/或后置摄像头。当装置200处于操作模式,如拍摄模式或视频模式时,前置摄像头和/或后置摄像头可以接收外部的多媒体数据。每个前置摄像头和后置摄像头可以是一个固定的光学透镜系统或具有焦距和光学变焦能力。
音频组件210被配置为输出和/或输入音频信号。例如,音频组件210包括一个麦克风(MIC),当装置200处于操作模式,如呼叫模式、记录模式和语音识别模式时,麦克风被配置为接收外部音频信号。所接收的音频信号可以被进一步存储在存储器204或经由通信组件216发送。在一些实施例中,音频组件210还包括一个扬声器,用于输出音频信号。
I/O接口212为处理组件202和外围接口模块之间提供接口,上述外围接口模块可以是键盘,点击轮,按钮等。这些按钮可包括但不限于:主页按钮、音量按钮、启动按钮和锁定按钮。
传感器组件214包括一个或多个传感器,用于为装置200提供各个方面的状态评估。例如,传感器组件214可以检测到装置200的打开/关闭状态,组件的相对定位,例如所述组件为装置200的显示器和小键盘,传感器组件214还可以检测装置200或装置200一个组件的位置改变,用户与装置200接触的存在或不存在,装置200方位或加速/减速和装置200的温度变化。传感器组件214可以包括接近传感器,被配置用来在没有任何的物理接触时检测附近物体的存在。传感器组件214还可以包括光传感器,如CMOS或CCD图像传感器,用于在成像应用中使用。在一些实施例中,该传感器组件214还可以包括加速度传感器,陀螺仪传感器,磁传感器,压力传感器或温度传感器。
通信组件216被配置为便于装置200和其他设备之间有线或无线方式的通信。装置200可以接入基于通信标准的无线网络,如WiFi,2G或3G,或它们的组合。在一个示例性实施例中,通信组件216经由广播信道接收来自外部广播管理系统的广播信号或广播相关信息。在一个示例性实施例中,所述通信组件216还包括近场通信(NFC)模块,以促进短程通信。例如,在NFC模块可基于射频识别(RFID)技术,红外数据协会(IrDA)技术,超宽带(UWB)技术,蓝牙(BT)技术和其他技术来实现。
在示例性实施例中,装置200可以被一个或多个应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,用于执行上述方法。
在示例性实施例中,还提供了一种包括指令的非临时性计算机可读存储介质,例如包括指令的存储器204,上述指令可由装置200的处理器220执行以完成上述方法。例如,所述非临时性计算机可读存储介质可以是ROM、随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。
进一步可以理解的是,本公开中“多个”是指两个或两个以上,其它量词与之类似。“和/或”,描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。字符“/”一般表示前后关联对象是一种“或”的关系。单数形式的“一种”、“所述”和“该”也旨在包括多数形式,除非上下文清楚地表示其他含义。
进一步可以理解的是,术语“第一”、“第二”等用于描述各种信息,但这些信息不应限于这些术语。这些术语仅用来将同一类型的信息彼此区分开,并不表示特定的顺序或者重要程度。实际上,“第一”、“第二”等表述完全可以互换使用。例如,在不脱离本公开范围的情况下,第一信息也可以被称为第二信息,类似地,第二信息也可以被称为第一信息。
进一步可以理解的是,本公开实施例中尽管在附图中以特定的顺序描述操作,但是不应将其理解为要求按照所示的特定顺序或是串行顺序来执行这些操作,或是要求执行全部所示的操作以得到期望的结果。在特定环境中,多任务和并行处理可能是有利的。
本领域技术人员在考虑说明书及实践这里公开的发明后,将容易想到本公开的其它实施方案。本申请旨在涵盖本公开的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本公开的一般性原理并包括本公开未公开的本技术领域中的公知常识或惯用技术手段。
应当理解的是,本公开并不局限于上面已经描述并在附图中示出的精确结构,并且可以在不脱离其范围进行各种修改和改变。本公开的范围仅由所附的权利范围来限制。

Claims (14)

  1. 一种对焦方法,其特征在于,包括:
    确定成像区域;
    对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体;
    基于所述目标对焦物体确定目标对焦区域,并基于所述目标对焦区域进行对焦。
  2. 根据权利要求1所述的方法,其特征在于,所述对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体,包括:
    获取用户选择的对焦模式,所述对焦模式用于表征待对焦的目标对焦物体;
    对所述成像区域进行分块,并对分块后的成像区域进行语义分割,以识别得到所述成像区域中包括的全部物体;
    在识别到的所述全部物体中确定与所述对焦模式匹配的物体,作为目标对焦物体。
  3. 根据权利要求1或2所述的方法,其特征在于,所述基于所述目标对焦物体确定目标对焦区域,包括:
    若所述目标对焦物体为单一的目标对焦物体,则将所述单一的目标对焦物体所在的分块作为目标对焦区域;
    若所述目标对焦物体的数量为至少两个,则基于所述至少两个目标对焦物体所属分块,确定目标对焦区域。
  4. 根据权利要求3所述的方法,其特征在于,所述基于所述至少两个目标对焦物体所属分块,确定目标对焦区域,包括:
    若所述至少两个目标对焦物体所属分块为同一分块,则将所述分块作为目标对焦区域;
    若所述至少两个目标对焦物体所属分块为不同分块,则基于所述不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域。
  5. 根据权利要求4所述的方法,其特征在于,所述基于所述不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域,包括:
    针对所述不同分块中各分块,分别确定目标对焦物体的像素数在所述成像区域的像素占比;
    确定像素占比大于预设阈值的目标分块,并确定所述目标分块的数量;
    若所述目标分块的数量为一个,则将所述目标分块作为目标对焦区域;
    若所述目标分块的数量为多个,则确定多个目标分块的颜色动态变化范围,并选择颜色动态变化范围最大的分块作为目标对焦区域。
  6. 根据权利要求1至5中任意一项所述的方法,其特征在于,所述基于所述目标对焦区域进行对焦,包括:
    若在所述目标对焦区域的设定范围内存在目标对焦物体;
    则以所述目标对焦区域中心为中心点,以设定比例扩展所述目标对焦区域,以使扩展后的目标对焦区域包括区域外的目标对焦物体;
    基于扩展后的目标对焦区域进行对焦。
  7. 一种对焦装置,其特征在于,包括:
    确定单元,用于确定成像区域;
    识别单元,用于对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体;
    对焦单元,用于基于所述目标对焦物体确定目标对焦区域,并基于所述目标对焦区域进行对焦。
  8. 根据权利要求7所述的装置,其特征在于,所述识别单元采用如下方式对所述成像区域进行物体识别,并在识别到的物体中确定目标对焦物体:
    获取用户选择的对焦模式,所述对焦模式用于表征待对焦的目标对焦物体;
    对所述成像区域进行分块,并对分块后的成像区域进行语义分割,以识别得到所述成像区域中包括的全部物体;
    在识别到的所述全部物体中确定与所述对焦模式匹配的物体,作为目标对焦物体。
  9. 根据权利要求7或8所述的装置,其特征在于,所述对焦单元采用如下方式基于所述目标对焦物体确定目标对焦区域:
    若所述目标对焦物体为单一的目标对焦物体,则将所述单一的目标对焦物体所在的分块作为目标对焦区域;
    若所述目标对焦物体的数量为至少两个,则基于所述至少两个目标对焦物体所属分块,确定目标对焦区域。
  10. 根据权利要求9所述的装置,其特征在于,所述对焦单元采用如下方式基于所述至少两个目标对焦物体所属分块,确定目标对焦区域:
    若所述至少两个目标对焦物体所属分块为同一分块,则将所述分块作为目标对焦区域;
    若所述至少两个目标对焦物体所属分块为不同分块,则基于所述不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域。
  11. 根据权利要求10所述的装置,其特征在于,所述对焦单元采用如下方式基于所述 不同分块中目标对焦物体的像素数在所述成像区域内的像素占比,确定目标对焦区域:
    针对所述不同分块中各分块,分别确定目标对焦物体的像素数在所述成像区域的像素占比;
    确定像素占比大于预设阈值的目标分块,并确定所述目标分块的数量;
    若所述目标分块的数量为一个,则将所述目标分块作为目标对焦区域;
    若所述目标分块的数量为多个,则确定多个所述分块的颜色动态变化范围,并选择颜色动态变化范围最大的分块作为目标对焦区域。
  12. 根据权利要求7至11中任意一项所述的装置,其特征在于,所述对焦单元采用如下方式基于所述目标对焦区域进行对焦:
    若在所述目标对焦区域的设定范围内存在目标对焦物体;
    则以所述目标对焦区域中心为中心点,以设定比例扩展所述目标对焦区域,以使扩展后的目标对焦区域包括区域外的目标对焦物体;
    基于扩展后的目标对焦区域进行对焦。
  13. 一种对焦装置,其特征在于,包括:
    处理器;
    用于存储处理器可执行指令的存储器;
    其中,所述处理器被配置为:执行权利要求1至6中任意一项所述的对焦方法。
  14. 一种非临时性计算机可读存储介质,当所述存储介质中的指令由移动终端的处理器执行时,使得移动终端能够执行权利要求1至6中任意一项所述的对焦方法。
PCT/CN2022/096934 2022-06-02 2022-06-02 一种对焦方法、装置及存储介质 WO2023231009A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2022/096934 WO2023231009A1 (zh) 2022-06-02 2022-06-02 一种对焦方法、装置及存储介质
CN202280004614.XA CN117652152A (zh) 2022-06-02 2022-06-02 一种对焦方法、装置及存储介质

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/096934 WO2023231009A1 (zh) 2022-06-02 2022-06-02 一种对焦方法、装置及存储介质

Publications (1)

Publication Number Publication Date
WO2023231009A1 true WO2023231009A1 (zh) 2023-12-07

Family

ID=89026766

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/096934 WO2023231009A1 (zh) 2022-06-02 2022-06-02 一种对焦方法、装置及存储介质

Country Status (2)

Country Link
CN (1) CN117652152A (zh)
WO (1) WO2023231009A1 (zh)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070104472A1 (en) * 2005-11-08 2007-05-10 Shuxue Quan Skin color prioritized automatic focus control via sensor-dependent skin color detection
CN104133339A (zh) * 2013-05-02 2014-11-05 聚晶半导体股份有限公司 自动对焦方法及自动对焦装置
US20140327743A1 (en) * 2013-05-02 2014-11-06 Altek Semiconductor Corp. Auto focus method and auto focus apparatus
CN107295252A (zh) * 2017-06-16 2017-10-24 广东欧珀移动通信有限公司 对焦区域显示方法、装置及终端设备
CN110572573A (zh) * 2019-09-17 2019-12-13 Oppo广东移动通信有限公司 对焦方法和装置、电子设备、计算机可读存储介质
CN112261292A (zh) * 2020-10-20 2021-01-22 Oppo广东移动通信有限公司 图像获取方法、终端、芯片及存储介质
CN113302907A (zh) * 2020-08-24 2021-08-24 深圳市大疆创新科技有限公司 拍摄方法、装置、设备及计算机可读存储介质

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070104472A1 (en) * 2005-11-08 2007-05-10 Shuxue Quan Skin color prioritized automatic focus control via sensor-dependent skin color detection
CN104133339A (zh) * 2013-05-02 2014-11-05 聚晶半导体股份有限公司 自动对焦方法及自动对焦装置
US20140327743A1 (en) * 2013-05-02 2014-11-06 Altek Semiconductor Corp. Auto focus method and auto focus apparatus
CN107295252A (zh) * 2017-06-16 2017-10-24 广东欧珀移动通信有限公司 对焦区域显示方法、装置及终端设备
CN110572573A (zh) * 2019-09-17 2019-12-13 Oppo广东移动通信有限公司 对焦方法和装置、电子设备、计算机可读存储介质
CN113302907A (zh) * 2020-08-24 2021-08-24 深圳市大疆创新科技有限公司 拍摄方法、装置、设备及计算机可读存储介质
CN112261292A (zh) * 2020-10-20 2021-01-22 Oppo广东移动通信有限公司 图像获取方法、终端、芯片及存储介质

Also Published As

Publication number Publication date
CN117652152A (zh) 2024-03-05

Similar Documents

Publication Publication Date Title
US10565763B2 (en) Method and camera device for processing image
CN108419016B (zh) 拍摄方法、装置及终端
CN109345485B (zh) 一种图像增强方法、装置、电子设备及存储介质
US9674395B2 (en) Methods and apparatuses for generating photograph
JP6335289B2 (ja) 画像フィルタを生成する方法及び装置
WO2016011747A1 (zh) 肤色调整方法和装置
WO2021051949A1 (zh) 一种图像处理方法及装置、电子设备和存储介质
CN108154465B (zh) 图像处理方法及装置
WO2017031901A1 (zh) 人脸识别方法、装置及终端
WO2020042826A1 (zh) 视频流降噪方法和装置、电子设备及存储介质
WO2020114087A1 (zh) 图像转换方法、装置、电子设备及存储介质
CN105631803B (zh) 滤镜处理的方法和装置
CN108154466B (zh) 图像处理方法及装置
WO2022077970A1 (zh) 特效添加方法及装置
CN110580688B (zh) 一种图像处理方法、装置、电子设备及存储介质
US11310443B2 (en) Video processing method, apparatus and storage medium
US11222235B2 (en) Method and apparatus for training image processing model, and storage medium
WO2020233201A1 (zh) 图标位置确定方法和装置
EP3211879A1 (en) Method and device for automatically capturing photograph, electronic device
CN107730443B (zh) 图像处理方法、装置及用户设备
CN111461950B (zh) 图像处理方法及装置
CN110913120B (zh) 图像拍摄方法及装置、电子设备、存储介质
CN106469446B (zh) 深度图像的分割方法和分割装置
WO2023231009A1 (zh) 一种对焦方法、装置及存储介质
CN114666490B (zh) 对焦方法、装置、电子设备和存储介质

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 202280004614.X

Country of ref document: CN