WO2021179830A1 - Procédé et appareil de guidage de composition d'image, et dispositif électronique - Google Patents

Procédé et appareil de guidage de composition d'image, et dispositif électronique Download PDF

Info

Publication number
WO2021179830A1
WO2021179830A1 PCT/CN2021/074203 CN2021074203W WO2021179830A1 WO 2021179830 A1 WO2021179830 A1 WO 2021179830A1 CN 2021074203 W CN2021074203 W CN 2021074203W WO 2021179830 A1 WO2021179830 A1 WO 2021179830A1
Authority
WO
WIPO (PCT)
Prior art keywords
type
composition
point
display area
electronic device
Prior art date
Application number
PCT/CN2021/074203
Other languages
English (en)
Chinese (zh)
Inventor
薛致远
李亚乾
蒋燚
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Publication of WO2021179830A1 publication Critical patent/WO2021179830A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/695Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects

Definitions

  • This application relates to the field of electronic technology, in particular to a composition guidance method, device and electronic equipment.
  • the embodiments of the present application provide a composition guidance method, device, and electronic equipment, which can guide a user to rationally compose a picture when shooting.
  • an embodiment of the present application provides a composition guidance method, which is applied to an electronic device, wherein the composition guidance method includes:
  • the scene to be photographed contains at least two types of objects of different categories, determining the first type of object and the second type of object in the scene to be photographed;
  • an embodiment of the present application provides a composition guidance device, which is applied to an electronic device, wherein the composition guidance device includes:
  • the object recognition module is used to perform object recognition on the scene to be photographed
  • the first determining module is configured to determine the first type of object and the second type of object in the to-be-shot scene when it is recognized that the to-be-shot scene contains at least two types of objects of different types;
  • a second determining module configured to determine a common composition positioning point for the first type of object and the second type of object according to the state information of the first type of object and the second type of object;
  • the composition guidance module is configured to output prompt information for adjusting the shooting posture of the electronic device when the position of the composition positioning point does not match the preset composition point.
  • an embodiment of the present application provides an electronic device, including a processor and a memory, and a computer program is stored in the memory, wherein the processor uses the computer program stored in the memory to To execute:
  • the scene to be photographed contains at least two types of objects of different categories, determining the first type of object and the second type of object in the scene to be photographed;
  • FIG. 1 is a schematic diagram of the first flow of a composition guidance method provided by an embodiment of this application.
  • FIG. 2 is a schematic diagram of a second flow of a composition guidance method provided by an embodiment of this application.
  • FIG. 3 is a schematic diagram of a preview image provided by an embodiment of the application.
  • FIG. 4 is a schematic diagram of a shooting preview screen provided by an embodiment of the application.
  • FIG. 5 is a schematic diagram of the first structure of a composition guidance device provided by an embodiment of the application.
  • FIG. 6 is a schematic diagram of a second structure of a composition guidance device provided by an embodiment of the application.
  • FIG. 7 is a schematic diagram of a third structure of a composition guidance device provided by an embodiment of the application.
  • FIG. 8 is a schematic diagram of the first structure of an electronic device provided by an embodiment of this application.
  • FIG. 9 is a schematic diagram of a second structure of an electronic device provided by an embodiment of the application.
  • the embodiment of the present application first provides a composition guidance method, which is applied to an electronic device.
  • the execution subject of the composition guidance method may be the composition guidance device provided in the embodiment of the present application, or an electronic device integrated with the composition guidance device.
  • the composition guidance device may be implemented in hardware or software.
  • the electronic device may be a smart phone, Tablet computers, handheld computers, notebook computers, or desktop computers are equipped with processors and have processing capabilities.
  • composition guidance method including:
  • the first type of object and the second type of object in the scene to be photographed are determined;
  • the performing object recognition on the scene to be photographed includes:
  • the object in the preview image is identified according to the object category to which each pixel in the preview image belongs.
  • the method before determining a common composition anchor point for the first type of object and the second type of object according to the state information of the first type of object and the second type of object, the method further includes:
  • a common composition positioning point is determined for the first-type object and the second-type object according to the state information of the first-type object and the second-type object.
  • the determining a common composition anchor point for the first-type object and the second-type object according to the state information of the first-type object and the second-type object includes:
  • the composition positioning point is determined according to the first center point, the second center point, and the area ratio, and the ratio of the distance from the composition positioning point to the first center point to the second center point is equal to The area ratio.
  • the determining the first center point of the first display area and determining the second center point of the second display area includes:
  • the second coordinate value average value of all pixels in the second display area is acquired, and the pixel point corresponding to the second coordinate value average value is determined as the second center point.
  • the obtaining the area ratio of the first display area to the second display area includes:
  • the ratio of the number of pixels is used as the area ratio.
  • the output of the prompt information for adjusting the shooting posture of the electronic device includes:
  • the movement direction of the composition positioning point to the preset composition point is displayed on the preview interface, and prompt information for adjusting the shooting posture of the electronic device in the movement direction is output.
  • the method further includes:
  • the category of each object and the probability of belonging to the category are respectively marked on the shooting preview interface.
  • the output of the prompt information for adjusting the shooting posture of the electronic device includes:
  • the voice prompt adjusts the shooting posture of the electronic device toward the relative position.
  • the method further includes:
  • the scene to be photographed is photographed to obtain an imaging image of the scene to be photographed.
  • FIG. 1 is a schematic flowchart of a composition guidance method provided by an embodiment of the application.
  • the composition guidance method is applied to the electronic device provided in the embodiment of the present application, and the composition guidance method provided in the embodiment of the present application may include the following steps:
  • the electronic device may include one or more cameras, through which the scene to be shot is photographed or previewed.
  • the electronic device may also include a viewfinder frame, and a picture of the scene to be photographed or previewed by the camera may be displayed in the viewfinder frame.
  • the image captured or previewed by the camera can also be displayed on the display screen.
  • the display screen of the electronic device serves as a viewfinder frame.
  • the electronic device when it performs object recognition on the scene to be shot, it can obtain a preview image of the scene to be shot, and recognize the object in the scene to be shot by recognizing the preview image.
  • object recognition can be performed on the scene to be shot by means of semantic segmentation.
  • semantic segmentation For example, before semantic segmentation is performed on the preview image, the semantic segmentation model is pre-trained on the semantic segmentation data set, and when the preview image is semantically segmented, the preview image is input into the pre-trained semantic segmentation model for semantic segmentation.
  • the semantic segmentation model can adopt a Bilateral Segmentation Network (BiSeNet) algorithm.
  • BiSeNet Bilateral Segmentation Network
  • the semantic segmentation model is trained first.
  • the BiSeNet algorithm can give a classification to each pixel of the input image, thereby accurately identifying the pixel range of each object in the image.
  • the image is input to the BiSeNet algorithm, and the BiSeNet algorithm sends the image to two paths for calculation at the same time.
  • One of the paths is a spatial path with a small step size, which is used to retain spatial information to generate high-resolution feature maps; the other path is a semantic path with fast down-sampling, which can maximize the receptive field and generate coarse-grained segmentation picture.
  • the spatial path preserves the high resolution of the input image, while the semantic path guarantees the speed of semantic segmentation.
  • the electronic device detects the infrared radiation energy of each object in the scene to be photographed, and converts the infrared radiation energy of each object in the scene to be photographed into an infrared thermal imaging image of the scene to be photographed by means of photoelectric conversion and signal processing.
  • the surface temperature of different objects is different, the infrared heat radiation conditions are also different, and the imaging conditions on the infrared thermal imaging map are also different. Therefore, different colors can be used to reflect the different temperature distributions on the surface of each object on the infrared thermal imaging map.
  • it can be represented by a grayscale image.
  • the objects in the scene to be photographed can be identified.
  • infrared thermal imaging technology the thermal radiation emitted by the object is converted into a visible light image, even in the dark night, the object in the scene to be photographed can be accurately identified without being affected by the brightness of the external environment.
  • the picture information of the scene to be shot can be displayed on the shooting preview interface, and the user observes the scene to be shot in real time through the shooting preview interface, so as to adjust the shooting posture of the camera to obtain a better shooting effect.
  • the screen information can be directly identified in the background.
  • the picture information of the scene to be photographed includes but is not limited to the preview image and the infrared thermal imaging map in the above examples.
  • each object may be marked on the photographing preview interface. For example, when two people, a dog, and a tree are identified from the scene to be shot, the two people are individually selected in the shooting preview interface and marked as "person”; and a dog is individually selected and marked “Dog”; select a tree by box and mark it as "tree”.
  • the probability that the object belongs to the category can also be labeled. For example, if there is an object recognized as a human being in two persons, the probability that the object belongs to "person” is 95% according to the algorithm.
  • the probability that the object belongs to a "person” is only 50%.
  • the probability is also marked "50%”. In this way, visualizing the results of object recognition can bring users a more intelligent experience.
  • the scene to be photographed includes at least two types of objects of different types, determine the first type of object and the second type of object in the scene to be photographed.
  • the category to which each object belongs is recognized.
  • the categories can include character categories and object categories, and the object categories can be divided into different object categories, for example, plant categories, building categories, and so on. It is understandable that when performing object recognition on the scene to be photographed, the object may not be recognized, or one, two or more types of objects may be recognized.
  • the two types of objects are selected as the first type of object and the second type of object in the scene to be photographed, and the electronic device combines the first type of object and the second type of object.
  • the second type of object is used as the main composition object to guide the composition.
  • the first type of object and the second type of object may be a person object and an object object, respectively, or they may be two different types of object objects.
  • the first type of object and the second type of object here are only used to distinguish objects of different types, and not used to describe a specific sequence. For example, a person object may be determined as a first type object, and a certain object object may be determined as a second type object, or a certain object object may be determined as a first type object, and a person object may be determined as a second type object.
  • a preview image or infrared thermal imaging image of the scene to be shot can be acquired, and the preview image or infrared thermal imaging image can be displayed on the shooting preview interface.
  • the user can download The first type of object and the second type of object are selected from each object. For example, two touch operations of the user on the shooting preview interface are received, and the objects corresponding to the two touch operations are determined as the first type of object and the second type of object.
  • the electronic device can also automatically select the first type of object and the second type of object based on the size and location of various types of objects.
  • the distance between various objects in the scene to be photographed and the electronic device is obtained (when multiple objects of the same type are included, the closest distance between the multiple objects and the electronic device is determined as the distance between the object and the electronic device. Distance), the two types of objects closest to the electronic device in the scene to be photographed are regarded as the first type of object and the second type of object.
  • a person object is determined as the first type of object, and the number of pixels corresponding to different types of object objects is acquired according to the preview image, and the object with the largest number of pixels is determined as the second type of object.
  • the object with the largest number of corresponding pixels is determined as the first type of object, and the object with the largest number of corresponding pixels other than the first type of object is determined as the second type of object, and so on.
  • the human object is determined as the first type of object, and the surface temperatures corresponding to different types of objects are obtained according to the infrared thermal imaging images, and the object with the highest surface temperature is determined as the second type of object.
  • the object with the highest corresponding surface temperature is determined as the first type of object, and the object with the highest corresponding surface temperature other than the first type of object is determined as the second type of object, and so on.
  • the status information may be the size of the first type of object and the second type of object, the distance from the electronic device, or the relative position of the first type of object and the second type of object, and so on. If an image such as a preview image or an infrared thermal imaging map of the scene to be photographed is acquired through the camera, the status information may also be information such as the position and area of the first type of object and the second type of object on the preview image or the infrared thermal imaging map.
  • a common composition anchor point is determined for the first type of object and the second type of object to represent the first type of object and the second type of object.
  • the position of the class object as a whole. After that, you can compare the position of the composition reference point with the preset composition point to determine whether it is necessary to guide the user in the composition when shooting.
  • the state information of the first-type object and the second-type object may be position information of the first-type object and the second-type object in the preview image.
  • Determine the first display area corresponding to the first type of object and the second display area corresponding to the second type of object in the preview image obtain the average value of the first coordinate values of all pixels in the first display area, and calculate the average value of the first coordinate value
  • the corresponding pixel point is determined as the first center point;
  • the second coordinate value average value of all pixels in the second display area is obtained, the pixel point corresponding to the second coordinate value average value is determined as the second center point, and the first center point and
  • the midpoint of the line connecting the second center point is determined as the positioning point of the composition of the first type of object and the second type of object.
  • the preview image use a rectangular frame to select the objects of the first category from top to bottom, and from left to right, and determine the geometric center point of the rectangle as the first center point, and determine it in the same way.
  • the second center point of the second type of object determines the midpoint of the line connecting the first center point and the second center point as the positioning point of the composition of the first type of object and the second type of object.
  • the preset composition point is a preset position reference point
  • the preset composition point may be the center point of the scene to be shot, that is, the geometric center point of the shooting preview interface, that is, the geometric center of the preview image or the infrared thermal image.
  • the point can also be a preset composition point in other positions, as long as the composition positioning point matches the position of the preset composition point, and the captured image has a beautiful composition.
  • the position matching can be that the two positions are the same, or the distance between the two is less than the preset threshold, or the relative position of the two meets certain conditions, for example, the composition positioning point is located at the preset distance above the preset composition point. Matches for the location. It should be noted that regardless of whether the position of the composition positioning point matches the preset composition point, the composition positioning point and the preset composition point can be displayed on the shooting preview interface of the electronic device for the user to see, of course, it may not be displayed in the On the shooting preview interface.
  • the composition positioning point is determined based on the status information of the first type of object and the second type of object in the image (such as preview image, infrared thermal imaging, etc.) collected by the camera, you can first Mark the composition anchor point on the image, and then mark the composition anchor point to the shooting preview interface according to the corresponding relationship between the image and the shooting preview interface, so that the user can see the entirety of the first type of object and the second type of object on the shooting preview interface Position the composition anchor point.
  • the electronic device compares the position of the composition anchor point and the preset composition point on the shooting preview interface, and when the position of the composition anchor point does not match the preset composition point, it outputs prompt information for adjusting the shooting posture of the electronic device to guide the user in the composition. That is, when the position of the composition positioning point does not match the position of the preset composition point, the step of outputting prompt information for adjusting the shooting posture of the electronic device may include:
  • the moving direction of the composition anchor point to the preset composition point is displayed on the preview interface, and the prompt information for adjusting the shooting posture of the electronic device in the moving direction is output.
  • the electronic device may prompt the user on the shooting preview interface how to adjust the shooting posture of the electronic device, so that the composition positioning point matches the position of the preset composition point. Displaying the moving direction of the composition positioning point to the preset composition point on the shooting preview interface enables the user to intuitively see how to adjust the shooting posture of the electronic device on the shooting preview interface. For example, the electronic device is moved in the moving direction, or the shooting angle of the electronic device is adjusted in the moving direction.
  • the prompt information for adjusting the shooting posture of the electronic device can also be voice prompt information.
  • the electronic device can also Instruct the user to compose the picture.
  • the electronic device can determine the relative position of the preset composition point relative to the composition positioning point in the background, and voice prompt the user to adjust the shooting posture of the electronic device toward the relative position.
  • the positions of the first type of object and the second type of object on the shooting preview interface will also change, correspondingly, the common one determined by the state information of the first type of object and the second type of object
  • the composition positioning point will also change in real time, and the electronic device can re-determine whether the composition positioning point matches the preset composition point according to the composition positioning point that changes in real time to decide whether to guide the user to compose the picture.
  • the electronic device can prompt the user to adjust the shooting direction to the right, determine the new composition anchor point in real time, and determine whether the new composition anchor point matches the preset composition point, if so , The shooting prompt information is output to remind the user that it is ready to shoot. If not, the relative position of the preset composition point relative to the new composition anchor point is re-judged at this time. For example, the user accidentally over-adjusts the preset composition point The left position of the positioning point of the new composition, the electronic device can prompt the user to adjust the shooting direction to the left.
  • the shooting posture adjusted by the electronic device can be judged in real time Whether the positioning point of the composition matches the position of the preset composition point, if it matches, the composition guidance will be stopped and the user will be prompted to shoot, and if it does not match, the user will continue to prompt the user to adjust the shooting posture.
  • the composition positioning point matches the position of the preset composition point, shooting the scene to be shot to obtain an imaging image of the scene to be shot.
  • the electronic device can prompt the user to shoot, and the user can choose to shoot according to the currently recommended composition to obtain an imaging image of the scene to be shot , You can also stick to your own composition and continue to adjust to your favorite composition before shooting.
  • composition guidance method provided by the embodiments of the present application can guide the user to rationally compose the picture when shooting, but does not limit the user's composition.
  • the user can still choose the composition method he likes, which provides guidance for the user while retaining the flexibility of the composition.
  • FIG. 2 is a schematic diagram of the second flow of the composition guidance method provided by an embodiment of the application.
  • the text display method is applied to the electronic device provided in the embodiment of the present application, and the composition guidance method may include the following steps:
  • the electronic device may include one or more cameras, through which the camera previews the scene to be shot, and obtains a preview image of the scene to be shot.
  • the electronic device realizes the preview function, the picture previewed by the camera can be displayed on the shooting preview interface.
  • the semantic segmentation model is pre-trained on the semantic segmentation data set, and when the preview image is semantically segmented, the preview image is input into the pre-trained semantic segmentation model for semantic segmentation.
  • the semantic segmentation model can use the BiSeNet algorithm.
  • the preview image is input into the semantic segmentation model for semantic segmentation, and the object category to which each pixel in the preview image belongs is obtained.
  • the object categories can include person categories and object categories, and the object categories can be divided into different object categories, for example, plant categories, building categories, and so on.
  • the pixels in the preview image that belong to the same object category constitute one type of object, and there are different display areas corresponding to different types of objects.
  • the scene to be photographed contains at least two types of objects of different types, determine the first type of object and the second type of object in the scene to be photographed.
  • the object when performing semantic segmentation on the preview image, the object may not be recognized, or one, two or more types of objects may be recognized.
  • the two types of objects are selected as the first type of object and the second type of object in the scene to be photographed, and the electronic device combines the first type of object and the second type of object.
  • the second type of object is used as the main composition object to guide the composition.
  • the first type of object and the second type of object may be a person object and an object object, respectively, or they may be two different types of object objects.
  • the first type of object and the second type of object here are only used to distinguish objects of different types, and not used to describe a specific sequence. For example, a person object may be determined as a first type object, and a certain object object may be determined as a second type object, or a certain object object may be determined as a first type object, and a person object may be determined as a second type object.
  • a preview image or infrared thermal imaging image of the scene to be shot can be acquired, and the preview image or infrared thermal imaging image can be displayed on the shooting preview interface.
  • the user can download The first type of object and the second type of object are selected from each object. For example, two touch operations of the user on the shooting preview interface are received, and the objects corresponding to the two touch operations are determined as the first type of object and the second type of object.
  • the electronic device can also automatically select the first type of object and the second type of object based on the size and location of various types of objects.
  • the distance between various objects in the scene to be photographed and the electronic device is obtained (when multiple objects of the same type are included, the closest distance between the multiple objects and the electronic device is determined as the distance between the object and the electronic device. Distance), the two types of objects closest to the electronic device in the scene to be photographed are regarded as the first type of object and the second type of object.
  • a person object is determined as the first type of object, and the number of pixels corresponding to different types of object objects is acquired according to the preview image, and the object with the largest number of pixels is determined as the second type of object.
  • the object with the largest number of corresponding pixels is determined as the first type of object, and the object with the largest number of corresponding pixels other than the first type of object is determined as the second type of object, and so on.
  • the human object is determined as the first type of object, and the surface temperatures corresponding to different types of objects are obtained according to the infrared thermal imaging images, and the object with the highest surface temperature is determined as the second type of object.
  • the object with the highest corresponding surface temperature is determined as the first type of object, and the object with the highest corresponding surface temperature other than the first type of object is determined as the second type of object, and so on.
  • the object category corresponding to each pixel in the preview image is identified through semantic segmentation, and the first type of object and the second type of object are determined.
  • the area occupied by the pixels corresponding to the objects of the first type in the preview image is determined as the first display area, and the area occupied by the pixels corresponding to the objects of the second type in the preview image is determined as the second display area.
  • step 206 Determine whether the proportion of the first area of the first display area in the preview image is greater than the first preset proportion. If yes, go to step 207, if not, go to step 209.
  • step 207 Determine whether the second area ratio of the second display area in the preview image is greater than the second preset ratio. If yes, go to step 208, if not, go to step 209.
  • the first display area and the second display area obtained by semantic segmentation it is determined whether the objects of the first type and the object of the second type in the scene to be photographed meet the preset conditions, if they do, go to step 208, if not, If yes, go to step 209.
  • the proportion of the first area of the first display area in the preview image is greater than the first preset proportion
  • the proportion of the second area of the second display area in the preview image is greater than the second preset proportion
  • the first preset ratio and the second preset ratio may be the same or different. For example, if there are human subjects in the shooting scene, it does not matter if the display area of the human subjects occupy a slightly lower area in the preview image. For a large building object, the proportion of the architectural object in the preview image is It should be higher to determine it as the focus of shooting.
  • the first type of image is a human object and the second type of object is a building object
  • the first type of object in the preview image accounts for more than 10% of the first area and the second type of object is the second in the preview image If the area accounts for more than 30%, it is determined that the objects of the first type and the objects of the second type in the scene to be photographed meet the preset conditions, and step 208 is entered.
  • the size of the first preset ratio and the second preset ratio can be set according to the actual volume of the object category and the user's photographing habits.
  • the average value of the first coordinate values of all pixels in the first display area is obtained, and the pixel point corresponding to the average value of the first coordinate value is determined as the first center point; and the first center point of all pixels in the second display area is obtained.
  • the average value of the two coordinate values, and the pixel point corresponding to the average value of the second coordinate value is determined as the second center point.
  • the proportion of the first area of the first display area in the preview image is not greater than the first preset proportion, and/or the proportion of the second area of the second display area in the preview image is not greater than the second preset proportion, it is indicated
  • the current scene to be photographed does not meet the preset conditions, and the object in the scene to be photographed may be too far away or too small, resulting in no suitable composition object. At this time, no composition guidance is given to the user.
  • the area ratio of the first display area to the second display area is calculated.
  • both the first display area and the second display area are pixel ranges composed of a plurality of pixels. Therefore, in order to calculate the area ratio, it is not necessary to know the specific area values of the first display area and the second display area, and only need to obtain the ratio of the number of pixels contained in the first display area and the second display area. That is, the step of obtaining the area ratio between the first display area and the second display area may include:
  • the ratio of the number of pixels is taken as the area ratio.
  • FIG. 3 is a schematic diagram of a preview image provided by an embodiment of the application.
  • the first type of object is a person object
  • the second type of object is a building object.
  • the first type of object in the preset image has a larger proportion of the first area S1 than the first preset ratio
  • the second type of object is in the preset image.
  • the proportion of the second area S 2 in the image is greater than the second preset proportion.
  • the object and the second type of object determine a common composition anchor point.
  • the coordinates of the composition anchor point in the preview image are Pr(X, Y), where X and Y represent the composition anchor point respectively.
  • the abscissa X and the ordinate Y of the composition anchor point can be determined by two coordinate determination formulas:
  • the distance from the composition positioning point to the first center point P1(x 1 , y 1) is:
  • the ratio of the distance between the positioning point of the composition and the first center point P1 (x 1 , y 1 ) and the second center point P2 (x 2 , y 2) is:
  • the ratio of the distance from the composition positioning point to the first center point P1 (x 1 , y 1 ) and the second center point P2 (x 2 , y 2 ) is equal to the area ratio S 1 /S 2 .
  • the ratio of the distance between the positioning point of the composition and the first center point P1 (x 1 , y 1 ) and the second center point P2 (x 2 , y 2) is:
  • the ratio of the distance from the composition positioning point to the first center point P1 (x 1 , y 1 ) and the second center point P2 (x 2 , y 2 ) is equal to the area ratio S 1 /S 2 .
  • the determined distance from the composition location point to the center point of the building is twice the distance from the center point of the person.
  • the ratio of the distance between the composition positioning point and the first center point P1 (x 1 , y 1 ) and the second center point P2 (x 2 , y 2 ) is equal to the area ratio S 1 / S 2 , it can be deduced that the position of the composition positioning point obtained according to the above coordinate determination formula satisfies: the composition positioning point to the first center point P1 (x 1 , y 1 ) and the second center point P2 (x 2 , y 2 ) The distance ratio is equal to the area ratio S 1 /S 2 .
  • step 212 Determine whether the position of the composition positioning point matches the preset composition point. When the composition positioning point does not match the position of the preset composition point, go to step 213. When the composition positioning point matches the position of the preset composition point, Go to step 215.
  • composition positioning point and the preset composition point are displayed on the shooting preview interface of the electronic device.
  • the position matching can be that the two positions are the same, or the distance between the two is less than the preset threshold, or the relative position of the two meets certain conditions, for example, the composition positioning point is located at the preset distance above the preset composition point. Matches for the location.
  • FIG. 4 is a schematic diagram of the shooting preview screen provided by this application.
  • the preset composition point can be the geometric center point on the preview image. Take the position matching as the same position as an example.
  • the preset composition point and the preset composition point are correspondingly displayed on the shooting preview interface, and the movement direction of the composition anchor point to the preset composition point is marked on the shooting preview interface to remind the user. How to adjust the shooting posture of the electronic device so that the positioning point of the composition matches the position of the preset composition point.
  • Displaying the moving direction of the composition positioning point to the preset composition point on the shooting preview interface enables the user to intuitively see how to adjust the shooting posture of the electronic device on the shooting preview interface. For example, the electronic device is moved in the moving direction, or the shooting angle of the electronic device is adjusted in the moving direction.
  • the preset composition point is at the upper right of the composition positioning point, and the electronic device can prompt the user to adjust the shooting direction to the upper right, determine the new composition positioning point in real time, and determine whether the new composition positioning point is consistent with the preset positioning point.
  • the electronic device may prompt the user to adjust the shooting direction to the lower left. Please continue to refer to Figure 4(b).
  • the electronic device can prompt the user to shoot, and the user can choose to follow the currently recommended composition Shoot, get the imaging image of the scene to be shot, you can also stick to your own composition and continue to adjust to your favorite composition before shooting.
  • composition guidance method provided by the embodiments of the present application can guide the user to rationally compose the picture when shooting, but does not limit the user's composition.
  • the user can still choose the composition method he likes, which provides guidance for the user while retaining the flexibility of the composition.
  • the present application is not limited by the order of execution of the various steps described, and certain steps may also be performed in other order or at the same time if there is no conflict.
  • the composition guidance method provided by the embodiments of the present application first performs object recognition on the scene to be photographed; when it is recognized that the scene to be photographed contains at least two types of objects in different categories, the first type of object in the scene to be photographed is determined And the second type of object; then according to the state information of the first type of object and the second type of object to determine the common composition positioning point for the first type of object and the second type of object; when the position of the composition positioning point and the preset composition point are not When matching, output prompt information for adjusting the shooting posture of the electronic device.
  • the embodiment of the application determines a common composition positioning point for two types of objects in the scene to be photographed, and can prompt the user to change the shooting posture of the electronic device when the composition positioning point does not match the preset composition point, thereby guiding the user in shooting Make a reasonable composition and take high-quality photos.
  • FIG. 5 is a schematic diagram of a first structure of a composition guiding device provided by an embodiment of the application.
  • the composition guidance device 300 is applied to electronic equipment, and the composition guidance device includes an object recognition module 301, a first determination module 302, a second determination module 303, and a composition guidance module 304, as follows:
  • the object recognition module 301 is used to perform object recognition on the scene to be photographed;
  • the first determining module 302 is configured to determine the first type of object and the second type of object in the to-be-shot scene when it is recognized that the to-be-shot scene contains at least two different types of objects;
  • the second determining module 303 is configured to determine a common composition positioning point for the first type of object and the second type of object according to the state information of the first type of object and the second type of object;
  • the composition guidance module 304 is configured to output prompt information for adjusting the shooting posture of the electronic device when the positioning point of the composition does not match the position of the preset composition point.
  • the object recognition module 301 may be used to:
  • FIG. 6 is a schematic diagram of a second structure of the composition guiding device 300 provided by an embodiment of the present application.
  • the composition guiding apparatus 300 may further include a third determining module 305, and the third determining module 305 is configured to:
  • the state information of the second-type object and the first-type object determine a common composition anchor point for the first-type object and the second-type object.
  • the second determining module 303 includes a first determining unit 3031, an acquiring unit 3032, and a second determining unit 3033.
  • the first determining unit 3031 determines the first center point of the first display area, and determines the second center point of the second display area;
  • the obtaining unit 3032 obtains the area ratio of the first display area to the second display area
  • the second determining unit 3033 determines the composition positioning point according to the first center point, the second center point, and the area ratio, and the ratio of the distance from the composition positioning point to the first center point and the second center point is equal to the area ratio.
  • the first determining unit 3031 may be used to:
  • the second coordinate value average value of all pixels in the second display area is acquired, and the pixel point corresponding to the second coordinate value average value is determined as the second center point.
  • the obtaining unit 3032 may be used to:
  • the ratio of the number of pixels is taken as the area ratio.
  • composition guidance module 304 may be used to:
  • the moving direction of the composition anchor point to the preset composition point is displayed on the preview interface, and the prompt information for adjusting the shooting posture of the electronic device in the moving direction is output.
  • the object recognition module 301 performs object recognition on the scene to be photographed; when it is recognized that the scene to be photographed contains at least two types of objects of different categories, the first determining module 302 determines The first type of object and the second type of object in the scene to be photographed; then the second determining module 303 determines a common composition location for the first type of object and the second type of object according to the state information of the first type of object and the second type of object Point; when the position of the composition positioning point and the preset composition point do not match, the composition guidance module 304 outputs prompt information for adjusting the shooting posture of the electronic device.
  • the embodiment of the application determines a common composition positioning point for two types of objects in the scene to be photographed, and can prompt the user to change the shooting posture of the electronic device when the composition positioning point does not match the preset composition point, thereby guiding the user in shooting Make a reasonable composition and take high-quality photos.
  • the embodiment of the present application also provides an electronic device.
  • Electronic equipment can be smart phones, tablet computers, gaming equipment, AR (Augmented Reality) equipment, automobiles, vehicle surrounding obstacle detection devices, audio playback devices, video playback devices, notebooks, desktop computing devices, wearable devices such as watches , Glasses, helmets, electronic bracelets, electronic necklaces, electronic clothing and other equipment.
  • FIG. 8 is a schematic diagram of a first structure of an electronic device 400 according to an embodiment of the application.
  • the electronic device 400 includes a processor 401 and a memory 402.
  • the processor 401 is electrically connected to the memory 402.
  • the processor 401 is the control center of the electronic device 400. It uses various interfaces and lines to connect various parts of the entire electronic device. Various functions and processing data of the equipment, so as to monitor the electronic equipment as a whole.
  • the processor 401 in the electronic device 400 can load the instructions corresponding to the process of one or more computer programs into the memory 402 according to the steps in the above-mentioned composition guidance method, and the processor 401 can run the instructions.
  • the computer program stored in the memory 402, so as to realize the steps in the above-mentioned composition guidance method for example:
  • the first type of object and the second type of object in the scene to be photographed are determined;
  • the processor 401 when performing object recognition on the scene to be photographed, performs the following steps:
  • the processor 401 before determining a common composition anchor point for the first type of object and the second type of object according to the state information of the first type of object and the second type of object, the processor 401 performs the following steps:
  • the state information of the second-type object and the first-type object determine a common composition anchor point for the first-type object and the second-type object.
  • the processor 401 when a common composition anchor point is determined for the first-type object and the second-type object according to the state information of the first-type object and the second-type object, the processor 401 performs the following steps:
  • the composition positioning point is determined according to the first center point, the second center point and the area ratio, and the ratio of the distance from the composition positioning point to the first center point to the second center point is equal to the area ratio.
  • the processor 401 when the first center point of the first display area is determined and the second center point of the second display area is determined, the processor 401 further performs the following steps:
  • the second coordinate value average value of all pixels in the second display area is acquired, and the pixel point corresponding to the second coordinate value average value is determined as the second center point.
  • the processor 401 when acquiring the area ratio of the first display area to the second display area, the processor 401 further executes the following steps:
  • the ratio of the number of pixels is taken as the area ratio.
  • the processor 401 when outputting prompt information for adjusting the shooting posture of the electronic device, the processor 401 further executes the following steps:
  • the moving direction of the composition anchor point to the preset composition point is displayed on the preview interface, and the prompt information for adjusting the shooting posture of the electronic device in the moving direction is output.
  • FIG. 9 is a schematic diagram of the second structure of the electronic device 400 according to an embodiment of the application.
  • the electronic device 400 further includes: a display screen 403, a control circuit 404, an input unit 405, a sensor 406, and a power supply 407.
  • the processor 401 is electrically connected to the display screen 403, the control circuit 404, the input unit 405, the sensor 406, and the power supply 407, respectively.
  • the display screen 403 can be used to display information input by the user or information provided to the user and various graphical user interfaces of the electronic device. These graphical user interfaces can be composed of images, text, icons, videos, and any combination thereof.
  • the control circuit 404 is electrically connected to the display screen 403 for controlling the display screen 403 to display information.
  • the input unit 405 can be used to receive inputted numbers, character information, or user characteristic information (such as fingerprints), and generate keyboard, mouse, joystick, optical or trackball signal input related to user settings and function control.
  • the input unit 405 may include a fingerprint recognition module.
  • the sensor 406 is used to collect the information of the electronic device itself or the information of the user or the external environment information.
  • the sensor 406 may include multiple sensors such as a distance sensor, a magnetic field sensor, a light sensor, an acceleration sensor, a fingerprint sensor, a Hall sensor, a position sensor, a gyroscope, an inertial sensor, a posture sensor, a barometer, and a heart rate sensor.
  • the power supply 407 is used to supply power to various components of the electronic device 400.
  • the power supply 407 may be logically connected to the processor 401 through a power management system, so that functions such as charging, discharging, and power consumption management can be managed through the power management system.
  • the electronic device 400 may also include a camera, a Bluetooth module, etc., which will not be repeated here.
  • an embodiment of the present application provides an electronic device.
  • the processor in the electronic device performs the following steps: performing object recognition on the scene to be photographed; Identify the first type of object and the second type of object in the scene to be photographed; determine the common composition positioning point for the first type of object and the second type of object according to the state information of the first type of object and the second type of object; when the composition is positioned When the position of the point does not match the preset composition point, a prompt message for adjusting the shooting posture of the electronic device is output.
  • the embodiment of the application determines a common composition positioning point for two types of objects in the scene to be photographed, and can prompt the user to change the shooting posture of the electronic device when the composition positioning point does not match the preset composition point, thereby guiding the user in shooting Make a reasonable composition and take high-quality photos.
  • An embodiment of the present application also provides a storage medium in which a computer program is stored.
  • the computer program When the computer program is run on a computer, the computer executes the composition guidance method of any one of the foregoing embodiments.
  • the computer program when the computer program runs on the computer, the computer performs the following steps:
  • the state information of the object determines a common composition positioning point for the first type of object and the second type of object; when the composition positioning point does not match the position of the preset composition point, a prompt message for adjusting the shooting posture of the electronic device is output.
  • the storage medium may include, but is not limited to: read only memory (ROM, Read Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks, and so on.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)

Abstract

Les modes de réalisation de l'invention concernent un procédé et un appareil de composition d'image, ainsi qu'un dispositif électronique. Le procédé de guidage de composition d'image consiste à : effectuer une reconnaissance d'objet sur une scène à photographier ; déterminer un objet de premier type et un objet de second type à partir d'au moins deux types d'objets différents reconnus ; déterminer un point de positionnement de composition d'image commun en fonction des informations d'état de l'objet de premier type et de l'objet de second type ; et si le point de positionnement de composition d'image ne correspond pas à la position d'un point de composition d'image prédéfini, générer des informations d'invite permettant d'ajuster une posture de photographie d'un dispositif électronique.
PCT/CN2021/074203 2020-03-09 2021-01-28 Procédé et appareil de guidage de composition d'image, et dispositif électronique WO2021179830A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010159225.7A CN111327829B (zh) 2020-03-09 2020-03-09 构图指导方法、装置、电子设备及存储介质
CN202010159225.7 2020-03-09

Publications (1)

Publication Number Publication Date
WO2021179830A1 true WO2021179830A1 (fr) 2021-09-16

Family

ID=71165671

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/074203 WO2021179830A1 (fr) 2020-03-09 2021-01-28 Procédé et appareil de guidage de composition d'image, et dispositif électronique

Country Status (2)

Country Link
CN (2) CN111327829B (fr)
WO (1) WO2021179830A1 (fr)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111327829B (zh) * 2020-03-09 2021-12-17 Oppo广东移动通信有限公司 构图指导方法、装置、电子设备及存储介质
WO2022178724A1 (fr) * 2021-02-24 2022-09-01 深圳市大疆创新科技有限公司 Procédé de photographie d'image, dispositif terminal, appareil photographique et support de stockage
CN113704526B (zh) * 2021-07-29 2023-08-04 福建榕基软件工程有限公司 一种拍摄构图引导方法与终端
CN116567385A (zh) * 2023-06-14 2023-08-08 深圳市宗匠科技有限公司 图像采集方法及图像采集装置

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101415077A (zh) * 2007-10-17 2009-04-22 索尼株式会社 构图判定设备、构图判定方法和程序
JP2011223599A (ja) * 2011-05-31 2011-11-04 Casio Comput Co Ltd 撮影装置及びプログラム
CN107613203A (zh) * 2017-09-22 2018-01-19 维沃移动通信有限公司 一种图像处理方法及移动终端
CN109547694A (zh) * 2018-11-29 2019-03-29 维沃移动通信有限公司 一种图像显示方法及终端设备
CN111327829A (zh) * 2020-03-09 2020-06-23 Oppo广东移动通信有限公司 构图指导方法、装置、电子设备及存储介质

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8045007B2 (en) * 2004-12-24 2011-10-25 Fujifilm Corporation Image capturing system and image capturing method
JP4508257B2 (ja) * 2008-03-19 2010-07-21 ソニー株式会社 構図判定装置、構図判定方法、プログラム
CN103559717A (zh) * 2013-11-14 2014-02-05 上海华勤通讯技术有限公司 拍照设备的拍照预览构图辅助方法及装置
US9667860B2 (en) * 2014-02-13 2017-05-30 Google Inc. Photo composition and position guidance in a camera or augmented reality system
KR20180098059A (ko) * 2017-02-24 2018-09-03 엘지전자 주식회사 이동 단말기
US10218901B2 (en) * 2017-04-05 2019-02-26 International Business Machines Corporation Picture composition adjustment
CN107734142B (zh) * 2017-09-15 2020-05-05 维沃移动通信有限公司 一种拍照方法、移动终端及服务器
CN107820020A (zh) * 2017-12-06 2018-03-20 广东欧珀移动通信有限公司 拍摄参数的调整方法、装置、存储介质及移动终端
CN108366203B (zh) * 2018-03-01 2020-10-13 北京金山安全软件有限公司 一种构图方法、装置、电子设备及存储介质
CN108513073A (zh) * 2018-04-13 2018-09-07 朱钢 一种具备摄影师构图意识的手机拍照机能的实现方法
CN108377339A (zh) * 2018-05-07 2018-08-07 维沃移动通信有限公司 一种拍照方法及拍照装置
CN109040596B (zh) * 2018-08-27 2020-08-28 Oppo广东移动通信有限公司 一种调整摄像头的方法、移动终端及存储介质
CN109587394A (zh) * 2018-10-23 2019-04-05 广东智媒云图科技股份有限公司 一种智能构图方法、电子设备及存储介质
CN109872340B (zh) * 2019-01-03 2023-06-27 广东智媒云图科技股份有限公司 构图方法及其电子设备、计算机可读存储介质
CN110830712A (zh) * 2019-09-16 2020-02-21 幻想动力(上海)文化传播有限公司 一种自主摄影系统和方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101415077A (zh) * 2007-10-17 2009-04-22 索尼株式会社 构图判定设备、构图判定方法和程序
JP2011223599A (ja) * 2011-05-31 2011-11-04 Casio Comput Co Ltd 撮影装置及びプログラム
CN107613203A (zh) * 2017-09-22 2018-01-19 维沃移动通信有限公司 一种图像处理方法及移动终端
CN109547694A (zh) * 2018-11-29 2019-03-29 维沃移动通信有限公司 一种图像显示方法及终端设备
CN111327829A (zh) * 2020-03-09 2020-06-23 Oppo广东移动通信有限公司 构图指导方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
CN111327829B (zh) 2021-12-17
CN111327829A (zh) 2020-06-23
CN114025097B (zh) 2023-12-12
CN114025097A (zh) 2022-02-08

Similar Documents

Publication Publication Date Title
WO2021179830A1 (fr) Procédé et appareil de guidage de composition d'image, et dispositif électronique
CN110647865B (zh) 人脸姿态的识别方法、装置、设备及存储介质
CN111263066B (zh) 构图指导方法、装置、电子设备及存储介质
WO2021135601A1 (fr) Procédé et appareil de photographie auxiliaire, équipement terminal, et support d'enregistrement
WO2021023059A1 (fr) Procédé de photographie et terminal mobile
EP3547218B1 (fr) Dispositif et procédé de traitement de fichiers, et interface utilisateur graphique
CN112585566B (zh) 用于与具有内置摄像头的设备进行交互的手遮脸输入感测
US10523916B2 (en) Modifying images with simulated light sources
US11070717B2 (en) Context-aware image filtering
US11030733B2 (en) Method, electronic device and storage medium for processing image
WO2022227393A1 (fr) Procédé et appareil de prise d'image photographique, dispositif électronique et support de stockage lisible par ordinateur
CN110290426B (zh) 展示资源的方法、装置、设备及存储介质
WO2024021742A1 (fr) Procédé d'estimation de point de fixation et dispositif associé
CN110851062A (zh) 一种绘图方法及电子设备
WO2021162201A1 (fr) Interface utilisateur d'appareil photographique à zoom à encliquetage
JP2012238293A (ja) 入力装置
EP4047549A1 (fr) Procédé et dispositif de détection d'image, et dispositif électronique
US9195310B2 (en) Camera cursor system
EP4072121A1 (fr) Procédé et appareil de photographie, support d'enregistrement et dispositif électronique
Lo et al. Augmediated reality system based on 3D camera selfgesture sensing
US11756302B1 (en) Managing presentation of subject-based segmented video feed on a receiving device
CN114697530B (zh) 一种智能取景推荐的拍照方法及装置
WO2021143628A1 (fr) Procédé de traitement d'image et appareil électronique
JP7293362B2 (ja) 撮影方法、装置、電子機器及び記憶媒体
KR101720607B1 (ko) 영상 촬영 장치 및 그 동작 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21768158

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21768158

Country of ref document: EP

Kind code of ref document: A1