WO2020093866A1 - Photography guiding method and apparatus, mobile terminal and storage medium - Google Patents

Photography guiding method and apparatus, mobile terminal and storage medium Download PDF

Info

Publication number
WO2020093866A1
WO2020093866A1 PCT/CN2019/112566 CN2019112566W WO2020093866A1 WO 2020093866 A1 WO2020093866 A1 WO 2020093866A1 CN 2019112566 W CN2019112566 W CN 2019112566W WO 2020093866 A1 WO2020093866 A1 WO 2020093866A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
area
sample
target
neural network
Prior art date
Application number
PCT/CN2019/112566
Other languages
French (fr)
Chinese (zh)
Inventor
张渊
郑文
Original Assignee
北京达佳互联信息技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京达佳互联信息技术有限公司 filed Critical 北京达佳互联信息技术有限公司
Publication of WO2020093866A1 publication Critical patent/WO2020093866A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces

Definitions

  • the present application relates to the technical field of image processing, and in particular to a shooting guidance method, device, mobile terminal, and storage medium.
  • composition is an important factor in the quality of the captured images.
  • the salient target area is an area containing a saliency target.
  • the saliency target is a target that human visually pays more attention to, such as: cars driving on roads, behaviors on snow, flowers in lush green leaves, and so on.
  • the saliency target area of the image is determined based on the gray value of each pixel in the image, that is, the saliency target area of the image is determined based on the gray scale feature of the image.
  • the inventor found that if the gray value of each pixel of the image is not significantly different, the accuracy of the determined saliency target area is low, which in turn makes the accuracy of shooting guidance based on the saliency target area low.
  • the present application provides a shooting guidance method, device, mobile terminal, and storage medium.
  • a shooting guidance method including:
  • the convolutional neural network Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled images corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
  • guide information for guiding the user to compose a picture is output.
  • a shooting guidance device including:
  • the first acquisition module is configured to acquire the indication image containing the target in the shooting scene
  • the first determining module is configured to determine the saliency target area in the indication image based on the pre-trained convolutional neural network; wherein, the convolutional neural network is based on multiple sample images and each sample image corresponds to From the training of labeled images, the marked image corresponding to each sample image is marked with the saliency target area in the sample image;
  • the instruction module is configured to output guide information for guiding the user to compose a composition based on the determined saliency target area.
  • a mobile terminal including:
  • Memory for storing processor executable instructions
  • the processor is configured to:
  • the convolutional neural network Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
  • guide information for guiding the user to compose a picture is output.
  • a non-transitory computer-readable storage medium When instructions in the storage medium are executed by a processor of a mobile terminal, the mobile terminal can perform a shooting guidance method.
  • a computer program product which, when run on a computer, enables the computer to execute a shooting guidance method.
  • a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output is used for Guidance information that guides the user to composition. Since this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since this scheme can be applied to various grayscale situations, this scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
  • Fig. 1 is a flow chart showing a method for shooting guidance according to an exemplary embodiment.
  • Fig. 2 (a) is a schematic diagram showing a detection result of a salient target area according to an exemplary embodiment.
  • Fig. 2 (b) is another schematic diagram showing the detection result of the salient target area according to an exemplary embodiment.
  • Fig. 2 (c) is another schematic diagram showing the detection result of the salient target area according to an exemplary embodiment.
  • Fig. 3 (a) is a schematic diagram of a shooting interface according to an exemplary embodiment.
  • Fig. 3 (b) is a schematic diagram of guiding the main body region to the target shooting region according to an exemplary embodiment.
  • Fig. 3 (c) is another schematic diagram of guiding the body region to the target shooting region according to an exemplary embodiment.
  • Fig. 3 (d) is another schematic diagram of guiding the body area to the target shooting area according to an exemplary embodiment.
  • Fig. 3 (e) is another schematic diagram of guiding the body area to the target shooting area according to an exemplary embodiment.
  • Fig. 4 is a flowchart illustrating training a convolutional neural network according to an exemplary embodiment.
  • Fig. 5 is a schematic structural diagram of a preset convolutional neural network according to an exemplary embodiment.
  • Fig. 6 is a block diagram of a shooting guide device according to an exemplary embodiment.
  • Fig. 7 is a block diagram of a mobile terminal according to an exemplary embodiment.
  • Fig. 8 is a block diagram of a device according to an exemplary embodiment.
  • Fig. 1 is a flowchart of a shooting guidance method according to an exemplary embodiment. As shown in Fig. 1, the shooting guidance method may be used in a mobile terminal, and may include the following steps.
  • step S11 an instruction image containing the target in the shooting scene is acquired.
  • the mobile terminal When the user is to take a picture through the mobile terminal, the mobile terminal obtains an indication image containing the target in the shooting scene. Specifically, the mobile terminal can acquire the indication image in real time. It should be emphasized that since this solution is to instruct the user to compose the composition during shooting, then the indicated image is the framing picture generated by the mobile terminal during the shooting process, that is, the user ’s display screen after the user turns on the camera function. The picture presented.
  • the target in the shooting scene may be a target that users, animals, vehicles, buildings, etc. usually pay more attention to.
  • the target in the shooting scene is the saliency target
  • the image area where the target in the indication image is located is the saliency target area.
  • the mobile terminal may be a smart phone, a tablet computer, a camera, a video camera, and other devices that have a photo function.
  • step S12 based on the pre-trained convolutional neural network, a saliency target area in the indication image is determined.
  • the convolutional neural network is a network for identifying the salient target area in the image.
  • the convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, and the marked image corresponding to each sample image is marked with the saliency target area of the sample image.
  • a convolutional neural network is obtained by training in advance based on multiple sample images and the labeled images corresponding to each sample image.
  • the saliency target area of the indication image can be determined based on the trained convolutional neural network.
  • the determined saliency target area may be one or more.
  • the instruction image is input to the trained convolutional neural network to obtain the saliency target area in the instruction image.
  • the saliency target area in the indication image can be embodied in the indication image through the area frame.
  • the shooting scene may include different categories of targets.
  • not only the marked target area is marked in the marked image corresponding to each sample image, but also the target category of the target contained in the marked target area may be marked.
  • the trained convolutional neural network can detect the saliency target area in the indication image, and can mark the target category corresponding to the saliency target area.
  • the three white line boxes show the three saliency target regions based on the convolutional neural network, and the target categories corresponding to the three saliency target regions, such as person /0.7, sheep / 0.951, sheep / 0.9; as shown in four white lines in Figure 2 (b), four saliency target regions based on convolutional neural network are shown, and four saliency can be marked
  • the target categories corresponding to the sexual target areas such as person / 0.752, person / 0.561, horse / 0.959, horse / 0.916; as shown by the two white lines in Figure 2 (c), they are obtained based on the convolutional neural network.
  • Two saliency target areas, and the target categories corresponding to the two saliency target areas can be marked, such as sheep / 0.918, sheep / 0.871.
  • step S13 based on the determined saliency target area, guidance information for guiding the user to compose a picture is output.
  • the guide information is information for guiding the user to compose a picture, that is, information for guiding the user to move the determined saliency target area to a certain position on the screen.
  • the guidance information can be displayed in a floating manner on the viewfinder screen.
  • the guide information may include: a first mark marked for the salient target area to be moved, a second mark marked for the position to be placed, and a composition description for moving the first mark and the second mark to coincide .
  • the number of salient target regions may be one or more.
  • output guide information for guiding the user to compose a picture which may specifically include:
  • the one saliency target area can be directly selected as the subject area.
  • the subject area can be selected according to the areas of different saliency target areas, for example, the saliency target area with the largest area is selected as the subject area.
  • the subject area can be selected according to the target category corresponding to different distinctive target areas. For example, when shooting an image that includes a person, it is generally expected that the person is located in a more important position in the image. At this time, it can be judged The target category corresponding to the saliency target area, the saliency target area whose target category is human is determined as the subject area, and so on.
  • the target category is a human significant target area is the subject area 1
  • the target category is an animal significant target area is the subject area 2, and so on.
  • the preset composition method may include a nine-square lattice composition, a symmetric composition, a guide line composition, a three-point composition, and the like.
  • the target shooting area can be different.
  • the target shooting area may be the middle square area of the nine-square grid
  • the target shooting area may be the left-side composition of the symmetric composition, and so on.
  • the movement of the main body area to the target shooting area may be: the main body area all falls within the target shooting area; or, the main body area and the target shooting area completely overlap; or, the main body area and the target shooting area overlap a predetermined ratio, such as 80%, 90% and many more.
  • the user can move the mobile terminal until the determined saliency target area or subject area is moved to the target shooting area.
  • Auxiliary shooting is performed under the shooting interface shown in FIG. 3 (a).
  • the shooting interface shown in FIG. 3 (a) includes smart composition, flash, low light, and variable speed options; and also includes albums and live streaming alongside the shooting options And K song options; as shown in Figure 3 (b), Figure 3 (c), Figure 3 (d), and Figure 3 (e), in order to move the subject area to the target shooting area, an open circle 301 annotating the subject area The change in the positional relationship with the solid area 302 marked with the target shooting area.
  • the open circle 301 moves down as the marked main body area moves down, that is, the open center circle 301 always marks the main body area, so that the virtual circle 301 falls into the solid circle 302 At this time, the subject area marked by the open circle also falls into the target shooting area.
  • the target shooting area is determined by the preset composition method
  • the preset composition method is the composition method used when shooting the instruction image.
  • the technical solutions provided by the embodiments of the present application may include the following beneficial effects: through a convolutional neural network, a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output is used for Guidance information that guides the user to composition.
  • this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since the scheme can be applied to various grayscale situations, the scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
  • the method provided in the embodiment of the present application may further include: instructing the subject area to move to After the target shooting area, it can also include:
  • the user When it is detected whether the subject area moves to the target shooting area, the user is notified that the composition is successful. Specifically, as shown in FIG. 3 (e), text information can be displayed in the shooting interface, such as: "the composition is successful, and the picture can be taken.” Or you can issue a voice message prompt, and so on.
  • the user can intuitively know that the composition is successful, and take a picture when the composition is successful, so that the target in the shooting scene can be taken when it is in the best shooting position, so that the user is not required to have professional photography knowledge to guide the user to compose the composition, thereby improving The quality of the captured image.
  • the saliency target area in the framing screen guides the user to compose a composition, for example, the user moves the mobile phone, so that the subject area selected from the saliency target area is located in the target shooting area, instructing the user to take a higher quality image.
  • the embodiments of the present application may further include: a step of training a convolutional neural network. Specifically, as shown in FIG. 4, it may include:
  • the electronic device acquires multiple sample images, such as 10,000, 20,000, etc.
  • the plurality of sample images include sample images of different categories, and the sample images of different categories include targets of different categories.
  • a data set may be constructed in advance, and then sample images are obtained from the data set.
  • images corresponding to natural scenes can be obtained, such as selecting 7 target categories in a realizable way to construct a data set, such as person (person), cat (cat), dog, (dog), horse (horse), sheep (sheep), cow (cow), bird (bird).
  • part of the sample image can be used as the training set and part of the test set.
  • the total training data in the data set is 1328415, and 1904 data is used as the test set to verify the convolutional neural network. effect.
  • the embodiments of the present application propose the following ways to determine the sampling weights of images of different categories, the main steps are as follows:
  • the regional convolutional neural network can ensure that the images of each category in a batch (batch processing) are balanced during network model training, and prevent deviations in model training.
  • the saliency target area corresponding to each sample image can be marked by manual marking; or the saliency target area corresponding to each sample image can be marked through detection of target characteristics, and so on.
  • the preset convolutional nerve may include a parameter to be tested, input the sample image to the preset convolutional neural network, and adjust the parameter to be tested, so that the output of the preset convolutional neural network is infinitely close to the labeled image corresponding to the sample image
  • the marked area when the difference between the output of the preset convolutional neural network and the image marked by the marked image corresponding to the sample image converges, the parameter to be measured is determined, and the obtained preset convolution including the determined parameter to be measured is obtained
  • the neural network is the convolutional neural network trained.
  • the parameters to be tested may include: batch size, learning rate, and / or number of iterations, and so on.
  • the preset convolutional neural network may include: a basic module, a feature module, a positioning module, and a classification module.
  • the basic module can be composed of 5 convolutional layers, such as Conv2d (3-> 8), Conv2d (8-> 16), Conv2d (16-> 32), Conv2d (32-> 64), Conv2d (64-> 128) ), Where Conv2d (3-> 8) is understood to convert a 3-channel RGB format image into an 8-channel image, and the meaning of other convolutional layers is similar;
  • the feature module can be composed of 3 convolutional layers, such as Conv2d (128- > 256), Conv2d (256-> 128), Conv2d (128-> 256);
  • the positioning module can be composed of a convolutional layer, such as Conv2d (*> 8);
  • the classification module can be composed of a convolutional layer, Such as Conv2d (*> 4).
  • the basic module can also be called Base Model
  • the feature module can also be called Extra Model
  • the positioning module can also be called Location Layer
  • the classification module can also be called Confidence Layer.
  • Base is mainly used to process the features from the bottom to the top of the image, and is used to provide features for Extra Model.
  • Extra Model mainly extracts feature maps of different scales, mainly designed for targets of different sizes in the image. Each pixel position of the feature map is mapped to the corresponding size of the bounding box on the original image, which can be Location and Layer Provide characteristic information of different size targets.
  • Location layer is used to return the bounding box coordinates of the target on the original image for each pixel of the feature map.
  • Confidence layer is used to calculate the target category of the bounding box on the original image for each pixel of the feature map.
  • the trained network may be optimized and evaluated.
  • the preset convolutional neural network After the preset convolutional neural network is constructed, it needs to be trained on the training data set. The preset convolutional neural network training depends on the loss function and the optimizer. Finally, the evaluation of the preset convolutional neural network needs to set an appropriate evaluation index.
  • the loss function uses the multi-task loss function in the optimization process of the convolutional neural network, uses the cross-entropy loss function for the classification task, and uses the L1 distance loss function for the bounding box.
  • Optimizer network optimization uses stochastic gradient descent method to update the parameters of convolutional neural network.
  • mean, accuracy, and average precision can be used in the embodiments of the present application to evaluate the detection result.
  • the acquiring multiple sample images includes:
  • the original image after interference is determined as the sample image.
  • the interference may include horizontal flip, vertical flip, random cropping, image pixel disturbance, lighting processing, occlusion, low contrast processing, and so on.
  • Fig. 6 is a block diagram of a shooting guidance device according to an exemplary embodiment.
  • the apparatus may include a first acquisition module 601, a first determination module 602, and an instruction module 603;
  • the first obtaining module 601 is configured to obtain an indication image containing the target in the shooting scene
  • the first determination module 602 is configured to determine a saliency target area in the indicated image based on the pre-trained convolutional neural network; wherein, the convolutional neural network is based on multiple sample images and each sample image corresponds to From the training of labeled images, the marked image corresponding to each sample image is marked with the saliency target area in the sample image;
  • the instruction module 603 is configured to output guide information for guiding the user to compose a composition based on the determined saliency target area.
  • the technical solutions provided by the embodiments of the present application may include the following beneficial effects: through a convolutional neural network, a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output Guidance information that guides the user to composition. Since this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since the scheme can be applied to various grayscale situations, the scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
  • the instruction module 603 is specifically configured to: select a subject area from the determined salient target areas; output guide information for guiding the user to move the subject area to the target shooting area; wherein, The target shooting area is determined based on a preset composition method, and the preset composition method is a composition method used when shooting the instruction image.
  • the device also includes:
  • the prompting module is configured to, after the instruction module outputs guidance information for guiding the user to move the subject area to the target shooting area, when it is detected whether the subject area moves to the target shooting area, to the user Prompt composition success.
  • the device also includes:
  • the second acquisition module is configured to acquire multiple sample images
  • the second determining module is configured to determine the mark image corresponding to each sample image respectively;
  • the training module is configured to take each sample image as input content and the labeled image corresponding to each sample image as supervising content to train the preset convolutional neural network to obtain a trained convolutional neural network.
  • the second acquisition module is specifically configured to: acquire multiple original images; add interference to the multiple original images respectively to obtain multiple original images after interference; combine the multiple original images after interference Both are determined as sample images.
  • the plurality of sample images include sample images of different categories, and the sample images of different categories include targets of different categories.
  • the preset convolutional neural network includes: a basic module, a feature module, a positioning module, and a classification module.
  • Fig. 7 is a block diagram of a mobile terminal 700 for shooting guidance according to an exemplary embodiment.
  • the mobile terminal 700 may be a mobile phone, a computer, a digital broadcasting terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
  • the mobile terminal 700 may include one or more of the following components: a processing component 702, a memory 704, a power component 706, a multimedia component 708, an audio component 710, an input / output (I / O) interface 712, and a sensor component 714 , ⁇ ⁇ ⁇ 716.
  • the processing component 702 generally controls the overall operations of the mobile terminal 700, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
  • the processing component 702 may include one or more processors 720 to execute instructions to complete all or part of the steps of the above-mentioned shooting guidance method.
  • the processing component 702 may include one or more modules to facilitate interaction between the processing component 702 and other components.
  • the processing component 702 may include a multimedia module to facilitate interaction between the multimedia component 708 and the processing component 702.
  • the memory 704 is configured to store various types of data to support operations at the mobile terminal 700. Examples of these data include instructions for any application or method operating on the mobile terminal 700, contact data, phone book data, messages, pictures, videos, and so on.
  • the memory 704 may be implemented by any type of volatile or nonvolatile storage device or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable and removable Programmable read only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory, magnetic disk or optical disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read only memory
  • EPROM erasable and removable Programmable read only memory
  • PROM programmable read only memory
  • ROM read only memory
  • magnetic memory flash memory
  • flash memory magnetic disk or optical disk.
  • the power supply component 706 provides power to various components of the mobile terminal 700.
  • the power supply component 706 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the mobile terminal 700.
  • the multimedia component 708 includes a screen between the mobile terminal 700 and the user that provides an output interface.
  • the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touch, swipe, and gestures on the touch panel. The touch sensor may not only sense the boundary of the touch or sliding action, but also detect the duration and pressure related to the touch or sliding operation.
  • the multimedia component 708 includes a front camera and / or a rear camera. When the mobile terminal 700 is in an operation mode, such as a shooting mode or a video mode, the front camera and / or the rear camera may receive external multimedia data. Each front camera and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
  • the audio component 710 is configured to output and / or input audio signals.
  • the audio component 710 includes a microphone (MIC), and when the mobile terminal 700 is in an operation mode, such as a call mode, a recording mode, and a voice recognition mode, the microphone is configured to receive an external audio signal.
  • the received audio signal may be further stored in the memory 704 or sent via the communication component 716.
  • the audio component 710 further includes a speaker for outputting audio signals.
  • the I / O interface 712 provides an interface between the processing component 702 and a peripheral interface module.
  • the peripheral interface module may be a keyboard, a click wheel, or a button. These buttons may include, but are not limited to: home button, volume button, start button, and lock button.
  • the sensor component 714 includes one or more sensors for providing the mobile terminal 700 with status evaluation in various aspects.
  • the sensor component 714 can detect the on / off state of the mobile terminal 700, and the relative positioning of the components, such as the display and the keypad of the mobile terminal 700, the sensor component 714 can also detect the mobile terminal 700 or the mobile terminal 700 The position of the component changes, the presence or absence of user contact with the mobile terminal 700, the orientation or acceleration / deceleration of the mobile terminal 700, and the temperature change of the mobile terminal 700.
  • the sensor assembly 714 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • the sensor component 714 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 714 may further include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • the communication component 716 is configured to facilitate wired or wireless communication between the mobile terminal 700 and other devices.
  • the mobile terminal 700 may access a wireless network based on a communication standard, such as WiFi, an operator network (such as 2G, 3G, 4G, or 5G), or a combination thereof.
  • the communication component 716 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel.
  • the communication component 716 further includes a near field communication (NFC) module to facilitate short-range communication.
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • the mobile terminal 700 may be used by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field Programming gate array (FPGA), controller, microcontroller, microprocessor or other electronic components are used to implement the above method.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field Programming gate array
  • controller microcontroller, microprocessor or other electronic components are used to implement the above method.
  • a non-transitory computer-readable storage medium including instructions is also provided, for example, a memory 704 including instructions, which can be executed by the processor 720 of the mobile terminal 700 to complete the above method.
  • the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, or the like.
  • a computer program product which, when run on a computer, enables the computer to execute the above-mentioned shooting guidance method.
  • Fig. 8 is a block diagram of a device 800 for shooting guidance according to an exemplary embodiment.
  • the device 800 may be provided as a server.
  • the device 800 includes a processing component 822, which further includes one or more processors, and memory resources represented by the memory 832, for storing instructions executable by the processing component 822, such as application programs.
  • the application programs stored in the memory 832 may include one or more modules each corresponding to a set of instructions.
  • the processing component 822 is configured to execute instructions to perform the method steps of the above-mentioned shooting guidance method.
  • the device 800 may also include a power component 826 configured to perform power management of the device 800, a wired or wireless network interface 850 configured to connect the device 800 to the network, and an input output (I / O) interface 858.
  • the device 800 can operate based on an operating system stored in the memory 832, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or the like.

Abstract

A photography guiding method and apparatus, a mobile terminal and a storage medium. The method comprises: acquiring an indication image including an object in a photography scenario; determining, based on a pre-trained convolutional neural network, a salient object region in the indication image, wherein the convolutional neural network is obtained by means of training according to multiple sample images and a marking image corresponding to each of the sample images, and the marking image corresponding to each of the sample images marks a salient object region in the sample image; and outputting, based on the determined salient object region, guidance information for guiding a user to compose a picture. By means of the photography guiding method and apparatus, the mobile terminal and the storage medium provided in the embodiments of the present application, an improvement in the accuracy of photography guidance can be ensured with regard to various gray differences.

Description

拍摄引导方法、装置及移动终端和存储介质Shooting guidance method, device, mobile terminal and storage medium
本申请要求于2018年11月05日提交中国专利局、申请号为201811307419.6发明名称为“拍摄引导方法、装置及移动终端和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application requires the priority of the Chinese patent application filed on November 05, 2018 in the Chinese Patent Office with the application number 201811307419.6 and the invention titled "Photographic guidance method, device and mobile terminal and storage medium" In this application.
技术领域Technical field
本申请涉及图像处理技术领域,特别是涉及一种拍摄引导方法、装置及移动终端和存储介质。The present application relates to the technical field of image processing, and in particular to a shooting guidance method, device, mobile terminal, and storage medium.
背景技术Background technique
随着智能手机、平板电脑等移动终端的发展,移动终端的拍照功能被广泛地使用。在使用移动终端拍照或者拍视频过程中,为了使得拍摄的图像的质量更好,拍摄者需要了解一定的拍摄知识,但是很多用户并不具备专业的拍摄知识,拍摄的图像不尽如人意。With the development of mobile terminals such as smart phones and tablet computers, the camera function of mobile terminals is widely used. In the process of taking pictures or taking videos with a mobile terminal, in order to make the quality of the captured images better, the photographer needs to understand certain shooting knowledge, but many users do not have professional shooting knowledge and the captured images are not satisfactory.
而构图是造成拍摄的图像质量的重要因素,为了能够提高不具备专业拍摄知识的用户拍摄的图像的质量,可以通过提取图像中的显著性目标区域,即提取取景画面中的显著性目标区域,进而基于该显著性目标区域引导构图,即拍摄引导。其中,显著性目标区域为包含显著性目标的区域,该显著性目标为人类视觉上较为关注的目标,例如:道路上行驶的汽车、雪地上的行为、繁茂绿叶中的花朵等等。相关技术中,基于图像中各个像素点的灰度值确定图像的显著性目标区域,即基于图像的灰度特征确定图像的显著性目标区域。发明人发现,若图像各个像素点的灰度值差别不大,则确定的显著性目标区域的准确度较低,进而会使得基于显著性目标区域进行拍摄引导的准确性较低。Composition is an important factor in the quality of the captured images. In order to improve the quality of the images captured by users who do not have professional shooting knowledge, you can extract the salient target area in the image, that is, the salient target area in the viewfinder screen. Then, the composition is guided based on the salient target area, that is, the shooting guidance. Among them, the saliency target area is an area containing a saliency target. The saliency target is a target that human visually pays more attention to, such as: cars driving on roads, behaviors on snow, flowers in lush green leaves, and so on. In the related art, the saliency target area of the image is determined based on the gray value of each pixel in the image, that is, the saliency target area of the image is determined based on the gray scale feature of the image. The inventor found that if the gray value of each pixel of the image is not significantly different, the accuracy of the determined saliency target area is low, which in turn makes the accuracy of shooting guidance based on the saliency target area low.
发明内容Summary of the invention
为克服相关技术中存在的问题,本申请提供一种拍摄引导方法、装置及移动终端和存储介质。To overcome the problems in the related art, the present application provides a shooting guidance method, device, mobile terminal, and storage medium.
根据本申请实施例的第一方面,提供一种拍摄引导方法,包括:According to a first aspect of the embodiments of the present application, a shooting guidance method is provided, including:
获取包含拍摄场景中目标的指示图像;Obtain an indication image containing the target in the shooting scene;
基于预先训练的卷积神经网络,确定所述指示图像中的显著性目标区域; 其中,所述卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像中的显著性目标区域;Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled images corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。Based on the determined saliency target area, guide information for guiding the user to compose a picture is output.
根据本申请实施例的第二方面,提供一种拍摄引导装置,包括:According to a second aspect of the embodiments of the present application, a shooting guidance device is provided, including:
第一获取模块,被配置为获取包含拍摄场景中目标的指示图像;The first acquisition module is configured to acquire the indication image containing the target in the shooting scene;
第一确定模块,被配置为基于预先训练的卷积神经网络,确定所述指示图像中的显著性目标区域;其中,所述卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像中的显著性目标区域;The first determining module is configured to determine the saliency target area in the indication image based on the pre-trained convolutional neural network; wherein, the convolutional neural network is based on multiple sample images and each sample image corresponds to From the training of labeled images, the marked image corresponding to each sample image is marked with the saliency target area in the sample image;
指示模块,被配置为基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。The instruction module is configured to output guide information for guiding the user to compose a composition based on the determined saliency target area.
根据本申请实施例的第三方面,提供一种移动终端,包括:According to a third aspect of the embodiments of the present application, a mobile terminal is provided, including:
处理器;processor;
用于存储处理器可执行指令的存储器;Memory for storing processor executable instructions;
其中,所述处理器被配置为:Wherein, the processor is configured to:
获取包含拍摄场景中目标的指示图像;Obtain an indication image containing the target in the shooting scene;
基于预先训练的卷积神经网络,确定所述指示图像中的显著性目标区域;其中,所述卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像中的显著性目标区域;Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。Based on the determined saliency target area, guide information for guiding the user to compose a picture is output.
根据本申请实施例的第四方面,提供一种非临时性计算机可读存储介质,当所述存储介质中的指令由移动终端的处理器执行时,使得移动终端能够执行一种拍摄引导方法。According to a fourth aspect of the embodiments of the present application, there is provided a non-transitory computer-readable storage medium. When instructions in the storage medium are executed by a processor of a mobile terminal, the mobile terminal can perform a shooting guidance method.
根据本申请实施例的第五方面,提供一种计算机程序产品,当其在计算机上运行时,使得计算机能够执行一种拍摄引导方法。According to a fifth aspect of the embodiments of the present application, there is provided a computer program product which, when run on a computer, enables the computer to execute a shooting guidance method.
本申请的实施例提供的技术方案可以包括以下有益效果:通过卷积神经网络,确定包括拍摄场景中目标的指示图像的显著性目标区域,进而,基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。由于本方案无需考虑各个像素点的灰度差异,因此,针对各种灰度差异情况,本方案均能够保证显著性目标区域确定的准确性,进而保证拍摄引导的准确性。而由 于本方案可以应用于各种灰度情况,因此,本方案对于不同情况的鲁棒性较高。且无需根据人为定义的确定特征提取显著性目标区域,使得应用场景更加广泛。The technical solutions provided by the embodiments of the present application may include the following beneficial effects: through a convolutional neural network, a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output is used for Guidance information that guides the user to composition. Since this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since this scheme can be applied to various grayscale situations, this scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,并不能限制本申请。It should be understood that the above general description and the following detailed description are only exemplary and explanatory, and cannot limit the present application.
附图说明BRIEF DESCRIPTION
为了更清楚地说明本申请实施例和现有技术的技术方案,下面对实施例和现有技术中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly explain the embodiments of the present application and the technical solutions of the prior art, the following briefly introduces the drawings required in the embodiments and the prior art. Obviously, the drawings in the following description are only For some embodiments of the application, for those of ordinary skill in the art, without paying any creative labor, other drawings may be obtained based on these drawings.
图1是根据一示例性实施例示出的一种拍摄引导方法的流程图。Fig. 1 is a flow chart showing a method for shooting guidance according to an exemplary embodiment.
图2(a)是根据一示例性实施例示出的显著性目标区域检测结果的一种示意图。Fig. 2 (a) is a schematic diagram showing a detection result of a salient target area according to an exemplary embodiment.
图2(b)是根据一示例性实施例示出的显著性目标区域检测结果的另一种示意图。Fig. 2 (b) is another schematic diagram showing the detection result of the salient target area according to an exemplary embodiment.
图2(c)是根据一示例性实施例示出的显著性目标区域检测结果的另一种示意图。Fig. 2 (c) is another schematic diagram showing the detection result of the salient target area according to an exemplary embodiment.
图3(a)是根据一示例性实施例示出的拍摄界面的示意图。Fig. 3 (a) is a schematic diagram of a shooting interface according to an exemplary embodiment.
图3(b)是根据一示例性实施例示出的引导主体区域移动至目标拍摄区域的一种示意图。Fig. 3 (b) is a schematic diagram of guiding the main body region to the target shooting region according to an exemplary embodiment.
图3(c)是根据一示例性实施例示出的引导主体区域移动至目标拍摄区域的另一种示意图。Fig. 3 (c) is another schematic diagram of guiding the body region to the target shooting region according to an exemplary embodiment.
图3(d)是根据一示例性实施例示出的引导主体区域移动至目标拍摄区域的另一种示意图。Fig. 3 (d) is another schematic diagram of guiding the body area to the target shooting area according to an exemplary embodiment.
图3(e)是根据一示例性实施例示出的引导主体区域移动至目标拍摄区域的另一种示意图。Fig. 3 (e) is another schematic diagram of guiding the body area to the target shooting area according to an exemplary embodiment.
图4是根据一示例性实施例示出的训练卷积神经网络的流程图。Fig. 4 is a flowchart illustrating training a convolutional neural network according to an exemplary embodiment.
图5是根据一示例性实施例示出的预设卷积神经网络的结构示意图。Fig. 5 is a schematic structural diagram of a preset convolutional neural network according to an exemplary embodiment.
图6是根据一示例性实施例示出的一种拍摄引导装置的框图。Fig. 6 is a block diagram of a shooting guide device according to an exemplary embodiment.
图7是根据一示例性实施例示出的一种移动终端的框图。Fig. 7 is a block diagram of a mobile terminal according to an exemplary embodiment.
图8是根据一示例性实施例示出的一种装置的框图。Fig. 8 is a block diagram of a device according to an exemplary embodiment.
具体实施方式detailed description
为使本申请的目的、技术方案、及优点更加清楚明白,以下参照附图并举实施例,对本申请进一步详细说明。显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。图1是根据一示例性实施例示出的一种拍摄引导方法的流程图,如图1所示,拍摄引导方法可以用于移动终端中,可以包括以下步骤。In order to make the purpose, technical solutions and advantages of the present application more clear, the following describes the present application in further detail with reference to the accompanying drawings and embodiments. Obviously, the described embodiments are only a part of the embodiments of the present application, but not all the embodiments. Based on the embodiments in the present application, all other embodiments obtained by a person of ordinary skill in the art without creative work fall within the protection scope of the present application. Fig. 1 is a flowchart of a shooting guidance method according to an exemplary embodiment. As shown in Fig. 1, the shooting guidance method may be used in a mobile terminal, and may include the following steps.
在步骤S11中,获取包含拍摄场景中目标的指示图像。In step S11, an instruction image containing the target in the shooting scene is acquired.
用户在待通过移动终端拍照时,移动终端获取包含拍摄场景中目标的指示图像。具体地,移动终端可以实时获取指示图像。需要强调的是,由于本方案为指示用户拍摄过程中的构图,那么,该指示图像为在拍摄过程中,移动终端所生成的取景画面,即用户打开相机功能后,移动终端的显示屏中所呈现的画面。When the user is to take a picture through the mobile terminal, the mobile terminal obtains an indication image containing the target in the shooting scene. Specifically, the mobile terminal can acquire the indication image in real time. It should be emphasized that since this solution is to instruct the user to compose the composition during shooting, then the indicated image is the framing picture generated by the mobile terminal during the shooting process, that is, the user ’s display screen after the user turns on the camera function. The picture presented.
其中,拍摄场景中的目标可以是人、动物、车辆、建筑物等用户通常较为关注的目标。该拍摄场景中的目标即为显著性目标,而该指示图像中该目标所在的图像区域即为显著性目标区域。Among them, the target in the shooting scene may be a target that users, animals, vehicles, buildings, etc. usually pay more attention to. The target in the shooting scene is the saliency target, and the image area where the target in the indication image is located is the saliency target area.
另外,移动终端可以智能手机、平板电脑、照相机、摄像机等等具有拍照功能的设备。In addition, the mobile terminal may be a smart phone, a tablet computer, a camera, a video camera, and other devices that have a photo function.
在步骤S12中,基于预先训练的卷积神经网络,确定指示图像中的显著性目标区域。In step S12, based on the pre-trained convolutional neural network, a saliency target area in the indication image is determined.
其中,该卷积神经网络为用于识别图像中显著性目标区域的网络。其中,卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像的显著性目标区域。本方案中,预先根据多个样本图像和每个样本图像对应的标记图像训练得到卷积神经网络。如此,针对指示图像,可以基于该训练好的卷积神经网络,确定该指示图像的显著性目标区域。确定出的显著性目标区域可以是一个,也可以是多个。具体地,将该指示图像输入至训练好的卷积神经网络,得到该指示图像中的显著性目标区域。其中,该指示图像中的显著性目标区域可以通过区域框体现于指示图像中。Among them, the convolutional neural network is a network for identifying the salient target area in the image. The convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, and the marked image corresponding to each sample image is marked with the saliency target area of the sample image. In this solution, a convolutional neural network is obtained by training in advance based on multiple sample images and the labeled images corresponding to each sample image. In this way, for the indication image, the saliency target area of the indication image can be determined based on the trained convolutional neural network. The determined saliency target area may be one or more. Specifically, the instruction image is input to the trained convolutional neural network to obtain the saliency target area in the instruction image. Wherein, the saliency target area in the indication image can be embodied in the indication image through the area frame.
因为拍摄场景中可能包括不同类别的目标。本申请实施例一种可选的方式中,每个样本图像对应的标记图像中不但标记有显著性目标区域,而且还可以标记显著性目标区域所包含目标的目标类别。这样,通过对多个样本图 像和样本图像对应的标记图像进行训练,使得训练得到的卷积神经网络能够检测指示图像中的显著性目标区域,且能够标注出显著性目标区域对应的目标类别。Because the shooting scene may include different categories of targets. In an optional manner of the embodiment of the present application, not only the marked target area is marked in the marked image corresponding to each sample image, but also the target category of the target contained in the marked target area may be marked. In this way, by training a plurality of sample images and labeled images corresponding to the sample images, the trained convolutional neural network can detect the saliency target area in the indication image, and can mark the target category corresponding to the saliency target area.
如图2(a)中三个白线框所示为基于卷积神经网络得到的三个显著性目标区域,且可以标注出三个显著性目标区域分别对应的目标类别,如人(person)/0.7、羊(sheep)/0.951、sheep/0.9;如图2(b)中四个白线框所示为基于卷积神经网络得到的四个显著性目标区域,且可以标注出四个显著性目标区域分别对应的目标类别,如person/0.752、person/0.561、马(horse)/0.959、horse/0.916;如图2(c)中两个白线框所示为基于卷积神经网络得到的两个显著性目标区域,且可以标注出两个显著性目标区域分别对应的目标类别,如sheep/0.918、sheep/0.871。As shown in Figure 2 (a), the three white line boxes show the three saliency target regions based on the convolutional neural network, and the target categories corresponding to the three saliency target regions, such as person /0.7, sheep / 0.951, sheep / 0.9; as shown in four white lines in Figure 2 (b), four saliency target regions based on convolutional neural network are shown, and four saliency can be marked The target categories corresponding to the sexual target areas, such as person / 0.752, person / 0.561, horse / 0.959, horse / 0.916; as shown by the two white lines in Figure 2 (c), they are obtained based on the convolutional neural network. Two saliency target areas, and the target categories corresponding to the two saliency target areas can be marked, such as sheep / 0.918, sheep / 0.871.
在步骤S13中,基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。In step S13, based on the determined saliency target area, guidance information for guiding the user to compose a picture is output.
其中,该引导信息为用于指引用户构图的信息,即指引用户将所确定的显著性目标区域移动至画面中的某一位置的信息。并且,该引导信息可以悬浮显示于取景画面。举例而言,该引导信息可以包括:为待移动的显著性目标区域标注的第一标识、为待放置位置所标注的第二标识,以及将第一标识和第二标识移动至重合的构图说明。Wherein, the guide information is information for guiding the user to compose a picture, that is, information for guiding the user to move the determined saliency target area to a certain position on the screen. In addition, the guidance information can be displayed in a floating manner on the viewfinder screen. For example, the guide information may include: a first mark marked for the salient target area to be moved, a second mark marked for the position to be placed, and a composition description for moving the first mark and the second mark to coincide .
其中,显著性目标区域的数量可以为一个或多个。Among them, the number of salient target regions may be one or more.
在一种实现方式中,基于所确定的显著性目标区域,输出用于指引用户构图的引导信息,具体可以包括:In one implementation, based on the determined saliency target area, output guide information for guiding the user to compose a picture, which may specifically include:
从所确定的显著性目标区域中,选择主体区域;From the determined significant target area, select the subject area;
输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息;其中,所述目标拍摄区域是预设构图方式所确定的,该预设构图方式为在拍摄该指示图像时所利用的构图方式。Output guide information for guiding the user to move the subject area to the target shooting area; wherein the target shooting area is determined by a preset composition method that is used when shooting the instruction image Composition method.
当确定出的显著性目标区域仅有一个时,可以直接选择该一个显著性目标区域作为主体区域。When there is only one saliency target area determined, the one saliency target area can be directly selected as the subject area.
当确定出的显著性目标区域有多个时,从所确定的显著性目标区域中,选择主体区域的实现方式存在多种。When there are multiple saliency target areas determined, from the determined saliency target areas, there are multiple ways of selecting the subject area.
一种可实现方式中,可以根据不同显著性目标区域的面积来选取主体区域,如选择面积最大的显著性目标区域为主体区域。In a realizable manner, the subject area can be selected according to the areas of different saliency target areas, for example, the saliency target area with the largest area is selected as the subject area.
另一种可实现方式中,可以根据不同显著性目标区域对应的目标类别选 择主体区域,如当拍摄包括人的图像时,一般情况下期望人位于图像中比较重要的位置,此时,可以判断显著性目标区域对应的目标类别,将目标类别为人的显著性目标区域确定为主体区域,等等。In another achievable way, the subject area can be selected according to the target category corresponding to different distinctive target areas. For example, when shooting an image that includes a person, it is generally expected that the person is located in a more important position in the image. At this time, it can be judged The target category corresponding to the saliency target area, the saliency target area whose target category is human is determined as the subject area, and so on.
本申请实施例一种可选的实现方式中,主体区域可以是多个。如确定目标类别为人的显著性目标区域为主体区域1,确定目标类别为动物的显著性目标区域为主体区域2,等等。In an optional implementation manner of the embodiment of the present application, there may be multiple body regions. For example, it is determined that the target category is a human significant target area is the subject area 1, the target category is an animal significant target area is the subject area 2, and so on.
并且,预设构图方式可以包括九宫格构图、对称构图、引导线构图、三分法构图等构图方式。不同的预设构图方式,目标拍摄区域可以不同。举例而言:针对九宫格构图而言,目标拍摄区域可以为九宫格的中间方格区域;而针对对称构图而言,目标拍摄区域可以为对称构图的左侧构图,等等。In addition, the preset composition method may include a nine-square lattice composition, a symmetric composition, a guide line composition, a three-point composition, and the like. Different preset composition methods, the target shooting area can be different. For example: for the nine-square grid composition, the target shooting area may be the middle square area of the nine-square grid; for symmetric composition, the target shooting area may be the left-side composition of the symmetric composition, and so on.
另外,主体区域移动至目标拍摄区域可以是:主体区域全部落入目标拍摄区域;或者,主体区域与目标拍摄区域完全重合;或者,主体区域与目标拍摄区域重合预定比例,如80%、90%等等。In addition, the movement of the main body area to the target shooting area may be: the main body area all falls within the target shooting area; or, the main body area and the target shooting area completely overlap; or, the main body area and the target shooting area overlap a predetermined ratio, such as 80%, 90% and many more.
另外,在输出引导信息后,用户可以移动移动终端,直至使得所确定的显著性目标区域或主体区域移动至目标拍摄区域。如图3所示。在图3(a)所示的拍摄界面下进行辅助拍摄,图3(a)所示的拍摄界面中包括智能构图、闪光灯、弱光、变速选项;且还包括与拍摄选项并列的相册、直播以及K歌选项;如图3(b)、图3(c)、图3(d)和图3(e),给出了为了将主体区域移动至目标拍摄区域,标注主体区域的虚心圆圈301与标注目标拍摄区域的实心区域302的位置关系所需发生的变化情况。需要强调的是,在实际应用时,虚心圆圈301随着所标注的主体区域的下移而下移,也就是说,虚心圆圈301始终标注主体区域,这样,在虚拟圆圈301落入实心圆圈302时,虚心圆圈标注的主体区域也落入到目标拍摄区域中。In addition, after outputting the guidance information, the user can move the mobile terminal until the determined saliency target area or subject area is moved to the target shooting area. As shown in Figure 3. Auxiliary shooting is performed under the shooting interface shown in FIG. 3 (a). The shooting interface shown in FIG. 3 (a) includes smart composition, flash, low light, and variable speed options; and also includes albums and live streaming alongside the shooting options And K song options; as shown in Figure 3 (b), Figure 3 (c), Figure 3 (d), and Figure 3 (e), in order to move the subject area to the target shooting area, an open circle 301 annotating the subject area The change in the positional relationship with the solid area 302 marked with the target shooting area. It should be emphasized that, in actual application, the open circle 301 moves down as the marked main body area moves down, that is, the open center circle 301 always marks the main body area, so that the virtual circle 301 falls into the solid circle 302 At this time, the subject area marked by the open circle also falls into the target shooting area.
另外,需要强调的是,基于所确定的显著性目标区域,输出用于指引用户构图的引导信息的具体实现方式存在多种,并不局限于上述的实现方式。举例而言:在另一种实现方式中,可以输出用于指引用户将所关注的显著性目标区域移动至目标拍摄区域的引导信息,其中,所述目标拍摄区域是预设构图方式所确定的,该预设构图方式为在拍摄该指示图像时所利用的构图方式。本申请的实施例提供的技术方案可以包括以下有益效果:通过卷积神经网络,确定包括拍摄场景中目标的指示图像的显著性目标区域,进而,基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。由于本方案无需考虑各个像素点的灰度差异,因此,针对各种灰度差异情况,本方案均 能够保证显著性目标区域确定的准确性,进而保证拍摄引导的准确性。而由于本方案可以应用于各种灰度情况,因此,本方案对于不同情况的鲁棒性较高。且无需根据人为定义的确定特征提取显著性目标区域,使得应用场景更加广泛。In addition, it should be emphasized that, based on the determined saliency target area, there are many specific implementations of outputting guide information for guiding the user to compose a composition, and are not limited to the above implementations. For example, in another implementation manner, guidance information for guiding the user to move the significant target area of interest to the target shooting area, where the target shooting area is determined by the preset composition method The preset composition method is the composition method used when shooting the instruction image. The technical solutions provided by the embodiments of the present application may include the following beneficial effects: through a convolutional neural network, a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output is used for Guidance information that guides the user to composition. Since this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since the scheme can be applied to various grayscale situations, the scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
本申请一种可选的实施例中,在输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息之后,本申请实施例所提供的方法还可以包括:在指示主体区域移动至目标拍摄区域之后,还可以包括:In an optional embodiment of the present application, after outputting guide information for guiding the user to move the subject area to the target shooting area, the method provided in the embodiment of the present application may further include: instructing the subject area to move to After the target shooting area, it can also include:
当检测到所述主体区域是否移动至所述目标拍摄区域时,向用户提示构图成功。具体地,如图3(e)所示,可以在拍摄界面中显示文字信息,如:“构图成功,可以拍照”。或者可以发出语音信息提示,等等。When it is detected whether the subject area moves to the target shooting area, the user is notified that the composition is successful. Specifically, as shown in FIG. 3 (e), text information can be displayed in the shooting interface, such as: "the composition is successful, and the picture can be taken." Or you can issue a voice message prompt, and so on.
如此,用户可以直观地知道构图成功,并在构图成功时进行拍照,可以使拍摄场景中的目标处于最佳拍摄位置时进行拍摄,使得无需用户具备专业的摄影知识,引导用户进行构图,进而提高拍摄的图像的质量。In this way, the user can intuitively know that the composition is successful, and take a picture when the composition is successful, so that the target in the shooting scene can be taken when it is in the best shooting position, so that the user is not required to have professional photography knowledge to guide the user to compose the composition, thereby improving The quality of the captured image.
通过取景画面中的显著性目标区域引导用户进行构图,如用户移动手机,使得从显著性目标区域中选择的主体区域位于目标拍摄区域,指示用户拍摄出质量更高的图像。The saliency target area in the framing screen guides the user to compose a composition, for example, the user moves the mobile phone, so that the subject area selected from the saliency target area is located in the target shooting area, instructing the user to take a higher quality image.
在上述实施例的基础上,本申请实施例还可以包括:训练卷积神经网络的步骤,具体地,如图4所示,可以包括:Based on the above embodiments, the embodiments of the present application may further include: a step of training a convolutional neural network. Specifically, as shown in FIG. 4, it may include:
S41,获取多个样本图像。S41. Acquire multiple sample images.
为了提高训练的准确性,电子设备获取多个样本图像,如10000个、20000个等等。In order to improve the accuracy of training, the electronic device acquires multiple sample images, such as 10,000, 20,000, etc.
一种可实现方式中,多个样本图像包括不同类别的样本图像,不同类别的样本图像包括不同类别目标。In a possible implementation manner, the plurality of sample images include sample images of different categories, and the sample images of different categories include targets of different categories.
具体地,可以预先构建数据集,然后从数据集中获取样本图像。具体地,可以获取自然场景对应的图像,如一种可实现方式中选择7种目标类别进行数据集的构建,如person(人),cat(猫),dog,(狗),horse(马),sheep(羊),cow(牛),bird(鸟)。且为了后续验证训练的卷积神经网络,可以将样本图像中的一部分作为训练集、一部分作为测试集,如数据集中训练数据总量为1328415,另外有1904数据作为测试集验证卷积神经网络的效果。Specifically, a data set may be constructed in advance, and then sample images are obtained from the data set. Specifically, images corresponding to natural scenes can be obtained, such as selecting 7 target categories in a realizable way to construct a data set, such as person (person), cat (cat), dog, (dog), horse (horse), sheep (sheep), cow (cow), bird (bird). And for subsequent verification of the trained convolutional neural network, part of the sample image can be used as the training set and part of the test set. For example, the total training data in the data set is 1328415, and 1904 data is used as the test set to verify the convolutional neural network. effect.
其中,不同类别之间,数据的数量有一定的差别,如果按照传统方法随机从数据集中选取图像进行训练,会出现样本不均衡问题,导致模型训练不准确。因此需要针对数据不同类之间存在的不均衡问题,设计合适的采样权 重。鉴于此,本申请实施例提出如下方式来确定不同类别图像的采样权重,主要步骤如下:Among them, there is a certain difference in the amount of data between different categories. If the images are randomly selected from the data set for training according to the traditional method, sample imbalance will occur, resulting in inaccurate model training. Therefore, it is necessary to design appropriate sampling weights for the imbalance between different types of data. In view of this, the embodiments of the present application propose the following ways to determine the sampling weights of images of different categories, the main steps are as follows:
(1)、统计各个类别图像总数k_i,及所有图像的总数K;(1). Count the total number of images in each category k_i, and the total number of all images K;
(2)、确定各个类别采样权重为K/k_i。(2) Determine the sampling weight of each category as K / k_i.
数量越多的图像权重越小,数量越少的图像权重越大,根据确定的不同类别图像对应的采样权重,从数据集中获取样本图像,以根据样本图像进行训练得到用于检测图像显著性目标区域的卷积神经网络,如此能够保证在网络模型训练时一个batch(批处理)中各个类别图像均衡,防止模型训练出现偏差。The larger the number of images, the smaller the weight, and the smaller the number of images, the greater the weight. According to the determined sampling weights corresponding to different categories of images, obtain sample images from the data set to train based on the sample images to obtain image saliency targets The regional convolutional neural network can ensure that the images of each category in a batch (batch processing) are balanced during network model training, and prevent deviations in model training.
S42,确定每个样本图像分别对应的标记图像;S42: Determine the label images corresponding to each sample image respectively;
S43,以每个样本图像作为输入内容,以每个样本图像对应的标记图像作为监督内容,对所述预设卷积神经网络进行训练,得到训练好的卷积神经网络。S43, taking each sample image as input content and using the labeled image corresponding to each sample image as supervising content, training the preset convolutional neural network to obtain a trained convolutional neural network.
具体地,可以通过人工标记的方式标记出各个样本图像对应的显著性目标区域;或者可以通过目标特征的检测标注出各个样本图像对应的显著性目标区域,等等。Specifically, the saliency target area corresponding to each sample image can be marked by manual marking; or the saliency target area corresponding to each sample image can be marked through detection of target characteristics, and so on.
具体地,预设卷积神经可以包括待测参数,将样本图像输入至预设卷积神经网络,调整待测参数,以使预设卷积神经网络的输出无限逼近于样本图像对应的标记图像所标记的区域,当预设卷积神经网络的输出与样本图像对应的标记图像所标记的图像之间的差异收敛时,确定待测参数,得到的包括该确定待测参数的预设卷积神经网络即为训练得到的卷积神经网络。其中,待测参数可以包括:批尺寸,学习速率,和/或迭代次数,等等。Specifically, the preset convolutional nerve may include a parameter to be tested, input the sample image to the preset convolutional neural network, and adjust the parameter to be tested, so that the output of the preset convolutional neural network is infinitely close to the labeled image corresponding to the sample image The marked area, when the difference between the output of the preset convolutional neural network and the image marked by the marked image corresponding to the sample image converges, the parameter to be measured is determined, and the obtained preset convolution including the determined parameter to be measured is obtained The neural network is the convolutional neural network trained. The parameters to be tested may include: batch size, learning rate, and / or number of iterations, and so on.
本申请一种可选的实施例中,预设卷积神经网络可以包括:基础模块、特征模块、定位模块以及分类模块。In an optional embodiment of the present application, the preset convolutional neural network may include: a basic module, a feature module, a positioning module, and a classification module.
如图5所示。基础模块可以由5层卷积层组成,如Conv2d(3->8),Conv2d(8->16),Conv2d(16->32),Conv2d(32->64),Conv2d(64->128),其中,Conv2d(3->8)理解为将3通道的RGB格式图像转换为8通道的图像,其他卷积层含义类似;特征模块可以由3层卷积层组成,如Conv2d(128->256),Conv2d(256->128),Conv2d(128->256);定位模块可以由一层卷积层组成,如Conv2d(*>8);分类模块可以由一层卷积层组成,如Conv2d(*>4)。As shown in Figure 5. The basic module can be composed of 5 convolutional layers, such as Conv2d (3-> 8), Conv2d (8-> 16), Conv2d (16-> 32), Conv2d (32-> 64), Conv2d (64-> 128) ), Where Conv2d (3-> 8) is understood to convert a 3-channel RGB format image into an 8-channel image, and the meaning of other convolutional layers is similar; the feature module can be composed of 3 convolutional layers, such as Conv2d (128- > 256), Conv2d (256-> 128), Conv2d (128-> 256); the positioning module can be composed of a convolutional layer, such as Conv2d (*> 8); the classification module can be composed of a convolutional layer, Such as Conv2d (*> 4).
其中,基础模块也可以称之为Base Model,特征模块也可以称之为Extra Model,定位模块也可以称之为Location layer,分类模块也可以称之为 Confidence layer。Among them, the basic module can also be called Base Model, the feature module can also be called Extra Model, the positioning module can also be called Location Layer, and the classification module can also be called Confidence Layer.
Base Model主要用于对图像进行从底层到高层的特征处理,用于为Extra Model提供特征。Base is mainly used to process the features from the bottom to the top of the image, and is used to provide features for Extra Model.
Extra Model主要来提取不同尺度的特征图,主要是为了图像中不同大小的目标而设计,特征图的每一个像素位置映射到原图上有对应大小的bounding box,这样可以为Location layer和Confidence layer提供不同大小目标的特征信息。Extra Model mainly extracts feature maps of different scales, mainly designed for targets of different sizes in the image. Each pixel position of the feature map is mapped to the corresponding size of the bounding box on the original image, which can be Location and Layer Provide characteristic information of different size targets.
Location layer用来针对特征图的每一个像素点,回归出原图上目标的bounding box(边框)坐标。Location layer is used to return the bounding box coordinates of the target on the original image for each pixel of the feature map.
Confidence layer用来针对特征图的每一个像素点,计算出原图上该bounding box的目标类别。Confidence layer is used to calculate the target category of the bounding box on the original image for each pixel of the feature map.
为了提高卷积神经网络训练的准确性,本申请实施例一种可选的实现方式中,可以对训练的网络进行优化和评估。In order to improve the accuracy of convolutional neural network training, in an optional implementation manner of the embodiments of the present application, the trained network may be optimized and evaluated.
预设卷积神经网络构建好后,需要在训练数据集上进行训练,预设卷积神经网络训练依赖损失函数及优化器,最后预设卷积神经网络的评价需要设置合适的评价指标。After the preset convolutional neural network is constructed, it needs to be trained on the training data set. The preset convolutional neural network training depends on the loss function and the optimizer. Finally, the evaluation of the preset convolutional neural network needs to set an appropriate evaluation index.
其中,损失函数,在卷积神经网络的优化过程中使用多任务损失函数,针对分类任务采用交叉熵损失函数,针对bounding box采用L1距离损失函数。Among them, the loss function uses the multi-task loss function in the optimization process of the convolutional neural network, uses the cross-entropy loss function for the classification task, and uses the L1 distance loss function for the bounding box.
优化器,网络优化采用随机梯度下降方法来更新卷积神经网络的参数。Optimizer, network optimization uses stochastic gradient descent method to update the parameters of convolutional neural network.
评价指标,本申请实施例中可以采用mean Average Precision(平均精度均值)来针对检测结果进行评价。For evaluation indicators, mean, accuracy, and average precision can be used in the embodiments of the present application to evaluate the detection result.
可选地,本申请一种可选的实施例中,所述获取多个样本图像,包括:Optionally, in an optional embodiment of the present application, the acquiring multiple sample images includes:
获取多个原始图像;Acquire multiple original images;
对所述多个原始图像分别添加干扰,得到干扰后的原始图像;Adding interference to the plurality of original images respectively to obtain the original image after interference;
将干扰后的原始图像确定为样本图像。The original image after interference is determined as the sample image.
其中,干扰可以包括水平翻转、垂直翻转、随机裁剪、图像像素扰动、光照处理、遮挡、低对比度处理等等。Among them, the interference may include horizontal flip, vertical flip, random cropping, image pixel disturbance, lighting processing, occlusion, low contrast processing, and so on.
通过在训练网络的过程中,添加不同类型的干扰,实现数据增强,使得训练好的卷积神经网络受外界干扰因素的影响较小,针对不同场景,如遮挡、形变、光照等外界干扰时更加鲁棒,即针对不同场景具有更好地鲁棒性,进一步提高确定的显著性目标区域的准确性。In the process of training the network, add different types of interference to achieve data enhancement, so that the trained convolutional neural network is less affected by external interference factors, especially for different scenarios, such as occlusion, deformation, lighting and other external interference Robust, that is, it has better robustness for different scenes, and further improves the accuracy of the determined saliency target area.
图6是根据一示例性实施例示出的一种拍摄引导装置框图。参照图6,该 装置可以包括第一获取模块601、第一确定模块602和指示模块603;Fig. 6 is a block diagram of a shooting guidance device according to an exemplary embodiment. Referring to FIG. 6, the apparatus may include a first acquisition module 601, a first determination module 602, and an instruction module 603;
该第一获取模块601,被配置为获取包含拍摄场景中目标的指示图像;The first obtaining module 601 is configured to obtain an indication image containing the target in the shooting scene;
该第一确定模块602,被配置为基于预先训练的卷积神经网络,确定指示图像中的显著性目标区域;其中,所述卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像中的显著性目标区域;The first determination module 602 is configured to determine a saliency target area in the indicated image based on the pre-trained convolutional neural network; wherein, the convolutional neural network is based on multiple sample images and each sample image corresponds to From the training of labeled images, the marked image corresponding to each sample image is marked with the saliency target area in the sample image;
该指示模块603,被配置为基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。The instruction module 603 is configured to output guide information for guiding the user to compose a composition based on the determined saliency target area.
本申请的实施例提供的技术方案可以包括以下有益效果:通过卷积神经网络,确定包括拍摄场景中目标的指示图像的显著性目标区域,进而,基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。由于本方案无需考虑各个像素点的灰度差异,因此,针对各种灰度差异情况,本方案均能够保证显著性目标区域确定的准确性,进而保证拍摄引导的准确性。而由于本方案可以应用于各种灰度情况,因此,本方案对于不同情况的鲁棒性较高。且无需根据人为定义的确定特征提取显著性目标区域,使得应用场景更加广泛。The technical solutions provided by the embodiments of the present application may include the following beneficial effects: through a convolutional neural network, a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output Guidance information that guides the user to composition. Since this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since the scheme can be applied to various grayscale situations, the scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
可选地,所述指示模块603,具体被配置为:从所确定的显著性目标区域中,选择主体区域;输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息;其中,所述目标拍摄区域是基于预设构图方式确定的,所述预设构图方式为在拍摄所述指示图像时所利用的构图方式。Optionally, the instruction module 603 is specifically configured to: select a subject area from the determined salient target areas; output guide information for guiding the user to move the subject area to the target shooting area; wherein, The target shooting area is determined based on a preset composition method, and the preset composition method is a composition method used when shooting the instruction image.
可选的,该装置还包括:Optionally, the device also includes:
提示模块,被配置为在所述指示模块输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息之后,当检测到所述主体区域是否移动至所述目标拍摄区域时,向用户提示构图成功。The prompting module is configured to, after the instruction module outputs guidance information for guiding the user to move the subject area to the target shooting area, when it is detected whether the subject area moves to the target shooting area, to the user Prompt composition success.
可选的,该装置还包括:Optionally, the device also includes:
第二获取模块,被配置为获取多个样本图像;The second acquisition module is configured to acquire multiple sample images;
第二确定模块,被配置为确定每个样本图像分别对应的标记图像;The second determining module is configured to determine the mark image corresponding to each sample image respectively;
训练模块,被配置为以每个样本图像作为输入内容,以每个样本图像对应的标记图像作为监督内容,对所述预设卷积神经网络进行训练,得到训练好的卷积神经网络。The training module is configured to take each sample image as input content and the labeled image corresponding to each sample image as supervising content to train the preset convolutional neural network to obtain a trained convolutional neural network.
可选的,所述第二获取模块,具体被配置为:获取多个原始图像;对所述多个原始图像分别添加干扰,得到多个干扰后的原始图像;将多个干扰后 的原始图像均确定为样本图像。Optionally, the second acquisition module is specifically configured to: acquire multiple original images; add interference to the multiple original images respectively to obtain multiple original images after interference; combine the multiple original images after interference Both are determined as sample images.
可选的,多个样本图像包括不同类别的样本图像,不同类别的样本图像包括不同类别目标。Optionally, the plurality of sample images include sample images of different categories, and the sample images of different categories include targets of different categories.
可选的,预设卷积神经网络,包括:基础模块、特征模块、定位模块以及分类模块。Optionally, the preset convolutional neural network includes: a basic module, a feature module, a positioning module, and a classification module.
关于上述实施例中的装置,其中各个模块执行操作的具体方式已经在有关该方法的实施例中进行了详细描述,此处将不做详细阐述说明。Regarding the device in the above embodiment, the specific manner in which each module performs operations has been described in detail in the embodiment related to the method, and will not be elaborated here.
图7是根据一示例性实施例示出的一种用于拍摄引导的移动终端700的框图。例如,移动终端700可以是移动电话,计算机,数字广播终端,消息收发设备,游戏控制台,平板设备,医疗设备,健身设备,个人数字助理等。Fig. 7 is a block diagram of a mobile terminal 700 for shooting guidance according to an exemplary embodiment. For example, the mobile terminal 700 may be a mobile phone, a computer, a digital broadcasting terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
参照图7,移动终端700可以包括以下一个或多个组件:处理组件702,存储器704,电源组件706,多媒体组件708,音频组件710,输入/输出(I/O)的接口712,传感器组件714,以及通信组件716。7, the mobile terminal 700 may include one or more of the following components: a processing component 702, a memory 704, a power component 706, a multimedia component 708, an audio component 710, an input / output (I / O) interface 712, and a sensor component 714 , 和 通信 组 716.
处理组件702通常控制移动终端700的整体操作,诸如与显示,电话呼叫,数据通信,相机操作和记录操作相关联的操作。处理组件702可以包括一个或多个处理器720来执行指令,以完成上述的拍摄引导方法的全部或部分步骤。此外,处理组件702可以包括一个或多个模块,便于处理组件702和其他组件之间的交互。例如,处理组件702可以包括多媒体模块,以方便多媒体组件708和处理组件702之间的交互。The processing component 702 generally controls the overall operations of the mobile terminal 700, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing component 702 may include one or more processors 720 to execute instructions to complete all or part of the steps of the above-mentioned shooting guidance method. In addition, the processing component 702 may include one or more modules to facilitate interaction between the processing component 702 and other components. For example, the processing component 702 may include a multimedia module to facilitate interaction between the multimedia component 708 and the processing component 702.
存储器704被配置为存储各种类型的数据以支持在移动终端700的操作。这些数据的示例包括用于在移动终端700上操作的任何应用程序或方法的指令,联系人数据,电话簿数据,消息,图片,视频等。存储器704可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。The memory 704 is configured to store various types of data to support operations at the mobile terminal 700. Examples of these data include instructions for any application or method operating on the mobile terminal 700, contact data, phone book data, messages, pictures, videos, and so on. The memory 704 may be implemented by any type of volatile or nonvolatile storage device or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable and removable Programmable read only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory, magnetic disk or optical disk.
电源组件706为移动终端700的各种组件提供电力。电源组件706可以包括电源管理系统,一个或多个电源,及其他与为移动终端700生成、管理和分配电力相关联的组件。The power supply component 706 provides power to various components of the mobile terminal 700. The power supply component 706 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the mobile terminal 700.
多媒体组件708包括在移动终端700和用户之间的提供一个输出接口的屏幕。在一些实施例中,屏幕可以包括液晶显示器(LCD)和触摸面板(TP)。如果屏幕包括触摸面板,屏幕可以被实现为触摸屏,以接收来自用户的输入 信号。触摸面板包括一个或多个触摸传感器以感测触摸、滑动和触摸面板上的手势。所述触摸传感器可以不仅感测触摸或滑动动作的边界,而且还检测与所述触摸或滑动操作相关的持续时间和压力。在一些实施例中,多媒体组件708包括一个前置摄像头和/或后置摄像头。当移动终端700处于操作模式,如拍摄模式或视频模式时,前置摄像头和/或后置摄像头可以接收外部的多媒体数据。每个前置摄像头和后置摄像头可以是一个固定的光学透镜系统或具有焦距和光学变焦能力。The multimedia component 708 includes a screen between the mobile terminal 700 and the user that provides an output interface. In some embodiments, the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user. The touch panel includes one or more touch sensors to sense touch, swipe, and gestures on the touch panel. The touch sensor may not only sense the boundary of the touch or sliding action, but also detect the duration and pressure related to the touch or sliding operation. In some embodiments, the multimedia component 708 includes a front camera and / or a rear camera. When the mobile terminal 700 is in an operation mode, such as a shooting mode or a video mode, the front camera and / or the rear camera may receive external multimedia data. Each front camera and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
音频组件710被配置为输出和/或输入音频信号。例如,音频组件710包括一个麦克风(MIC),当移动终端700处于操作模式,如呼叫模式、记录模式和语音识别模式时,麦克风被配置为接收外部音频信号。所接收的音频信号可以被进一步存储在存储器704或经由通信组件716发送。在一些实施例中,音频组件710还包括一个扬声器,用于输出音频信号。The audio component 710 is configured to output and / or input audio signals. For example, the audio component 710 includes a microphone (MIC), and when the mobile terminal 700 is in an operation mode, such as a call mode, a recording mode, and a voice recognition mode, the microphone is configured to receive an external audio signal. The received audio signal may be further stored in the memory 704 or sent via the communication component 716. In some embodiments, the audio component 710 further includes a speaker for outputting audio signals.
I/O接口712为处理组件702和外围接口模块之间提供接口,上述外围接口模块可以是键盘,点击轮,按钮等。这些按钮可包括但不限于:主页按钮、音量按钮、启动按钮和锁定按钮。The I / O interface 712 provides an interface between the processing component 702 and a peripheral interface module. The peripheral interface module may be a keyboard, a click wheel, or a button. These buttons may include, but are not limited to: home button, volume button, start button, and lock button.
传感器组件714包括一个或多个传感器,用于为移动终端700提供各个方面的状态评估。例如,传感器组件714可以检测到移动终端700的打开/关闭状态,组件的相对定位,例如所述组件为移动终端700的显示器和小键盘,传感器组件714还可以检测移动终端700或移动终端700一个组件的位置改变,用户与移动终端700接触的存在或不存在,移动终端700方位或加速/减速和移动终端700的温度变化。传感器组件714可以包括接近传感器,被配置用来在没有任何的物理接触时检测附近物体的存在。传感器组件714还可以包括光传感器,如CMOS或CCD图像传感器,用于在成像应用中使用。在一些实施例中,该传感器组件714还可以包括加速度传感器,陀螺仪传感器,磁传感器,压力传感器或温度传感器。The sensor component 714 includes one or more sensors for providing the mobile terminal 700 with status evaluation in various aspects. For example, the sensor component 714 can detect the on / off state of the mobile terminal 700, and the relative positioning of the components, such as the display and the keypad of the mobile terminal 700, the sensor component 714 can also detect the mobile terminal 700 or the mobile terminal 700 The position of the component changes, the presence or absence of user contact with the mobile terminal 700, the orientation or acceleration / deceleration of the mobile terminal 700, and the temperature change of the mobile terminal 700. The sensor assembly 714 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact. The sensor component 714 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor component 714 may further include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
通信组件716被配置为便于移动终端700和其他设备之间有线或无线方式的通信。移动终端700可以接入基于通信标准的无线网络,如WiFi,运营商网络(如2G、3G、4G或5G),或它们的组合。在一个示例性实施例中,通信组件716经由广播信道接收来自外部广播管理系统的广播信号或广播相关信息。在一个示例性实施例中,所述通信组件716还包括近场通信(NFC)模块,以促进短程通信。例如,在NFC模块可基于射频识别(RFID)技术,红外数据协会(IrDA)技术,超宽带(UWB)技术,蓝牙(BT)技术和其他 技术来实现。The communication component 716 is configured to facilitate wired or wireless communication between the mobile terminal 700 and other devices. The mobile terminal 700 may access a wireless network based on a communication standard, such as WiFi, an operator network (such as 2G, 3G, 4G, or 5G), or a combination thereof. In an exemplary embodiment, the communication component 716 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 716 further includes a near field communication (NFC) module to facilitate short-range communication. For example, the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
在示例性实施例中,移动终端700可以被一个或多个应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,用于执行上述方法。In an exemplary embodiment, the mobile terminal 700 may be used by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field Programming gate array (FPGA), controller, microcontroller, microprocessor or other electronic components are used to implement the above method.
在示例性实施例中,还提供了一种包括指令的非临时性计算机可读存储介质,例如包括指令的存储器704,上述指令可由移动终端700的处理器720执行以完成上述方法。例如,所述非临时性计算机可读存储介质可以是ROM、随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。In an exemplary embodiment, a non-transitory computer-readable storage medium including instructions is also provided, for example, a memory 704 including instructions, which can be executed by the processor 720 of the mobile terminal 700 to complete the above method. For example, the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, or the like.
根据本申请实施例又一方面,提供一种计算机程序产品,当其在计算机上运行时,使得计算机能够执行上述拍摄引导方法。According to still another aspect of the embodiments of the present application, there is provided a computer program product which, when run on a computer, enables the computer to execute the above-mentioned shooting guidance method.
图8是根据一示例性实施例示出的一种用于拍摄引导的装置800的框图。例如,装置800可以被提供为一服务器。参照图8,装置800包括处理组件822,其进一步包括一个或多个处理器,以及由存储器832所代表的存储器资源,用于存储可由处理组件822的执行的指令,例如应用程序。存储器832中存储的应用程序可以包括一个或一个以上的每一个对应于一组指令的模块。此外,处理组件822被配置为执行指令,以执行上述拍摄引导方法的方法步骤。Fig. 8 is a block diagram of a device 800 for shooting guidance according to an exemplary embodiment. For example, the device 800 may be provided as a server. 8, the device 800 includes a processing component 822, which further includes one or more processors, and memory resources represented by the memory 832, for storing instructions executable by the processing component 822, such as application programs. The application programs stored in the memory 832 may include one or more modules each corresponding to a set of instructions. In addition, the processing component 822 is configured to execute instructions to perform the method steps of the above-mentioned shooting guidance method.
装置800还可以包括一个电源组件826被配置为执行装置800的电源管理,一个有线或无线网络接口850被配置为将装置800连接到网络,和一个输入输出(I/O)接口858。装置800可以操作基于存储在存储器832的操作系统,例如Windows ServerTM,Mac OS XTM,UnixTM,LinuxTM,FreeBSDTM或类似。The device 800 may also include a power component 826 configured to perform power management of the device 800, a wired or wireless network interface 850 configured to connect the device 800 to the network, and an input output (I / O) interface 858. The device 800 can operate based on an operating system stored in the memory 832, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or the like.
以上所述仅为本申请的较佳实施例而已,并不用以限制本申请,凡在本申请的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本申请保护的范围之内。The above are only the preferred embodiments of this application and are not intended to limit this application. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of this application should be included in this application Within the scope of protection.

Claims (22)

  1. 一种拍摄引导方法,包括:A shooting guide method, including:
    获取包含拍摄场景中目标的指示图像;Obtain an indication image containing the target in the shooting scene;
    基于预先训练的卷积神经网络,确定所述指示图像中的显著性目标区域;其中,所述卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像中的显著性目标区域;Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
    基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。Based on the determined saliency target area, guide information for guiding the user to compose a picture is output.
  2. 根据权利要求1所述的方法,所述基于所确定的显著性目标区域,输出用于指引用户构图的引导信息,包括:The method according to claim 1, the outputting guide information for guiding the user to compose a composition based on the determined saliency target area includes:
    从所确定的显著性目标区域中,选择主体区域;From the determined significant target area, select the subject area;
    输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息;其中,所述目标拍摄区域是基于预设构图方式确定的,所述预设构图方式为在拍摄所述指示图像时所利用的构图方式。Output guide information for guiding the user to move the subject area to the target shooting area; wherein, the target shooting area is determined based on a preset composition method, which is used when the instruction image is captured Use the composition method.
  3. 根据权利要求2所述的方法,在输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息之后,所述方法还包括:The method according to claim 2, after outputting guide information for guiding the user to move the subject area to the target shooting area, the method further comprises:
    当检测到所述主体区域是否移动至所述目标拍摄区域时,向用户提示构图成功。When it is detected whether the subject area moves to the target shooting area, the user is notified that the composition is successful.
  4. 根据权利要求1-3任一项所述的方法,训练所述卷积神经网络的步骤,包括:The method according to any one of claims 1-3, the step of training the convolutional neural network includes:
    获取多个样本图像;Acquire multiple sample images;
    确定每个样本图像分别对应的标记图像;Determine the corresponding labeled images for each sample image;
    以每个样本图像作为输入内容,以每个样本图像对应的标记图像作为监督内容,对所述预设卷积神经网络进行训练,得到训练好的卷积神经网络。Taking each sample image as input content and using the labeled image corresponding to each sample image as supervising content, the preset convolutional neural network is trained to obtain a trained convolutional neural network.
  5. 根据权利要求4所述的方法,所述获取多个样本图像,包括:The method according to claim 4, said acquiring a plurality of sample images, comprising:
    获取多个原始图像;Acquire multiple original images;
    对所述多个原始图像分别添加干扰,得到多个干扰后的原始图像;Adding interference to the multiple original images respectively to obtain multiple original images after interference;
    将多个干扰后的原始图像均确定为样本图像。The original images after multiple interferences are determined as sample images.
  6. 根据权利要求4所述的方法,多个样本图像包括不同类别的样本图像, 不同类别的样本图像包括不同类别目标。According to the method of claim 4, the plurality of sample images include different types of sample images, and the different types of sample images include different types of targets.
  7. 根据权利要求4所述的方法,所述预设卷积神经网络,包括:基础模块、特征模块、定位模块以及分类模块。According to the method of claim 4, the preset convolutional neural network includes: a basic module, a feature module, a positioning module, and a classification module.
  8. 一种拍摄引导装置,包括:A shooting guide device, including:
    第一获取模块,被配置为获取包含拍摄场景中目标的指示图像;The first acquisition module is configured to acquire the indication image containing the target in the shooting scene;
    第一确定模块,被配置为基于预先训练的卷积神经网络,确定所述指示图像中的显著性目标区域;其中,所述卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像中的显著性目标区域;The first determining module is configured to determine the saliency target area in the indication image based on the pre-trained convolutional neural network; wherein, the convolutional neural network is based on multiple sample images and each sample image corresponds to From the training of labeled images, the marked image corresponding to each sample image is marked with the saliency target area in the sample image;
    指示模块,被配置为基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。The instruction module is configured to output guide information for guiding the user to compose a composition based on the determined saliency target area.
  9. 根据权利要求8所述的装置,所述指示模块,具体被配置为:The apparatus according to claim 8, the indication module is specifically configured to:
    从所确定的显著性目标区域中,选择主体区域;From the determined significant target area, select the subject area;
    输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息;其中,所述目标拍摄区域是基于预设构图方式确定的,所述预设构图方式为在拍摄所述指示图像时所利用的构图方式。Output guide information for guiding the user to move the subject area to the target shooting area; wherein, the target shooting area is determined based on a preset composition method, which is used when the instruction image is captured Use the composition method.
  10. 根据权利要求9所述的装置,所述装置还包括:The device according to claim 9, further comprising:
    提示模块,被配置为在所述指示模块输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息之后,当检测到所述主体区域是否移动至所述目标拍摄区域时,向用户提示构图成功。The prompting module is configured to, after the instruction module outputs guidance information for guiding the user to move the subject area to the target shooting area, when it is detected whether the subject area moves to the target shooting area, to the user Prompt composition success.
  11. 根据权利要求8-10任一项所述的装置,所述装置还包括:第二获取模块,被配置为获取多个样本图像;The apparatus according to any one of claims 8-10, the apparatus further comprising: a second acquisition module configured to acquire a plurality of sample images;
    第二确定模块,被配置为确定每个样本图像分别对应的标记图像;The second determining module is configured to determine the mark image corresponding to each sample image respectively;
    训练模块,被配置为以每个样本图像作为输入内容,以每个样本图像对应的标记图像作为监督内容,对所述预设卷积神经网络进行训练,得到训练好的卷积神经网络。The training module is configured to take each sample image as input content and the labeled image corresponding to each sample image as supervising content to train the preset convolutional neural network to obtain a trained convolutional neural network.
  12. 根据权利要求11所述的装置,所述第二获取模块,具体被配置为:According to the apparatus of claim 11, the second acquisition module is specifically configured to:
    获取多个原始图像;Acquire multiple original images;
    对所述多个原始图像分别添加干扰,得到多个干扰后的原始图像;Adding interference to the multiple original images respectively to obtain multiple original images after interference;
    将多个干扰后的原始图像均确定为样本图像。The original images after multiple interferences are determined as sample images.
  13. 根据权利要求11所述的装置,多个样本图像包括不同类别的样本图像,不同类别的样本图像包括不同类别目标。According to the apparatus of claim 11, the plurality of sample images include different types of sample images, and the different types of sample images include different types of targets.
  14. 根据权利要求11所述的装置,所述预设卷积神经网络,包括:基础模块、特征模块、定位模块以及分类模块。The apparatus according to claim 11, the preset convolutional neural network includes: a basic module, a feature module, a positioning module, and a classification module.
  15. 一种移动终端,包括:A mobile terminal, including:
    处理器;processor;
    用于存储处理器可执行指令的存储器;Memory for storing processor executable instructions;
    其中,所述处理器被配置为:Wherein, the processor is configured to:
    获取包含拍摄场景中目标的指示图像;Obtain an indication image containing the target in the shooting scene;
    基于预先训练的卷积神经网络,确定所述指示图像中的显著性目标区域;其中,所述卷积神经网络是根据多个样本图像以及每个样本图像对应的标记图像训练得到的,每个样本图像对应的标记图像中标记有该样本图像中的显著性目标区域;Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
    基于所确定的显著性目标区域,输出用于指引用户构图的引导信息。Based on the determined saliency target area, guide information for guiding the user to compose a picture is output.
  16. 根据权利要求15所述的移动终端,所述处理器基于所确定的显著性目标区域,输出用于指引用户构图的引导信息,包括:The mobile terminal of claim 15, the processor outputting guide information for guiding the user to compose a composition based on the determined saliency target area, including:
    从所确定的显著性目标区域中,选择主体区域;From the determined significant target area, select the subject area;
    输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息;其中,所述目标拍摄区域是基于预设构图方式确定的,所述预设构图方式为在拍摄所述指示图像时所利用的构图方式。Output guide information for guiding the user to move the subject area to the target shooting area; wherein, the target shooting area is determined based on a preset composition method, which is used when the instruction image is captured Use the composition method.
  17. 根据权利要求16所述的移动终端,所述处理器还被配置为:在输出用于指引用户将所述主体区域移动至目标拍摄区域的引导信息之后,当检测到所述主体区域是否移动至所述目标拍摄区域时,向用户提示构图成功。The mobile terminal of claim 16, the processor is further configured to: after outputting guide information for guiding the user to move the subject area to the target shooting area, when it is detected whether the subject area moves to When the target shooting area is displayed, the user is informed that the composition is successful.
  18. 根据权利要求15-17任一项所述的移动终端,所述处理器还被配置为:The mobile terminal according to any one of claims 15-17, the processor is further configured to:
    获取多个样本图像;Acquire multiple sample images;
    确定每个样本图像分别对应的标记图像;Determine the corresponding labeled images for each sample image;
    以每个样本图像作为输入内容,以每个样本图像对应的标记图像作为监督内容,对所述预设卷积神经网络进行训练,得到训练好的卷积神经网络。Taking each sample image as input content and using the labeled image corresponding to each sample image as supervising content, the preset convolutional neural network is trained to obtain a trained convolutional neural network.
  19. 根据权利要求18所述的移动终端,所述处理器获取多个样本图像,包括:The mobile terminal of claim 18, the processor acquiring a plurality of sample images, including:
    获取多个原始图像;Acquire multiple original images;
    对所述多个原始图像分别添加干扰,得到多个干扰后的原始图像;Adding interference to the multiple original images respectively to obtain multiple original images after interference;
    将多个干扰后的原始图像均确定为样本图像。The original images after multiple interferences are determined as sample images.
  20. 根据权利要求18所述的移动终端,多个样本图像包括不同类别的样本图像,不同类别的样本图像包括不同类别目标。According to the mobile terminal of claim 18, the plurality of sample images include different types of sample images, and the different types of sample images include different types of targets.
  21. 根据权利要求18所述的移动终端,所述预设卷积神经网络,包括:基础模块、特征模块、定位模块以及分类模块。The mobile terminal according to claim 18, the preset convolutional neural network includes: a basic module, a feature module, a positioning module, and a classification module.
  22. 一种非临时性计算机可读存储介质,当所述存储介质中的指令由移动终端的处理器执行时,使得移动终端能够执行权利要求1-7任一项所述的拍摄引导方法。A non-transitory computer-readable storage medium, when instructions in the storage medium are executed by a processor of a mobile terminal, enable the mobile terminal to perform the shooting guidance method of any one of claims 1-7.
PCT/CN2019/112566 2018-11-05 2019-10-22 Photography guiding method and apparatus, mobile terminal and storage medium WO2020093866A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811307419.6A CN109040605A (en) 2018-11-05 2018-11-05 Shoot bootstrap technique, device and mobile terminal and storage medium
CN201811307419.6 2018-11-05

Publications (1)

Publication Number Publication Date
WO2020093866A1 true WO2020093866A1 (en) 2020-05-14

Family

ID=64614404

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/112566 WO2020093866A1 (en) 2018-11-05 2019-10-22 Photography guiding method and apparatus, mobile terminal and storage medium

Country Status (2)

Country Link
CN (1) CN109040605A (en)
WO (1) WO2020093866A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860242A (en) * 2020-07-07 2020-10-30 北京海益同展信息科技有限公司 Robot inspection method and device and computer readable medium
CN112766285A (en) * 2021-01-26 2021-05-07 北京有竹居网络技术有限公司 Image sample generation method and device and electronic equipment

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109040605A (en) * 2018-11-05 2018-12-18 北京达佳互联信息技术有限公司 Shoot bootstrap technique, device and mobile terminal and storage medium
CN111246098B (en) * 2020-01-19 2022-02-22 深圳市人工智能与机器人研究院 Robot photographing method and device, computer equipment and storage medium
CN111327833B (en) * 2020-03-31 2021-06-01 厦门美图之家科技有限公司 Auxiliary shooting method and device, electronic equipment and readable storage medium
CN111464743A (en) * 2020-04-09 2020-07-28 上海城诗信息科技有限公司 Photographic composition matching method and system
CN111461248A (en) * 2020-04-09 2020-07-28 上海城诗信息科技有限公司 Photographic composition line matching method, device, equipment and storage medium
CN112581446A (en) * 2020-12-15 2021-03-30 影石创新科技股份有限公司 Method, device and equipment for detecting salient object of image and storage medium
WO2022183443A1 (en) * 2021-03-04 2022-09-09 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method of suggesting shooting position for electronic device and electronic device
CN113329175A (en) * 2021-05-21 2021-08-31 浙江大华技术股份有限公司 Snapshot method, device, electronic device and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104717413A (en) * 2013-12-12 2015-06-17 北京三星通信技术研究有限公司 Shooting assistance method and equipment
CN107247930A (en) * 2017-05-26 2017-10-13 西安电子科技大学 SAR image object detection method based on CNN and Selective Attention Mechanism
CN107423747A (en) * 2017-04-13 2017-12-01 中国人民解放军国防科学技术大学 A kind of conspicuousness object detection method based on depth convolutional network
CN107909109A (en) * 2017-11-17 2018-04-13 西安电子科技大学 SAR image sorting technique based on conspicuousness and multiple dimensioned depth network model
US20180314943A1 (en) * 2017-04-27 2018-11-01 Jianming Liang Systems, methods, and/or media, for selecting candidates for annotation for use in training a classifier
CN109040605A (en) * 2018-11-05 2018-12-18 北京达佳互联信息技术有限公司 Shoot bootstrap technique, device and mobile terminal and storage medium

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105931255A (en) * 2016-05-18 2016-09-07 天津工业大学 Method for locating target in image based on obviousness and deep convolutional neural network
US10579860B2 (en) * 2016-06-06 2020-03-03 Samsung Electronics Co., Ltd. Learning model for salient facial region detection

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104717413A (en) * 2013-12-12 2015-06-17 北京三星通信技术研究有限公司 Shooting assistance method and equipment
CN107423747A (en) * 2017-04-13 2017-12-01 中国人民解放军国防科学技术大学 A kind of conspicuousness object detection method based on depth convolutional network
US20180314943A1 (en) * 2017-04-27 2018-11-01 Jianming Liang Systems, methods, and/or media, for selecting candidates for annotation for use in training a classifier
CN107247930A (en) * 2017-05-26 2017-10-13 西安电子科技大学 SAR image object detection method based on CNN and Selective Attention Mechanism
CN107909109A (en) * 2017-11-17 2018-04-13 西安电子科技大学 SAR image sorting technique based on conspicuousness and multiple dimensioned depth network model
CN109040605A (en) * 2018-11-05 2018-12-18 北京达佳互联信息技术有限公司 Shoot bootstrap technique, device and mobile terminal and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860242A (en) * 2020-07-07 2020-10-30 北京海益同展信息科技有限公司 Robot inspection method and device and computer readable medium
CN112766285A (en) * 2021-01-26 2021-05-07 北京有竹居网络技术有限公司 Image sample generation method and device and electronic equipment
CN112766285B (en) * 2021-01-26 2024-03-19 北京有竹居网络技术有限公司 Image sample generation method and device and electronic equipment

Also Published As

Publication number Publication date
CN109040605A (en) 2018-12-18

Similar Documents

Publication Publication Date Title
WO2020093866A1 (en) Photography guiding method and apparatus, mobile terminal and storage medium
US10534972B2 (en) Image processing method, device and medium
KR101649596B1 (en) Method, apparatus, program, and recording medium for skin color adjustment
RU2659746C2 (en) Method and device for image processing
CN104270565B (en) Image capturing method, device and equipment
RU2577188C1 (en) Method, apparatus and device for image segmentation
US20170064181A1 (en) Method and apparatus for controlling photography of unmanned aerial vehicle
EP3226204A1 (en) Method and apparatus for intelligently capturing image
US10110800B2 (en) Method and apparatus for setting image capturing parameters
CN104918107B (en) The identification processing method and device of video file
US10248855B2 (en) Method and apparatus for identifying gesture
CN105959587B (en) Shutter speed acquisition methods and device
CN109889730B (en) Prompting method and device for adjusting shooting angle and electronic equipment
CN105758319B (en) The method and apparatus for measuring target object height by mobile terminal
CN106408603A (en) Camera method and device
CN111182212B (en) Image processing method, image processing apparatus, storage medium, and electronic device
CN110751659B (en) Image segmentation method and device, terminal and storage medium
CN106664366A (en) Image processing device, image capturing apparatus, image processing method, and program
CN109670458A (en) A kind of licence plate recognition method and device
US20180150722A1 (en) Photo synthesizing method, device, and medium
CN103731599A (en) Photographing method and camera
WO2020119254A1 (en) Method and device for filter recommendation, electronic equipment, and storage medium
US11574415B2 (en) Method and apparatus for determining an icon position
CN108717542B (en) Method and device for recognizing character area and computer readable storage medium
WO2018036010A1 (en) Method and apparatus for controlling balancing vehicle

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19882219

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19882219

Country of ref document: EP

Kind code of ref document: A1