WO2020093866A1 - Procédé et appareil d'aide à la photographie, terminal mobile et support de stockage - Google Patents

Procédé et appareil d'aide à la photographie, terminal mobile et support de stockage Download PDF

Info

Publication number
WO2020093866A1
WO2020093866A1 PCT/CN2019/112566 CN2019112566W WO2020093866A1 WO 2020093866 A1 WO2020093866 A1 WO 2020093866A1 CN 2019112566 W CN2019112566 W CN 2019112566W WO 2020093866 A1 WO2020093866 A1 WO 2020093866A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
area
sample
target
neural network
Prior art date
Application number
PCT/CN2019/112566
Other languages
English (en)
Chinese (zh)
Inventor
张渊
郑文
Original Assignee
北京达佳互联信息技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京达佳互联信息技术有限公司 filed Critical 北京达佳互联信息技术有限公司
Publication of WO2020093866A1 publication Critical patent/WO2020093866A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces

Definitions

  • the present application relates to the technical field of image processing, and in particular to a shooting guidance method, device, mobile terminal, and storage medium.
  • composition is an important factor in the quality of the captured images.
  • the salient target area is an area containing a saliency target.
  • the saliency target is a target that human visually pays more attention to, such as: cars driving on roads, behaviors on snow, flowers in lush green leaves, and so on.
  • the saliency target area of the image is determined based on the gray value of each pixel in the image, that is, the saliency target area of the image is determined based on the gray scale feature of the image.
  • the inventor found that if the gray value of each pixel of the image is not significantly different, the accuracy of the determined saliency target area is low, which in turn makes the accuracy of shooting guidance based on the saliency target area low.
  • the present application provides a shooting guidance method, device, mobile terminal, and storage medium.
  • a shooting guidance method including:
  • the convolutional neural network Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled images corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
  • guide information for guiding the user to compose a picture is output.
  • a shooting guidance device including:
  • the first acquisition module is configured to acquire the indication image containing the target in the shooting scene
  • the first determining module is configured to determine the saliency target area in the indication image based on the pre-trained convolutional neural network; wherein, the convolutional neural network is based on multiple sample images and each sample image corresponds to From the training of labeled images, the marked image corresponding to each sample image is marked with the saliency target area in the sample image;
  • the instruction module is configured to output guide information for guiding the user to compose a composition based on the determined saliency target area.
  • a mobile terminal including:
  • Memory for storing processor executable instructions
  • the processor is configured to:
  • the convolutional neural network Based on the pre-trained convolutional neural network, determine the saliency target area in the indicator image; wherein, the convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, each The marked image corresponding to the sample image is marked with the saliency target area in the sample image;
  • guide information for guiding the user to compose a picture is output.
  • a non-transitory computer-readable storage medium When instructions in the storage medium are executed by a processor of a mobile terminal, the mobile terminal can perform a shooting guidance method.
  • a computer program product which, when run on a computer, enables the computer to execute a shooting guidance method.
  • a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output is used for Guidance information that guides the user to composition. Since this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since this scheme can be applied to various grayscale situations, this scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
  • Fig. 1 is a flow chart showing a method for shooting guidance according to an exemplary embodiment.
  • Fig. 2 (a) is a schematic diagram showing a detection result of a salient target area according to an exemplary embodiment.
  • Fig. 2 (b) is another schematic diagram showing the detection result of the salient target area according to an exemplary embodiment.
  • Fig. 2 (c) is another schematic diagram showing the detection result of the salient target area according to an exemplary embodiment.
  • Fig. 3 (a) is a schematic diagram of a shooting interface according to an exemplary embodiment.
  • Fig. 3 (b) is a schematic diagram of guiding the main body region to the target shooting region according to an exemplary embodiment.
  • Fig. 3 (c) is another schematic diagram of guiding the body region to the target shooting region according to an exemplary embodiment.
  • Fig. 3 (d) is another schematic diagram of guiding the body area to the target shooting area according to an exemplary embodiment.
  • Fig. 3 (e) is another schematic diagram of guiding the body area to the target shooting area according to an exemplary embodiment.
  • Fig. 4 is a flowchart illustrating training a convolutional neural network according to an exemplary embodiment.
  • Fig. 5 is a schematic structural diagram of a preset convolutional neural network according to an exemplary embodiment.
  • Fig. 6 is a block diagram of a shooting guide device according to an exemplary embodiment.
  • Fig. 7 is a block diagram of a mobile terminal according to an exemplary embodiment.
  • Fig. 8 is a block diagram of a device according to an exemplary embodiment.
  • Fig. 1 is a flowchart of a shooting guidance method according to an exemplary embodiment. As shown in Fig. 1, the shooting guidance method may be used in a mobile terminal, and may include the following steps.
  • step S11 an instruction image containing the target in the shooting scene is acquired.
  • the mobile terminal When the user is to take a picture through the mobile terminal, the mobile terminal obtains an indication image containing the target in the shooting scene. Specifically, the mobile terminal can acquire the indication image in real time. It should be emphasized that since this solution is to instruct the user to compose the composition during shooting, then the indicated image is the framing picture generated by the mobile terminal during the shooting process, that is, the user ’s display screen after the user turns on the camera function. The picture presented.
  • the target in the shooting scene may be a target that users, animals, vehicles, buildings, etc. usually pay more attention to.
  • the target in the shooting scene is the saliency target
  • the image area where the target in the indication image is located is the saliency target area.
  • the mobile terminal may be a smart phone, a tablet computer, a camera, a video camera, and other devices that have a photo function.
  • step S12 based on the pre-trained convolutional neural network, a saliency target area in the indication image is determined.
  • the convolutional neural network is a network for identifying the salient target area in the image.
  • the convolutional neural network is trained based on multiple sample images and the labeled image corresponding to each sample image, and the marked image corresponding to each sample image is marked with the saliency target area of the sample image.
  • a convolutional neural network is obtained by training in advance based on multiple sample images and the labeled images corresponding to each sample image.
  • the saliency target area of the indication image can be determined based on the trained convolutional neural network.
  • the determined saliency target area may be one or more.
  • the instruction image is input to the trained convolutional neural network to obtain the saliency target area in the instruction image.
  • the saliency target area in the indication image can be embodied in the indication image through the area frame.
  • the shooting scene may include different categories of targets.
  • not only the marked target area is marked in the marked image corresponding to each sample image, but also the target category of the target contained in the marked target area may be marked.
  • the trained convolutional neural network can detect the saliency target area in the indication image, and can mark the target category corresponding to the saliency target area.
  • the three white line boxes show the three saliency target regions based on the convolutional neural network, and the target categories corresponding to the three saliency target regions, such as person /0.7, sheep / 0.951, sheep / 0.9; as shown in four white lines in Figure 2 (b), four saliency target regions based on convolutional neural network are shown, and four saliency can be marked
  • the target categories corresponding to the sexual target areas such as person / 0.752, person / 0.561, horse / 0.959, horse / 0.916; as shown by the two white lines in Figure 2 (c), they are obtained based on the convolutional neural network.
  • Two saliency target areas, and the target categories corresponding to the two saliency target areas can be marked, such as sheep / 0.918, sheep / 0.871.
  • step S13 based on the determined saliency target area, guidance information for guiding the user to compose a picture is output.
  • the guide information is information for guiding the user to compose a picture, that is, information for guiding the user to move the determined saliency target area to a certain position on the screen.
  • the guidance information can be displayed in a floating manner on the viewfinder screen.
  • the guide information may include: a first mark marked for the salient target area to be moved, a second mark marked for the position to be placed, and a composition description for moving the first mark and the second mark to coincide .
  • the number of salient target regions may be one or more.
  • output guide information for guiding the user to compose a picture which may specifically include:
  • the one saliency target area can be directly selected as the subject area.
  • the subject area can be selected according to the areas of different saliency target areas, for example, the saliency target area with the largest area is selected as the subject area.
  • the subject area can be selected according to the target category corresponding to different distinctive target areas. For example, when shooting an image that includes a person, it is generally expected that the person is located in a more important position in the image. At this time, it can be judged The target category corresponding to the saliency target area, the saliency target area whose target category is human is determined as the subject area, and so on.
  • the target category is a human significant target area is the subject area 1
  • the target category is an animal significant target area is the subject area 2, and so on.
  • the preset composition method may include a nine-square lattice composition, a symmetric composition, a guide line composition, a three-point composition, and the like.
  • the target shooting area can be different.
  • the target shooting area may be the middle square area of the nine-square grid
  • the target shooting area may be the left-side composition of the symmetric composition, and so on.
  • the movement of the main body area to the target shooting area may be: the main body area all falls within the target shooting area; or, the main body area and the target shooting area completely overlap; or, the main body area and the target shooting area overlap a predetermined ratio, such as 80%, 90% and many more.
  • the user can move the mobile terminal until the determined saliency target area or subject area is moved to the target shooting area.
  • Auxiliary shooting is performed under the shooting interface shown in FIG. 3 (a).
  • the shooting interface shown in FIG. 3 (a) includes smart composition, flash, low light, and variable speed options; and also includes albums and live streaming alongside the shooting options And K song options; as shown in Figure 3 (b), Figure 3 (c), Figure 3 (d), and Figure 3 (e), in order to move the subject area to the target shooting area, an open circle 301 annotating the subject area The change in the positional relationship with the solid area 302 marked with the target shooting area.
  • the open circle 301 moves down as the marked main body area moves down, that is, the open center circle 301 always marks the main body area, so that the virtual circle 301 falls into the solid circle 302 At this time, the subject area marked by the open circle also falls into the target shooting area.
  • the target shooting area is determined by the preset composition method
  • the preset composition method is the composition method used when shooting the instruction image.
  • the technical solutions provided by the embodiments of the present application may include the following beneficial effects: through a convolutional neural network, a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output is used for Guidance information that guides the user to composition.
  • this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since the scheme can be applied to various grayscale situations, the scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
  • the method provided in the embodiment of the present application may further include: instructing the subject area to move to After the target shooting area, it can also include:
  • the user When it is detected whether the subject area moves to the target shooting area, the user is notified that the composition is successful. Specifically, as shown in FIG. 3 (e), text information can be displayed in the shooting interface, such as: "the composition is successful, and the picture can be taken.” Or you can issue a voice message prompt, and so on.
  • the user can intuitively know that the composition is successful, and take a picture when the composition is successful, so that the target in the shooting scene can be taken when it is in the best shooting position, so that the user is not required to have professional photography knowledge to guide the user to compose the composition, thereby improving The quality of the captured image.
  • the saliency target area in the framing screen guides the user to compose a composition, for example, the user moves the mobile phone, so that the subject area selected from the saliency target area is located in the target shooting area, instructing the user to take a higher quality image.
  • the embodiments of the present application may further include: a step of training a convolutional neural network. Specifically, as shown in FIG. 4, it may include:
  • the electronic device acquires multiple sample images, such as 10,000, 20,000, etc.
  • the plurality of sample images include sample images of different categories, and the sample images of different categories include targets of different categories.
  • a data set may be constructed in advance, and then sample images are obtained from the data set.
  • images corresponding to natural scenes can be obtained, such as selecting 7 target categories in a realizable way to construct a data set, such as person (person), cat (cat), dog, (dog), horse (horse), sheep (sheep), cow (cow), bird (bird).
  • part of the sample image can be used as the training set and part of the test set.
  • the total training data in the data set is 1328415, and 1904 data is used as the test set to verify the convolutional neural network. effect.
  • the embodiments of the present application propose the following ways to determine the sampling weights of images of different categories, the main steps are as follows:
  • the regional convolutional neural network can ensure that the images of each category in a batch (batch processing) are balanced during network model training, and prevent deviations in model training.
  • the saliency target area corresponding to each sample image can be marked by manual marking; or the saliency target area corresponding to each sample image can be marked through detection of target characteristics, and so on.
  • the preset convolutional nerve may include a parameter to be tested, input the sample image to the preset convolutional neural network, and adjust the parameter to be tested, so that the output of the preset convolutional neural network is infinitely close to the labeled image corresponding to the sample image
  • the marked area when the difference between the output of the preset convolutional neural network and the image marked by the marked image corresponding to the sample image converges, the parameter to be measured is determined, and the obtained preset convolution including the determined parameter to be measured is obtained
  • the neural network is the convolutional neural network trained.
  • the parameters to be tested may include: batch size, learning rate, and / or number of iterations, and so on.
  • the preset convolutional neural network may include: a basic module, a feature module, a positioning module, and a classification module.
  • the basic module can be composed of 5 convolutional layers, such as Conv2d (3-> 8), Conv2d (8-> 16), Conv2d (16-> 32), Conv2d (32-> 64), Conv2d (64-> 128) ), Where Conv2d (3-> 8) is understood to convert a 3-channel RGB format image into an 8-channel image, and the meaning of other convolutional layers is similar;
  • the feature module can be composed of 3 convolutional layers, such as Conv2d (128- > 256), Conv2d (256-> 128), Conv2d (128-> 256);
  • the positioning module can be composed of a convolutional layer, such as Conv2d (*> 8);
  • the classification module can be composed of a convolutional layer, Such as Conv2d (*> 4).
  • the basic module can also be called Base Model
  • the feature module can also be called Extra Model
  • the positioning module can also be called Location Layer
  • the classification module can also be called Confidence Layer.
  • Base is mainly used to process the features from the bottom to the top of the image, and is used to provide features for Extra Model.
  • Extra Model mainly extracts feature maps of different scales, mainly designed for targets of different sizes in the image. Each pixel position of the feature map is mapped to the corresponding size of the bounding box on the original image, which can be Location and Layer Provide characteristic information of different size targets.
  • Location layer is used to return the bounding box coordinates of the target on the original image for each pixel of the feature map.
  • Confidence layer is used to calculate the target category of the bounding box on the original image for each pixel of the feature map.
  • the trained network may be optimized and evaluated.
  • the preset convolutional neural network After the preset convolutional neural network is constructed, it needs to be trained on the training data set. The preset convolutional neural network training depends on the loss function and the optimizer. Finally, the evaluation of the preset convolutional neural network needs to set an appropriate evaluation index.
  • the loss function uses the multi-task loss function in the optimization process of the convolutional neural network, uses the cross-entropy loss function for the classification task, and uses the L1 distance loss function for the bounding box.
  • Optimizer network optimization uses stochastic gradient descent method to update the parameters of convolutional neural network.
  • mean, accuracy, and average precision can be used in the embodiments of the present application to evaluate the detection result.
  • the acquiring multiple sample images includes:
  • the original image after interference is determined as the sample image.
  • the interference may include horizontal flip, vertical flip, random cropping, image pixel disturbance, lighting processing, occlusion, low contrast processing, and so on.
  • Fig. 6 is a block diagram of a shooting guidance device according to an exemplary embodiment.
  • the apparatus may include a first acquisition module 601, a first determination module 602, and an instruction module 603;
  • the first obtaining module 601 is configured to obtain an indication image containing the target in the shooting scene
  • the first determination module 602 is configured to determine a saliency target area in the indicated image based on the pre-trained convolutional neural network; wherein, the convolutional neural network is based on multiple sample images and each sample image corresponds to From the training of labeled images, the marked image corresponding to each sample image is marked with the saliency target area in the sample image;
  • the instruction module 603 is configured to output guide information for guiding the user to compose a composition based on the determined saliency target area.
  • the technical solutions provided by the embodiments of the present application may include the following beneficial effects: through a convolutional neural network, a saliency target area including an indication image of a target in a shooting scene is determined, and further, based on the determined saliency target area, output Guidance information that guides the user to composition. Since this scheme does not need to consider the grayscale difference of each pixel, therefore, for various grayscale differences, this scheme can ensure the accuracy of determining the salient target area, and then ensure the accuracy of the shooting guidance. Since the scheme can be applied to various grayscale situations, the scheme is more robust to different situations. And there is no need to extract the salient target area according to the artificially defined determined features, which makes the application scene more extensive.
  • the instruction module 603 is specifically configured to: select a subject area from the determined salient target areas; output guide information for guiding the user to move the subject area to the target shooting area; wherein, The target shooting area is determined based on a preset composition method, and the preset composition method is a composition method used when shooting the instruction image.
  • the device also includes:
  • the prompting module is configured to, after the instruction module outputs guidance information for guiding the user to move the subject area to the target shooting area, when it is detected whether the subject area moves to the target shooting area, to the user Prompt composition success.
  • the device also includes:
  • the second acquisition module is configured to acquire multiple sample images
  • the second determining module is configured to determine the mark image corresponding to each sample image respectively;
  • the training module is configured to take each sample image as input content and the labeled image corresponding to each sample image as supervising content to train the preset convolutional neural network to obtain a trained convolutional neural network.
  • the second acquisition module is specifically configured to: acquire multiple original images; add interference to the multiple original images respectively to obtain multiple original images after interference; combine the multiple original images after interference Both are determined as sample images.
  • the plurality of sample images include sample images of different categories, and the sample images of different categories include targets of different categories.
  • the preset convolutional neural network includes: a basic module, a feature module, a positioning module, and a classification module.
  • Fig. 7 is a block diagram of a mobile terminal 700 for shooting guidance according to an exemplary embodiment.
  • the mobile terminal 700 may be a mobile phone, a computer, a digital broadcasting terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
  • the mobile terminal 700 may include one or more of the following components: a processing component 702, a memory 704, a power component 706, a multimedia component 708, an audio component 710, an input / output (I / O) interface 712, and a sensor component 714 , ⁇ ⁇ ⁇ 716.
  • the processing component 702 generally controls the overall operations of the mobile terminal 700, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
  • the processing component 702 may include one or more processors 720 to execute instructions to complete all or part of the steps of the above-mentioned shooting guidance method.
  • the processing component 702 may include one or more modules to facilitate interaction between the processing component 702 and other components.
  • the processing component 702 may include a multimedia module to facilitate interaction between the multimedia component 708 and the processing component 702.
  • the memory 704 is configured to store various types of data to support operations at the mobile terminal 700. Examples of these data include instructions for any application or method operating on the mobile terminal 700, contact data, phone book data, messages, pictures, videos, and so on.
  • the memory 704 may be implemented by any type of volatile or nonvolatile storage device or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable and removable Programmable read only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory, magnetic disk or optical disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read only memory
  • EPROM erasable and removable Programmable read only memory
  • PROM programmable read only memory
  • ROM read only memory
  • magnetic memory flash memory
  • flash memory magnetic disk or optical disk.
  • the power supply component 706 provides power to various components of the mobile terminal 700.
  • the power supply component 706 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the mobile terminal 700.
  • the multimedia component 708 includes a screen between the mobile terminal 700 and the user that provides an output interface.
  • the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touch, swipe, and gestures on the touch panel. The touch sensor may not only sense the boundary of the touch or sliding action, but also detect the duration and pressure related to the touch or sliding operation.
  • the multimedia component 708 includes a front camera and / or a rear camera. When the mobile terminal 700 is in an operation mode, such as a shooting mode or a video mode, the front camera and / or the rear camera may receive external multimedia data. Each front camera and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
  • the audio component 710 is configured to output and / or input audio signals.
  • the audio component 710 includes a microphone (MIC), and when the mobile terminal 700 is in an operation mode, such as a call mode, a recording mode, and a voice recognition mode, the microphone is configured to receive an external audio signal.
  • the received audio signal may be further stored in the memory 704 or sent via the communication component 716.
  • the audio component 710 further includes a speaker for outputting audio signals.
  • the I / O interface 712 provides an interface between the processing component 702 and a peripheral interface module.
  • the peripheral interface module may be a keyboard, a click wheel, or a button. These buttons may include, but are not limited to: home button, volume button, start button, and lock button.
  • the sensor component 714 includes one or more sensors for providing the mobile terminal 700 with status evaluation in various aspects.
  • the sensor component 714 can detect the on / off state of the mobile terminal 700, and the relative positioning of the components, such as the display and the keypad of the mobile terminal 700, the sensor component 714 can also detect the mobile terminal 700 or the mobile terminal 700 The position of the component changes, the presence or absence of user contact with the mobile terminal 700, the orientation or acceleration / deceleration of the mobile terminal 700, and the temperature change of the mobile terminal 700.
  • the sensor assembly 714 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • the sensor component 714 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 714 may further include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • the communication component 716 is configured to facilitate wired or wireless communication between the mobile terminal 700 and other devices.
  • the mobile terminal 700 may access a wireless network based on a communication standard, such as WiFi, an operator network (such as 2G, 3G, 4G, or 5G), or a combination thereof.
  • the communication component 716 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel.
  • the communication component 716 further includes a near field communication (NFC) module to facilitate short-range communication.
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • the mobile terminal 700 may be used by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field Programming gate array (FPGA), controller, microcontroller, microprocessor or other electronic components are used to implement the above method.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field Programming gate array
  • controller microcontroller, microprocessor or other electronic components are used to implement the above method.
  • a non-transitory computer-readable storage medium including instructions is also provided, for example, a memory 704 including instructions, which can be executed by the processor 720 of the mobile terminal 700 to complete the above method.
  • the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, or the like.
  • a computer program product which, when run on a computer, enables the computer to execute the above-mentioned shooting guidance method.
  • Fig. 8 is a block diagram of a device 800 for shooting guidance according to an exemplary embodiment.
  • the device 800 may be provided as a server.
  • the device 800 includes a processing component 822, which further includes one or more processors, and memory resources represented by the memory 832, for storing instructions executable by the processing component 822, such as application programs.
  • the application programs stored in the memory 832 may include one or more modules each corresponding to a set of instructions.
  • the processing component 822 is configured to execute instructions to perform the method steps of the above-mentioned shooting guidance method.
  • the device 800 may also include a power component 826 configured to perform power management of the device 800, a wired or wireless network interface 850 configured to connect the device 800 to the network, and an input output (I / O) interface 858.
  • the device 800 can operate based on an operating system stored in the memory 832, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Studio Devices (AREA)

Abstract

L'invention concerne un procédé et un appareil d'aide à la photographie, ainsi qu'un terminal mobile et un support de stockage. Le procédé consiste à : acquérir une image d'indication comprenant un objet dans un scénario de photographie; déterminer, sur la base d'un réseau neuronal convolutif pré-entraîné, une région d'objet en saillie dans l'image d'indication, le réseau neuronal convolutif étant obtenu au moyen d'un apprentissage en fonction de multiples images d'échantillon et d'une image de marquage correspondant à chacune des images d'échantillon, et l'image de marquage correspondant à chacune des images d'échantillon marquant une région d'objet en saillie dans l'image d'échantillon; et délivrer, sur la base de la région d'objet en saillie déterminée, des informations d'aide pour amener un utilisateur à composer une photographie. Le procédé et l'appareil d'aide à la photographie, le terminal mobile et le support de stockage décrits dans les modes de réalisation de la présente invention permettent de garantir une amélioration de la précision d'aide à la photographie en tenant compte de diverses nuances de gris.
PCT/CN2019/112566 2018-11-05 2019-10-22 Procédé et appareil d'aide à la photographie, terminal mobile et support de stockage WO2020093866A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811307419.6 2018-11-05
CN201811307419.6A CN109040605A (zh) 2018-11-05 2018-11-05 拍摄引导方法、装置及移动终端和存储介质

Publications (1)

Publication Number Publication Date
WO2020093866A1 true WO2020093866A1 (fr) 2020-05-14

Family

ID=64614404

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/112566 WO2020093866A1 (fr) 2018-11-05 2019-10-22 Procédé et appareil d'aide à la photographie, terminal mobile et support de stockage

Country Status (2)

Country Link
CN (1) CN109040605A (fr)
WO (1) WO2020093866A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860242A (zh) * 2020-07-07 2020-10-30 北京海益同展信息科技有限公司 一种机器人巡检方法、装置和计算机可读介质
CN112766285A (zh) * 2021-01-26 2021-05-07 北京有竹居网络技术有限公司 图像样本生成方法、装置和电子设备

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109040605A (zh) * 2018-11-05 2018-12-18 北京达佳互联信息技术有限公司 拍摄引导方法、装置及移动终端和存储介质
CN111246098B (zh) * 2020-01-19 2022-02-22 深圳市人工智能与机器人研究院 机器人拍照方法、装置、计算机设备和存储介质
CN111327833B (zh) * 2020-03-31 2021-06-01 厦门美图之家科技有限公司 辅助拍摄方法、装置、电子设备和可读存储介质
CN111461248A (zh) * 2020-04-09 2020-07-28 上海城诗信息科技有限公司 一种摄影构图线匹配方法、装置、设备及存储介质
CN111464743A (zh) * 2020-04-09 2020-07-28 上海城诗信息科技有限公司 一种摄影构图匹配方法及系统
CN112581446A (zh) * 2020-12-15 2021-03-30 影石创新科技股份有限公司 一种图像的显著性物体检测方法、装置、设备及存储介质
CN116998159A (zh) * 2021-03-04 2023-11-03 Oppo广东移动通信有限公司 建议电子设备的拍摄位置的方法及电子设备
CN113329175A (zh) * 2021-05-21 2021-08-31 浙江大华技术股份有限公司 抓拍方法、装置、电子装置和存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104717413A (zh) * 2013-12-12 2015-06-17 北京三星通信技术研究有限公司 拍照辅助方法及设备
CN107247930A (zh) * 2017-05-26 2017-10-13 西安电子科技大学 基于cnn和选择性注意机制的sar图像目标检测方法
CN107423747A (zh) * 2017-04-13 2017-12-01 中国人民解放军国防科学技术大学 一种基于深度卷积网络的显著性目标检测方法
CN107909109A (zh) * 2017-11-17 2018-04-13 西安电子科技大学 基于显著性和多尺度深度网络模型的sar图像分类方法
US20180314943A1 (en) * 2017-04-27 2018-11-01 Jianming Liang Systems, methods, and/or media, for selecting candidates for annotation for use in training a classifier
CN109040605A (zh) * 2018-11-05 2018-12-18 北京达佳互联信息技术有限公司 拍摄引导方法、装置及移动终端和存储介质

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105931255A (zh) * 2016-05-18 2016-09-07 天津工业大学 基于显著性和深度卷积神经网络的图像中目标定位方法
US10579860B2 (en) * 2016-06-06 2020-03-03 Samsung Electronics Co., Ltd. Learning model for salient facial region detection

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104717413A (zh) * 2013-12-12 2015-06-17 北京三星通信技术研究有限公司 拍照辅助方法及设备
CN107423747A (zh) * 2017-04-13 2017-12-01 中国人民解放军国防科学技术大学 一种基于深度卷积网络的显著性目标检测方法
US20180314943A1 (en) * 2017-04-27 2018-11-01 Jianming Liang Systems, methods, and/or media, for selecting candidates for annotation for use in training a classifier
CN107247930A (zh) * 2017-05-26 2017-10-13 西安电子科技大学 基于cnn和选择性注意机制的sar图像目标检测方法
CN107909109A (zh) * 2017-11-17 2018-04-13 西安电子科技大学 基于显著性和多尺度深度网络模型的sar图像分类方法
CN109040605A (zh) * 2018-11-05 2018-12-18 北京达佳互联信息技术有限公司 拍摄引导方法、装置及移动终端和存储介质

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860242A (zh) * 2020-07-07 2020-10-30 北京海益同展信息科技有限公司 一种机器人巡检方法、装置和计算机可读介质
CN111860242B (zh) * 2020-07-07 2024-05-21 京东科技信息技术有限公司 一种机器人巡检方法、装置和计算机可读介质
CN112766285A (zh) * 2021-01-26 2021-05-07 北京有竹居网络技术有限公司 图像样本生成方法、装置和电子设备
CN112766285B (zh) * 2021-01-26 2024-03-19 北京有竹居网络技术有限公司 图像样本生成方法、装置和电子设备

Also Published As

Publication number Publication date
CN109040605A (zh) 2018-12-18

Similar Documents

Publication Publication Date Title
WO2020093866A1 (fr) Procédé et appareil d'aide à la photographie, terminal mobile et support de stockage
US10534972B2 (en) Image processing method, device and medium
KR101649596B1 (ko) 피부색 조절방법, 장치, 프로그램 및 기록매체
CN104270565B (zh) 图像拍摄方法、装置及设备
RU2577188C1 (ru) Способ, аппарат и устройство для сегментации изображения
US20170064181A1 (en) Method and apparatus for controlling photography of unmanned aerial vehicle
EP3226204A1 (fr) Procédé et appareil de capture d'image intelligente
US10110800B2 (en) Method and apparatus for setting image capturing parameters
CN104918107B (zh) 视频文件的标识处理方法及装置
CN109889730B (zh) 调整拍摄角度的提示方法、装置及电子设备
JP2016533066A (ja) 撮影制御方法、撮影制御装置、端末、プログラム及び記録媒体
US10248855B2 (en) Method and apparatus for identifying gesture
CN105959587B (zh) 快门速度获取方法和装置
CN105758319B (zh) 通过移动终端测量目标物体高度的方法和装置
CN106408603A (zh) 拍摄方法及装置
CN111182212B (zh) 图像处理方法、装置、存储介质以及电子设备
CN110751659B (zh) 图像分割方法及装置、终端、存储介质
CN106664366A (zh) 图像处理装置、摄像装置、图像处理方法以及程序
US20180150722A1 (en) Photo synthesizing method, device, and medium
CN109670458A (zh) 一种车牌识别方法及装置
WO2020119254A1 (fr) Procédé et dispositif de recommandation de filtre, équipement électronique et support de stockage
US11574415B2 (en) Method and apparatus for determining an icon position
CN108717542B (zh) 识别文字区域的方法、装置及计算机可读存储介质
WO2018036010A1 (fr) Procédé et appareil de commande de véhicule à équilibrage
CN112115894A (zh) 手部关键点检测模型的训练方法、装置及电子设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19882219

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19882219

Country of ref document: EP

Kind code of ref document: A1