WO2021057267A1 - Procédé de traitement d'image et dispositif terminal - Google Patents

Procédé de traitement d'image et dispositif terminal Download PDF

Info

Publication number
WO2021057267A1
WO2021057267A1 PCT/CN2020/106777 CN2020106777W WO2021057267A1 WO 2021057267 A1 WO2021057267 A1 WO 2021057267A1 CN 2020106777 W CN2020106777 W CN 2020106777W WO 2021057267 A1 WO2021057267 A1 WO 2021057267A1
Authority
WO
WIPO (PCT)
Prior art keywords
target image
terminal device
module
target
training sample
Prior art date
Application number
PCT/CN2020/106777
Other languages
English (en)
Chinese (zh)
Inventor
任帅
Original Assignee
维沃移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 维沃移动通信有限公司 filed Critical 维沃移动通信有限公司
Publication of WO2021057267A1 publication Critical patent/WO2021057267A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • H04N23/632Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof

Definitions

  • the embodiments of the present disclosure relate to the field of communication technologies, and in particular, to an image processing method and terminal equipment.
  • terminal device technology users use terminal devices more and more frequently, and users have higher and higher requirements for the shooting function of terminal devices.
  • the camera in the terminal device uses optical zoom and digital zoom to increase the zoom factor of the camera, which can enlarge the distant scene during shooting, and then can shoot high-quality distant images.
  • the embodiments of the present disclosure provide an image processing method and terminal device to solve the problem of privacy information leakage in the process of remote shooting.
  • an embodiment of the present disclosure provides an image processing method, the method including:
  • the first zoom factor of the camera of the terminal device is greater than or equal to the predetermined threshold, acquiring the target image in the preview interface of the camera;
  • the above-mentioned target image is recognized, and if it is recognized that the target image contains preset privacy information, the target area in the target image is blurred, and the target area is the area where the preset privacy information is located.
  • the embodiments of the present disclosure also provide a terminal device, which includes:
  • An acquiring module configured to acquire the target image in the preview interface of the camera when the first zoom factor of the camera of the terminal device is greater than or equal to a predetermined threshold
  • the recognition module is used to recognize the above-mentioned target image acquired by the above-mentioned acquisition module
  • the blurring module is configured to blur the target area in the target image if the recognition module recognizes that the target image contains preset privacy information, and the target area is the area where the preset privacy information is located.
  • the embodiments of the present disclosure provide a terminal device, including a processor, a memory, and a computer program stored on the memory and capable of running on the processor.
  • the computer program is executed by the processor to achieve the following On the one hand, the steps of the image processing method.
  • embodiments of the present disclosure provide a computer-readable storage medium on which a computer program is stored.
  • the computer program is executed by a processor, the steps of the image processing method as described in the first aspect are implemented. .
  • the terminal device when the first zoom factor of the camera of the terminal device is greater than or equal to the predetermined threshold, the terminal device acquires the target image in the preview interface of the camera, and recognizes the target image. In the case that the target image contains preset private information, the terminal device blurs the area where the preset private information is located in the target image to avoid unintentional or malicious disclosure of other people's private information during remote shooting.
  • FIG. 1 is a schematic structural diagram of a possible Android operating system provided by an embodiment of the present disclosure
  • FIG. 2 is a schematic flowchart of an image processing method provided by an embodiment of the disclosure
  • FIG. 3 is one of the schematic diagrams of an interface applied by an image processing method provided by an embodiment of the disclosure
  • FIG. 4 is a second schematic diagram of an interface applied by an image processing method provided by an embodiment of the disclosure.
  • FIG. 5 is one of the schematic structural diagrams of a terminal device provided by an embodiment of the disclosure.
  • FIG. 6 is the second structural diagram of a terminal device provided by an embodiment of the disclosure.
  • A/B can mean A or B
  • the "and/or" in this article is only an association relationship describing associated objects, indicating that there may be three A relationship, for example, A and/or B, can mean: A alone exists, A and B exist at the same time, and B exists alone.
  • words such as “first” and “second” are used to describe the same items or similar items with basically the same function or effect. To distinguish, those skilled in the art can understand that words such as “first” and “second” do not limit the number and execution order.
  • the first zoom factor and the second zoom factor are used to distinguish different zoom factors, not to describe a specific sequence of zoom factors.
  • the execution subject of the image processing method provided by the embodiments of the present disclosure may be the aforementioned terminal device (including mobile terminal devices and non-mobile terminal devices), or may be a functional module and/or function in the terminal device that can implement the image processing method.
  • the entity can be specifically determined according to actual usage requirements, and the embodiment of the present disclosure does not limit it.
  • the following takes a terminal device as an example to illustrate the image processing method provided by the embodiment of the present disclosure.
  • the terminal device in the embodiment of the present disclosure may be a mobile terminal device or a non-mobile terminal device.
  • the mobile terminal device can be a mobile phone, a tablet computer, a notebook computer, a palmtop computer, a vehicle-mounted terminal device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook, or a personal digital assistant (PDA)
  • the non-mobile terminal device may be a personal computer (PC), a television (television, TV), a teller machine, or a self-service machine, etc.; the embodiment of the present disclosure does not specifically limit it.
  • the terminal device in the embodiment of the present disclosure may be a terminal device with an operating system.
  • the operating system may be an Android operating system, an ios operating system, or other possible operating systems, which are not specifically limited in the embodiments of the present disclosure.
  • the following takes the Android operating system as an example to introduce the software environment to which the image processing method provided in the embodiments of the present disclosure is applied.
  • FIG. 1 it is a schematic structural diagram of a possible Android operating system provided by an embodiment of the present disclosure.
  • the architecture of the Android operating system includes 4 layers, namely: application layer, application framework layer, system runtime library layer, and kernel layer (specifically, it may be the Linux kernel layer).
  • the application layer includes various applications (including system applications and third-party applications) in the Android operating system.
  • the application framework layer is the framework of the application. Developers can develop some applications based on the application framework layer while complying with the development principles of the application framework.
  • the system runtime library layer includes a library (also called a system library) and an Android operating system runtime environment.
  • the library mainly provides various resources needed by the Android operating system.
  • the Android operating system operating environment is used to provide a software environment for the Android operating system.
  • the kernel layer is the operating system layer of the Android operating system and belongs to the lowest level of the Android operating system software.
  • the kernel layer is based on the Linux kernel to provide core system services and hardware-related drivers for the Android operating system.
  • developers can develop software programs that implement the image processing method provided by the embodiments of the present disclosure based on the system architecture of the Android operating system as shown in FIG.
  • the processing method can be run based on the Android operating system as shown in FIG. 1. That is, the processor or the terminal device can implement the image processing method provided by the embodiment of the present disclosure by running the software program in the Android operating system.
  • FIG. 2 is a schematic flowchart of an image processing method provided by an embodiment of the present disclosure, including step 201 and step 202:
  • Step 201 When the first zoom factor of the camera of the terminal device is greater than or equal to a predetermined threshold, the terminal device acquires the target image in the preview interface of the camera.
  • the aforementioned zoom may include: optical zoom, digital zoom, and dual-camera zoom.
  • the terminal device may increase the zoom factor of the camera to enlarge the distant scene during shooting, so as to take a high-quality distant scene image.
  • the terminal device can automatically obtain the first zoom factor of the camera.
  • the first zoom factor of the camera is greater than or equal to a predetermined threshold (for example, the zoom factor is 3 times), this At this time, it can be considered that the camera of the terminal device enters the remote shooting mode, and the terminal device obtains the target image in the preview interface of the camera.
  • Step 202 The terminal device recognizes the above-mentioned target image, and if it recognizes that the target image contains preset privacy information, the terminal device blurs the target area in the target image.
  • the above-mentioned target area is the area where the preset private information is located, and the terminal device may determine the target area by acquiring the coordinates of the preset private information in the target image.
  • the aforementioned preset privacy information may include: improperly dressed residents in a distant building, payment codes (such as QR codes or barcodes) of customers in stores, and payment passwords of customers in stores. Specifically, It is set according to actual needs, and this disclosure does not limit this.
  • the terminal device may use a blur algorithm to blur the target area in the target image. It can be understood that blurring the target area in the target image is blurring the target area in the target image.
  • the above-mentioned fuzzy algorithm includes at least one of the following: Gaussian fuzzy filter algorithm, mean fuzzy filter algorithm, and Laplacian fuzzy filter algorithm, which can be specifically set according to actual needs, and the present disclosure does not limit this.
  • the terminal device acquires the camera The image of the distant building in the preview interface (that is, the above-mentioned target image, as shown in Figure 3, 31), when the terminal device recognizes that the image 31 shows an improperly dressed resident (that is, the above-mentioned preset privacy information) , As shown in Figure 4, the terminal device uses the Gaussian blur filter algorithm to blur the area (ie, the above-mentioned target area, as 42 in Figure 4) in the image of the distant building collected by the camera (ie 41 in Figure 4). ). It should be noted that the oblique line in 42 in FIG. 4 is used to indicate blur.
  • the terminal device can exit the remote shooting mode after recognizing that the above-mentioned target image contains preset privacy information.
  • the terminal device may adjust the first zoom factor of the aforementioned camera. Wherein, the adjusted zoom factor is less than the aforementioned predetermined threshold. It can be considered that the terminal device has exited the long-distance shooting mode, and no processing is performed on the above-mentioned target image at this time.
  • the terminal device may adjust the degree of blurring of the target area in the target image according to the aforementioned first zoom factor.
  • step 202 specifically includes the following steps 202a and 202b:
  • Step 202a The terminal device determines a target blur factor corresponding to the first zoom factor of the camera.
  • Step 202b The terminal device blurs the target area in the target image according to the target blur factor.
  • the blur factor of the image of the target area in the target image is proportional to the first zoom factor of the camera.
  • the blur factor can be adjusted by the size of the Gaussian blur kernel, and the blur factor is proportional to the first zoom factor of the camera, which can be understood as .
  • the size of the Gaussian blur kernel is proportional to the first zoom factor of the camera, that is, the larger the first zoom factor, the larger the corresponding Gaussian blur kernel, and the greater the blur degree of the target area.
  • the terminal device when the first zoom factor of the camera of the terminal device is greater than or equal to a predetermined threshold, the terminal device acquires the target image in the preview interface of the camera, and recognizes the target image, In the case of recognizing that the target image contains preset private information, the terminal device blurs the area where the preset private information is located in the target image to avoid unintentional or malicious disclosure of other people's private information during long-distance shooting .
  • the terminal device may also remind or warn the user that the target image contains preset privacy information in other ways.
  • the method further includes the following step 203:
  • Step 203 The terminal device sends out a warning message.
  • the terminal device can directly remind the user that the target image contains preset private information by means of vibration and text.
  • the terminal device recognizes that the above-mentioned target image contains preset privacy information, there is no obvious order in the execution process of the terminal device blurring the target area in the target image and the terminal device issuing the warning message.
  • the terminal device may issue a warning message before blurring the target area in the target image, or the terminal device may issue a warning message while blurring the target area in the target image, or the terminal device may blur the target image
  • a warning message is issued after the target area in the target area, which is not limited in the embodiment of the present disclosure.
  • the terminal device after the terminal device sends out a vibration warning signal to remind the user that the target image contains preset privacy information, the user can manually switch the shooting scene. At this time, the terminal device can reacquire the image in the preview interface of the camera. The image is the target image.
  • the terminal device when the terminal device blurs the above-mentioned target area, the terminal device can exit the remote shooting mode, which can be achieved in at least two possible implementation manners.
  • the method further includes the following step 204:
  • Step 204 The terminal device adjusts the first zoom factor of the aforementioned camera to the second zoom factor.
  • the adjusted second zoom factor is less than the aforementioned predetermined threshold.
  • the terminal device can actively adjust the focus. After blurring the above-mentioned target area, the terminal device directly adjusts the focus according to the predetermined second zoom factor (less than the above-mentioned predetermined threshold), that is, the terminal device exits remote shooting and no longer performs blurring processing on the target image.
  • the predetermined second zoom factor less than the above-mentioned predetermined threshold
  • the user can manually adjust the focus. After the terminal device blurs the aforementioned target area, the user can manually lower the first zoom factor of the camera of the terminal device. At this time, when the terminal device obtains that the second zoom factor of the camera is less than the predetermined threshold, it can be considered as the terminal device The long-distance shooting mode has been exited, and the target image is no longer blurred at this time, that is, the terminal device enters normal shooting.
  • the zoom factor is directly proportional to the degree of blur, in the process of manually lowering the first zoom factor of the camera of the terminal device, the degree of blurring of the target area of the terminal device can change with the zoom factor. It is smaller and lowered until the terminal device no longer blurs the target image.
  • the target image is recognized. If it is recognized that the target image contains preset privacy information, the zoom factor of the camera can be directly reduced, that is, the first zoom factor is reduced to the second zoom factor, and the second zoom factor is less than The predetermined threshold may not blur the target area where the preset private information is located.
  • the terminal device after the terminal device obtains the target image currently collected by the camera, it needs to identify whether the target image contains preset privacy information.
  • step 202 specifically includes the following step A1:
  • Step A1 The terminal device uses the image detection model to recognize the above-mentioned target image, and obtains the recognition result.
  • the aforementioned recognition result is used to indicate whether the aforementioned target image contains preset privacy information.
  • the method further includes the following step A2:
  • Step A2 The terminal device trains the above-mentioned image detection model according to the training samples in the training sample library.
  • the aforementioned training sample library includes at least one training sample, and each training sample in the aforementioned at least one training sample contains preset privacy information.
  • the aforementioned training sample may be an image collected by a user through a terminal device with an image collection function (for example, a camera, a mobile phone, etc.).
  • an image collection function for example, a camera, a mobile phone, etc.
  • the method further includes the following step B1:
  • Step B1 If the terminal device recognizes that the target image contains preset privacy information, the terminal device uses the target image as a training sample and stores it in the training sample library.
  • the terminal device recognizes that the above-mentioned target image contains preset privacy information, and can store the target image in a sample library, so that the training samples in the training sample library can be enriched, and the recognition result of the image detection model can be improved. accuracy.
  • the image detection model recognizes the above-mentioned target image and finds that the recognition result is incorrect (for example, the target image contains preset privacy information, but the recognition result indicates that it does not contain preset privacy information), in this case, the The target image is used as a training sample and stored in the above-mentioned training sample library, and the above-mentioned image detection model is retrained with the training samples in the supplemented training sample library to improve the accuracy of the recognition result obtained after the image detection model recognizes the target image Sex.
  • FIG. 5 is a schematic diagram of a possible structure for implementing a terminal device provided by an embodiment of the present disclosure.
  • the terminal device 600 includes: an acquisition module 601, an identification module 602, and a blurring module 603, where:
  • the acquiring module 601 is configured to acquire the target image in the preview interface of the camera when the first zoom factor of the camera of the terminal device is greater than or equal to a predetermined threshold.
  • the recognition module 602 is configured to recognize the above-mentioned target image acquired by the above-mentioned acquisition module.
  • the blurring module 603 is configured to blur the target area in the target image if the recognition module 602 recognizes that the target image contains preset privacy information, and the target area is the area where the preset privacy information is located.
  • the terminal device further includes: an adjustment module 604, wherein: the adjustment module 604 is configured to adjust the first zoom factor of the camera to a second zoom factor, and the second zoom factor is smaller than the aforementioned The predetermined threshold.
  • the recognition module 602 is specifically configured to use an image detection model to recognize the target image acquired by the acquisition module 601 to obtain a recognition result, and the recognition result is used to indicate whether the target image contains preset privacy information.
  • the terminal device further includes: a training module 605 and a storage module 606, wherein: the above-mentioned training module 605 is configured to train the above-mentioned image detection model according to the training samples in the training sample library;
  • the training sample library includes at least one training sample, and each training sample in the at least one training sample contains preset privacy information.
  • the storage module 606 is configured to, if it is recognized that the target image contains preset privacy information, use the target image acquired by the acquisition module 601 as a training sample and store it in the training sample library.
  • the terminal device further includes: a determining module 607, wherein: the determining module 607 is configured to determine a target blur factor corresponding to the first zoom factor; the blurring module 602 specifically uses According to the target blur factor determined by the determination module 607, the target area in the target image is blurred.
  • the terminal device further includes: a sending module 608, wherein: the above-mentioned sending module 608 is configured to send a warning message.
  • the terminal device when the first zoom factor of the camera of the terminal device is greater than or equal to a predetermined threshold, the terminal device acquires the target image in the preview interface of the camera, and recognizes the target image. When it is recognized that the target image contains preset private information, the terminal device blurs the area where the preset private information is located in the target image to avoid unintentional or malicious disclosure of other people's private information during remote shooting.
  • the terminal device provided in the embodiment of the present disclosure can implement each process implemented by the terminal device in the foregoing method embodiment, and to avoid repetition, details are not described herein again.
  • the terminal device 100 includes but is not limited to: a radio frequency unit 101, a network module 102, an audio output unit 103, an input unit 104, a sensor 105, and a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111 and other components.
  • a radio frequency unit 101 for example, a radio frequency unit
  • a network module 102 includes but is not limited to: a radio frequency unit 101, a network module 102, an audio output unit 103, an input unit 104, a sensor 105, and a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111 and other components.
  • a radio frequency unit 101 includes but is not limited to: a radio frequency unit 101, a network module 102, an audio output unit 103, an input unit 104, a sensor 105, and a display unit 106, a
  • the terminal device 100 includes, but is not limited to, a mobile phone, a tablet computer, a notebook computer, a palmtop computer, a vehicle-mounted terminal device, a wearable device, a pedometer, and the like.
  • the processor 110 is configured to obtain the target image in the preview interface of the camera when the first zoom factor of the camera of the terminal device is greater than or equal to a predetermined threshold; it is also used to recognize the target image. If the target image contains preset privacy information, the target area in the target image is blurred, and the target area is the area where the preset privacy information is located.
  • the terminal device when the zoom factor of the first camera of the terminal device is greater than or equal to a predetermined threshold, the terminal device acquires a target image in the preview interface of the camera, and recognizes the target image. When it is recognized that the target image contains preset private information, the terminal device blurs the area where the preset private information is located in the target image to avoid unintentional or malicious disclosure of other people's private information during remote shooting.
  • the radio frequency unit 101 can be used for receiving and sending signals in the process of sending and receiving information or talking. Specifically, the downlink data from the base station is received and processed by the processor 110; in addition, Uplink data is sent to the base station.
  • the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
  • the radio frequency unit 101 can also communicate with the network and other devices through a wireless communication system.
  • the terminal device 100 provides users with wireless broadband Internet access through the network module 102, such as helping users to send and receive emails, browse web pages, and access streaming media.
  • the audio output unit 103 can convert the audio data received by the radio frequency unit 101 or the network module 102 or stored in the memory 109 into an audio signal and output it as sound. Moreover, the audio output unit 103 may also provide audio output related to a specific function performed by the terminal device 100 (for example, call signal reception sound, message reception sound, etc.).
  • the audio output unit 103 includes a speaker, a buzzer, a receiver, and the like.
  • the input unit 104 is used to receive audio or video signals.
  • the input unit 104 may include a graphics processing unit (GPU) 1041 and a microphone 1042, and the graphics processor 1041 is configured to monitor images of still pictures or videos obtained by an image capture device (such as a camera) in a video capture mode or an image capture mode.
  • the data is processed.
  • the processed image frame can be displayed on the display unit 106.
  • the image frame processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or sent via the radio frequency unit 101 or the network module 102.
  • the microphone 1042 can receive sound, and can process such sound into audio data.
  • the processed audio data can be converted into a format that can be sent to a mobile communication base station via the radio frequency unit 101 for output in the case of a telephone call mode.
  • the terminal device 100 also includes at least one sensor 105, such as a light sensor, a motion sensor, and other sensors.
  • the light sensor includes an ambient light sensor and a proximity sensor.
  • the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of the ambient light.
  • the proximity sensor can close the display panel 1061 and the display panel 1061 when the terminal device 100 is moved to the ear. / Or backlight.
  • the accelerometer sensor can detect the magnitude of acceleration in various directions (usually three-axis), and can detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of the terminal device (such as horizontal and vertical screen switching, related games) , Magnetometer attitude calibration), vibration recognition related functions (such as pedometer, percussion), etc.; sensor 105 can also include fingerprint sensors, pressure sensors, iris sensors, molecular sensors, gyroscopes, barometers, hygrometers, thermometers, Infrared sensors, etc., will not be repeated here.
  • the display unit 106 is used to display information input by the user or information provided to the user.
  • the display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (LCD), an organic light-emitting diode (OLED), etc.
  • LCD liquid crystal display
  • OLED organic light-emitting diode
  • the user input unit 107 may be used to receive input numeric or character information, and generate key signal input related to user settings and function control of the terminal device 100.
  • the user input unit 107 includes a touch panel 1071 and other input devices 1072.
  • the touch panel 1071 also called a touch screen, can collect the user's touch operations on or near it (for example, the user uses any suitable objects or accessories such as fingers, stylus, etc.) on the touch panel 1071 or near the touch panel 1071. operating).
  • the touch panel 1071 may include two parts: a touch detection device and a touch controller.
  • the touch detection device detects the user's touch position, detects the signal brought by the touch operation, and transmits the signal to the touch controller; the touch controller receives the touch information from the touch detection device, converts it into contact coordinates, and then sends it To the processor 110, the command sent by the processor 110 is received and executed.
  • the touch panel 1071 can be implemented in multiple types such as resistive, capacitive, infrared, and surface acoustic wave.
  • the user input unit 107 may also include other input devices 1072.
  • other input devices 1072 may include, but are not limited to, a physical keyboard, function keys (such as volume control buttons, switch buttons, etc.), trackball, mouse, and joystick, which will not be repeated here.
  • the touch panel 1071 can be overlaid on the display panel 1061.
  • the touch panel 1071 detects a touch operation on or near it, it is transmitted to the processor 110 to determine the type of the touch event, and then the processor 110 according to The type of touch event provides corresponding visual output on the display panel 1061.
  • the touch panel 1071 and the display panel 1061 are used as two independent components to implement the input and output functions of the terminal device 100, in some embodiments, the touch panel 1071 and the display panel 1061 can be combined.
  • the input and output functions of the terminal device 100 are realized by integration, which is not specifically limited here.
  • the interface unit 108 is an interface for connecting an external device with the terminal device 100.
  • the external device may include a wired or wireless headset port, an external power source (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device with an identification module, audio input/output (I/O) port, video I/O port, headphone port, etc.
  • the interface unit 108 can be used to receive input from an external device (for example, data information, power, etc.) and transmit the received input to one or more elements in the terminal device 100 or can be used to connect to the terminal device 100 and external devices. Transfer data between devices.
  • the memory 109 can be used to store software programs and various data.
  • the memory 109 may mainly include a program storage area and a data storage area.
  • the program storage area may store an operating system, an application program required by at least one function (such as a sound playback function, an image playback function, etc.), etc.; Data created by the use of mobile phones (such as audio data, phone book, etc.), etc.
  • the memory 109 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, a flash memory device, or other volatile solid-state storage devices.
  • the processor 110 is the control center of the terminal device 100. It uses various interfaces and lines to connect the various parts of the entire terminal device 100, runs or executes software programs and/or modules stored in the memory 109, and calls and stores them in the memory 109.
  • the data of the terminal device 100 performs various functions and processing data, so as to monitor the terminal device 100 as a whole.
  • the processor 110 may include one or more processing units; optionally, the processor 110 may integrate an application processor and a modem processor, where the application processor mainly processes the operating system, user interface, and application programs, etc.
  • the adjustment processor mainly deals with wireless communication. It can be understood that the foregoing modem processor may not be integrated into the processor 110.
  • the terminal device 100 may also include a power source 111 (such as a battery) for supplying power to various components.
  • a power source 111 such as a battery
  • the power source 111 may be logically connected to the processor 110 through a power management system, so as to manage charging, discharging, and power consumption through the power management system. Management and other functions.
  • the terminal device 100 includes some functional modules not shown, which will not be repeated here.
  • an embodiment of the present disclosure further provides a terminal device, including a processor, a memory, and a computer program stored in the memory and running on the processor 110.
  • the computer program realizes the above-mentioned image when executed by the processor.
  • Each process of the embodiment of the processing method can achieve the same technical effect. In order to avoid repetition, it will not be repeated here.
  • the embodiments of the present disclosure also provide a computer-readable storage medium, and a computer program is stored on the computer-readable storage medium.
  • a computer program is stored on the computer-readable storage medium.
  • the computer program is executed by a processor, each process of the above-mentioned image processing method embodiment is realized, and the same technology can be achieved. The effect, in order to avoid repetition, will not be repeated here.
  • the computer-readable storage medium such as read-only memory (Read-Only Memory, ROM for short), random access memory (Random Access Memory, RAM for short), magnetic disk, or optical disk, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Telephone Function (AREA)
  • Studio Devices (AREA)

Abstract

La présente invention concerne, selon certains modes de réalisation, un procédé de traitement d'image et un dispositif terminal. Le procédé consiste à : obtenir, si un premier rapport de zoom d'une caméra d'un dispositif terminal est supérieur ou égal à un seuil prédéfini, une image cible dans une interface de prévisualisation de la caméra; et identifier l'image cible, et s'il est identifié que l'image cible comprend des informations de confidentialité prédéfinies, rendre flou une zone cible dans l'image cible, la zone cible étant une zone où les informations de confidentialité prédéfinies sont situées.
PCT/CN2020/106777 2019-09-24 2020-08-04 Procédé de traitement d'image et dispositif terminal WO2021057267A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910907111.3 2019-09-24
CN201910907111.3A CN110719402B (zh) 2019-09-24 2019-09-24 图像处理方法及终端设备

Publications (1)

Publication Number Publication Date
WO2021057267A1 true WO2021057267A1 (fr) 2021-04-01

Family

ID=69210032

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/106777 WO2021057267A1 (fr) 2019-09-24 2020-08-04 Procédé de traitement d'image et dispositif terminal

Country Status (2)

Country Link
CN (1) CN110719402B (fr)
WO (1) WO2021057267A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114463264A (zh) * 2021-12-28 2022-05-10 浙江大华技术股份有限公司 高空抛物监测中隐私保护方法及装置
CN114692202A (zh) * 2022-03-31 2022-07-01 马上消费金融股份有限公司 图像处理方法、装置、电子设备及存储介质

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110719402B (zh) * 2019-09-24 2021-07-06 维沃移动通信(杭州)有限公司 图像处理方法及终端设备
CN111339831B (zh) * 2020-01-23 2023-08-18 深圳市大拿科技有限公司 一种照明灯控制方法和系统
CN113742183A (zh) * 2020-05-29 2021-12-03 青岛海信移动通信技术股份有限公司 一种录屏方法、终端及存储介质
CN112040145B (zh) * 2020-08-28 2023-04-07 维沃移动通信有限公司 图像处理方法、装置及电子设备
CN112257123B (zh) * 2020-09-07 2024-07-26 西安万像电子科技有限公司 图像的处理方法及系统
CN112057874A (zh) * 2020-09-10 2020-12-11 重庆五诶科技有限公司 具备隐私保护的游戏辅助系统及方法
CN112182648A (zh) * 2020-09-18 2021-01-05 支付宝(杭州)信息技术有限公司 一种隐私图像、人脸隐私的处理方法、装置及设备
CN114815638A (zh) * 2021-08-25 2022-07-29 北京京东方技术开发有限公司 设备配置方法、电子装置和计算机可读存储介质
CN113722735A (zh) * 2021-08-31 2021-11-30 维沃移动通信有限公司 隐私保护方法、装置、电子设备及可读存储介质
CN113987602A (zh) * 2021-10-28 2022-01-28 维沃移动通信有限公司 图像数据的处理方法和电子设备
CN117135307A (zh) * 2022-08-08 2023-11-28 惠州Tcl移动通信有限公司 摄像头监控方法及装置

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105653041A (zh) * 2016-01-29 2016-06-08 北京小米移动软件有限公司 显示状态调整方法及装置
CN107122679A (zh) * 2017-05-16 2017-09-01 北京小米移动软件有限公司 图像处理方法及装置
US9848167B1 (en) * 2016-06-21 2017-12-19 Amazon Technologies, Inc. Low bandwidth video
CN107958161A (zh) * 2017-11-30 2018-04-24 维沃移动通信有限公司 一种多任务显示方法及移动终端
CN108366196A (zh) * 2018-01-25 2018-08-03 西安中科创达软件有限公司 一种保护图片隐私的方法
CN110719402A (zh) * 2019-09-24 2020-01-21 维沃移动通信(杭州)有限公司 图像处理方法及终端设备

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100771137B1 (ko) * 2005-02-21 2007-10-30 삼성전자주식회사 프라이버시영역을 마스크 처리하는 감시시스템 및 마스크영역 설정방법
JP5088161B2 (ja) * 2008-02-15 2012-12-05 ソニー株式会社 画像処理装置、カメラ装置、通信システム、画像処理方法、およびプログラム
US8970697B2 (en) * 2010-07-06 2015-03-03 Panasonic Intellectual Property Corporation Of America Image distribution apparatus
WO2013061505A1 (fr) * 2011-10-25 2013-05-02 Sony Corporation Appareil, procédé et produit programme d'ordinateur de traitement d'image
JP6687488B2 (ja) * 2015-12-24 2020-04-22 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America 無人飛行体及びその制御方法
JP6910772B2 (ja) * 2016-09-08 2021-07-28 キヤノン株式会社 撮像装置、撮像装置の制御方法およびプログラム
CN106803930A (zh) * 2017-02-10 2017-06-06 上海斐讯数据通信技术有限公司 一种基于路由器的智能视频监控方法及智能路由器
CN109413323A (zh) * 2017-08-15 2019-03-01 联发科技(新加坡)私人有限公司 图像处理方法、拍照设备、及存储介质
CN108848334A (zh) * 2018-07-11 2018-11-20 广东小天才科技有限公司 一种视频处理的方法、装置、终端和存储介质
CN109040594B (zh) * 2018-08-24 2020-12-18 创新先进技术有限公司 拍照方法及装置

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105653041A (zh) * 2016-01-29 2016-06-08 北京小米移动软件有限公司 显示状态调整方法及装置
US9848167B1 (en) * 2016-06-21 2017-12-19 Amazon Technologies, Inc. Low bandwidth video
CN107122679A (zh) * 2017-05-16 2017-09-01 北京小米移动软件有限公司 图像处理方法及装置
CN107958161A (zh) * 2017-11-30 2018-04-24 维沃移动通信有限公司 一种多任务显示方法及移动终端
CN108366196A (zh) * 2018-01-25 2018-08-03 西安中科创达软件有限公司 一种保护图片隐私的方法
CN110719402A (zh) * 2019-09-24 2020-01-21 维沃移动通信(杭州)有限公司 图像处理方法及终端设备

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114463264A (zh) * 2021-12-28 2022-05-10 浙江大华技术股份有限公司 高空抛物监测中隐私保护方法及装置
CN114692202A (zh) * 2022-03-31 2022-07-01 马上消费金融股份有限公司 图像处理方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
CN110719402A (zh) 2020-01-21
CN110719402B (zh) 2021-07-06

Similar Documents

Publication Publication Date Title
WO2021057267A1 (fr) Procédé de traitement d'image et dispositif terminal
WO2021104197A1 (fr) Procédé de poursuite d'objet et dispositif électronique
WO2021098678A1 (fr) Procédé de commande de vidéocapture d'écran et dispositif électronique
CN108513070B (zh) 一种图像处理方法、移动终端及计算机可读存储介质
WO2021104195A1 (fr) Procédé d'affichage d'images et dispositif électronique
CN111562896B (zh) 投屏方法及电子设备
WO2021098603A1 (fr) Procédé d'affichage d'image de prévisualisation et dispositif électronique
CN109523253B (zh) 一种支付方法和装置
CN108307106B (zh) 一种图像处理方法、装置及移动终端
WO2021082772A1 (fr) Procédé de capture d'écran et dispositif électronique
WO2021057712A1 (fr) Procédé de détection de proximité et dispositif terminal
CN110944113B (zh) 对象显示方法及电子设备
CN110138967B (zh) 一种终端的操作控制方法及终端
CN110830713A (zh) 一种变焦方法及电子设备
US20220272275A1 (en) Photographing method and electronic device
EP3816768A1 (fr) Procédé de reconnaissance d'objets et terminal mobile
CN109871253A (zh) 一种显示方法及终端
CN108881721A (zh) 一种显示方法及终端
CN109992192B (zh) 一种界面显示方法及终端设备
CN109104573B (zh) 一种确定对焦点的方法及终端设备
CN109005337B (zh) 一种拍照方法及终端
WO2021104265A1 (fr) Dispositif électronique et procédé de mise au point
CN108243489B (zh) 一种拍照控制方法及移动终端
CN109756630A (zh) 一种屏幕控制方法及终端设备
WO2021036504A1 (fr) Procédé de suppression d'image et dispositif terminal

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20870238

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20870238

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 20870238

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 13/10/2022)

122 Ep: pct application non-entry in european phase

Ref document number: 20870238

Country of ref document: EP

Kind code of ref document: A1